var/home/core/zuul-output/0000755000175000017500000000000015116374063014533 5ustar corecorevar/home/core/zuul-output/logs/0000755000175000017500000000000015116405275015477 5ustar corecorevar/home/core/zuul-output/logs/kubelet.log0000644000000000000000005170651715116405267017720 0ustar rootrootDec 10 22:49:24 crc systemd[1]: Starting Kubernetes Kubelet... Dec 10 22:49:24 crc restorecon[4680]: Relabeled /var/lib/kubelet/config.json from system_u:object_r:unlabeled_t:s0 to system_u:object_r:container_var_lib_t:s0 Dec 10 22:49:24 crc restorecon[4680]: /var/lib/kubelet/device-plugins not reset as customized by admin to system_u:object_r:container_file_t:s0 Dec 10 22:49:24 crc restorecon[4680]: /var/lib/kubelet/device-plugins/kubelet.sock not reset as customized by admin to system_u:object_r:container_file_t:s0 Dec 10 22:49:24 crc restorecon[4680]: /var/lib/kubelet/pods/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8/volumes/kubernetes.io~configmap/nginx-conf/..2025_02_23_05_40_35.4114275528/nginx.conf not reset as customized by admin to system_u:object_r:container_file_t:s0:c15,c25 Dec 10 22:49:24 crc restorecon[4680]: /var/lib/kubelet/pods/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c15,c25 Dec 10 22:49:24 crc restorecon[4680]: /var/lib/kubelet/pods/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8/containers/networking-console-plugin/22e96971 not reset as customized by admin to system_u:object_r:container_file_t:s0:c15,c25 Dec 10 22:49:24 crc restorecon[4680]: /var/lib/kubelet/pods/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8/containers/networking-console-plugin/21c98286 not reset as customized by admin to system_u:object_r:container_file_t:s0:c15,c25 Dec 10 22:49:24 crc restorecon[4680]: /var/lib/kubelet/pods/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8/containers/networking-console-plugin/0f1869e1 not reset as customized by admin to system_u:object_r:container_file_t:s0:c15,c25 Dec 10 22:49:24 crc restorecon[4680]: /var/lib/kubelet/pods/d1b160f5dda77d281dd8e69ec8d817f9/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c215,c682 Dec 10 22:49:24 crc restorecon[4680]: /var/lib/kubelet/pods/d1b160f5dda77d281dd8e69ec8d817f9/containers/setup/46889d52 not reset as customized by admin to system_u:object_r:container_file_t:s0:c225,c458 Dec 10 22:49:24 crc restorecon[4680]: /var/lib/kubelet/pods/d1b160f5dda77d281dd8e69ec8d817f9/containers/setup/5b6a5969 not reset as customized by admin to system_u:object_r:container_file_t:s0:c666,c963 Dec 10 22:49:24 crc restorecon[4680]: /var/lib/kubelet/pods/d1b160f5dda77d281dd8e69ec8d817f9/containers/setup/6c7921f5 not reset as customized by admin to system_u:object_r:container_file_t:s0:c215,c682 Dec 10 22:49:24 crc restorecon[4680]: /var/lib/kubelet/pods/d1b160f5dda77d281dd8e69ec8d817f9/containers/kube-rbac-proxy-crio/4804f443 not reset as customized by admin to system_u:object_r:container_file_t:s0:c225,c458 Dec 10 22:49:24 crc restorecon[4680]: /var/lib/kubelet/pods/d1b160f5dda77d281dd8e69ec8d817f9/containers/kube-rbac-proxy-crio/2a46b283 not reset as customized by admin to system_u:object_r:container_file_t:s0:c225,c458 Dec 10 22:49:24 crc restorecon[4680]: /var/lib/kubelet/pods/d1b160f5dda77d281dd8e69ec8d817f9/containers/kube-rbac-proxy-crio/a6b5573e not reset as customized by admin to system_u:object_r:container_file_t:s0:c225,c458 Dec 10 22:49:24 crc restorecon[4680]: /var/lib/kubelet/pods/d1b160f5dda77d281dd8e69ec8d817f9/containers/kube-rbac-proxy-crio/4f88ee5b not reset as customized by admin to system_u:object_r:container_file_t:s0:c225,c458 Dec 10 22:49:24 crc restorecon[4680]: /var/lib/kubelet/pods/d1b160f5dda77d281dd8e69ec8d817f9/containers/kube-rbac-proxy-crio/5a4eee4b not reset as customized by admin to system_u:object_r:container_file_t:s0:c666,c963 Dec 10 22:49:24 crc restorecon[4680]: /var/lib/kubelet/pods/d1b160f5dda77d281dd8e69ec8d817f9/containers/kube-rbac-proxy-crio/cd87c521 not reset as customized by admin to system_u:object_r:container_file_t:s0:c215,c682 Dec 10 22:49:24 crc restorecon[4680]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/volumes/kubernetes.io~configmap/service-ca-bundle not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Dec 10 22:49:24 crc restorecon[4680]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/volumes/kubernetes.io~configmap/service-ca-bundle/..2025_02_23_05_33_42.2574241751 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Dec 10 22:49:24 crc restorecon[4680]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/volumes/kubernetes.io~configmap/service-ca-bundle/..2025_02_23_05_33_42.2574241751/service-ca.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Dec 10 22:49:24 crc restorecon[4680]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/volumes/kubernetes.io~configmap/service-ca-bundle/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Dec 10 22:49:24 crc restorecon[4680]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/volumes/kubernetes.io~configmap/service-ca-bundle/service-ca.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Dec 10 22:49:24 crc restorecon[4680]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Dec 10 22:49:24 crc restorecon[4680]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/containers/router/38602af4 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Dec 10 22:49:24 crc restorecon[4680]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/containers/router/1483b002 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Dec 10 22:49:24 crc restorecon[4680]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/containers/router/0346718b not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Dec 10 22:49:24 crc restorecon[4680]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/containers/router/d3ed4ada not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Dec 10 22:49:24 crc restorecon[4680]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/containers/router/3bb473a5 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Dec 10 22:49:24 crc restorecon[4680]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/containers/router/8cd075a9 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Dec 10 22:49:24 crc restorecon[4680]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/containers/router/00ab4760 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Dec 10 22:49:24 crc restorecon[4680]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/containers/router/54a21c09 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Dec 10 22:49:24 crc restorecon[4680]: /var/lib/kubelet/pods/37a5e44f-9a88-4405-be8a-b645485e7312/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c589,c726 Dec 10 22:49:24 crc restorecon[4680]: /var/lib/kubelet/pods/37a5e44f-9a88-4405-be8a-b645485e7312/containers/network-operator/70478888 not reset as customized by admin to system_u:object_r:container_file_t:s0:c176,c499 Dec 10 22:49:24 crc restorecon[4680]: /var/lib/kubelet/pods/37a5e44f-9a88-4405-be8a-b645485e7312/containers/network-operator/43802770 not reset as customized by admin to system_u:object_r:container_file_t:s0:c176,c499 Dec 10 22:49:24 crc restorecon[4680]: /var/lib/kubelet/pods/37a5e44f-9a88-4405-be8a-b645485e7312/containers/network-operator/955a0edc not reset as customized by admin to system_u:object_r:container_file_t:s0:c176,c499 Dec 10 22:49:24 crc restorecon[4680]: /var/lib/kubelet/pods/37a5e44f-9a88-4405-be8a-b645485e7312/containers/network-operator/bca2d009 not reset as customized by admin to system_u:object_r:container_file_t:s0:c140,c1009 Dec 10 22:49:24 crc restorecon[4680]: /var/lib/kubelet/pods/37a5e44f-9a88-4405-be8a-b645485e7312/containers/network-operator/b295f9bd not reset as customized by admin to system_u:object_r:container_file_t:s0:c589,c726 Dec 10 22:49:24 crc restorecon[4680]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/volumes/kubernetes.io~configmap/cni-binary-copy not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Dec 10 22:49:24 crc restorecon[4680]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/volumes/kubernetes.io~configmap/cni-binary-copy/..2025_02_23_05_21_22.3617465230 not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Dec 10 22:49:24 crc restorecon[4680]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/volumes/kubernetes.io~configmap/cni-binary-copy/..2025_02_23_05_21_22.3617465230/cnibincopy.sh not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Dec 10 22:49:24 crc restorecon[4680]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/volumes/kubernetes.io~configmap/cni-binary-copy/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Dec 10 22:49:24 crc restorecon[4680]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/volumes/kubernetes.io~configmap/cni-binary-copy/cnibincopy.sh not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Dec 10 22:49:24 crc restorecon[4680]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/volumes/kubernetes.io~configmap/cni-sysctl-allowlist not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Dec 10 22:49:24 crc restorecon[4680]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/volumes/kubernetes.io~configmap/cni-sysctl-allowlist/..2025_02_23_05_21_22.2050650026 not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Dec 10 22:49:24 crc restorecon[4680]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/volumes/kubernetes.io~configmap/cni-sysctl-allowlist/..2025_02_23_05_21_22.2050650026/allowlist.conf not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Dec 10 22:49:24 crc restorecon[4680]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/volumes/kubernetes.io~configmap/cni-sysctl-allowlist/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Dec 10 22:49:24 crc restorecon[4680]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/volumes/kubernetes.io~configmap/cni-sysctl-allowlist/allowlist.conf not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Dec 10 22:49:24 crc restorecon[4680]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Dec 10 22:49:24 crc restorecon[4680]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/egress-router-binary-copy/bc46ea27 not reset as customized by admin to system_u:object_r:container_file_t:s0:c203,c924 Dec 10 22:49:24 crc restorecon[4680]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/egress-router-binary-copy/5731fc1b not reset as customized by admin to system_u:object_r:container_file_t:s0:c138,c778 Dec 10 22:49:24 crc restorecon[4680]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/egress-router-binary-copy/5e1b2a3c not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Dec 10 22:49:24 crc restorecon[4680]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/cni-plugins/943f0936 not reset as customized by admin to system_u:object_r:container_file_t:s0:c203,c924 Dec 10 22:49:24 crc restorecon[4680]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/cni-plugins/3f764ee4 not reset as customized by admin to system_u:object_r:container_file_t:s0:c138,c778 Dec 10 22:49:24 crc restorecon[4680]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/cni-plugins/8695e3f9 not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Dec 10 22:49:24 crc restorecon[4680]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/bond-cni-plugin/aed7aa86 not reset as customized by admin to system_u:object_r:container_file_t:s0:c203,c924 Dec 10 22:49:24 crc restorecon[4680]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/bond-cni-plugin/c64d7448 not reset as customized by admin to system_u:object_r:container_file_t:s0:c138,c778 Dec 10 22:49:24 crc restorecon[4680]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/bond-cni-plugin/0ba16bd2 not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Dec 10 22:49:24 crc restorecon[4680]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/routeoverride-cni/207a939f not reset as customized by admin to system_u:object_r:container_file_t:s0:c203,c924 Dec 10 22:49:24 crc restorecon[4680]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/routeoverride-cni/54aa8cdb not reset as customized by admin to system_u:object_r:container_file_t:s0:c138,c778 Dec 10 22:49:24 crc restorecon[4680]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/routeoverride-cni/1f5fa595 not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Dec 10 22:49:24 crc restorecon[4680]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/whereabouts-cni-bincopy/bf9c8153 not reset as customized by admin to system_u:object_r:container_file_t:s0:c203,c924 Dec 10 22:49:24 crc restorecon[4680]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/whereabouts-cni-bincopy/47fba4ea not reset as customized by admin to system_u:object_r:container_file_t:s0:c138,c778 Dec 10 22:49:24 crc restorecon[4680]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/whereabouts-cni-bincopy/7ae55ce9 not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Dec 10 22:49:24 crc restorecon[4680]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/whereabouts-cni/7906a268 not reset as customized by admin to system_u:object_r:container_file_t:s0:c203,c924 Dec 10 22:49:24 crc restorecon[4680]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/whereabouts-cni/ce43fa69 not reset as customized by admin to system_u:object_r:container_file_t:s0:c138,c778 Dec 10 22:49:24 crc restorecon[4680]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/whereabouts-cni/7fc7ea3a not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Dec 10 22:49:24 crc restorecon[4680]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/kube-multus-additional-cni-plugins/d8c38b7d not reset as customized by admin to system_u:object_r:container_file_t:s0:c203,c924 Dec 10 22:49:24 crc restorecon[4680]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/kube-multus-additional-cni-plugins/9ef015fb not reset as customized by admin to system_u:object_r:container_file_t:s0:c138,c778 Dec 10 22:49:24 crc restorecon[4680]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/kube-multus-additional-cni-plugins/b9db6a41 not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Dec 10 22:49:24 crc restorecon[4680]: /var/lib/kubelet/pods/5b88f790-22fa-440e-b583-365168c0b23d/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c432,c991 Dec 10 22:49:24 crc restorecon[4680]: /var/lib/kubelet/pods/5b88f790-22fa-440e-b583-365168c0b23d/containers/network-metrics-daemon/b1733d79 not reset as customized by admin to system_u:object_r:container_file_t:s0:c476,c820 Dec 10 22:49:24 crc restorecon[4680]: /var/lib/kubelet/pods/5b88f790-22fa-440e-b583-365168c0b23d/containers/network-metrics-daemon/afccd338 not reset as customized by admin to system_u:object_r:container_file_t:s0:c272,c818 Dec 10 22:49:24 crc restorecon[4680]: /var/lib/kubelet/pods/5b88f790-22fa-440e-b583-365168c0b23d/containers/network-metrics-daemon/9df0a185 not reset as customized by admin to system_u:object_r:container_file_t:s0:c432,c991 Dec 10 22:49:24 crc restorecon[4680]: /var/lib/kubelet/pods/5b88f790-22fa-440e-b583-365168c0b23d/containers/kube-rbac-proxy/18938cf8 not reset as customized by admin to system_u:object_r:container_file_t:s0:c476,c820 Dec 10 22:49:24 crc restorecon[4680]: /var/lib/kubelet/pods/5b88f790-22fa-440e-b583-365168c0b23d/containers/kube-rbac-proxy/7ab4eb23 not reset as customized by admin to system_u:object_r:container_file_t:s0:c272,c818 Dec 10 22:49:24 crc restorecon[4680]: /var/lib/kubelet/pods/5b88f790-22fa-440e-b583-365168c0b23d/containers/kube-rbac-proxy/56930be6 not reset as customized by admin to system_u:object_r:container_file_t:s0:c432,c991 Dec 10 22:49:24 crc restorecon[4680]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/volumes/kubernetes.io~configmap/env-overrides not reset as customized by admin to system_u:object_r:container_file_t:s0:c440,c975 Dec 10 22:49:24 crc restorecon[4680]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/volumes/kubernetes.io~configmap/env-overrides/..2025_02_23_05_21_35.630010865 not reset as customized by admin to system_u:object_r:container_file_t:s0:c440,c975 Dec 10 22:49:24 crc restorecon[4680]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/volumes/kubernetes.io~configmap/env-overrides/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c440,c975 Dec 10 22:49:24 crc restorecon[4680]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/volumes/kubernetes.io~configmap/ovnkube-config not reset as customized by admin to system_u:object_r:container_file_t:s0:c440,c975 Dec 10 22:49:24 crc restorecon[4680]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/volumes/kubernetes.io~configmap/ovnkube-config/..2025_02_23_05_21_35.1088506337 not reset as customized by admin to system_u:object_r:container_file_t:s0:c440,c975 Dec 10 22:49:24 crc restorecon[4680]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/volumes/kubernetes.io~configmap/ovnkube-config/..2025_02_23_05_21_35.1088506337/ovnkube.conf not reset as customized by admin to system_u:object_r:container_file_t:s0:c440,c975 Dec 10 22:49:24 crc restorecon[4680]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/volumes/kubernetes.io~configmap/ovnkube-config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c440,c975 Dec 10 22:49:24 crc restorecon[4680]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/volumes/kubernetes.io~configmap/ovnkube-config/ovnkube.conf not reset as customized by admin to system_u:object_r:container_file_t:s0:c440,c975 Dec 10 22:49:24 crc restorecon[4680]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c440,c975 Dec 10 22:49:24 crc restorecon[4680]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/containers/kube-rbac-proxy/0d8e3722 not reset as customized by admin to system_u:object_r:container_file_t:s0:c89,c211 Dec 10 22:49:24 crc restorecon[4680]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/containers/kube-rbac-proxy/d22b2e76 not reset as customized by admin to system_u:object_r:container_file_t:s0:c382,c850 Dec 10 22:49:24 crc restorecon[4680]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/containers/kube-rbac-proxy/e036759f not reset as customized by admin to system_u:object_r:container_file_t:s0:c440,c975 Dec 10 22:49:24 crc restorecon[4680]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/containers/ovnkube-cluster-manager/2734c483 not reset as customized by admin to system_u:object_r:container_file_t:s0:c89,c211 Dec 10 22:49:24 crc restorecon[4680]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/containers/ovnkube-cluster-manager/57878fe7 not reset as customized by admin to system_u:object_r:container_file_t:s0:c89,c211 Dec 10 22:49:24 crc restorecon[4680]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/containers/ovnkube-cluster-manager/3f3c2e58 not reset as customized by admin to system_u:object_r:container_file_t:s0:c89,c211 Dec 10 22:49:24 crc restorecon[4680]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/containers/ovnkube-cluster-manager/375bec3e not reset as customized by admin to system_u:object_r:container_file_t:s0:c382,c850 Dec 10 22:49:24 crc restorecon[4680]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/containers/ovnkube-cluster-manager/7bc41e08 not reset as customized by admin to system_u:object_r:container_file_t:s0:c440,c975 Dec 10 22:49:24 crc restorecon[4680]: /var/lib/kubelet/pods/cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Dec 10 22:49:24 crc restorecon[4680]: /var/lib/kubelet/pods/cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d/containers/download-server/48c7a72d not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Dec 10 22:49:24 crc restorecon[4680]: /var/lib/kubelet/pods/cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d/containers/download-server/4b66701f not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Dec 10 22:49:24 crc restorecon[4680]: /var/lib/kubelet/pods/cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d/containers/download-server/a5a1c202 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Dec 10 22:49:24 crc restorecon[4680]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/volumes/kubernetes.io~configmap/ovnkube-identity-cm not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Dec 10 22:49:24 crc restorecon[4680]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/volumes/kubernetes.io~configmap/ovnkube-identity-cm/..2025_02_23_05_21_40.3350632666 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Dec 10 22:49:24 crc restorecon[4680]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/volumes/kubernetes.io~configmap/ovnkube-identity-cm/..2025_02_23_05_21_40.3350632666/additional-cert-acceptance-cond.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Dec 10 22:49:24 crc restorecon[4680]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/volumes/kubernetes.io~configmap/ovnkube-identity-cm/..2025_02_23_05_21_40.3350632666/additional-pod-admission-cond.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Dec 10 22:49:24 crc restorecon[4680]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/volumes/kubernetes.io~configmap/ovnkube-identity-cm/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Dec 10 22:49:24 crc restorecon[4680]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/volumes/kubernetes.io~configmap/ovnkube-identity-cm/additional-cert-acceptance-cond.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Dec 10 22:49:24 crc restorecon[4680]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/volumes/kubernetes.io~configmap/ovnkube-identity-cm/additional-pod-admission-cond.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Dec 10 22:49:24 crc restorecon[4680]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/volumes/kubernetes.io~configmap/env-overrides not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Dec 10 22:49:24 crc restorecon[4680]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/volumes/kubernetes.io~configmap/env-overrides/..2025_02_23_05_21_40.1388695756 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Dec 10 22:49:24 crc restorecon[4680]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/volumes/kubernetes.io~configmap/env-overrides/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Dec 10 22:49:24 crc restorecon[4680]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Dec 10 22:49:24 crc restorecon[4680]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/containers/webhook/26f3df5b not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Dec 10 22:49:24 crc restorecon[4680]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/containers/webhook/6d8fb21d not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Dec 10 22:49:24 crc restorecon[4680]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/containers/webhook/50e94777 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Dec 10 22:49:24 crc restorecon[4680]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/containers/approver/208473b3 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Dec 10 22:49:24 crc restorecon[4680]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/containers/approver/ec9e08ba not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Dec 10 22:49:24 crc restorecon[4680]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/containers/approver/3b787c39 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Dec 10 22:49:24 crc restorecon[4680]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/containers/approver/208eaed5 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Dec 10 22:49:24 crc restorecon[4680]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/containers/approver/93aa3a2b not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Dec 10 22:49:24 crc restorecon[4680]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/containers/approver/3c697968 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Dec 10 22:49:24 crc restorecon[4680]: /var/lib/kubelet/pods/3b6479f0-333b-4a96-9adf-2099afdc2447/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c20,c21 Dec 10 22:49:24 crc restorecon[4680]: /var/lib/kubelet/pods/3b6479f0-333b-4a96-9adf-2099afdc2447/containers/network-check-target-container/ba950ec9 not reset as customized by admin to system_u:object_r:container_file_t:s0:c20,c21 Dec 10 22:49:24 crc restorecon[4680]: /var/lib/kubelet/pods/3b6479f0-333b-4a96-9adf-2099afdc2447/containers/network-check-target-container/cb5cdb37 not reset as customized by admin to system_u:object_r:container_file_t:s0:c20,c21 Dec 10 22:49:24 crc restorecon[4680]: /var/lib/kubelet/pods/3b6479f0-333b-4a96-9adf-2099afdc2447/containers/network-check-target-container/f2df9827 not reset as customized by admin to system_u:object_r:container_file_t:s0:c20,c21 Dec 10 22:49:24 crc restorecon[4680]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/volumes/kubernetes.io~configmap/images not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Dec 10 22:49:24 crc restorecon[4680]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/volumes/kubernetes.io~configmap/images/..2025_02_23_05_22_30.473230615 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Dec 10 22:49:24 crc restorecon[4680]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/volumes/kubernetes.io~configmap/images/..2025_02_23_05_22_30.473230615/images.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Dec 10 22:49:24 crc restorecon[4680]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/volumes/kubernetes.io~configmap/images/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Dec 10 22:49:24 crc restorecon[4680]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/volumes/kubernetes.io~configmap/images/images.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Dec 10 22:49:24 crc restorecon[4680]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/volumes/kubernetes.io~configmap/auth-proxy-config not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Dec 10 22:49:24 crc restorecon[4680]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/volumes/kubernetes.io~configmap/auth-proxy-config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Dec 10 22:49:24 crc restorecon[4680]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/volumes/kubernetes.io~configmap/auth-proxy-config/config-file.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Dec 10 22:49:24 crc restorecon[4680]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/volumes/kubernetes.io~configmap/auth-proxy-config/..2025_02_24_06_22_02.1904938450 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Dec 10 22:49:24 crc restorecon[4680]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/volumes/kubernetes.io~configmap/auth-proxy-config/..2025_02_24_06_22_02.1904938450/config-file.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Dec 10 22:49:24 crc restorecon[4680]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Dec 10 22:49:24 crc restorecon[4680]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/containers/machine-config-operator/fedaa673 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Dec 10 22:49:24 crc restorecon[4680]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/containers/machine-config-operator/9ca2df95 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Dec 10 22:49:24 crc restorecon[4680]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/containers/machine-config-operator/b2d7460e not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Dec 10 22:49:24 crc restorecon[4680]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/containers/kube-rbac-proxy/2207853c not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Dec 10 22:49:24 crc restorecon[4680]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/containers/kube-rbac-proxy/241c1c29 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Dec 10 22:49:24 crc restorecon[4680]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/containers/kube-rbac-proxy/2d910eaf not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Dec 10 22:49:24 crc restorecon[4680]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/etcd-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Dec 10 22:49:24 crc restorecon[4680]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/etcd-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Dec 10 22:49:24 crc restorecon[4680]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/etcd-ca/..2025_02_23_05_23_49.3726007728 not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Dec 10 22:49:24 crc restorecon[4680]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/etcd-ca/..2025_02_23_05_23_49.3726007728/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Dec 10 22:49:24 crc restorecon[4680]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/etcd-ca/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Dec 10 22:49:24 crc restorecon[4680]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/etcd-service-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Dec 10 22:49:24 crc restorecon[4680]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/etcd-service-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Dec 10 22:49:24 crc restorecon[4680]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/etcd-service-ca/..2025_02_23_05_23_49.841175008 not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Dec 10 22:49:24 crc restorecon[4680]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/etcd-service-ca/..2025_02_23_05_23_49.841175008/service-ca.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Dec 10 22:49:24 crc restorecon[4680]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/etcd-service-ca/service-ca.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Dec 10 22:49:24 crc restorecon[4680]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Dec 10 22:49:24 crc restorecon[4680]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.843437178 not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Dec 10 22:49:24 crc restorecon[4680]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.843437178/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Dec 10 22:49:24 crc restorecon[4680]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Dec 10 22:49:24 crc restorecon[4680]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/config/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Dec 10 22:49:24 crc restorecon[4680]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Dec 10 22:49:24 crc restorecon[4680]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/containers/etcd-operator/c6c0f2e7 not reset as customized by admin to system_u:object_r:container_file_t:s0:c263,c871 Dec 10 22:49:24 crc restorecon[4680]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/containers/etcd-operator/399edc97 not reset as customized by admin to system_u:object_r:container_file_t:s0:c263,c871 Dec 10 22:49:24 crc restorecon[4680]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/containers/etcd-operator/8049f7cc not reset as customized by admin to system_u:object_r:container_file_t:s0:c263,c871 Dec 10 22:49:24 crc restorecon[4680]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/containers/etcd-operator/0cec5484 not reset as customized by admin to system_u:object_r:container_file_t:s0:c263,c871 Dec 10 22:49:24 crc restorecon[4680]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/containers/etcd-operator/312446d0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c406,c828 Dec 10 22:49:24 crc restorecon[4680]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/containers/etcd-operator/8e56a35d not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Dec 10 22:49:24 crc restorecon[4680]: /var/lib/kubelet/pods/1386a44e-36a2-460c-96d0-0359d2b6f0f5/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c108,c511 Dec 10 22:49:24 crc restorecon[4680]: /var/lib/kubelet/pods/1386a44e-36a2-460c-96d0-0359d2b6f0f5/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.133159589 not reset as customized by admin to system_u:object_r:container_file_t:s0:c108,c511 Dec 10 22:49:24 crc restorecon[4680]: /var/lib/kubelet/pods/1386a44e-36a2-460c-96d0-0359d2b6f0f5/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.133159589/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c108,c511 Dec 10 22:49:24 crc restorecon[4680]: /var/lib/kubelet/pods/1386a44e-36a2-460c-96d0-0359d2b6f0f5/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c108,c511 Dec 10 22:49:24 crc restorecon[4680]: /var/lib/kubelet/pods/1386a44e-36a2-460c-96d0-0359d2b6f0f5/volumes/kubernetes.io~configmap/config/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c108,c511 Dec 10 22:49:24 crc restorecon[4680]: /var/lib/kubelet/pods/1386a44e-36a2-460c-96d0-0359d2b6f0f5/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c108,c511 Dec 10 22:49:24 crc restorecon[4680]: /var/lib/kubelet/pods/1386a44e-36a2-460c-96d0-0359d2b6f0f5/containers/kube-controller-manager-operator/2d30ddb9 not reset as customized by admin to system_u:object_r:container_file_t:s0:c380,c909 Dec 10 22:49:24 crc restorecon[4680]: /var/lib/kubelet/pods/1386a44e-36a2-460c-96d0-0359d2b6f0f5/containers/kube-controller-manager-operator/eca8053d not reset as customized by admin to system_u:object_r:container_file_t:s0:c380,c909 Dec 10 22:49:24 crc restorecon[4680]: /var/lib/kubelet/pods/1386a44e-36a2-460c-96d0-0359d2b6f0f5/containers/kube-controller-manager-operator/c3a25c9a not reset as customized by admin to system_u:object_r:container_file_t:s0:c168,c522 Dec 10 22:49:24 crc restorecon[4680]: /var/lib/kubelet/pods/1386a44e-36a2-460c-96d0-0359d2b6f0f5/containers/kube-controller-manager-operator/b9609c22 not reset as customized by admin to system_u:object_r:container_file_t:s0:c108,c511 Dec 10 22:49:24 crc restorecon[4680]: /var/lib/kubelet/pods/96b93a3a-6083-4aea-8eab-fe1aa8245ad9/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c968,c969 Dec 10 22:49:24 crc restorecon[4680]: /var/lib/kubelet/pods/96b93a3a-6083-4aea-8eab-fe1aa8245ad9/containers/dns-operator/e8b0eca9 not reset as customized by admin to system_u:object_r:container_file_t:s0:c106,c418 Dec 10 22:49:24 crc restorecon[4680]: /var/lib/kubelet/pods/96b93a3a-6083-4aea-8eab-fe1aa8245ad9/containers/dns-operator/b36a9c3f not reset as customized by admin to system_u:object_r:container_file_t:s0:c529,c711 Dec 10 22:49:24 crc restorecon[4680]: /var/lib/kubelet/pods/96b93a3a-6083-4aea-8eab-fe1aa8245ad9/containers/dns-operator/38af7b07 not reset as customized by admin to system_u:object_r:container_file_t:s0:c968,c969 Dec 10 22:49:24 crc restorecon[4680]: /var/lib/kubelet/pods/96b93a3a-6083-4aea-8eab-fe1aa8245ad9/containers/kube-rbac-proxy/ae821620 not reset as customized by admin to system_u:object_r:container_file_t:s0:c106,c418 Dec 10 22:49:24 crc restorecon[4680]: /var/lib/kubelet/pods/96b93a3a-6083-4aea-8eab-fe1aa8245ad9/containers/kube-rbac-proxy/baa23338 not reset as customized by admin to system_u:object_r:container_file_t:s0:c529,c711 Dec 10 22:49:24 crc restorecon[4680]: /var/lib/kubelet/pods/96b93a3a-6083-4aea-8eab-fe1aa8245ad9/containers/kube-rbac-proxy/2c534809 not reset as customized by admin to system_u:object_r:container_file_t:s0:c968,c969 Dec 10 22:49:24 crc restorecon[4680]: /var/lib/kubelet/pods/496e6271-fb68-4057-954e-a0d97a4afa3f/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c661,c999 Dec 10 22:49:24 crc restorecon[4680]: /var/lib/kubelet/pods/496e6271-fb68-4057-954e-a0d97a4afa3f/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.3532625537 not reset as customized by admin to system_u:object_r:container_file_t:s0:c661,c999 Dec 10 22:49:24 crc restorecon[4680]: /var/lib/kubelet/pods/496e6271-fb68-4057-954e-a0d97a4afa3f/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.3532625537/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c661,c999 Dec 10 22:49:24 crc restorecon[4680]: /var/lib/kubelet/pods/496e6271-fb68-4057-954e-a0d97a4afa3f/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c661,c999 Dec 10 22:49:24 crc restorecon[4680]: /var/lib/kubelet/pods/496e6271-fb68-4057-954e-a0d97a4afa3f/volumes/kubernetes.io~configmap/config/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c661,c999 Dec 10 22:49:24 crc restorecon[4680]: /var/lib/kubelet/pods/496e6271-fb68-4057-954e-a0d97a4afa3f/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c661,c999 Dec 10 22:49:24 crc restorecon[4680]: /var/lib/kubelet/pods/496e6271-fb68-4057-954e-a0d97a4afa3f/containers/kube-scheduler-operator-container/59b29eae not reset as customized by admin to system_u:object_r:container_file_t:s0:c338,c381 Dec 10 22:49:24 crc restorecon[4680]: /var/lib/kubelet/pods/496e6271-fb68-4057-954e-a0d97a4afa3f/containers/kube-scheduler-operator-container/c91a8e4f not reset as customized by admin to system_u:object_r:container_file_t:s0:c338,c381 Dec 10 22:49:24 crc restorecon[4680]: /var/lib/kubelet/pods/496e6271-fb68-4057-954e-a0d97a4afa3f/containers/kube-scheduler-operator-container/4d87494a not reset as customized by admin to system_u:object_r:container_file_t:s0:c442,c857 Dec 10 22:49:24 crc restorecon[4680]: /var/lib/kubelet/pods/496e6271-fb68-4057-954e-a0d97a4afa3f/containers/kube-scheduler-operator-container/1e33ca63 not reset as customized by admin to system_u:object_r:container_file_t:s0:c661,c999 Dec 10 22:49:24 crc restorecon[4680]: /var/lib/kubelet/pods/3ab1a177-2de0-46d9-b765-d0d0649bb42e/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Dec 10 22:49:24 crc restorecon[4680]: /var/lib/kubelet/pods/3ab1a177-2de0-46d9-b765-d0d0649bb42e/containers/kube-rbac-proxy/8dea7be2 not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Dec 10 22:49:24 crc restorecon[4680]: /var/lib/kubelet/pods/3ab1a177-2de0-46d9-b765-d0d0649bb42e/containers/kube-rbac-proxy/d0b04a99 not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Dec 10 22:49:24 crc restorecon[4680]: /var/lib/kubelet/pods/3ab1a177-2de0-46d9-b765-d0d0649bb42e/containers/kube-rbac-proxy/d84f01e7 not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Dec 10 22:49:24 crc restorecon[4680]: /var/lib/kubelet/pods/3ab1a177-2de0-46d9-b765-d0d0649bb42e/containers/package-server-manager/4109059b not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Dec 10 22:49:24 crc restorecon[4680]: /var/lib/kubelet/pods/3ab1a177-2de0-46d9-b765-d0d0649bb42e/containers/package-server-manager/a7258a3e not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Dec 10 22:49:24 crc restorecon[4680]: /var/lib/kubelet/pods/3ab1a177-2de0-46d9-b765-d0d0649bb42e/containers/package-server-manager/05bdf2b6 not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Dec 10 22:49:24 crc restorecon[4680]: /var/lib/kubelet/pods/6731426b-95fe-49ff-bb5f-40441049fde2/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Dec 10 22:49:24 crc restorecon[4680]: /var/lib/kubelet/pods/6731426b-95fe-49ff-bb5f-40441049fde2/containers/control-plane-machine-set-operator/f3261b51 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Dec 10 22:49:24 crc restorecon[4680]: /var/lib/kubelet/pods/6731426b-95fe-49ff-bb5f-40441049fde2/containers/control-plane-machine-set-operator/315d045e not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Dec 10 22:49:24 crc restorecon[4680]: /var/lib/kubelet/pods/6731426b-95fe-49ff-bb5f-40441049fde2/containers/control-plane-machine-set-operator/5fdcf278 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Dec 10 22:49:24 crc restorecon[4680]: /var/lib/kubelet/pods/6731426b-95fe-49ff-bb5f-40441049fde2/containers/control-plane-machine-set-operator/d053f757 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Dec 10 22:49:24 crc restorecon[4680]: /var/lib/kubelet/pods/6731426b-95fe-49ff-bb5f-40441049fde2/containers/control-plane-machine-set-operator/c2850dc7 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Dec 10 22:49:24 crc restorecon[4680]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/volumes/kubernetes.io~configmap/marketplace-trusted-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 22:49:24 crc restorecon[4680]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/volumes/kubernetes.io~configmap/marketplace-trusted-ca/..2025_02_23_05_22_30.2390596521 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 22:49:24 crc restorecon[4680]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/volumes/kubernetes.io~configmap/marketplace-trusted-ca/..2025_02_23_05_22_30.2390596521/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 22:49:24 crc restorecon[4680]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/volumes/kubernetes.io~configmap/marketplace-trusted-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 22:49:24 crc restorecon[4680]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/volumes/kubernetes.io~configmap/marketplace-trusted-ca/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 22:49:24 crc restorecon[4680]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 22:49:24 crc restorecon[4680]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/containers/marketplace-operator/fcfb0b2b not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 22:49:24 crc restorecon[4680]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/containers/marketplace-operator/c7ac9b7d not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 22:49:24 crc restorecon[4680]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/containers/marketplace-operator/fa0c0d52 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 22:49:24 crc restorecon[4680]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/containers/marketplace-operator/c609b6ba not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 22:49:24 crc restorecon[4680]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/containers/marketplace-operator/2be6c296 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 22:49:24 crc restorecon[4680]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/containers/marketplace-operator/89a32653 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 22:49:24 crc restorecon[4680]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/containers/marketplace-operator/4eb9afeb not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 22:49:24 crc restorecon[4680]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/containers/marketplace-operator/13af6efa not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 22:49:24 crc restorecon[4680]: /var/lib/kubelet/pods/b6312bbd-5731-4ea0-a20f-81d5a57df44a/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Dec 10 22:49:24 crc restorecon[4680]: /var/lib/kubelet/pods/b6312bbd-5731-4ea0-a20f-81d5a57df44a/containers/olm-operator/b03f9724 not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Dec 10 22:49:24 crc restorecon[4680]: /var/lib/kubelet/pods/b6312bbd-5731-4ea0-a20f-81d5a57df44a/containers/olm-operator/e3d105cc not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Dec 10 22:49:24 crc restorecon[4680]: /var/lib/kubelet/pods/b6312bbd-5731-4ea0-a20f-81d5a57df44a/containers/olm-operator/3aed4d83 not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Dec 10 22:49:24 crc restorecon[4680]: /var/lib/kubelet/pods/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c18 Dec 10 22:49:24 crc restorecon[4680]: /var/lib/kubelet/pods/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.1906041176 not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c18 Dec 10 22:49:24 crc restorecon[4680]: /var/lib/kubelet/pods/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.1906041176/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c18 Dec 10 22:49:24 crc restorecon[4680]: /var/lib/kubelet/pods/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c18 Dec 10 22:49:24 crc restorecon[4680]: /var/lib/kubelet/pods/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c/volumes/kubernetes.io~configmap/config/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c18 Dec 10 22:49:24 crc restorecon[4680]: /var/lib/kubelet/pods/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c18 Dec 10 22:49:24 crc restorecon[4680]: /var/lib/kubelet/pods/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c/containers/kube-storage-version-migrator-operator/0765fa6e not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c18 Dec 10 22:49:24 crc restorecon[4680]: /var/lib/kubelet/pods/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c/containers/kube-storage-version-migrator-operator/2cefc627 not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c18 Dec 10 22:49:24 crc restorecon[4680]: /var/lib/kubelet/pods/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c/containers/kube-storage-version-migrator-operator/3dcc6345 not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c18 Dec 10 22:49:24 crc restorecon[4680]: /var/lib/kubelet/pods/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c/containers/kube-storage-version-migrator-operator/365af391 not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c18 Dec 10 22:49:24 crc restorecon[4680]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/volumes/kubernetes.io~empty-dir/available-featuregates not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Dec 10 22:49:24 crc restorecon[4680]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/volumes/kubernetes.io~empty-dir/available-featuregates/featureGate-SelfManagedHA-Default.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Dec 10 22:49:24 crc restorecon[4680]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/volumes/kubernetes.io~empty-dir/available-featuregates/featureGate-SelfManagedHA-TechPreviewNoUpgrade.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Dec 10 22:49:24 crc restorecon[4680]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/volumes/kubernetes.io~empty-dir/available-featuregates/featureGate-SelfManagedHA-DevPreviewNoUpgrade.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Dec 10 22:49:24 crc restorecon[4680]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/volumes/kubernetes.io~empty-dir/available-featuregates/featureGate-Hypershift-TechPreviewNoUpgrade.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Dec 10 22:49:24 crc restorecon[4680]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/volumes/kubernetes.io~empty-dir/available-featuregates/featureGate-Hypershift-DevPreviewNoUpgrade.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Dec 10 22:49:24 crc restorecon[4680]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/volumes/kubernetes.io~empty-dir/available-featuregates/featureGate-Hypershift-Default.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Dec 10 22:49:24 crc restorecon[4680]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Dec 10 22:49:24 crc restorecon[4680]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/containers/openshift-api/b1130c0f not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Dec 10 22:49:24 crc restorecon[4680]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/containers/openshift-api/236a5913 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Dec 10 22:49:24 crc restorecon[4680]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/containers/openshift-api/b9432e26 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Dec 10 22:49:24 crc restorecon[4680]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/containers/openshift-config-operator/5ddb0e3f not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Dec 10 22:49:24 crc restorecon[4680]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/containers/openshift-config-operator/986dc4fd not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Dec 10 22:49:24 crc restorecon[4680]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/containers/openshift-config-operator/8a23ff9a not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Dec 10 22:49:24 crc restorecon[4680]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/containers/openshift-config-operator/9728ae68 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Dec 10 22:49:24 crc restorecon[4680]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/containers/openshift-config-operator/665f31d0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Dec 10 22:49:24 crc restorecon[4680]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Dec 10 22:49:24 crc restorecon[4680]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.1255385357 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Dec 10 22:49:24 crc restorecon[4680]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.1255385357/operator-config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Dec 10 22:49:24 crc restorecon[4680]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Dec 10 22:49:24 crc restorecon[4680]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/config/operator-config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Dec 10 22:49:24 crc restorecon[4680]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/service-ca-bundle not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Dec 10 22:49:24 crc restorecon[4680]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/service-ca-bundle/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Dec 10 22:49:24 crc restorecon[4680]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/service-ca-bundle/..2025_02_23_05_23_57.573792656 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Dec 10 22:49:24 crc restorecon[4680]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/service-ca-bundle/..2025_02_23_05_23_57.573792656/service-ca.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Dec 10 22:49:24 crc restorecon[4680]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/service-ca-bundle/service-ca.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Dec 10 22:49:24 crc restorecon[4680]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/trusted-ca-bundle not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Dec 10 22:49:24 crc restorecon[4680]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/trusted-ca-bundle/..2025_02_23_05_22_30.3254245399 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Dec 10 22:49:24 crc restorecon[4680]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/trusted-ca-bundle/..2025_02_23_05_22_30.3254245399/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Dec 10 22:49:24 crc restorecon[4680]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/trusted-ca-bundle/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Dec 10 22:49:24 crc restorecon[4680]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/trusted-ca-bundle/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Dec 10 22:49:24 crc restorecon[4680]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Dec 10 22:49:24 crc restorecon[4680]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/containers/authentication-operator/136c9b42 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Dec 10 22:49:24 crc restorecon[4680]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/containers/authentication-operator/98a1575b not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Dec 10 22:49:24 crc restorecon[4680]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/containers/authentication-operator/cac69136 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Dec 10 22:49:24 crc restorecon[4680]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/containers/authentication-operator/5deb77a7 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Dec 10 22:49:24 crc restorecon[4680]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/containers/authentication-operator/2ae53400 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Dec 10 22:49:24 crc restorecon[4680]: /var/lib/kubelet/pods/01ab3dd5-8196-46d0-ad33-122e2ca51def/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c16 Dec 10 22:49:24 crc restorecon[4680]: /var/lib/kubelet/pods/01ab3dd5-8196-46d0-ad33-122e2ca51def/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.3608339744 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c16 Dec 10 22:49:24 crc restorecon[4680]: /var/lib/kubelet/pods/01ab3dd5-8196-46d0-ad33-122e2ca51def/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.3608339744/operator-config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c16 Dec 10 22:49:24 crc restorecon[4680]: /var/lib/kubelet/pods/01ab3dd5-8196-46d0-ad33-122e2ca51def/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c16 Dec 10 22:49:24 crc restorecon[4680]: /var/lib/kubelet/pods/01ab3dd5-8196-46d0-ad33-122e2ca51def/volumes/kubernetes.io~configmap/config/operator-config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c16 Dec 10 22:49:24 crc restorecon[4680]: /var/lib/kubelet/pods/01ab3dd5-8196-46d0-ad33-122e2ca51def/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c16 Dec 10 22:49:24 crc restorecon[4680]: /var/lib/kubelet/pods/01ab3dd5-8196-46d0-ad33-122e2ca51def/containers/service-ca-operator/e46f2326 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c16 Dec 10 22:49:24 crc restorecon[4680]: /var/lib/kubelet/pods/01ab3dd5-8196-46d0-ad33-122e2ca51def/containers/service-ca-operator/dc688d3c not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c16 Dec 10 22:49:24 crc restorecon[4680]: /var/lib/kubelet/pods/01ab3dd5-8196-46d0-ad33-122e2ca51def/containers/service-ca-operator/3497c3cd not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c16 Dec 10 22:49:24 crc restorecon[4680]: /var/lib/kubelet/pods/01ab3dd5-8196-46d0-ad33-122e2ca51def/containers/service-ca-operator/177eb008 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c16 Dec 10 22:49:24 crc restorecon[4680]: /var/lib/kubelet/pods/7539238d-5fe0-46ed-884e-1c3b566537ec/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c13 Dec 10 22:49:24 crc restorecon[4680]: /var/lib/kubelet/pods/7539238d-5fe0-46ed-884e-1c3b566537ec/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.3819292994 not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c13 Dec 10 22:49:24 crc restorecon[4680]: /var/lib/kubelet/pods/7539238d-5fe0-46ed-884e-1c3b566537ec/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.3819292994/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c13 Dec 10 22:49:24 crc restorecon[4680]: /var/lib/kubelet/pods/7539238d-5fe0-46ed-884e-1c3b566537ec/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c13 Dec 10 22:49:24 crc restorecon[4680]: /var/lib/kubelet/pods/7539238d-5fe0-46ed-884e-1c3b566537ec/volumes/kubernetes.io~configmap/config/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c13 Dec 10 22:49:24 crc restorecon[4680]: /var/lib/kubelet/pods/7539238d-5fe0-46ed-884e-1c3b566537ec/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c13 Dec 10 22:49:24 crc restorecon[4680]: /var/lib/kubelet/pods/7539238d-5fe0-46ed-884e-1c3b566537ec/containers/openshift-apiserver-operator/af5a2afa not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c13 Dec 10 22:49:24 crc restorecon[4680]: /var/lib/kubelet/pods/7539238d-5fe0-46ed-884e-1c3b566537ec/containers/openshift-apiserver-operator/d780cb1f not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c13 Dec 10 22:49:24 crc restorecon[4680]: /var/lib/kubelet/pods/7539238d-5fe0-46ed-884e-1c3b566537ec/containers/openshift-apiserver-operator/49b0f374 not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c13 Dec 10 22:49:24 crc restorecon[4680]: /var/lib/kubelet/pods/7539238d-5fe0-46ed-884e-1c3b566537ec/containers/openshift-apiserver-operator/26fbb125 not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c13 Dec 10 22:49:24 crc restorecon[4680]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/volumes/kubernetes.io~configmap/trusted-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Dec 10 22:49:24 crc restorecon[4680]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/volumes/kubernetes.io~configmap/trusted-ca/..2025_02_23_05_22_30.3244779536 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Dec 10 22:49:24 crc restorecon[4680]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/volumes/kubernetes.io~configmap/trusted-ca/..2025_02_23_05_22_30.3244779536/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Dec 10 22:49:24 crc restorecon[4680]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/volumes/kubernetes.io~configmap/trusted-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Dec 10 22:49:24 crc restorecon[4680]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/volumes/kubernetes.io~configmap/trusted-ca/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Dec 10 22:49:24 crc restorecon[4680]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Dec 10 22:49:24 crc restorecon[4680]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/containers/ingress-operator/cf14125a not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Dec 10 22:49:24 crc restorecon[4680]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/containers/ingress-operator/b7f86972 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Dec 10 22:49:24 crc restorecon[4680]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/containers/ingress-operator/e51d739c not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Dec 10 22:49:24 crc restorecon[4680]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/containers/ingress-operator/88ba6a69 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Dec 10 22:49:24 crc restorecon[4680]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/containers/ingress-operator/669a9acf not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Dec 10 22:49:24 crc restorecon[4680]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/containers/ingress-operator/5cd51231 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Dec 10 22:49:24 crc restorecon[4680]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/containers/ingress-operator/75349ec7 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Dec 10 22:49:24 crc restorecon[4680]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/containers/ingress-operator/15c26839 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Dec 10 22:49:24 crc restorecon[4680]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/containers/ingress-operator/45023dcd not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Dec 10 22:49:24 crc restorecon[4680]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/containers/ingress-operator/2bb66a50 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Dec 10 22:49:24 crc restorecon[4680]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/containers/kube-rbac-proxy/64d03bdd not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Dec 10 22:49:24 crc restorecon[4680]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/containers/kube-rbac-proxy/ab8e7ca0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Dec 10 22:49:24 crc restorecon[4680]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/containers/kube-rbac-proxy/bb9be25f not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Dec 10 22:49:24 crc restorecon[4680]: /var/lib/kubelet/pods/bf126b07-da06-4140-9a57-dfd54fc6b486/volumes/kubernetes.io~configmap/trusted-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 10 22:49:24 crc restorecon[4680]: /var/lib/kubelet/pods/bf126b07-da06-4140-9a57-dfd54fc6b486/volumes/kubernetes.io~configmap/trusted-ca/..2025_02_23_05_22_30.2034221258 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 10 22:49:24 crc restorecon[4680]: /var/lib/kubelet/pods/bf126b07-da06-4140-9a57-dfd54fc6b486/volumes/kubernetes.io~configmap/trusted-ca/..2025_02_23_05_22_30.2034221258/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 10 22:49:24 crc restorecon[4680]: /var/lib/kubelet/pods/bf126b07-da06-4140-9a57-dfd54fc6b486/volumes/kubernetes.io~configmap/trusted-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 10 22:49:24 crc restorecon[4680]: /var/lib/kubelet/pods/bf126b07-da06-4140-9a57-dfd54fc6b486/volumes/kubernetes.io~configmap/trusted-ca/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 10 22:49:24 crc restorecon[4680]: /var/lib/kubelet/pods/bf126b07-da06-4140-9a57-dfd54fc6b486/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 10 22:49:24 crc restorecon[4680]: /var/lib/kubelet/pods/bf126b07-da06-4140-9a57-dfd54fc6b486/containers/cluster-image-registry-operator/9a0b61d3 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 10 22:49:24 crc restorecon[4680]: /var/lib/kubelet/pods/bf126b07-da06-4140-9a57-dfd54fc6b486/containers/cluster-image-registry-operator/d471b9d2 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 10 22:49:24 crc restorecon[4680]: /var/lib/kubelet/pods/bf126b07-da06-4140-9a57-dfd54fc6b486/containers/cluster-image-registry-operator/8cb76b8e not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 10 22:49:24 crc restorecon[4680]: /var/lib/kubelet/pods/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Dec 10 22:49:24 crc restorecon[4680]: /var/lib/kubelet/pods/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9/containers/catalog-operator/11a00840 not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Dec 10 22:49:24 crc restorecon[4680]: /var/lib/kubelet/pods/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9/containers/catalog-operator/ec355a92 not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Dec 10 22:49:24 crc restorecon[4680]: /var/lib/kubelet/pods/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9/containers/catalog-operator/992f735e not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Dec 10 22:49:24 crc restorecon[4680]: /var/lib/kubelet/pods/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c14 Dec 10 22:49:24 crc restorecon[4680]: /var/lib/kubelet/pods/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.1782968797 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c14 Dec 10 22:49:24 crc restorecon[4680]: /var/lib/kubelet/pods/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.1782968797/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c14 Dec 10 22:49:24 crc restorecon[4680]: /var/lib/kubelet/pods/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c14 Dec 10 22:49:24 crc restorecon[4680]: /var/lib/kubelet/pods/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b/volumes/kubernetes.io~configmap/config/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c14 Dec 10 22:49:24 crc restorecon[4680]: /var/lib/kubelet/pods/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c14 Dec 10 22:49:24 crc restorecon[4680]: /var/lib/kubelet/pods/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b/containers/openshift-controller-manager-operator/d59cdbbc not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c14 Dec 10 22:49:24 crc restorecon[4680]: /var/lib/kubelet/pods/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b/containers/openshift-controller-manager-operator/72133ff0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c14 Dec 10 22:49:24 crc restorecon[4680]: /var/lib/kubelet/pods/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b/containers/openshift-controller-manager-operator/c56c834c not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c14 Dec 10 22:49:24 crc restorecon[4680]: /var/lib/kubelet/pods/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b/containers/openshift-controller-manager-operator/d13724c7 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c14 Dec 10 22:49:24 crc restorecon[4680]: /var/lib/kubelet/pods/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b/containers/openshift-controller-manager-operator/0a498258 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c14 Dec 10 22:49:24 crc restorecon[4680]: /var/lib/kubelet/pods/5fe579f8-e8a6-4643-bce5-a661393c4dde/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Dec 10 22:49:24 crc restorecon[4680]: /var/lib/kubelet/pods/5fe579f8-e8a6-4643-bce5-a661393c4dde/containers/machine-config-server/fa471982 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Dec 10 22:49:24 crc restorecon[4680]: /var/lib/kubelet/pods/5fe579f8-e8a6-4643-bce5-a661393c4dde/containers/machine-config-server/fc900d92 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Dec 10 22:49:24 crc restorecon[4680]: /var/lib/kubelet/pods/5fe579f8-e8a6-4643-bce5-a661393c4dde/containers/machine-config-server/fa7d68da not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Dec 10 22:49:24 crc restorecon[4680]: /var/lib/kubelet/pods/49ef4625-1d3a-4a9f-b595-c2433d32326d/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c22 Dec 10 22:49:24 crc restorecon[4680]: /var/lib/kubelet/pods/49ef4625-1d3a-4a9f-b595-c2433d32326d/containers/migrator/4bacf9b4 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c22 Dec 10 22:49:24 crc restorecon[4680]: /var/lib/kubelet/pods/49ef4625-1d3a-4a9f-b595-c2433d32326d/containers/migrator/424021b1 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c22 Dec 10 22:49:24 crc restorecon[4680]: /var/lib/kubelet/pods/49ef4625-1d3a-4a9f-b595-c2433d32326d/containers/migrator/fc2e31a3 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c22 Dec 10 22:49:24 crc restorecon[4680]: /var/lib/kubelet/pods/49ef4625-1d3a-4a9f-b595-c2433d32326d/containers/graceful-termination/f51eefac not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c22 Dec 10 22:49:24 crc restorecon[4680]: /var/lib/kubelet/pods/49ef4625-1d3a-4a9f-b595-c2433d32326d/containers/graceful-termination/c8997f2f not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c22 Dec 10 22:49:24 crc restorecon[4680]: /var/lib/kubelet/pods/49ef4625-1d3a-4a9f-b595-c2433d32326d/containers/graceful-termination/7481f599 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c22 Dec 10 22:49:24 crc restorecon[4680]: /var/lib/kubelet/pods/25e176fe-21b4-4974-b1ed-c8b94f112a7f/volumes/kubernetes.io~configmap/signing-cabundle not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c22 Dec 10 22:49:24 crc restorecon[4680]: /var/lib/kubelet/pods/25e176fe-21b4-4974-b1ed-c8b94f112a7f/volumes/kubernetes.io~configmap/signing-cabundle/..2025_02_23_05_22_49.2255460704 not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c22 Dec 10 22:49:24 crc restorecon[4680]: /var/lib/kubelet/pods/25e176fe-21b4-4974-b1ed-c8b94f112a7f/volumes/kubernetes.io~configmap/signing-cabundle/..2025_02_23_05_22_49.2255460704/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c22 Dec 10 22:49:24 crc restorecon[4680]: /var/lib/kubelet/pods/25e176fe-21b4-4974-b1ed-c8b94f112a7f/volumes/kubernetes.io~configmap/signing-cabundle/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c22 Dec 10 22:49:24 crc restorecon[4680]: /var/lib/kubelet/pods/25e176fe-21b4-4974-b1ed-c8b94f112a7f/volumes/kubernetes.io~configmap/signing-cabundle/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c22 Dec 10 22:49:24 crc restorecon[4680]: /var/lib/kubelet/pods/25e176fe-21b4-4974-b1ed-c8b94f112a7f/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c22 Dec 10 22:49:24 crc restorecon[4680]: /var/lib/kubelet/pods/25e176fe-21b4-4974-b1ed-c8b94f112a7f/containers/service-ca-controller/fdafea19 not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c22 Dec 10 22:49:24 crc restorecon[4680]: /var/lib/kubelet/pods/25e176fe-21b4-4974-b1ed-c8b94f112a7f/containers/service-ca-controller/d0e1c571 not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c22 Dec 10 22:49:24 crc restorecon[4680]: /var/lib/kubelet/pods/25e176fe-21b4-4974-b1ed-c8b94f112a7f/containers/service-ca-controller/ee398915 not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c22 Dec 10 22:49:24 crc restorecon[4680]: /var/lib/kubelet/pods/25e176fe-21b4-4974-b1ed-c8b94f112a7f/containers/service-ca-controller/682bb6b8 not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c22 Dec 10 22:49:24 crc restorecon[4680]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c666,c920 Dec 10 22:49:24 crc restorecon[4680]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/setup/a3e67855 not reset as customized by admin to system_u:object_r:container_file_t:s0:c294,c884 Dec 10 22:49:24 crc restorecon[4680]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/setup/a989f289 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c1016 Dec 10 22:49:24 crc restorecon[4680]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/setup/915431bd not reset as customized by admin to system_u:object_r:container_file_t:s0:c666,c920 Dec 10 22:49:24 crc restorecon[4680]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-ensure-env-vars/7796fdab not reset as customized by admin to system_u:object_r:container_file_t:s0:c294,c884 Dec 10 22:49:24 crc restorecon[4680]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-ensure-env-vars/dcdb5f19 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c1016 Dec 10 22:49:24 crc restorecon[4680]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-ensure-env-vars/a3aaa88c not reset as customized by admin to system_u:object_r:container_file_t:s0:c666,c920 Dec 10 22:49:24 crc restorecon[4680]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-resources-copy/5508e3e6 not reset as customized by admin to system_u:object_r:container_file_t:s0:c294,c884 Dec 10 22:49:24 crc restorecon[4680]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-resources-copy/160585de not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c1016 Dec 10 22:49:24 crc restorecon[4680]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-resources-copy/e99f8da3 not reset as customized by admin to system_u:object_r:container_file_t:s0:c666,c920 Dec 10 22:49:24 crc restorecon[4680]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcdctl/8bc85570 not reset as customized by admin to system_u:object_r:container_file_t:s0:c294,c884 Dec 10 22:49:24 crc restorecon[4680]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcdctl/a5861c91 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c1016 Dec 10 22:49:24 crc restorecon[4680]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcdctl/84db1135 not reset as customized by admin to system_u:object_r:container_file_t:s0:c666,c920 Dec 10 22:49:24 crc restorecon[4680]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd/9e1a6043 not reset as customized by admin to system_u:object_r:container_file_t:s0:c294,c884 Dec 10 22:49:24 crc restorecon[4680]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd/c1aba1c2 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c1016 Dec 10 22:49:24 crc restorecon[4680]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd/d55ccd6d not reset as customized by admin to system_u:object_r:container_file_t:s0:c666,c920 Dec 10 22:49:24 crc restorecon[4680]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-metrics/971cc9f6 not reset as customized by admin to system_u:object_r:container_file_t:s0:c294,c884 Dec 10 22:49:24 crc restorecon[4680]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-metrics/8f2e3dcf not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c1016 Dec 10 22:49:24 crc restorecon[4680]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-metrics/ceb35e9c not reset as customized by admin to system_u:object_r:container_file_t:s0:c666,c920 Dec 10 22:49:24 crc restorecon[4680]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-readyz/1c192745 not reset as customized by admin to system_u:object_r:container_file_t:s0:c294,c884 Dec 10 22:49:24 crc restorecon[4680]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-readyz/5209e501 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c1016 Dec 10 22:49:24 crc restorecon[4680]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-readyz/f83de4df not reset as customized by admin to system_u:object_r:container_file_t:s0:c666,c920 Dec 10 22:49:24 crc restorecon[4680]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-rev/e7b978ac not reset as customized by admin to system_u:object_r:container_file_t:s0:c294,c884 Dec 10 22:49:24 crc restorecon[4680]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-rev/c64304a1 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c1016 Dec 10 22:49:24 crc restorecon[4680]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-rev/5384386b not reset as customized by admin to system_u:object_r:container_file_t:s0:c666,c920 Dec 10 22:49:24 crc restorecon[4680]: /var/lib/kubelet/pods/efdd0498-1daa-4136-9a4a-3b948c2293fc/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c268,c620 Dec 10 22:49:24 crc restorecon[4680]: /var/lib/kubelet/pods/efdd0498-1daa-4136-9a4a-3b948c2293fc/containers/multus-admission-controller/cce3e3ff not reset as customized by admin to system_u:object_r:container_file_t:s0:c435,c756 Dec 10 22:49:24 crc restorecon[4680]: /var/lib/kubelet/pods/efdd0498-1daa-4136-9a4a-3b948c2293fc/containers/multus-admission-controller/8fb75465 not reset as customized by admin to system_u:object_r:container_file_t:s0:c268,c620 Dec 10 22:49:24 crc restorecon[4680]: /var/lib/kubelet/pods/efdd0498-1daa-4136-9a4a-3b948c2293fc/containers/kube-rbac-proxy/740f573e not reset as customized by admin to system_u:object_r:container_file_t:s0:c435,c756 Dec 10 22:49:24 crc restorecon[4680]: /var/lib/kubelet/pods/efdd0498-1daa-4136-9a4a-3b948c2293fc/containers/kube-rbac-proxy/32fd1134 not reset as customized by admin to system_u:object_r:container_file_t:s0:c268,c620 Dec 10 22:49:24 crc restorecon[4680]: /var/lib/kubelet/pods/20b0d48f-5fd6-431c-a545-e3c800c7b866/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c24 Dec 10 22:49:24 crc restorecon[4680]: /var/lib/kubelet/pods/20b0d48f-5fd6-431c-a545-e3c800c7b866/containers/serve-healthcheck-canary/0a861bd3 not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c24 Dec 10 22:49:24 crc restorecon[4680]: /var/lib/kubelet/pods/20b0d48f-5fd6-431c-a545-e3c800c7b866/containers/serve-healthcheck-canary/80363026 not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c24 Dec 10 22:49:24 crc restorecon[4680]: /var/lib/kubelet/pods/20b0d48f-5fd6-431c-a545-e3c800c7b866/containers/serve-healthcheck-canary/bfa952a8 not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c24 Dec 10 22:49:24 crc restorecon[4680]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/volumes/kubernetes.io~configmap/auth-proxy-config not reset as customized by admin to system_u:object_r:container_file_t:s0:c129,c158 Dec 10 22:49:24 crc restorecon[4680]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/volumes/kubernetes.io~configmap/auth-proxy-config/..2025_02_23_05_33_31.2122464563 not reset as customized by admin to system_u:object_r:container_file_t:s0:c129,c158 Dec 10 22:49:24 crc restorecon[4680]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/volumes/kubernetes.io~configmap/auth-proxy-config/..2025_02_23_05_33_31.2122464563/config-file.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c129,c158 Dec 10 22:49:24 crc restorecon[4680]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/volumes/kubernetes.io~configmap/auth-proxy-config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c129,c158 Dec 10 22:49:24 crc restorecon[4680]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/volumes/kubernetes.io~configmap/auth-proxy-config/config-file.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c129,c158 Dec 10 22:49:24 crc restorecon[4680]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c129,c158 Dec 10 22:49:24 crc restorecon[4680]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/volumes/kubernetes.io~configmap/config/..2025_02_23_05_33_31.333075221 not reset as customized by admin to system_u:object_r:container_file_t:s0:c129,c158 Dec 10 22:49:24 crc restorecon[4680]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c129,c158 Dec 10 22:49:24 crc restorecon[4680]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c129,c158 Dec 10 22:49:24 crc restorecon[4680]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/containers/kube-rbac-proxy/793bf43d not reset as customized by admin to system_u:object_r:container_file_t:s0:c381,c387 Dec 10 22:49:24 crc restorecon[4680]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/containers/kube-rbac-proxy/7db1bb6e not reset as customized by admin to system_u:object_r:container_file_t:s0:c142,c438 Dec 10 22:49:24 crc restorecon[4680]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/containers/kube-rbac-proxy/4f6a0368 not reset as customized by admin to system_u:object_r:container_file_t:s0:c129,c158 Dec 10 22:49:24 crc restorecon[4680]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/containers/machine-approver-controller/c12c7d86 not reset as customized by admin to system_u:object_r:container_file_t:s0:c381,c387 Dec 10 22:49:24 crc restorecon[4680]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/containers/machine-approver-controller/36c4a773 not reset as customized by admin to system_u:object_r:container_file_t:s0:c142,c438 Dec 10 22:49:24 crc restorecon[4680]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/containers/machine-approver-controller/4c1e98ae not reset as customized by admin to system_u:object_r:container_file_t:s0:c142,c438 Dec 10 22:49:24 crc restorecon[4680]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/containers/machine-approver-controller/a4c8115c not reset as customized by admin to system_u:object_r:container_file_t:s0:c129,c158 Dec 10 22:49:24 crc restorecon[4680]: /var/lib/kubelet/pods/f4b27818a5e8e43d0dc095d08835c792/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c97,c980 Dec 10 22:49:24 crc restorecon[4680]: /var/lib/kubelet/pods/f4b27818a5e8e43d0dc095d08835c792/containers/setup/7db1802e not reset as customized by admin to system_u:object_r:container_file_t:s0:c97,c980 Dec 10 22:49:24 crc restorecon[4680]: /var/lib/kubelet/pods/f4b27818a5e8e43d0dc095d08835c792/containers/kube-apiserver/a008a7ab not reset as customized by admin to system_u:object_r:container_file_t:s0:c97,c980 Dec 10 22:49:24 crc restorecon[4680]: /var/lib/kubelet/pods/f4b27818a5e8e43d0dc095d08835c792/containers/kube-apiserver-cert-syncer/2c836bac not reset as customized by admin to system_u:object_r:container_file_t:s0:c97,c980 Dec 10 22:49:24 crc restorecon[4680]: /var/lib/kubelet/pods/f4b27818a5e8e43d0dc095d08835c792/containers/kube-apiserver-cert-regeneration-controller/0ce62299 not reset as customized by admin to system_u:object_r:container_file_t:s0:c97,c980 Dec 10 22:49:24 crc restorecon[4680]: /var/lib/kubelet/pods/f4b27818a5e8e43d0dc095d08835c792/containers/kube-apiserver-insecure-readyz/945d2457 not reset as customized by admin to system_u:object_r:container_file_t:s0:c97,c980 Dec 10 22:49:24 crc restorecon[4680]: /var/lib/kubelet/pods/f4b27818a5e8e43d0dc095d08835c792/containers/kube-apiserver-check-endpoints/7d5c1dd8 not reset as customized by admin to system_u:object_r:container_file_t:s0:c97,c980 Dec 10 22:49:24 crc restorecon[4680]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/utilities not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 22:49:24 crc restorecon[4680]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/utilities/copy-content not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 22:49:24 crc restorecon[4680]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 22:49:24 crc restorecon[4680]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 22:49:24 crc restorecon[4680]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/3scale-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 22:49:24 crc restorecon[4680]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/3scale-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 22:49:24 crc restorecon[4680]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/advanced-cluster-management not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 22:49:24 crc restorecon[4680]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/advanced-cluster-management/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 22:49:24 crc restorecon[4680]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/amq-broker-rhel8 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 22:49:24 crc restorecon[4680]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/amq-broker-rhel8/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 22:49:24 crc restorecon[4680]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/amq-online not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 22:49:24 crc restorecon[4680]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/amq-online/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 22:49:24 crc restorecon[4680]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/amq-streams not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 22:49:24 crc restorecon[4680]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/amq-streams/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 22:49:24 crc restorecon[4680]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/amq-streams-console not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 22:49:24 crc restorecon[4680]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/amq-streams-console/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 22:49:24 crc restorecon[4680]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/amq7-interconnect-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 22:49:24 crc restorecon[4680]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/amq7-interconnect-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 22:49:24 crc restorecon[4680]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ansible-automation-platform-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 22:49:24 crc restorecon[4680]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ansible-automation-platform-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 22:49:24 crc restorecon[4680]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ansible-cloud-addons-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 22:49:24 crc restorecon[4680]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ansible-cloud-addons-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 22:49:24 crc restorecon[4680]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apicast-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 22:49:24 crc restorecon[4680]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apicast-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 22:49:24 crc restorecon[4680]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apicurio-registry-3 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 22:49:24 crc restorecon[4680]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apicurio-registry-3/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 22:49:24 crc restorecon[4680]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/authorino-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 22:49:24 crc restorecon[4680]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/authorino-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 22:49:24 crc restorecon[4680]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aws-load-balancer-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 22:49:24 crc restorecon[4680]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aws-load-balancer-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 22:49:24 crc restorecon[4680]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/bamoe-businessautomation-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 22:49:24 crc restorecon[4680]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/bamoe-businessautomation-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 22:49:24 crc restorecon[4680]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/bamoe-kogito-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 22:49:24 crc restorecon[4680]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/bamoe-kogito-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 22:49:24 crc restorecon[4680]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/bpfman-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 22:49:24 crc restorecon[4680]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/bpfman-operator/index.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 22:49:24 crc restorecon[4680]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/businessautomation-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 22:49:24 crc restorecon[4680]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/businessautomation-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 22:49:24 crc restorecon[4680]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cephcsi-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 22:49:24 crc restorecon[4680]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cephcsi-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 22:49:24 crc restorecon[4680]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cincinnati-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 22:49:24 crc restorecon[4680]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cincinnati-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 22:49:24 crc restorecon[4680]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cluster-kube-descheduler-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 22:49:24 crc restorecon[4680]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cluster-kube-descheduler-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 22:49:24 crc restorecon[4680]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cluster-logging not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 22:49:24 crc restorecon[4680]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cluster-logging/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 22:49:24 crc restorecon[4680]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cluster-observability-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 22:49:24 crc restorecon[4680]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cluster-observability-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 22:49:24 crc restorecon[4680]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/compliance-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 22:49:24 crc restorecon[4680]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/compliance-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 22:49:24 crc restorecon[4680]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/container-security-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 22:49:24 crc restorecon[4680]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/container-security-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 22:49:24 crc restorecon[4680]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/costmanagement-metrics-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 22:49:24 crc restorecon[4680]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/costmanagement-metrics-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 22:49:24 crc restorecon[4680]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cryostat-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 22:49:24 crc restorecon[4680]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cryostat-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 22:49:24 crc restorecon[4680]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/datagrid not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 22:49:24 crc restorecon[4680]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/datagrid/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 22:49:24 crc restorecon[4680]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/devspaces not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 22:49:24 crc restorecon[4680]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/devspaces/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 22:49:24 crc restorecon[4680]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/devworkspace-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 22:49:24 crc restorecon[4680]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/devworkspace-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 22:49:24 crc restorecon[4680]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dpu-network-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 22:49:24 crc restorecon[4680]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dpu-network-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 22:49:24 crc restorecon[4680]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/eap not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 22:49:24 crc restorecon[4680]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/eap/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 22:49:24 crc restorecon[4680]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/elasticsearch-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 22:49:24 crc restorecon[4680]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/elasticsearch-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 22:49:24 crc restorecon[4680]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/external-dns-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 22:49:24 crc restorecon[4680]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/external-dns-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 22:49:24 crc restorecon[4680]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fence-agents-remediation not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 22:49:24 crc restorecon[4680]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fence-agents-remediation/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 22:49:24 crc restorecon[4680]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/file-integrity-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 22:49:24 crc restorecon[4680]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/file-integrity-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 22:49:24 crc restorecon[4680]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fuse-apicurito not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 22:49:24 crc restorecon[4680]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fuse-apicurito/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 22:49:24 crc restorecon[4680]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fuse-console not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 22:49:24 crc restorecon[4680]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fuse-console/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 22:49:24 crc restorecon[4680]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fuse-online not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 22:49:24 crc restorecon[4680]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fuse-online/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 22:49:24 crc restorecon[4680]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/gatekeeper-operator-product not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 22:49:24 crc restorecon[4680]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/gatekeeper-operator-product/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 22:49:24 crc restorecon[4680]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/jaeger-product not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 22:49:24 crc restorecon[4680]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/jaeger-product/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 22:49:24 crc restorecon[4680]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/jws-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 22:49:24 crc restorecon[4680]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/jws-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 22:49:24 crc restorecon[4680]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kernel-module-management not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 22:49:24 crc restorecon[4680]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kernel-module-management/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 22:49:24 crc restorecon[4680]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kernel-module-management-hub not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 22:49:24 crc restorecon[4680]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kernel-module-management-hub/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 22:49:24 crc restorecon[4680]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kiali-ossm not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 22:49:24 crc restorecon[4680]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kiali-ossm/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 22:49:24 crc restorecon[4680]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubevirt-hyperconverged not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 22:49:24 crc restorecon[4680]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubevirt-hyperconverged/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 22:49:24 crc restorecon[4680]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/logic-operator-rhel8 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 22:49:24 crc restorecon[4680]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/logic-operator-rhel8/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 22:49:24 crc restorecon[4680]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/loki-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 22:49:24 crc restorecon[4680]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/loki-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 22:49:24 crc restorecon[4680]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/lvms-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 22:49:24 crc restorecon[4680]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/lvms-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 22:49:24 crc restorecon[4680]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/machine-deletion-remediation not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 22:49:24 crc restorecon[4680]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/machine-deletion-remediation/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 22:49:24 crc restorecon[4680]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mcg-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 22:49:24 crc restorecon[4680]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mcg-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 22:49:24 crc restorecon[4680]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mta-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 22:49:24 crc restorecon[4680]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mta-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 22:49:24 crc restorecon[4680]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mtc-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 22:49:24 crc restorecon[4680]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mtc-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 22:49:24 crc restorecon[4680]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mtr-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 22:49:24 crc restorecon[4680]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mtr-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 22:49:24 crc restorecon[4680]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mtv-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 22:49:24 crc restorecon[4680]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mtv-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 22:49:24 crc restorecon[4680]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/multicluster-engine not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 22:49:24 crc restorecon[4680]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/multicluster-engine/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 22:49:24 crc restorecon[4680]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/netobserv-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 22:49:24 crc restorecon[4680]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/netobserv-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 22:49:24 crc restorecon[4680]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/node-healthcheck-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 22:49:24 crc restorecon[4680]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/node-healthcheck-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 22:49:24 crc restorecon[4680]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/node-maintenance-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 22:49:24 crc restorecon[4680]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/node-maintenance-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 22:49:24 crc restorecon[4680]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/node-observability-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 22:49:24 crc restorecon[4680]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/node-observability-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 22:49:24 crc restorecon[4680]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ocs-client-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 22:49:24 crc restorecon[4680]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ocs-client-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 22:49:24 crc restorecon[4680]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ocs-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 22:49:24 crc restorecon[4680]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ocs-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 22:49:24 crc restorecon[4680]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odf-csi-addons-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 22:49:24 crc restorecon[4680]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odf-csi-addons-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 22:49:24 crc restorecon[4680]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odf-multicluster-orchestrator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 22:49:24 crc restorecon[4680]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odf-multicluster-orchestrator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 22:49:24 crc restorecon[4680]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odf-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 22:49:24 crc restorecon[4680]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odf-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 22:49:24 crc restorecon[4680]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odf-prometheus-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 22:49:24 crc restorecon[4680]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odf-prometheus-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 22:49:24 crc restorecon[4680]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odr-cluster-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 22:49:24 crc restorecon[4680]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odr-cluster-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 22:49:24 crc restorecon[4680]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odr-hub-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 22:49:24 crc restorecon[4680]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odr-hub-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 22:49:24 crc restorecon[4680]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-cert-manager-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 22:49:24 crc restorecon[4680]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-cert-manager-operator/bundle-v1.15.0.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 22:49:24 crc restorecon[4680]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-cert-manager-operator/channel.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 22:49:24 crc restorecon[4680]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-cert-manager-operator/package.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 22:49:24 crc restorecon[4680]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-custom-metrics-autoscaler-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 22:49:24 crc restorecon[4680]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-custom-metrics-autoscaler-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 22:49:24 crc restorecon[4680]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-gitops-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 22:49:24 crc restorecon[4680]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-gitops-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 22:49:24 crc restorecon[4680]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-pipelines-operator-rh not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 22:49:24 crc restorecon[4680]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-pipelines-operator-rh/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 22:49:24 crc restorecon[4680]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-secondary-scheduler-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 22:49:24 crc restorecon[4680]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-secondary-scheduler-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 22:49:24 crc restorecon[4680]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/opentelemetry-product not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 22:49:24 crc restorecon[4680]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/opentelemetry-product/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 22:49:24 crc restorecon[4680]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/quay-bridge-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 22:49:24 crc restorecon[4680]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/quay-bridge-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 22:49:24 crc restorecon[4680]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/quay-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 22:49:24 crc restorecon[4680]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/quay-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 22:49:24 crc restorecon[4680]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/recipe not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 22:49:24 crc restorecon[4680]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/recipe/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 22:49:24 crc restorecon[4680]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/red-hat-camel-k not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 22:49:24 crc restorecon[4680]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/red-hat-camel-k/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 22:49:24 crc restorecon[4680]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/red-hat-hawtio-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 22:49:24 crc restorecon[4680]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/red-hat-hawtio-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 22:49:24 crc restorecon[4680]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/redhat-oadp-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 22:49:24 crc restorecon[4680]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/redhat-oadp-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 22:49:24 crc restorecon[4680]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rh-service-binding-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 22:49:24 crc restorecon[4680]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rh-service-binding-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 22:49:24 crc restorecon[4680]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhacs-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 22:49:24 crc restorecon[4680]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhacs-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 22:49:24 crc restorecon[4680]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhbk-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 22:49:24 crc restorecon[4680]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhbk-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 22:49:24 crc restorecon[4680]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhdh not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 22:49:24 crc restorecon[4680]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhdh/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 22:49:24 crc restorecon[4680]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhods-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 22:49:24 crc restorecon[4680]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhods-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 22:49:24 crc restorecon[4680]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhods-prometheus-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 22:49:24 crc restorecon[4680]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhods-prometheus-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 22:49:24 crc restorecon[4680]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhpam-kogito-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 22:49:24 crc restorecon[4680]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhpam-kogito-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 22:49:24 crc restorecon[4680]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhsso-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 22:49:24 crc restorecon[4680]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhsso-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 22:49:24 crc restorecon[4680]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rook-ceph-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 22:49:24 crc restorecon[4680]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rook-ceph-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 22:49:24 crc restorecon[4680]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/run-once-duration-override-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 22:49:24 crc restorecon[4680]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/run-once-duration-override-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 22:49:24 crc restorecon[4680]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sandboxed-containers-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 22:49:24 crc restorecon[4680]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sandboxed-containers-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 22:49:24 crc restorecon[4680]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/security-profiles-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 22:49:24 crc restorecon[4680]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/security-profiles-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 22:49:24 crc restorecon[4680]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/self-node-remediation not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 22:49:24 crc restorecon[4680]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/self-node-remediation/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 22:49:24 crc restorecon[4680]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/serverless-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 22:49:24 crc restorecon[4680]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/serverless-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 22:49:24 crc restorecon[4680]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/service-registry-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 22:49:24 crc restorecon[4680]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/service-registry-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 22:49:24 crc restorecon[4680]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/servicemeshoperator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 22:49:24 crc restorecon[4680]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/servicemeshoperator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 22:49:24 crc restorecon[4680]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/servicemeshoperator3 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 22:49:24 crc restorecon[4680]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/servicemeshoperator3/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 22:49:24 crc restorecon[4680]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/skupper-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 22:49:24 crc restorecon[4680]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/skupper-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 22:49:24 crc restorecon[4680]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/submariner not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 22:49:24 crc restorecon[4680]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/submariner/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 22:49:24 crc restorecon[4680]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tang-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 22:49:24 crc restorecon[4680]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tang-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 22:49:24 crc restorecon[4680]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tempo-product not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 22:49:24 crc restorecon[4680]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tempo-product/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 22:49:24 crc restorecon[4680]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/trustee-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 22:49:24 crc restorecon[4680]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/trustee-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 22:49:24 crc restorecon[4680]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/volsync-product not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 22:49:24 crc restorecon[4680]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/volsync-product/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 22:49:24 crc restorecon[4680]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/web-terminal not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 22:49:24 crc restorecon[4680]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/web-terminal/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 22:49:24 crc restorecon[4680]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/cache not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 22:49:24 crc restorecon[4680]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 22:49:24 crc restorecon[4680]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 22:49:24 crc restorecon[4680]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/00000-1.psg not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 22:49:24 crc restorecon[4680]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/00000-1.psg.pmt not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 22:49:24 crc restorecon[4680]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/db.pmt not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 22:49:24 crc restorecon[4680]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/index.pmt not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 22:49:24 crc restorecon[4680]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/main.pix not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 22:49:24 crc restorecon[4680]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/overflow.pix not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 22:49:24 crc restorecon[4680]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/digest not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 22:49:24 crc restorecon[4680]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 22:49:24 crc restorecon[4680]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/containers/extract-utilities/bc8d0691 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 22:49:24 crc restorecon[4680]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/containers/extract-utilities/6b76097a not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 22:49:24 crc restorecon[4680]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/containers/extract-utilities/34d1af30 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 22:49:24 crc restorecon[4680]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/containers/extract-content/312ba61c not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 22:49:24 crc restorecon[4680]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/containers/extract-content/645d5dd1 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 22:49:24 crc restorecon[4680]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/containers/extract-content/16e825f0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 22:49:24 crc restorecon[4680]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/containers/registry-server/4cf51fc9 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 22:49:24 crc restorecon[4680]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/containers/registry-server/2a23d348 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 22:49:24 crc restorecon[4680]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/containers/registry-server/075dbd49 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 22:49:24 crc restorecon[4680]: /var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/volumes/kubernetes.io~configmap/serviceca not reset as customized by admin to system_u:object_r:container_file_t:s0:c842,c986 Dec 10 22:49:24 crc restorecon[4680]: /var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/volumes/kubernetes.io~configmap/serviceca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c842,c986 Dec 10 22:49:24 crc restorecon[4680]: /var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/volumes/kubernetes.io~configmap/serviceca/..2025_02_24_06_09_13.3521195566 not reset as customized by admin to system_u:object_r:container_file_t:s0:c842,c986 Dec 10 22:49:24 crc restorecon[4680]: /var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/volumes/kubernetes.io~configmap/serviceca/..2025_02_24_06_09_13.3521195566/image-registry.openshift-image-registry.svc..5000 not reset as customized by admin to system_u:object_r:container_file_t:s0:c842,c986 Dec 10 22:49:24 crc restorecon[4680]: /var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/volumes/kubernetes.io~configmap/serviceca/..2025_02_24_06_09_13.3521195566/image-registry.openshift-image-registry.svc.cluster.local..5000 not reset as customized by admin to system_u:object_r:container_file_t:s0:c842,c986 Dec 10 22:49:24 crc restorecon[4680]: /var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/volumes/kubernetes.io~configmap/serviceca/..2025_02_24_06_09_13.3521195566/default-route-openshift-image-registry.apps-crc.testing not reset as customized by admin to system_u:object_r:container_file_t:s0:c842,c986 Dec 10 22:49:24 crc restorecon[4680]: /var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/volumes/kubernetes.io~configmap/serviceca/default-route-openshift-image-registry.apps-crc.testing not reset as customized by admin to system_u:object_r:container_file_t:s0:c842,c986 Dec 10 22:49:24 crc restorecon[4680]: /var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/volumes/kubernetes.io~configmap/serviceca/image-registry.openshift-image-registry.svc..5000 not reset as customized by admin to system_u:object_r:container_file_t:s0:c842,c986 Dec 10 22:49:24 crc restorecon[4680]: /var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/volumes/kubernetes.io~configmap/serviceca/image-registry.openshift-image-registry.svc.cluster.local..5000 not reset as customized by admin to system_u:object_r:container_file_t:s0:c842,c986 Dec 10 22:49:24 crc restorecon[4680]: /var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c842,c986 Dec 10 22:49:24 crc restorecon[4680]: /var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/containers/node-ca/dd585ddd not reset as customized by admin to system_u:object_r:container_file_t:s0:c377,c642 Dec 10 22:49:24 crc restorecon[4680]: /var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/containers/node-ca/17ebd0ab not reset as customized by admin to system_u:object_r:container_file_t:s0:c338,c343 Dec 10 22:49:24 crc restorecon[4680]: /var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/containers/node-ca/005579f4 not reset as customized by admin to system_u:object_r:container_file_t:s0:c842,c986 Dec 10 22:49:24 crc restorecon[4680]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes/kubernetes.io~configmap/etcd-serving-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Dec 10 22:49:24 crc restorecon[4680]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes/kubernetes.io~configmap/etcd-serving-ca/..2025_02_23_05_23_11.449897510 not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Dec 10 22:49:24 crc restorecon[4680]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes/kubernetes.io~configmap/etcd-serving-ca/..2025_02_23_05_23_11.449897510/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Dec 10 22:49:24 crc restorecon[4680]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes/kubernetes.io~configmap/etcd-serving-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Dec 10 22:49:24 crc restorecon[4680]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes/kubernetes.io~configmap/etcd-serving-ca/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Dec 10 22:49:24 crc restorecon[4680]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes/kubernetes.io~configmap/trusted-ca-bundle not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Dec 10 22:49:24 crc restorecon[4680]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes/kubernetes.io~configmap/trusted-ca-bundle/..2025_02_23_05_23_11.1287037894 not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Dec 10 22:49:24 crc restorecon[4680]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes/kubernetes.io~configmap/trusted-ca-bundle/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Dec 10 22:49:24 crc restorecon[4680]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes/kubernetes.io~configmap/audit-policies not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Dec 10 22:49:24 crc restorecon[4680]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes/kubernetes.io~configmap/audit-policies/..2025_02_23_05_23_11.1301053334 not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Dec 10 22:49:24 crc restorecon[4680]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes/kubernetes.io~configmap/audit-policies/..2025_02_23_05_23_11.1301053334/policy.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Dec 10 22:49:24 crc restorecon[4680]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes/kubernetes.io~configmap/audit-policies/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Dec 10 22:49:24 crc restorecon[4680]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes/kubernetes.io~configmap/audit-policies/policy.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Dec 10 22:49:24 crc restorecon[4680]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Dec 10 22:49:24 crc restorecon[4680]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/containers/fix-audit-permissions/bf5f3b9c not reset as customized by admin to system_u:object_r:container_file_t:s0:c49,c263 Dec 10 22:49:24 crc restorecon[4680]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/containers/fix-audit-permissions/af276eb7 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c701 Dec 10 22:49:24 crc restorecon[4680]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/containers/fix-audit-permissions/ea28e322 not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Dec 10 22:49:24 crc restorecon[4680]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/containers/oauth-apiserver/692e6683 not reset as customized by admin to system_u:object_r:container_file_t:s0:c49,c263 Dec 10 22:49:24 crc restorecon[4680]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/containers/oauth-apiserver/871746a7 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c701 Dec 10 22:49:24 crc restorecon[4680]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/containers/oauth-apiserver/4eb2e958 not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Dec 10 22:49:24 crc restorecon[4680]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/console-config not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Dec 10 22:49:24 crc restorecon[4680]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/console-config/..2025_02_24_06_09_06.2875086261 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Dec 10 22:49:24 crc restorecon[4680]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/console-config/..2025_02_24_06_09_06.2875086261/console-config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Dec 10 22:49:24 crc restorecon[4680]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/console-config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Dec 10 22:49:24 crc restorecon[4680]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/console-config/console-config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Dec 10 22:49:24 crc restorecon[4680]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/trusted-ca-bundle not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Dec 10 22:49:24 crc restorecon[4680]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/trusted-ca-bundle/..2025_02_24_06_09_06.286118152 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Dec 10 22:49:24 crc restorecon[4680]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/trusted-ca-bundle/..2025_02_24_06_09_06.286118152/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Dec 10 22:49:24 crc restorecon[4680]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/trusted-ca-bundle/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Dec 10 22:49:24 crc restorecon[4680]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/trusted-ca-bundle/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Dec 10 22:49:24 crc restorecon[4680]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/oauth-serving-cert not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Dec 10 22:49:24 crc restorecon[4680]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/oauth-serving-cert/..2025_02_24_06_09_06.3865795478 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Dec 10 22:49:24 crc restorecon[4680]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/oauth-serving-cert/..2025_02_24_06_09_06.3865795478/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Dec 10 22:49:24 crc restorecon[4680]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/oauth-serving-cert/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Dec 10 22:49:24 crc restorecon[4680]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/oauth-serving-cert/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Dec 10 22:49:24 crc restorecon[4680]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/service-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Dec 10 22:49:24 crc restorecon[4680]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/service-ca/..2025_02_24_06_09_06.584414814 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Dec 10 22:49:24 crc restorecon[4680]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/service-ca/..2025_02_24_06_09_06.584414814/service-ca.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Dec 10 22:49:24 crc restorecon[4680]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/service-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Dec 10 22:49:24 crc restorecon[4680]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/service-ca/service-ca.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Dec 10 22:49:24 crc restorecon[4680]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Dec 10 22:49:24 crc restorecon[4680]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/containers/console/ca9b62da not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Dec 10 22:49:24 crc restorecon[4680]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/containers/console/0edd6fce not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Dec 10 22:49:24 crc restorecon[4680]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Dec 10 22:49:24 crc restorecon[4680]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/config/..2025_02_24_06_20_07.2406383837 not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Dec 10 22:49:24 crc restorecon[4680]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/config/..2025_02_24_06_20_07.2406383837/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Dec 10 22:49:24 crc restorecon[4680]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/config/..2025_02_24_06_20_07.2406383837/openshift-controller-manager.client-ca.configmap not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Dec 10 22:49:24 crc restorecon[4680]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/config/..2025_02_24_06_20_07.2406383837/openshift-controller-manager.openshift-global-ca.configmap not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Dec 10 22:49:24 crc restorecon[4680]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/config/..2025_02_24_06_20_07.2406383837/openshift-controller-manager.serving-cert.secret not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Dec 10 22:49:24 crc restorecon[4680]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Dec 10 22:49:24 crc restorecon[4680]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/config/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Dec 10 22:49:24 crc restorecon[4680]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/config/openshift-controller-manager.client-ca.configmap not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Dec 10 22:49:24 crc restorecon[4680]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/config/openshift-controller-manager.openshift-global-ca.configmap not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Dec 10 22:49:24 crc restorecon[4680]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/config/openshift-controller-manager.serving-cert.secret not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Dec 10 22:49:24 crc restorecon[4680]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/client-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Dec 10 22:49:24 crc restorecon[4680]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/client-ca/..2025_02_24_06_20_07.1071801880 not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Dec 10 22:49:24 crc restorecon[4680]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/client-ca/..2025_02_24_06_20_07.1071801880/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Dec 10 22:49:24 crc restorecon[4680]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/client-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Dec 10 22:49:24 crc restorecon[4680]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/client-ca/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Dec 10 22:49:24 crc restorecon[4680]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/proxy-ca-bundles not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Dec 10 22:49:24 crc restorecon[4680]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/proxy-ca-bundles/..2025_02_24_06_20_07.2494444877 not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Dec 10 22:49:24 crc restorecon[4680]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/proxy-ca-bundles/..2025_02_24_06_20_07.2494444877/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Dec 10 22:49:24 crc restorecon[4680]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/proxy-ca-bundles/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Dec 10 22:49:24 crc restorecon[4680]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/proxy-ca-bundles/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Dec 10 22:49:25 crc restorecon[4680]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Dec 10 22:49:25 crc restorecon[4680]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/containers/controller-manager/89b4555f not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Dec 10 22:49:25 crc restorecon[4680]: /var/lib/kubelet/pods/87cf06ed-a83f-41a7-828d-70653580a8cb/volumes/kubernetes.io~configmap/config-volume not reset as customized by admin to system_u:object_r:container_file_t:s0:c466,c972 Dec 10 22:49:25 crc restorecon[4680]: /var/lib/kubelet/pods/87cf06ed-a83f-41a7-828d-70653580a8cb/volumes/kubernetes.io~configmap/config-volume/..2025_02_23_05_23_22.4071100442 not reset as customized by admin to system_u:object_r:container_file_t:s0:c466,c972 Dec 10 22:49:25 crc restorecon[4680]: /var/lib/kubelet/pods/87cf06ed-a83f-41a7-828d-70653580a8cb/volumes/kubernetes.io~configmap/config-volume/..2025_02_23_05_23_22.4071100442/Corefile not reset as customized by admin to system_u:object_r:container_file_t:s0:c466,c972 Dec 10 22:49:25 crc restorecon[4680]: /var/lib/kubelet/pods/87cf06ed-a83f-41a7-828d-70653580a8cb/volumes/kubernetes.io~configmap/config-volume/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c466,c972 Dec 10 22:49:25 crc restorecon[4680]: /var/lib/kubelet/pods/87cf06ed-a83f-41a7-828d-70653580a8cb/volumes/kubernetes.io~configmap/config-volume/Corefile not reset as customized by admin to system_u:object_r:container_file_t:s0:c466,c972 Dec 10 22:49:25 crc restorecon[4680]: /var/lib/kubelet/pods/87cf06ed-a83f-41a7-828d-70653580a8cb/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c466,c972 Dec 10 22:49:25 crc restorecon[4680]: /var/lib/kubelet/pods/87cf06ed-a83f-41a7-828d-70653580a8cb/containers/dns/655fcd71 not reset as customized by admin to system_u:object_r:container_file_t:s0:c457,c841 Dec 10 22:49:25 crc restorecon[4680]: /var/lib/kubelet/pods/87cf06ed-a83f-41a7-828d-70653580a8cb/containers/dns/0d43c002 not reset as customized by admin to system_u:object_r:container_file_t:s0:c55,c1022 Dec 10 22:49:25 crc restorecon[4680]: /var/lib/kubelet/pods/87cf06ed-a83f-41a7-828d-70653580a8cb/containers/dns/e68efd17 not reset as customized by admin to system_u:object_r:container_file_t:s0:c466,c972 Dec 10 22:49:25 crc restorecon[4680]: /var/lib/kubelet/pods/87cf06ed-a83f-41a7-828d-70653580a8cb/containers/kube-rbac-proxy/9acf9b65 not reset as customized by admin to system_u:object_r:container_file_t:s0:c457,c841 Dec 10 22:49:25 crc restorecon[4680]: /var/lib/kubelet/pods/87cf06ed-a83f-41a7-828d-70653580a8cb/containers/kube-rbac-proxy/5ae3ff11 not reset as customized by admin to system_u:object_r:container_file_t:s0:c55,c1022 Dec 10 22:49:25 crc restorecon[4680]: /var/lib/kubelet/pods/87cf06ed-a83f-41a7-828d-70653580a8cb/containers/kube-rbac-proxy/1e59206a not reset as customized by admin to system_u:object_r:container_file_t:s0:c466,c972 Dec 10 22:49:25 crc restorecon[4680]: /var/lib/kubelet/pods/44663579-783b-4372-86d6-acf235a62d72/containers/dns-node-resolver/27af16d1 not reset as customized by admin to system_u:object_r:container_file_t:s0:c304,c1017 Dec 10 22:49:25 crc restorecon[4680]: /var/lib/kubelet/pods/44663579-783b-4372-86d6-acf235a62d72/containers/dns-node-resolver/7918e729 not reset as customized by admin to system_u:object_r:container_file_t:s0:c853,c893 Dec 10 22:49:25 crc restorecon[4680]: /var/lib/kubelet/pods/44663579-783b-4372-86d6-acf235a62d72/containers/dns-node-resolver/5d976d0e not reset as customized by admin to system_u:object_r:container_file_t:s0:c585,c981 Dec 10 22:49:25 crc restorecon[4680]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Dec 10 22:49:25 crc restorecon[4680]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/volumes/kubernetes.io~configmap/config/..2025_02_23_05_38_56.1112187283 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Dec 10 22:49:25 crc restorecon[4680]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/volumes/kubernetes.io~configmap/config/..2025_02_23_05_38_56.1112187283/controller-config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Dec 10 22:49:25 crc restorecon[4680]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Dec 10 22:49:25 crc restorecon[4680]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/volumes/kubernetes.io~configmap/config/controller-config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Dec 10 22:49:25 crc restorecon[4680]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/volumes/kubernetes.io~configmap/trusted-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Dec 10 22:49:25 crc restorecon[4680]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/volumes/kubernetes.io~configmap/trusted-ca/..2025_02_23_05_38_56.2839772658 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Dec 10 22:49:25 crc restorecon[4680]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/volumes/kubernetes.io~configmap/trusted-ca/..2025_02_23_05_38_56.2839772658/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Dec 10 22:49:25 crc restorecon[4680]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/volumes/kubernetes.io~configmap/trusted-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Dec 10 22:49:25 crc restorecon[4680]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/volumes/kubernetes.io~configmap/trusted-ca/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Dec 10 22:49:25 crc restorecon[4680]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Dec 10 22:49:25 crc restorecon[4680]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/containers/console-operator/d7f55cbb not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Dec 10 22:49:25 crc restorecon[4680]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/containers/console-operator/f0812073 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Dec 10 22:49:25 crc restorecon[4680]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/containers/console-operator/1a56cbeb not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Dec 10 22:49:25 crc restorecon[4680]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/containers/console-operator/7fdd437e not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Dec 10 22:49:25 crc restorecon[4680]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/containers/console-operator/cdfb5652 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Dec 10 22:49:25 crc restorecon[4680]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/etcd-serving-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Dec 10 22:49:25 crc restorecon[4680]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/etcd-serving-ca/..2025_02_24_06_17_29.3844392896 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Dec 10 22:49:25 crc restorecon[4680]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/etcd-serving-ca/..2025_02_24_06_17_29.3844392896/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Dec 10 22:49:25 crc restorecon[4680]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/etcd-serving-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Dec 10 22:49:25 crc restorecon[4680]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/etcd-serving-ca/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Dec 10 22:49:25 crc restorecon[4680]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Dec 10 22:49:25 crc restorecon[4680]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/config/..2025_02_24_06_17_29.848549803 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Dec 10 22:49:25 crc restorecon[4680]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/config/..2025_02_24_06_17_29.848549803/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Dec 10 22:49:25 crc restorecon[4680]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Dec 10 22:49:25 crc restorecon[4680]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/config/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Dec 10 22:49:25 crc restorecon[4680]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/audit not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Dec 10 22:49:25 crc restorecon[4680]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/audit/..2025_02_24_06_17_29.780046231 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Dec 10 22:49:25 crc restorecon[4680]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/audit/..2025_02_24_06_17_29.780046231/policy.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Dec 10 22:49:25 crc restorecon[4680]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/audit/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Dec 10 22:49:25 crc restorecon[4680]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/audit/policy.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Dec 10 22:49:25 crc restorecon[4680]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/image-import-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Dec 10 22:49:25 crc restorecon[4680]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/image-import-ca/..2025_02_24_06_17_29.2926008347 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Dec 10 22:49:25 crc restorecon[4680]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/image-import-ca/..2025_02_24_06_17_29.2926008347/image-registry.openshift-image-registry.svc..5000 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Dec 10 22:49:25 crc restorecon[4680]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/image-import-ca/..2025_02_24_06_17_29.2926008347/image-registry.openshift-image-registry.svc.cluster.local..5000 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Dec 10 22:49:25 crc restorecon[4680]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/image-import-ca/..2025_02_24_06_17_29.2926008347/default-route-openshift-image-registry.apps-crc.testing not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Dec 10 22:49:25 crc restorecon[4680]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/image-import-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Dec 10 22:49:25 crc restorecon[4680]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/image-import-ca/image-registry.openshift-image-registry.svc..5000 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Dec 10 22:49:25 crc restorecon[4680]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/image-import-ca/image-registry.openshift-image-registry.svc.cluster.local..5000 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Dec 10 22:49:25 crc restorecon[4680]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/image-import-ca/default-route-openshift-image-registry.apps-crc.testing not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Dec 10 22:49:25 crc restorecon[4680]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/trusted-ca-bundle not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Dec 10 22:49:25 crc restorecon[4680]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/trusted-ca-bundle/..2025_02_24_06_17_29.2729721485 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Dec 10 22:49:25 crc restorecon[4680]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/trusted-ca-bundle/..2025_02_24_06_17_29.2729721485/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Dec 10 22:49:25 crc restorecon[4680]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/trusted-ca-bundle/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Dec 10 22:49:25 crc restorecon[4680]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/trusted-ca-bundle/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Dec 10 22:49:25 crc restorecon[4680]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Dec 10 22:49:25 crc restorecon[4680]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/containers/fix-audit-permissions/fb93119e not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Dec 10 22:49:25 crc restorecon[4680]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/containers/openshift-apiserver/f1e8fc0e not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Dec 10 22:49:25 crc restorecon[4680]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/containers/openshift-apiserver-check-endpoints/218511f3 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Dec 10 22:49:25 crc restorecon[4680]: /var/lib/kubelet/pods/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b/volumes/kubernetes.io~empty-dir/tmpfs not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Dec 10 22:49:25 crc restorecon[4680]: /var/lib/kubelet/pods/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b/volumes/kubernetes.io~empty-dir/tmpfs/k8s-webhook-server not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Dec 10 22:49:25 crc restorecon[4680]: /var/lib/kubelet/pods/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b/volumes/kubernetes.io~empty-dir/tmpfs/k8s-webhook-server/serving-certs not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Dec 10 22:49:25 crc restorecon[4680]: /var/lib/kubelet/pods/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Dec 10 22:49:25 crc restorecon[4680]: /var/lib/kubelet/pods/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b/containers/packageserver/ca8af7b3 not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Dec 10 22:49:25 crc restorecon[4680]: /var/lib/kubelet/pods/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b/containers/packageserver/72cc8a75 not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Dec 10 22:49:25 crc restorecon[4680]: /var/lib/kubelet/pods/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b/containers/packageserver/6e8a3760 not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Dec 10 22:49:25 crc restorecon[4680]: /var/lib/kubelet/pods/0b78653f-4ff9-4508-8672-245ed9b561e3/volumes/kubernetes.io~configmap/service-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c6 Dec 10 22:49:25 crc restorecon[4680]: /var/lib/kubelet/pods/0b78653f-4ff9-4508-8672-245ed9b561e3/volumes/kubernetes.io~configmap/service-ca/..2025_02_23_05_27_30.557428972 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c6 Dec 10 22:49:25 crc restorecon[4680]: /var/lib/kubelet/pods/0b78653f-4ff9-4508-8672-245ed9b561e3/volumes/kubernetes.io~configmap/service-ca/..2025_02_23_05_27_30.557428972/service-ca.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c6 Dec 10 22:49:25 crc restorecon[4680]: /var/lib/kubelet/pods/0b78653f-4ff9-4508-8672-245ed9b561e3/volumes/kubernetes.io~configmap/service-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c6 Dec 10 22:49:25 crc restorecon[4680]: /var/lib/kubelet/pods/0b78653f-4ff9-4508-8672-245ed9b561e3/volumes/kubernetes.io~configmap/service-ca/service-ca.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c6 Dec 10 22:49:25 crc restorecon[4680]: /var/lib/kubelet/pods/0b78653f-4ff9-4508-8672-245ed9b561e3/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c6 Dec 10 22:49:25 crc restorecon[4680]: /var/lib/kubelet/pods/0b78653f-4ff9-4508-8672-245ed9b561e3/containers/cluster-version-operator/4c3455c0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c6 Dec 10 22:49:25 crc restorecon[4680]: /var/lib/kubelet/pods/0b78653f-4ff9-4508-8672-245ed9b561e3/containers/cluster-version-operator/2278acb0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c6 Dec 10 22:49:25 crc restorecon[4680]: /var/lib/kubelet/pods/0b78653f-4ff9-4508-8672-245ed9b561e3/containers/cluster-version-operator/4b453e4f not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c6 Dec 10 22:49:25 crc restorecon[4680]: /var/lib/kubelet/pods/0b78653f-4ff9-4508-8672-245ed9b561e3/containers/cluster-version-operator/3ec09bda not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c6 Dec 10 22:49:25 crc restorecon[4680]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/trusted-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 10 22:49:25 crc restorecon[4680]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/trusted-ca/..2025_02_24_06_25_03.422633132 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 10 22:49:25 crc restorecon[4680]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/trusted-ca/..2025_02_24_06_25_03.422633132/anchors not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 10 22:49:25 crc restorecon[4680]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/trusted-ca/..2025_02_24_06_25_03.422633132/anchors/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 10 22:49:25 crc restorecon[4680]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/trusted-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 10 22:49:25 crc restorecon[4680]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/trusted-ca/anchors not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 10 22:49:25 crc restorecon[4680]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/registry-certificates not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 10 22:49:25 crc restorecon[4680]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/registry-certificates/..2025_02_24_06_25_03.3594477318 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 10 22:49:25 crc restorecon[4680]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/registry-certificates/..2025_02_24_06_25_03.3594477318/image-registry.openshift-image-registry.svc..5000 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 10 22:49:25 crc restorecon[4680]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/registry-certificates/..2025_02_24_06_25_03.3594477318/image-registry.openshift-image-registry.svc.cluster.local..5000 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 10 22:49:25 crc restorecon[4680]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/registry-certificates/..2025_02_24_06_25_03.3594477318/default-route-openshift-image-registry.apps-crc.testing not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 10 22:49:25 crc restorecon[4680]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/registry-certificates/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 10 22:49:25 crc restorecon[4680]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/registry-certificates/image-registry.openshift-image-registry.svc..5000 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 10 22:49:25 crc restorecon[4680]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/registry-certificates/image-registry.openshift-image-registry.svc.cluster.local..5000 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 10 22:49:25 crc restorecon[4680]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/registry-certificates/default-route-openshift-image-registry.apps-crc.testing not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 10 22:49:25 crc restorecon[4680]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 10 22:49:25 crc restorecon[4680]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/edk2 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 10 22:49:25 crc restorecon[4680]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/edk2/cacerts.bin not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 10 22:49:25 crc restorecon[4680]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/java not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 10 22:49:25 crc restorecon[4680]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/java/cacerts not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 10 22:49:25 crc restorecon[4680]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/openssl not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 10 22:49:25 crc restorecon[4680]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/openssl/ca-bundle.trust.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 10 22:49:25 crc restorecon[4680]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 10 22:49:25 crc restorecon[4680]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 10 22:49:25 crc restorecon[4680]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/email-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 10 22:49:25 crc restorecon[4680]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/objsign-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 10 22:49:25 crc restorecon[4680]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 10 22:49:25 crc restorecon[4680]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/2ae6433e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 10 22:49:25 crc restorecon[4680]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/fde84897.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 10 22:49:25 crc restorecon[4680]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/75680d2e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 10 22:49:25 crc restorecon[4680]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/openshift-service-serving-signer_1740288168.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 10 22:49:25 crc restorecon[4680]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/facfc4fa.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 10 22:49:25 crc restorecon[4680]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/8f5a969c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 10 22:49:25 crc restorecon[4680]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/CFCA_EV_ROOT.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 10 22:49:25 crc restorecon[4680]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9ef4a08a.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 10 22:49:25 crc restorecon[4680]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ingress-operator_1740288202.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 10 22:49:25 crc restorecon[4680]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/2f332aed.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 10 22:49:25 crc restorecon[4680]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/248c8271.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 10 22:49:25 crc restorecon[4680]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/8d10a21f.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 10 22:49:25 crc restorecon[4680]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ACCVRAIZ1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 10 22:49:25 crc restorecon[4680]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/a94d09e5.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 10 22:49:25 crc restorecon[4680]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/3c9a4d3b.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 10 22:49:25 crc restorecon[4680]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/40193066.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 10 22:49:25 crc restorecon[4680]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/AC_RAIZ_FNMT-RCM.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 10 22:49:25 crc restorecon[4680]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/cd8c0d63.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 10 22:49:25 crc restorecon[4680]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b936d1c6.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 10 22:49:25 crc restorecon[4680]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/CA_Disig_Root_R2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 10 22:49:25 crc restorecon[4680]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/4fd49c6c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 10 22:49:25 crc restorecon[4680]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/AC_RAIZ_FNMT-RCM_SERVIDORES_SEGUROS.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 10 22:49:25 crc restorecon[4680]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b81b93f0.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 10 22:49:25 crc restorecon[4680]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5f9a69fa.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 10 22:49:25 crc restorecon[4680]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Certigna.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 10 22:49:25 crc restorecon[4680]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b30d5fda.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 10 22:49:25 crc restorecon[4680]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ANF_Secure_Server_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 10 22:49:25 crc restorecon[4680]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b433981b.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 10 22:49:25 crc restorecon[4680]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/93851c9e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 10 22:49:25 crc restorecon[4680]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9282e51c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 10 22:49:25 crc restorecon[4680]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e7dd1bc4.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 10 22:49:25 crc restorecon[4680]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Actalis_Authentication_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 10 22:49:25 crc restorecon[4680]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/930ac5d2.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 10 22:49:25 crc restorecon[4680]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5f47b495.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 10 22:49:25 crc restorecon[4680]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e113c810.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 10 22:49:25 crc restorecon[4680]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5931b5bc.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 10 22:49:25 crc restorecon[4680]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/AffirmTrust_Commercial.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 10 22:49:25 crc restorecon[4680]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/2b349938.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 10 22:49:25 crc restorecon[4680]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e48193cf.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 10 22:49:25 crc restorecon[4680]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/302904dd.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 10 22:49:25 crc restorecon[4680]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/a716d4ed.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 10 22:49:25 crc restorecon[4680]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/AffirmTrust_Networking.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 10 22:49:25 crc restorecon[4680]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/93bc0acc.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 10 22:49:25 crc restorecon[4680]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/86212b19.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 10 22:49:25 crc restorecon[4680]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Certigna_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 10 22:49:25 crc restorecon[4680]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/AffirmTrust_Premium.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 10 22:49:25 crc restorecon[4680]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b727005e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 10 22:49:25 crc restorecon[4680]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/dbc54cab.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 10 22:49:25 crc restorecon[4680]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/f51bb24c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 10 22:49:25 crc restorecon[4680]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/c28a8a30.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 10 22:49:25 crc restorecon[4680]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/AffirmTrust_Premium_ECC.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 10 22:49:25 crc restorecon[4680]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9c8dfbd4.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 10 22:49:25 crc restorecon[4680]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ccc52f49.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 10 22:49:25 crc restorecon[4680]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/cb1c3204.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 10 22:49:25 crc restorecon[4680]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Amazon_Root_CA_1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 10 22:49:25 crc restorecon[4680]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ce5e74ef.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 10 22:49:25 crc restorecon[4680]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/fd08c599.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 10 22:49:25 crc restorecon[4680]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Certum_Trusted_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 10 22:49:25 crc restorecon[4680]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Amazon_Root_CA_2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 10 22:49:25 crc restorecon[4680]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/6d41d539.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 10 22:49:25 crc restorecon[4680]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/fb5fa911.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 10 22:49:25 crc restorecon[4680]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e35234b1.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 10 22:49:25 crc restorecon[4680]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Amazon_Root_CA_3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 10 22:49:25 crc restorecon[4680]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/8cb5ee0f.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 10 22:49:25 crc restorecon[4680]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/7a7c655d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 10 22:49:25 crc restorecon[4680]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/f8fc53da.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 10 22:49:25 crc restorecon[4680]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Amazon_Root_CA_4.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 10 22:49:25 crc restorecon[4680]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/de6d66f3.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 10 22:49:25 crc restorecon[4680]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d41b5e2a.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 10 22:49:25 crc restorecon[4680]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/41a3f684.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 10 22:49:25 crc restorecon[4680]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/1df5a75f.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 10 22:49:25 crc restorecon[4680]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Atos_TrustedRoot_2011.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 10 22:49:25 crc restorecon[4680]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e36a6752.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 10 22:49:25 crc restorecon[4680]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b872f2b4.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 10 22:49:25 crc restorecon[4680]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9576d26b.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 10 22:49:25 crc restorecon[4680]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/228f89db.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 10 22:49:25 crc restorecon[4680]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Atos_TrustedRoot_Root_CA_ECC_TLS_2021.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 10 22:49:25 crc restorecon[4680]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/fb717492.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 10 22:49:25 crc restorecon[4680]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/2d21b73c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 10 22:49:25 crc restorecon[4680]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/0b1b94ef.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 10 22:49:25 crc restorecon[4680]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/595e996b.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 10 22:49:25 crc restorecon[4680]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Atos_TrustedRoot_Root_CA_RSA_TLS_2021.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 10 22:49:25 crc restorecon[4680]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9b46e03d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 10 22:49:25 crc restorecon[4680]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/128f4b91.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 10 22:49:25 crc restorecon[4680]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Buypass_Class_3_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 10 22:49:25 crc restorecon[4680]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/81f2d2b1.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 10 22:49:25 crc restorecon[4680]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Autoridad_de_Certificacion_Firmaprofesional_CIF_A62634068.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 10 22:49:25 crc restorecon[4680]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/3bde41ac.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 10 22:49:25 crc restorecon[4680]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d16a5865.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 10 22:49:25 crc restorecon[4680]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Certum_EC-384_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 10 22:49:25 crc restorecon[4680]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/BJCA_Global_Root_CA1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 10 22:49:25 crc restorecon[4680]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/0179095f.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 10 22:49:25 crc restorecon[4680]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ffa7f1eb.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 10 22:49:25 crc restorecon[4680]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9482e63a.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 10 22:49:25 crc restorecon[4680]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d4dae3dd.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 10 22:49:25 crc restorecon[4680]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/BJCA_Global_Root_CA2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 10 22:49:25 crc restorecon[4680]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/3e359ba6.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 10 22:49:25 crc restorecon[4680]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/7e067d03.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 10 22:49:25 crc restorecon[4680]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/95aff9e3.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 10 22:49:25 crc restorecon[4680]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d7746a63.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 10 22:49:25 crc restorecon[4680]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Baltimore_CyberTrust_Root.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 10 22:49:25 crc restorecon[4680]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/653b494a.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 10 22:49:25 crc restorecon[4680]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/3ad48a91.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 10 22:49:25 crc restorecon[4680]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Certum_Trusted_Network_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 10 22:49:25 crc restorecon[4680]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Buypass_Class_2_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 10 22:49:25 crc restorecon[4680]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/54657681.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 10 22:49:25 crc restorecon[4680]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/82223c44.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 10 22:49:25 crc restorecon[4680]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e8de2f56.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 10 22:49:25 crc restorecon[4680]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/2d9dafe4.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 10 22:49:25 crc restorecon[4680]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d96b65e2.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 10 22:49:25 crc restorecon[4680]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ee64a828.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 10 22:49:25 crc restorecon[4680]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/COMODO_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 10 22:49:25 crc restorecon[4680]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/40547a79.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 10 22:49:25 crc restorecon[4680]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5a3f0ff8.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 10 22:49:25 crc restorecon[4680]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/7a780d93.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 10 22:49:25 crc restorecon[4680]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/34d996fb.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 10 22:49:25 crc restorecon[4680]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/COMODO_ECC_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 10 22:49:25 crc restorecon[4680]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/eed8c118.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 10 22:49:25 crc restorecon[4680]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/89c02a45.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 10 22:49:25 crc restorecon[4680]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Certainly_Root_R1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 10 22:49:25 crc restorecon[4680]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b1159c4c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 10 22:49:25 crc restorecon[4680]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/COMODO_RSA_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 10 22:49:25 crc restorecon[4680]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d6325660.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 10 22:49:25 crc restorecon[4680]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d4c339cb.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 10 22:49:25 crc restorecon[4680]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/8312c4c1.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 10 22:49:25 crc restorecon[4680]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Certainly_Root_E1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 10 22:49:25 crc restorecon[4680]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/8508e720.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 10 22:49:25 crc restorecon[4680]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5fdd185d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 10 22:49:25 crc restorecon[4680]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/48bec511.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 10 22:49:25 crc restorecon[4680]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/69105f4f.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 10 22:49:25 crc restorecon[4680]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/GlobalSign.1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 10 22:49:25 crc restorecon[4680]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/0b9bc432.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 10 22:49:25 crc restorecon[4680]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Certum_Trusted_Network_CA_2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 10 22:49:25 crc restorecon[4680]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/GTS_Root_R3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 10 22:49:25 crc restorecon[4680]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/32888f65.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 10 22:49:25 crc restorecon[4680]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/CommScope_Public_Trust_ECC_Root-01.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 10 22:49:25 crc restorecon[4680]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/6b03dec0.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 10 22:49:25 crc restorecon[4680]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/219d9499.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 10 22:49:25 crc restorecon[4680]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/CommScope_Public_Trust_ECC_Root-02.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 10 22:49:25 crc restorecon[4680]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5acf816d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 10 22:49:25 crc restorecon[4680]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/cbf06781.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 10 22:49:25 crc restorecon[4680]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/CommScope_Public_Trust_RSA_Root-01.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 10 22:49:25 crc restorecon[4680]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/GTS_Root_R4.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 10 22:49:25 crc restorecon[4680]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/dc99f41e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 10 22:49:25 crc restorecon[4680]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/CommScope_Public_Trust_RSA_Root-02.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 10 22:49:25 crc restorecon[4680]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/GlobalSign.3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 10 22:49:25 crc restorecon[4680]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/AAA_Certificate_Services.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 10 22:49:25 crc restorecon[4680]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/985c1f52.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 10 22:49:25 crc restorecon[4680]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/8794b4e3.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 10 22:49:25 crc restorecon[4680]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/D-TRUST_BR_Root_CA_1_2020.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 10 22:49:25 crc restorecon[4680]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e7c037b4.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 10 22:49:25 crc restorecon[4680]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ef954a4e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 10 22:49:25 crc restorecon[4680]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/D-TRUST_EV_Root_CA_1_2020.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 10 22:49:25 crc restorecon[4680]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/2add47b6.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 10 22:49:25 crc restorecon[4680]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/90c5a3c8.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 10 22:49:25 crc restorecon[4680]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/D-TRUST_Root_Class_3_CA_2_2009.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 10 22:49:25 crc restorecon[4680]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b0f3e76e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 10 22:49:25 crc restorecon[4680]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/53a1b57a.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 10 22:49:25 crc restorecon[4680]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/D-TRUST_Root_Class_3_CA_2_EV_2009.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 10 22:49:25 crc restorecon[4680]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/GlobalSign_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 10 22:49:25 crc restorecon[4680]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/DigiCert_Assured_ID_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 10 22:49:25 crc restorecon[4680]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5ad8a5d6.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 10 22:49:25 crc restorecon[4680]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/68dd7389.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 10 22:49:25 crc restorecon[4680]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/DigiCert_Assured_ID_Root_G2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 10 22:49:25 crc restorecon[4680]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9d04f354.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 10 22:49:25 crc restorecon[4680]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/8d6437c3.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 10 22:49:25 crc restorecon[4680]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/062cdee6.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 10 22:49:25 crc restorecon[4680]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/bd43e1dd.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 10 22:49:25 crc restorecon[4680]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/DigiCert_Assured_ID_Root_G3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 10 22:49:25 crc restorecon[4680]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/7f3d5d1d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 10 22:49:25 crc restorecon[4680]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/c491639e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 10 22:49:25 crc restorecon[4680]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/GlobalSign_Root_E46.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 10 22:49:25 crc restorecon[4680]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/DigiCert_Global_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 10 22:49:25 crc restorecon[4680]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/3513523f.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 10 22:49:25 crc restorecon[4680]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/399e7759.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 10 22:49:25 crc restorecon[4680]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/feffd413.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 10 22:49:25 crc restorecon[4680]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d18e9066.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 10 22:49:25 crc restorecon[4680]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/DigiCert_Global_Root_G2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 10 22:49:25 crc restorecon[4680]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/607986c7.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 10 22:49:25 crc restorecon[4680]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/c90bc37d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 10 22:49:25 crc restorecon[4680]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/1b0f7e5c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 10 22:49:25 crc restorecon[4680]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/1e08bfd1.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 10 22:49:25 crc restorecon[4680]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/DigiCert_Global_Root_G3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 10 22:49:25 crc restorecon[4680]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/dd8e9d41.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 10 22:49:25 crc restorecon[4680]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ed39abd0.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 10 22:49:25 crc restorecon[4680]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/a3418fda.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 10 22:49:25 crc restorecon[4680]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/bc3f2570.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 10 22:49:25 crc restorecon[4680]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/DigiCert_High_Assurance_EV_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 10 22:49:25 crc restorecon[4680]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/244b5494.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 10 22:49:25 crc restorecon[4680]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/81b9768f.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 10 22:49:25 crc restorecon[4680]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/GlobalSign.2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 10 22:49:25 crc restorecon[4680]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/4be590e0.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 10 22:49:25 crc restorecon[4680]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/DigiCert_TLS_ECC_P384_Root_G5.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 10 22:49:25 crc restorecon[4680]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9846683b.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 10 22:49:25 crc restorecon[4680]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/252252d2.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 10 22:49:25 crc restorecon[4680]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/1e8e7201.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 10 22:49:25 crc restorecon[4680]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ISRG_Root_X1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 10 22:49:25 crc restorecon[4680]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/DigiCert_TLS_RSA4096_Root_G5.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 10 22:49:25 crc restorecon[4680]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d52c538d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 10 22:49:25 crc restorecon[4680]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/c44cc0c0.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 10 22:49:25 crc restorecon[4680]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/GlobalSign_Root_R46.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 10 22:49:25 crc restorecon[4680]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/DigiCert_Trusted_Root_G4.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 10 22:49:25 crc restorecon[4680]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/75d1b2ed.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 10 22:49:25 crc restorecon[4680]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/a2c66da8.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 10 22:49:25 crc restorecon[4680]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/GTS_Root_R2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 10 22:49:25 crc restorecon[4680]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ecccd8db.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 10 22:49:25 crc restorecon[4680]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Entrust.net_Certification_Authority__2048_.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 10 22:49:25 crc restorecon[4680]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/aee5f10d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 10 22:49:25 crc restorecon[4680]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/3e7271e8.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 10 22:49:25 crc restorecon[4680]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b0e59380.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 10 22:49:25 crc restorecon[4680]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/4c3982f2.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 10 22:49:25 crc restorecon[4680]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Entrust_Root_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 10 22:49:25 crc restorecon[4680]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/6b99d060.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 10 22:49:25 crc restorecon[4680]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/bf64f35b.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 10 22:49:25 crc restorecon[4680]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/0a775a30.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 10 22:49:25 crc restorecon[4680]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/002c0b4f.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 10 22:49:25 crc restorecon[4680]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/cc450945.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 10 22:49:25 crc restorecon[4680]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Entrust_Root_Certification_Authority_-_EC1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 10 22:49:25 crc restorecon[4680]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/106f3e4d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 10 22:49:25 crc restorecon[4680]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b3fb433b.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 10 22:49:25 crc restorecon[4680]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/GlobalSign.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 10 22:49:25 crc restorecon[4680]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/4042bcee.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 10 22:49:25 crc restorecon[4680]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Entrust_Root_Certification_Authority_-_G2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 10 22:49:25 crc restorecon[4680]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/02265526.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 10 22:49:25 crc restorecon[4680]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/455f1b52.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 10 22:49:25 crc restorecon[4680]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/0d69c7e1.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 10 22:49:25 crc restorecon[4680]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9f727ac7.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 10 22:49:25 crc restorecon[4680]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Entrust_Root_Certification_Authority_-_G4.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 10 22:49:25 crc restorecon[4680]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5e98733a.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 10 22:49:25 crc restorecon[4680]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/f0cd152c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 10 22:49:25 crc restorecon[4680]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/dc4d6a89.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 10 22:49:25 crc restorecon[4680]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/6187b673.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 10 22:49:25 crc restorecon[4680]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/FIRMAPROFESIONAL_CA_ROOT-A_WEB.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 10 22:49:25 crc restorecon[4680]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ba8887ce.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 10 22:49:25 crc restorecon[4680]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/068570d1.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 10 22:49:25 crc restorecon[4680]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/f081611a.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 10 22:49:25 crc restorecon[4680]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/48a195d8.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 10 22:49:25 crc restorecon[4680]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/GDCA_TrustAUTH_R5_ROOT.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 10 22:49:25 crc restorecon[4680]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/0f6fa695.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 10 22:49:25 crc restorecon[4680]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ab59055e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 10 22:49:25 crc restorecon[4680]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b92fd57f.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 10 22:49:25 crc restorecon[4680]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/GLOBALTRUST_2020.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 10 22:49:25 crc restorecon[4680]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/fa5da96b.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 10 22:49:25 crc restorecon[4680]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/1ec40989.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 10 22:49:25 crc restorecon[4680]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/7719f463.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 10 22:49:25 crc restorecon[4680]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/GTS_Root_R1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 10 22:49:25 crc restorecon[4680]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/1001acf7.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 10 22:49:25 crc restorecon[4680]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/f013ecaf.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 10 22:49:25 crc restorecon[4680]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/626dceaf.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 10 22:49:25 crc restorecon[4680]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/c559d742.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 10 22:49:25 crc restorecon[4680]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/1d3472b9.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 10 22:49:25 crc restorecon[4680]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9479c8c3.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 10 22:49:25 crc restorecon[4680]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/a81e292b.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 10 22:49:25 crc restorecon[4680]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/4bfab552.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 10 22:49:25 crc restorecon[4680]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Go_Daddy_Class_2_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 10 22:49:25 crc restorecon[4680]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Sectigo_Public_Server_Authentication_Root_E46.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 10 22:49:25 crc restorecon[4680]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Go_Daddy_Root_Certificate_Authority_-_G2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 10 22:49:25 crc restorecon[4680]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e071171e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 10 22:49:25 crc restorecon[4680]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/57bcb2da.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 10 22:49:25 crc restorecon[4680]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/HARICA_TLS_ECC_Root_CA_2021.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 10 22:49:25 crc restorecon[4680]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ab5346f4.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 10 22:49:25 crc restorecon[4680]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5046c355.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 10 22:49:25 crc restorecon[4680]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/HARICA_TLS_RSA_Root_CA_2021.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 10 22:49:25 crc restorecon[4680]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/865fbdf9.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 10 22:49:25 crc restorecon[4680]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/da0cfd1d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 10 22:49:25 crc restorecon[4680]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/85cde254.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 10 22:49:25 crc restorecon[4680]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Hellenic_Academic_and_Research_Institutions_ECC_RootCA_2015.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 10 22:49:25 crc restorecon[4680]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/cbb3f32b.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 10 22:49:25 crc restorecon[4680]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/SecureSign_RootCA11.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 10 22:49:25 crc restorecon[4680]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Hellenic_Academic_and_Research_Institutions_RootCA_2015.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 10 22:49:25 crc restorecon[4680]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5860aaa6.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 10 22:49:25 crc restorecon[4680]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/31188b5e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 10 22:49:25 crc restorecon[4680]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/HiPKI_Root_CA_-_G1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 10 22:49:25 crc restorecon[4680]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/c7f1359b.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 10 22:49:25 crc restorecon[4680]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5f15c80c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 10 22:49:25 crc restorecon[4680]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Hongkong_Post_Root_CA_3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 10 22:49:25 crc restorecon[4680]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/09789157.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 10 22:49:25 crc restorecon[4680]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ISRG_Root_X2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 10 22:49:25 crc restorecon[4680]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/18856ac4.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 10 22:49:25 crc restorecon[4680]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/1e09d511.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 10 22:49:25 crc restorecon[4680]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/IdenTrust_Commercial_Root_CA_1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 10 22:49:25 crc restorecon[4680]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/cf701eeb.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 10 22:49:25 crc restorecon[4680]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d06393bb.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 10 22:49:25 crc restorecon[4680]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/IdenTrust_Public_Sector_Root_CA_1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 10 22:49:25 crc restorecon[4680]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/10531352.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 10 22:49:25 crc restorecon[4680]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Izenpe.com.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 10 22:49:25 crc restorecon[4680]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/SecureTrust_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 10 22:49:25 crc restorecon[4680]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b0ed035a.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 10 22:49:25 crc restorecon[4680]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Microsec_e-Szigno_Root_CA_2009.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 10 22:49:25 crc restorecon[4680]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/8160b96c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 10 22:49:25 crc restorecon[4680]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e8651083.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 10 22:49:25 crc restorecon[4680]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/2c63f966.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 10 22:49:25 crc restorecon[4680]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Security_Communication_RootCA2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 10 22:49:25 crc restorecon[4680]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Microsoft_ECC_Root_Certificate_Authority_2017.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 10 22:49:25 crc restorecon[4680]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/8d89cda1.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 10 22:49:25 crc restorecon[4680]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/01419da9.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 10 22:49:25 crc restorecon[4680]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/SSL.com_TLS_RSA_Root_CA_2022.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 10 22:49:25 crc restorecon[4680]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b7a5b843.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 10 22:49:25 crc restorecon[4680]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Microsoft_RSA_Root_Certificate_Authority_2017.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 10 22:49:25 crc restorecon[4680]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/bf53fb88.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 10 22:49:25 crc restorecon[4680]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9591a472.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 10 22:49:25 crc restorecon[4680]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/3afde786.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 10 22:49:25 crc restorecon[4680]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/SwissSign_Gold_CA_-_G2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 10 22:49:25 crc restorecon[4680]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/NAVER_Global_Root_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 10 22:49:25 crc restorecon[4680]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/3fb36b73.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 10 22:49:25 crc restorecon[4680]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d39b0a2c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 10 22:49:25 crc restorecon[4680]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/a89d74c2.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 10 22:49:25 crc restorecon[4680]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/cd58d51e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 10 22:49:25 crc restorecon[4680]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b7db1890.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 10 22:49:25 crc restorecon[4680]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/NetLock_Arany__Class_Gold__F__tan__s__tv__ny.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 10 22:49:25 crc restorecon[4680]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/988a38cb.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 10 22:49:25 crc restorecon[4680]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/60afe812.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 10 22:49:25 crc restorecon[4680]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/f39fc864.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 10 22:49:25 crc restorecon[4680]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5443e9e3.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 10 22:49:25 crc restorecon[4680]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/OISTE_WISeKey_Global_Root_GB_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 10 22:49:25 crc restorecon[4680]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e73d606e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 10 22:49:25 crc restorecon[4680]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/dfc0fe80.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 10 22:49:25 crc restorecon[4680]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b66938e9.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 10 22:49:25 crc restorecon[4680]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/1e1eab7c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 10 22:49:25 crc restorecon[4680]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/OISTE_WISeKey_Global_Root_GC_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 10 22:49:25 crc restorecon[4680]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/773e07ad.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 10 22:49:25 crc restorecon[4680]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/3c899c73.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 10 22:49:25 crc restorecon[4680]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d59297b8.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 10 22:49:25 crc restorecon[4680]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ddcda989.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 10 22:49:25 crc restorecon[4680]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/QuoVadis_Root_CA_1_G3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 10 22:49:25 crc restorecon[4680]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/749e9e03.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 10 22:49:25 crc restorecon[4680]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/52b525c7.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 10 22:49:25 crc restorecon[4680]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Security_Communication_RootCA3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 10 22:49:25 crc restorecon[4680]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/QuoVadis_Root_CA_2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 10 22:49:25 crc restorecon[4680]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d7e8dc79.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 10 22:49:25 crc restorecon[4680]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/7a819ef2.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 10 22:49:25 crc restorecon[4680]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/08063a00.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 10 22:49:25 crc restorecon[4680]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/6b483515.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 10 22:49:25 crc restorecon[4680]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/QuoVadis_Root_CA_2_G3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 10 22:49:25 crc restorecon[4680]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/064e0aa9.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 10 22:49:25 crc restorecon[4680]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/1f58a078.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 10 22:49:25 crc restorecon[4680]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/6f7454b3.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 10 22:49:25 crc restorecon[4680]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/7fa05551.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 10 22:49:25 crc restorecon[4680]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/QuoVadis_Root_CA_3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 10 22:49:25 crc restorecon[4680]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/76faf6c0.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 10 22:49:25 crc restorecon[4680]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9339512a.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 10 22:49:25 crc restorecon[4680]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/f387163d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 10 22:49:25 crc restorecon[4680]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ee37c333.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 10 22:49:25 crc restorecon[4680]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/QuoVadis_Root_CA_3_G3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 10 22:49:25 crc restorecon[4680]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e18bfb83.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 10 22:49:25 crc restorecon[4680]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e442e424.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 10 22:49:25 crc restorecon[4680]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/fe8a2cd8.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 10 22:49:25 crc restorecon[4680]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/23f4c490.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 10 22:49:25 crc restorecon[4680]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5cd81ad7.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 10 22:49:25 crc restorecon[4680]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/SSL.com_EV_Root_Certification_Authority_ECC.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 10 22:49:25 crc restorecon[4680]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/f0c70a8d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 10 22:49:25 crc restorecon[4680]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/7892ad52.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 10 22:49:25 crc restorecon[4680]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/SZAFIR_ROOT_CA2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 10 22:49:25 crc restorecon[4680]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/4f316efb.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 10 22:49:25 crc restorecon[4680]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/SSL.com_EV_Root_Certification_Authority_RSA_R2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 10 22:49:25 crc restorecon[4680]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/06dc52d5.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 10 22:49:25 crc restorecon[4680]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/583d0756.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 10 22:49:25 crc restorecon[4680]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Sectigo_Public_Server_Authentication_Root_R46.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 10 22:49:25 crc restorecon[4680]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/SSL.com_Root_Certification_Authority_ECC.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 10 22:49:25 crc restorecon[4680]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/0bf05006.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 10 22:49:25 crc restorecon[4680]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/88950faa.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 10 22:49:25 crc restorecon[4680]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9046744a.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 10 22:49:25 crc restorecon[4680]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/3c860d51.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 10 22:49:25 crc restorecon[4680]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/SSL.com_Root_Certification_Authority_RSA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 10 22:49:25 crc restorecon[4680]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/6fa5da56.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 10 22:49:25 crc restorecon[4680]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/33ee480d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 10 22:49:25 crc restorecon[4680]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Secure_Global_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 10 22:49:25 crc restorecon[4680]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/63a2c897.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 10 22:49:25 crc restorecon[4680]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/SSL.com_TLS_ECC_Root_CA_2022.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 10 22:49:25 crc restorecon[4680]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/bdacca6f.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 10 22:49:25 crc restorecon[4680]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ff34af3f.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 10 22:49:25 crc restorecon[4680]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/dbff3a01.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 10 22:49:25 crc restorecon[4680]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Security_Communication_ECC_RootCA1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 10 22:49:25 crc restorecon[4680]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/emSign_Root_CA_-_C1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 10 22:49:25 crc restorecon[4680]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Starfield_Class_2_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 10 22:49:25 crc restorecon[4680]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/406c9bb1.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 10 22:49:25 crc restorecon[4680]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Starfield_Root_Certificate_Authority_-_G2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 10 22:49:25 crc restorecon[4680]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/emSign_ECC_Root_CA_-_C3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 10 22:49:25 crc restorecon[4680]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Starfield_Services_Root_Certificate_Authority_-_G2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 10 22:49:25 crc restorecon[4680]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/SwissSign_Silver_CA_-_G2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 10 22:49:25 crc restorecon[4680]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/99e1b953.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 10 22:49:25 crc restorecon[4680]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/T-TeleSec_GlobalRoot_Class_2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 10 22:49:25 crc restorecon[4680]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/vTrus_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 10 22:49:25 crc restorecon[4680]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/T-TeleSec_GlobalRoot_Class_3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 10 22:49:25 crc restorecon[4680]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/14bc7599.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 10 22:49:25 crc restorecon[4680]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/TUBITAK_Kamu_SM_SSL_Kok_Sertifikasi_-_Surum_1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 10 22:49:25 crc restorecon[4680]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/TWCA_Global_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 10 22:49:25 crc restorecon[4680]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/7a3adc42.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 10 22:49:25 crc restorecon[4680]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/TWCA_Root_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 10 22:49:25 crc restorecon[4680]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/f459871d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 10 22:49:25 crc restorecon[4680]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Telekom_Security_TLS_ECC_Root_2020.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 10 22:49:25 crc restorecon[4680]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/emSign_Root_CA_-_G1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 10 22:49:25 crc restorecon[4680]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Telekom_Security_TLS_RSA_Root_2023.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 10 22:49:25 crc restorecon[4680]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/TeliaSonera_Root_CA_v1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 10 22:49:25 crc restorecon[4680]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Telia_Root_CA_v2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 10 22:49:25 crc restorecon[4680]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/8f103249.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 10 22:49:25 crc restorecon[4680]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/f058632f.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 10 22:49:25 crc restorecon[4680]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ca-certificates.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 10 22:49:25 crc restorecon[4680]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/TrustAsia_Global_Root_CA_G3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 10 22:49:25 crc restorecon[4680]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9bf03295.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 10 22:49:25 crc restorecon[4680]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/98aaf404.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 10 22:49:25 crc restorecon[4680]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 10 22:49:25 crc restorecon[4680]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/TrustAsia_Global_Root_CA_G4.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 10 22:49:25 crc restorecon[4680]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/1cef98f5.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 10 22:49:25 crc restorecon[4680]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/073bfcc5.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 10 22:49:25 crc restorecon[4680]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/2923b3f9.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 10 22:49:25 crc restorecon[4680]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Trustwave_Global_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 10 22:49:25 crc restorecon[4680]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/f249de83.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 10 22:49:25 crc restorecon[4680]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/edcbddb5.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 10 22:49:25 crc restorecon[4680]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/emSign_ECC_Root_CA_-_G3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 10 22:49:25 crc restorecon[4680]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Trustwave_Global_ECC_P256_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 10 22:49:25 crc restorecon[4680]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9b5697b0.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 10 22:49:25 crc restorecon[4680]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/1ae85e5e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 10 22:49:25 crc restorecon[4680]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b74d2bd5.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 10 22:49:25 crc restorecon[4680]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Trustwave_Global_ECC_P384_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 10 22:49:25 crc restorecon[4680]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d887a5bb.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 10 22:49:25 crc restorecon[4680]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9aef356c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 10 22:49:25 crc restorecon[4680]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/TunTrust_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 10 22:49:25 crc restorecon[4680]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/fd64f3fc.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 10 22:49:25 crc restorecon[4680]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e13665f9.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 10 22:49:25 crc restorecon[4680]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/UCA_Extended_Validation_Root.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 10 22:49:25 crc restorecon[4680]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/0f5dc4f3.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 10 22:49:25 crc restorecon[4680]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/da7377f6.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 10 22:49:25 crc restorecon[4680]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/UCA_Global_G2_Root.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 10 22:49:25 crc restorecon[4680]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/c01eb047.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 10 22:49:25 crc restorecon[4680]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/304d27c3.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 10 22:49:25 crc restorecon[4680]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ed858448.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 10 22:49:25 crc restorecon[4680]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/USERTrust_ECC_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 10 22:49:25 crc restorecon[4680]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/f30dd6ad.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 10 22:49:25 crc restorecon[4680]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/04f60c28.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 10 22:49:25 crc restorecon[4680]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/vTrus_ECC_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 10 22:49:25 crc restorecon[4680]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/USERTrust_RSA_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 10 22:49:25 crc restorecon[4680]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/fc5a8f99.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 10 22:49:25 crc restorecon[4680]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/35105088.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 10 22:49:25 crc restorecon[4680]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ee532fd5.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 10 22:49:25 crc restorecon[4680]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/XRamp_Global_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 10 22:49:25 crc restorecon[4680]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/706f604c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 10 22:49:25 crc restorecon[4680]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/76579174.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 10 22:49:25 crc restorecon[4680]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/certSIGN_ROOT_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 10 22:49:25 crc restorecon[4680]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/8d86cdd1.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 10 22:49:25 crc restorecon[4680]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/882de061.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 10 22:49:25 crc restorecon[4680]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/certSIGN_ROOT_CA_G2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 10 22:49:25 crc restorecon[4680]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5f618aec.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 10 22:49:25 crc restorecon[4680]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/a9d40e02.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 10 22:49:25 crc restorecon[4680]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e-Szigno_Root_CA_2017.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 10 22:49:25 crc restorecon[4680]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e868b802.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 10 22:49:25 crc restorecon[4680]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/83e9984f.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 10 22:49:25 crc restorecon[4680]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ePKI_Root_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 10 22:49:25 crc restorecon[4680]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ca6e4ad9.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 10 22:49:25 crc restorecon[4680]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9d6523ce.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 10 22:49:25 crc restorecon[4680]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/4b718d9b.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 10 22:49:25 crc restorecon[4680]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/869fbf79.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 10 22:49:25 crc restorecon[4680]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 10 22:49:25 crc restorecon[4680]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/containers/registry/f8d22bdb not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 10 22:49:25 crc restorecon[4680]: /var/lib/kubelet/pods/a0128f3a-b052-44ed-a84e-c4c8aaf17c13/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c17 Dec 10 22:49:25 crc restorecon[4680]: /var/lib/kubelet/pods/a0128f3a-b052-44ed-a84e-c4c8aaf17c13/containers/cluster-samples-operator/6e8bbfac not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c17 Dec 10 22:49:25 crc restorecon[4680]: /var/lib/kubelet/pods/a0128f3a-b052-44ed-a84e-c4c8aaf17c13/containers/cluster-samples-operator/54dd7996 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c17 Dec 10 22:49:25 crc restorecon[4680]: /var/lib/kubelet/pods/a0128f3a-b052-44ed-a84e-c4c8aaf17c13/containers/cluster-samples-operator/a4f1bb05 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c17 Dec 10 22:49:25 crc restorecon[4680]: /var/lib/kubelet/pods/a0128f3a-b052-44ed-a84e-c4c8aaf17c13/containers/cluster-samples-operator-watch/207129da not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c17 Dec 10 22:49:25 crc restorecon[4680]: /var/lib/kubelet/pods/a0128f3a-b052-44ed-a84e-c4c8aaf17c13/containers/cluster-samples-operator-watch/c1df39e1 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c17 Dec 10 22:49:25 crc restorecon[4680]: /var/lib/kubelet/pods/a0128f3a-b052-44ed-a84e-c4c8aaf17c13/containers/cluster-samples-operator-watch/15b8f1cd not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c17 Dec 10 22:49:25 crc restorecon[4680]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Dec 10 22:49:25 crc restorecon[4680]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/volumes/kubernetes.io~configmap/config/..2025_02_23_05_27_49.3523263858 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Dec 10 22:49:25 crc restorecon[4680]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/volumes/kubernetes.io~configmap/config/..2025_02_23_05_27_49.3523263858/config-file.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Dec 10 22:49:25 crc restorecon[4680]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Dec 10 22:49:25 crc restorecon[4680]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/volumes/kubernetes.io~configmap/config/config-file.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Dec 10 22:49:25 crc restorecon[4680]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/volumes/kubernetes.io~configmap/images not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Dec 10 22:49:25 crc restorecon[4680]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/volumes/kubernetes.io~configmap/images/..2025_02_23_05_27_49.3256605594 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Dec 10 22:49:25 crc restorecon[4680]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/volumes/kubernetes.io~configmap/images/..2025_02_23_05_27_49.3256605594/images.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Dec 10 22:49:25 crc restorecon[4680]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/volumes/kubernetes.io~configmap/images/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Dec 10 22:49:25 crc restorecon[4680]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/volumes/kubernetes.io~configmap/images/images.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Dec 10 22:49:25 crc restorecon[4680]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Dec 10 22:49:25 crc restorecon[4680]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/containers/kube-rbac-proxy/77bd6913 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Dec 10 22:49:25 crc restorecon[4680]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/containers/kube-rbac-proxy/2382c1b1 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Dec 10 22:49:25 crc restorecon[4680]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/containers/kube-rbac-proxy/704ce128 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Dec 10 22:49:25 crc restorecon[4680]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/containers/machine-api-operator/70d16fe0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Dec 10 22:49:25 crc restorecon[4680]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/containers/machine-api-operator/bfb95535 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Dec 10 22:49:25 crc restorecon[4680]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/containers/machine-api-operator/57a8e8e2 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Dec 10 22:49:25 crc restorecon[4680]: /var/lib/kubelet/pods/e7e6199b-1264-4501-8953-767f51328d08/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c219,c404 Dec 10 22:49:25 crc restorecon[4680]: /var/lib/kubelet/pods/e7e6199b-1264-4501-8953-767f51328d08/volumes/kubernetes.io~configmap/config/..2025_02_23_05_27_49.3413793711 not reset as customized by admin to system_u:object_r:container_file_t:s0:c219,c404 Dec 10 22:49:25 crc restorecon[4680]: /var/lib/kubelet/pods/e7e6199b-1264-4501-8953-767f51328d08/volumes/kubernetes.io~configmap/config/..2025_02_23_05_27_49.3413793711/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c219,c404 Dec 10 22:49:25 crc restorecon[4680]: /var/lib/kubelet/pods/e7e6199b-1264-4501-8953-767f51328d08/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c219,c404 Dec 10 22:49:25 crc restorecon[4680]: /var/lib/kubelet/pods/e7e6199b-1264-4501-8953-767f51328d08/volumes/kubernetes.io~configmap/config/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c219,c404 Dec 10 22:49:25 crc restorecon[4680]: /var/lib/kubelet/pods/e7e6199b-1264-4501-8953-767f51328d08/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c219,c404 Dec 10 22:49:25 crc restorecon[4680]: /var/lib/kubelet/pods/e7e6199b-1264-4501-8953-767f51328d08/containers/kube-apiserver-operator/1b9d3e5e not reset as customized by admin to system_u:object_r:container_file_t:s0:c107,c917 Dec 10 22:49:25 crc restorecon[4680]: /var/lib/kubelet/pods/e7e6199b-1264-4501-8953-767f51328d08/containers/kube-apiserver-operator/fddb173c not reset as customized by admin to system_u:object_r:container_file_t:s0:c202,c983 Dec 10 22:49:25 crc restorecon[4680]: /var/lib/kubelet/pods/e7e6199b-1264-4501-8953-767f51328d08/containers/kube-apiserver-operator/95d3c6c4 not reset as customized by admin to system_u:object_r:container_file_t:s0:c219,c404 Dec 10 22:49:25 crc restorecon[4680]: /var/lib/kubelet/pods/9d751cbb-f2e2-430d-9754-c882a5e924a5/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c20,c21 Dec 10 22:49:25 crc restorecon[4680]: /var/lib/kubelet/pods/9d751cbb-f2e2-430d-9754-c882a5e924a5/containers/check-endpoints/bfb5fff5 not reset as customized by admin to system_u:object_r:container_file_t:s0:c20,c21 Dec 10 22:49:25 crc restorecon[4680]: /var/lib/kubelet/pods/9d751cbb-f2e2-430d-9754-c882a5e924a5/containers/check-endpoints/2aef40aa not reset as customized by admin to system_u:object_r:container_file_t:s0:c20,c21 Dec 10 22:49:25 crc restorecon[4680]: /var/lib/kubelet/pods/9d751cbb-f2e2-430d-9754-c882a5e924a5/containers/check-endpoints/c0391cad not reset as customized by admin to system_u:object_r:container_file_t:s0:c20,c21 Dec 10 22:49:25 crc restorecon[4680]: /var/lib/kubelet/pods/f614b9022728cf315e60c057852e563e/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c214,c928 Dec 10 22:49:25 crc restorecon[4680]: /var/lib/kubelet/pods/f614b9022728cf315e60c057852e563e/containers/kube-controller-manager/1119e69d not reset as customized by admin to system_u:object_r:container_file_t:s0:c776,c1007 Dec 10 22:49:25 crc restorecon[4680]: /var/lib/kubelet/pods/f614b9022728cf315e60c057852e563e/containers/kube-controller-manager/660608b4 not reset as customized by admin to system_u:object_r:container_file_t:s0:c214,c928 Dec 10 22:49:25 crc restorecon[4680]: /var/lib/kubelet/pods/f614b9022728cf315e60c057852e563e/containers/kube-controller-manager/8220bd53 not reset as customized by admin to system_u:object_r:container_file_t:s0:c214,c928 Dec 10 22:49:25 crc restorecon[4680]: /var/lib/kubelet/pods/f614b9022728cf315e60c057852e563e/containers/cluster-policy-controller/85f99d5c not reset as customized by admin to system_u:object_r:container_file_t:s0:c776,c1007 Dec 10 22:49:25 crc restorecon[4680]: /var/lib/kubelet/pods/f614b9022728cf315e60c057852e563e/containers/cluster-policy-controller/4b0225f6 not reset as customized by admin to system_u:object_r:container_file_t:s0:c214,c928 Dec 10 22:49:25 crc restorecon[4680]: /var/lib/kubelet/pods/f614b9022728cf315e60c057852e563e/containers/kube-controller-manager-cert-syncer/9c2a3394 not reset as customized by admin to system_u:object_r:container_file_t:s0:c776,c1007 Dec 10 22:49:25 crc restorecon[4680]: /var/lib/kubelet/pods/f614b9022728cf315e60c057852e563e/containers/kube-controller-manager-cert-syncer/e820b243 not reset as customized by admin to system_u:object_r:container_file_t:s0:c214,c928 Dec 10 22:49:25 crc restorecon[4680]: /var/lib/kubelet/pods/f614b9022728cf315e60c057852e563e/containers/kube-controller-manager-recovery-controller/1ca52ea0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c776,c1007 Dec 10 22:49:25 crc restorecon[4680]: /var/lib/kubelet/pods/f614b9022728cf315e60c057852e563e/containers/kube-controller-manager-recovery-controller/e6988e45 not reset as customized by admin to system_u:object_r:container_file_t:s0:c214,c928 Dec 10 22:49:25 crc restorecon[4680]: /var/lib/kubelet/pods/0b574797-001e-440a-8f4e-c0be86edad0f/volumes/kubernetes.io~configmap/mcc-auth-proxy-config not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Dec 10 22:49:25 crc restorecon[4680]: /var/lib/kubelet/pods/0b574797-001e-440a-8f4e-c0be86edad0f/volumes/kubernetes.io~configmap/mcc-auth-proxy-config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Dec 10 22:49:25 crc restorecon[4680]: /var/lib/kubelet/pods/0b574797-001e-440a-8f4e-c0be86edad0f/volumes/kubernetes.io~configmap/mcc-auth-proxy-config/config-file.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Dec 10 22:49:25 crc restorecon[4680]: /var/lib/kubelet/pods/0b574797-001e-440a-8f4e-c0be86edad0f/volumes/kubernetes.io~configmap/mcc-auth-proxy-config/..2025_02_24_06_09_21.2517297950 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Dec 10 22:49:25 crc restorecon[4680]: /var/lib/kubelet/pods/0b574797-001e-440a-8f4e-c0be86edad0f/volumes/kubernetes.io~configmap/mcc-auth-proxy-config/..2025_02_24_06_09_21.2517297950/config-file.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Dec 10 22:49:25 crc restorecon[4680]: /var/lib/kubelet/pods/0b574797-001e-440a-8f4e-c0be86edad0f/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Dec 10 22:49:25 crc restorecon[4680]: /var/lib/kubelet/pods/0b574797-001e-440a-8f4e-c0be86edad0f/containers/machine-config-controller/6655f00b not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Dec 10 22:49:25 crc restorecon[4680]: /var/lib/kubelet/pods/0b574797-001e-440a-8f4e-c0be86edad0f/containers/machine-config-controller/98bc3986 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Dec 10 22:49:25 crc restorecon[4680]: /var/lib/kubelet/pods/0b574797-001e-440a-8f4e-c0be86edad0f/containers/machine-config-controller/08e3458a not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Dec 10 22:49:25 crc restorecon[4680]: /var/lib/kubelet/pods/0b574797-001e-440a-8f4e-c0be86edad0f/containers/kube-rbac-proxy/2a191cb0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Dec 10 22:49:25 crc restorecon[4680]: /var/lib/kubelet/pods/0b574797-001e-440a-8f4e-c0be86edad0f/containers/kube-rbac-proxy/6c4eeefb not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Dec 10 22:49:25 crc restorecon[4680]: /var/lib/kubelet/pods/0b574797-001e-440a-8f4e-c0be86edad0f/containers/kube-rbac-proxy/f61a549c not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Dec 10 22:49:25 crc restorecon[4680]: /var/lib/kubelet/pods/bd23aa5c-e532-4e53-bccf-e79f130c5ae8/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c318,c553 Dec 10 22:49:25 crc restorecon[4680]: /var/lib/kubelet/pods/bd23aa5c-e532-4e53-bccf-e79f130c5ae8/containers/hostpath-provisioner/24891863 not reset as customized by admin to system_u:object_r:container_file_t:s0:c37,c572 Dec 10 22:49:25 crc restorecon[4680]: /var/lib/kubelet/pods/bd23aa5c-e532-4e53-bccf-e79f130c5ae8/containers/hostpath-provisioner/fbdfd89c not reset as customized by admin to system_u:object_r:container_file_t:s0:c318,c553 Dec 10 22:49:25 crc restorecon[4680]: /var/lib/kubelet/pods/bd23aa5c-e532-4e53-bccf-e79f130c5ae8/containers/liveness-probe/9b63b3bc not reset as customized by admin to system_u:object_r:container_file_t:s0:c37,c572 Dec 10 22:49:25 crc restorecon[4680]: /var/lib/kubelet/pods/bd23aa5c-e532-4e53-bccf-e79f130c5ae8/containers/liveness-probe/8acde6d6 not reset as customized by admin to system_u:object_r:container_file_t:s0:c318,c553 Dec 10 22:49:25 crc restorecon[4680]: /var/lib/kubelet/pods/bd23aa5c-e532-4e53-bccf-e79f130c5ae8/containers/node-driver-registrar/59ecbba3 not reset as customized by admin to system_u:object_r:container_file_t:s0:c318,c553 Dec 10 22:49:25 crc restorecon[4680]: /var/lib/kubelet/pods/bd23aa5c-e532-4e53-bccf-e79f130c5ae8/containers/csi-provisioner/685d4be3 not reset as customized by admin to system_u:object_r:container_file_t:s0:c318,c553 Dec 10 22:49:25 crc restorecon[4680]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Dec 10 22:49:25 crc restorecon[4680]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/config/..2025_02_24_06_20_07.341639300 not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Dec 10 22:49:25 crc restorecon[4680]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/config/..2025_02_24_06_20_07.341639300/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Dec 10 22:49:25 crc restorecon[4680]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/config/..2025_02_24_06_20_07.341639300/openshift-route-controller-manager.client-ca.configmap not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Dec 10 22:49:25 crc restorecon[4680]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/config/..2025_02_24_06_20_07.341639300/openshift-route-controller-manager.serving-cert.secret not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Dec 10 22:49:25 crc restorecon[4680]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Dec 10 22:49:25 crc restorecon[4680]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/config/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Dec 10 22:49:25 crc restorecon[4680]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/config/openshift-route-controller-manager.client-ca.configmap not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Dec 10 22:49:25 crc restorecon[4680]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/config/openshift-route-controller-manager.serving-cert.secret not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Dec 10 22:49:25 crc restorecon[4680]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/client-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Dec 10 22:49:25 crc restorecon[4680]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/client-ca/..2025_02_24_06_20_07.2950937851 not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Dec 10 22:49:25 crc restorecon[4680]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/client-ca/..2025_02_24_06_20_07.2950937851/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Dec 10 22:49:25 crc restorecon[4680]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/client-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Dec 10 22:49:25 crc restorecon[4680]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/client-ca/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Dec 10 22:49:25 crc restorecon[4680]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Dec 10 22:49:25 crc restorecon[4680]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/containers/route-controller-manager/feaea55e not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Dec 10 22:49:25 crc restorecon[4680]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 22:49:25 crc restorecon[4680]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 22:49:25 crc restorecon[4680]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/abinitio-runtime-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 22:49:25 crc restorecon[4680]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/abinitio-runtime-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 22:49:25 crc restorecon[4680]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/accuknox-operator-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 22:49:25 crc restorecon[4680]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/accuknox-operator-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 22:49:25 crc restorecon[4680]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aci-containers-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 22:49:25 crc restorecon[4680]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aci-containers-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 22:49:25 crc restorecon[4680]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aikit-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 22:49:25 crc restorecon[4680]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aikit-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 22:49:25 crc restorecon[4680]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/airlock-microgateway not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 22:49:25 crc restorecon[4680]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/airlock-microgateway/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 22:49:25 crc restorecon[4680]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ako-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 22:49:25 crc restorecon[4680]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ako-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 22:49:25 crc restorecon[4680]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/alloy not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 22:49:25 crc restorecon[4680]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/alloy/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 22:49:25 crc restorecon[4680]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anchore-engine not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 22:49:25 crc restorecon[4680]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anchore-engine/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 22:49:25 crc restorecon[4680]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anzo-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 22:49:25 crc restorecon[4680]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anzo-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 22:49:25 crc restorecon[4680]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anzograph-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 22:49:25 crc restorecon[4680]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anzograph-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 22:49:25 crc restorecon[4680]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anzounstructured-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 22:49:25 crc restorecon[4680]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anzounstructured-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 22:49:25 crc restorecon[4680]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/appdynamics-cloud-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 22:49:25 crc restorecon[4680]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/appdynamics-cloud-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 22:49:25 crc restorecon[4680]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/appdynamics-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 22:49:25 crc restorecon[4680]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/appdynamics-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 22:49:25 crc restorecon[4680]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aqua-operator-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 22:49:25 crc restorecon[4680]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aqua-operator-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 22:49:25 crc restorecon[4680]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cass-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 22:49:25 crc restorecon[4680]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cass-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 22:49:25 crc restorecon[4680]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ccm-node-agent-dcap-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 22:49:25 crc restorecon[4680]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ccm-node-agent-dcap-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 22:49:25 crc restorecon[4680]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ccm-node-agent-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 22:49:25 crc restorecon[4680]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ccm-node-agent-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 22:49:25 crc restorecon[4680]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cfm-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 22:49:25 crc restorecon[4680]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cfm-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 22:49:25 crc restorecon[4680]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cilium not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 22:49:25 crc restorecon[4680]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cilium/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 22:49:25 crc restorecon[4680]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cilium-enterprise not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 22:49:25 crc restorecon[4680]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cilium-enterprise/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 22:49:25 crc restorecon[4680]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cloud-native-postgresql not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 22:49:25 crc restorecon[4680]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cloud-native-postgresql/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 22:49:25 crc restorecon[4680]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cloudbees-ci not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 22:49:25 crc restorecon[4680]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cloudbees-ci/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 22:49:25 crc restorecon[4680]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cloudera-streams-messaging-kubernetes-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 22:49:25 crc restorecon[4680]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cloudera-streams-messaging-kubernetes-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 22:49:25 crc restorecon[4680]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cloudnative-pg not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 22:49:25 crc restorecon[4680]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cloudnative-pg/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 22:49:25 crc restorecon[4680]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cnfv-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 22:49:25 crc restorecon[4680]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cnfv-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 22:49:25 crc restorecon[4680]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cockroachdb-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 22:49:25 crc restorecon[4680]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cockroachdb-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 22:49:25 crc restorecon[4680]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/conjur-follower-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 22:49:25 crc restorecon[4680]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/conjur-follower-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 22:49:25 crc restorecon[4680]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/coroot-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 22:49:25 crc restorecon[4680]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/coroot-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 22:49:25 crc restorecon[4680]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/crunchy-postgres-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 22:49:25 crc restorecon[4680]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/crunchy-postgres-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 22:49:25 crc restorecon[4680]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cte-k8s-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 22:49:25 crc restorecon[4680]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cte-k8s-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 22:49:25 crc restorecon[4680]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/datadog-operator-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 22:49:25 crc restorecon[4680]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/datadog-operator-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 22:49:25 crc restorecon[4680]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dell-csm-operator-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 22:49:25 crc restorecon[4680]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dell-csm-operator-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 22:49:25 crc restorecon[4680]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/digitalai-deploy-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 22:49:25 crc restorecon[4680]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/digitalai-deploy-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 22:49:25 crc restorecon[4680]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/digitalai-release-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 22:49:25 crc restorecon[4680]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/digitalai-release-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 22:49:25 crc restorecon[4680]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dynatrace-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 22:49:25 crc restorecon[4680]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dynatrace-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 22:49:25 crc restorecon[4680]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/edb-hcp-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 22:49:25 crc restorecon[4680]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/edb-hcp-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 22:49:25 crc restorecon[4680]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/eginnovations-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 22:49:25 crc restorecon[4680]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/eginnovations-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 22:49:25 crc restorecon[4680]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/elasticsearch-eck-operator-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 22:49:25 crc restorecon[4680]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/elasticsearch-eck-operator-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 22:49:25 crc restorecon[4680]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/falcon-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 22:49:25 crc restorecon[4680]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/falcon-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 22:49:25 crc restorecon[4680]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/federatorai-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 22:49:25 crc restorecon[4680]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/federatorai-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 22:49:25 crc restorecon[4680]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fujitsu-enterprise-postgres-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 22:49:25 crc restorecon[4680]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fujitsu-enterprise-postgres-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 22:49:25 crc restorecon[4680]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/function-mesh not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 22:49:25 crc restorecon[4680]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/function-mesh/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 22:49:25 crc restorecon[4680]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/harness-gitops-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 22:49:25 crc restorecon[4680]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/harness-gitops-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 22:49:25 crc restorecon[4680]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hazelcast-platform-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 22:49:25 crc restorecon[4680]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hazelcast-platform-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 22:49:25 crc restorecon[4680]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hcp-terraform-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 22:49:25 crc restorecon[4680]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hcp-terraform-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 22:49:25 crc restorecon[4680]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hpe-ezmeral-csi-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 22:49:25 crc restorecon[4680]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hpe-ezmeral-csi-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 22:49:25 crc restorecon[4680]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-application-gateway-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 22:49:25 crc restorecon[4680]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-application-gateway-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 22:49:25 crc restorecon[4680]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-block-csi-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 22:49:25 crc restorecon[4680]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-block-csi-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 22:49:25 crc restorecon[4680]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-security-verify-access-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 22:49:25 crc restorecon[4680]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-security-verify-access-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 22:49:25 crc restorecon[4680]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-security-verify-directory-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 22:49:25 crc restorecon[4680]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-security-verify-directory-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 22:49:25 crc restorecon[4680]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-security-verify-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 22:49:25 crc restorecon[4680]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-security-verify-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 22:49:25 crc restorecon[4680]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/infoscale-dr-manager not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 22:49:25 crc restorecon[4680]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/infoscale-dr-manager/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 22:49:25 crc restorecon[4680]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/infoscale-licensing-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 22:49:25 crc restorecon[4680]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/infoscale-licensing-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 22:49:25 crc restorecon[4680]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/infoscale-sds-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 22:49:25 crc restorecon[4680]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/infoscale-sds-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 22:49:25 crc restorecon[4680]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/infrastructure-asset-orchestrator-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 22:49:25 crc restorecon[4680]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/infrastructure-asset-orchestrator-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 22:49:25 crc restorecon[4680]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/instana-agent-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 22:49:25 crc restorecon[4680]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/instana-agent-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 22:49:25 crc restorecon[4680]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/intel-device-plugins-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 22:49:25 crc restorecon[4680]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/intel-device-plugins-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 22:49:25 crc restorecon[4680]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/intel-kubernetes-power-manager not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 22:49:25 crc restorecon[4680]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/intel-kubernetes-power-manager/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 22:49:25 crc restorecon[4680]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/iomesh-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 22:49:25 crc restorecon[4680]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/iomesh-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 22:49:25 crc restorecon[4680]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/joget-dx-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 22:49:25 crc restorecon[4680]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/joget-dx-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 22:49:25 crc restorecon[4680]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/joget-dx8-openshift-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 22:49:25 crc restorecon[4680]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/joget-dx8-openshift-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 22:49:25 crc restorecon[4680]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/joget-dx8-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 22:49:25 crc restorecon[4680]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/joget-dx8-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 22:49:25 crc restorecon[4680]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/k8s-triliovault not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 22:49:25 crc restorecon[4680]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/k8s-triliovault/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 22:49:25 crc restorecon[4680]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-ati-updates not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 22:49:25 crc restorecon[4680]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-ati-updates/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 22:49:25 crc restorecon[4680]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-kcos-framework not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 22:49:25 crc restorecon[4680]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-kcos-framework/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 22:49:25 crc restorecon[4680]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-kcos-ingress not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 22:49:25 crc restorecon[4680]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-kcos-ingress/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 22:49:25 crc restorecon[4680]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-kcos-licensing not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 22:49:25 crc restorecon[4680]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-kcos-licensing/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 22:49:25 crc restorecon[4680]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-kcos-sso not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 22:49:25 crc restorecon[4680]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-kcos-sso/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 22:49:25 crc restorecon[4680]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-keycloak-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 22:49:25 crc restorecon[4680]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-keycloak-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 22:49:25 crc restorecon[4680]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-load-core not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 22:49:25 crc restorecon[4680]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-load-core/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 22:49:25 crc restorecon[4680]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-loadcore-agents not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 22:49:25 crc restorecon[4680]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-loadcore-agents/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 22:49:25 crc restorecon[4680]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-nats-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 22:49:25 crc restorecon[4680]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-nats-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 22:49:25 crc restorecon[4680]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-nimbusmosaic-dusim not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 22:49:25 crc restorecon[4680]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-nimbusmosaic-dusim/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 22:49:25 crc restorecon[4680]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-rest-api-browser-v1 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 22:49:25 crc restorecon[4680]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-rest-api-browser-v1/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 22:49:25 crc restorecon[4680]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-appsec not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 22:49:25 crc restorecon[4680]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-appsec/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 22:49:25 crc restorecon[4680]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-core not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 22:49:25 crc restorecon[4680]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-core/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 22:49:25 crc restorecon[4680]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-db not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 22:49:25 crc restorecon[4680]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-db/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 22:49:25 crc restorecon[4680]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-diagnostics not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 22:49:25 crc restorecon[4680]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-diagnostics/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 22:49:25 crc restorecon[4680]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-logging not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 22:49:25 crc restorecon[4680]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-logging/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 22:49:25 crc restorecon[4680]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-migration not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 22:49:25 crc restorecon[4680]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-migration/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 22:49:25 crc restorecon[4680]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-msg-broker not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 22:49:25 crc restorecon[4680]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-msg-broker/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 22:49:25 crc restorecon[4680]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-notifications not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 22:49:25 crc restorecon[4680]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-notifications/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 22:49:25 crc restorecon[4680]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-stats-dashboards not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 22:49:25 crc restorecon[4680]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-stats-dashboards/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 22:49:25 crc restorecon[4680]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-storage not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 22:49:25 crc restorecon[4680]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-storage/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 22:49:25 crc restorecon[4680]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-test-core not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 22:49:25 crc restorecon[4680]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-test-core/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 22:49:25 crc restorecon[4680]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-ui not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 22:49:25 crc restorecon[4680]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-ui/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 22:49:25 crc restorecon[4680]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-websocket-service not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 22:49:25 crc restorecon[4680]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-websocket-service/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 22:49:25 crc restorecon[4680]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kong-gateway-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 22:49:25 crc restorecon[4680]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kong-gateway-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 22:49:25 crc restorecon[4680]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubearmor-operator-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 22:49:25 crc restorecon[4680]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubearmor-operator-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 22:49:25 crc restorecon[4680]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubecost-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 22:49:25 crc restorecon[4680]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubecost-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 22:49:25 crc restorecon[4680]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubemq-operator-marketplace not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 22:49:25 crc restorecon[4680]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubemq-operator-marketplace/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 22:49:25 crc restorecon[4680]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubeturbo-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 22:49:25 crc restorecon[4680]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubeturbo-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 22:49:25 crc restorecon[4680]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/lenovo-locd-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 22:49:25 crc restorecon[4680]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/lenovo-locd-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 22:49:25 crc restorecon[4680]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/marketplace-games-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 22:49:25 crc restorecon[4680]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/marketplace-games-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 22:49:25 crc restorecon[4680]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/memcached-operator-ogaye not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 22:49:25 crc restorecon[4680]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/memcached-operator-ogaye/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 22:49:25 crc restorecon[4680]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/memory-machine-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 22:49:25 crc restorecon[4680]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/memory-machine-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 22:49:25 crc restorecon[4680]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/model-builder-for-vision-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 22:49:25 crc restorecon[4680]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/model-builder-for-vision-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 22:49:25 crc restorecon[4680]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mongodb-atlas-kubernetes not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 22:49:25 crc restorecon[4680]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mongodb-atlas-kubernetes/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 22:49:25 crc restorecon[4680]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mongodb-enterprise not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 22:49:25 crc restorecon[4680]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mongodb-enterprise/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 22:49:25 crc restorecon[4680]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/netapp-spark-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 22:49:25 crc restorecon[4680]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/netapp-spark-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 22:49:25 crc restorecon[4680]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/netscaler-adm-agent-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 22:49:25 crc restorecon[4680]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/netscaler-adm-agent-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 22:49:25 crc restorecon[4680]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/netscaler-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 22:49:25 crc restorecon[4680]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/netscaler-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 22:49:25 crc restorecon[4680]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/neuvector-certified-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 22:49:25 crc restorecon[4680]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/neuvector-certified-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 22:49:25 crc restorecon[4680]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nexus-repository-ha-operator-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 22:49:25 crc restorecon[4680]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nexus-repository-ha-operator-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 22:49:25 crc restorecon[4680]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nginx-ingress-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 22:49:25 crc restorecon[4680]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nginx-ingress-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 22:49:25 crc restorecon[4680]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pcc-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 22:49:25 crc restorecon[4680]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pcc-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 22:49:25 crc restorecon[4680]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nim-operator-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 22:49:25 crc restorecon[4680]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nim-operator-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 22:49:25 crc restorecon[4680]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nxiq-operator-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 22:49:25 crc restorecon[4680]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nxiq-operator-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 22:49:25 crc restorecon[4680]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nxrm-operator-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 22:49:25 crc restorecon[4680]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nxrm-operator-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 22:49:25 crc restorecon[4680]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odigos-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 22:49:25 crc restorecon[4680]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odigos-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 22:49:25 crc restorecon[4680]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/open-liberty-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 22:49:25 crc restorecon[4680]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/open-liberty-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 22:49:25 crc restorecon[4680]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshiftartifactoryha-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 22:49:25 crc restorecon[4680]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshiftartifactoryha-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 22:49:25 crc restorecon[4680]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshiftxray-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 22:49:25 crc restorecon[4680]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshiftxray-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 22:49:25 crc restorecon[4680]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/operator-certification-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 22:49:25 crc restorecon[4680]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/operator-certification-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 22:49:25 crc restorecon[4680]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ovms-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 22:49:25 crc restorecon[4680]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ovms-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 22:49:25 crc restorecon[4680]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pachyderm-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 22:49:25 crc restorecon[4680]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pachyderm-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 22:49:25 crc restorecon[4680]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pmem-csi-operator-os not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 22:49:25 crc restorecon[4680]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pmem-csi-operator-os/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 22:49:25 crc restorecon[4680]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/portworx-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 22:49:25 crc restorecon[4680]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/portworx-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 22:49:25 crc restorecon[4680]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/prometurbo-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 22:49:25 crc restorecon[4680]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/prometurbo-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 22:49:25 crc restorecon[4680]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pubsubplus-eventbroker-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 22:49:25 crc restorecon[4680]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pubsubplus-eventbroker-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 22:49:25 crc restorecon[4680]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/redis-enterprise-operator-cert not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 22:49:25 crc restorecon[4680]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/redis-enterprise-operator-cert/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 22:49:25 crc restorecon[4680]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/runtime-component-operator-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 22:49:25 crc restorecon[4680]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/runtime-component-operator-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 22:49:25 crc restorecon[4680]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/runtime-fabric-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 22:49:25 crc restorecon[4680]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/runtime-fabric-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 22:49:25 crc restorecon[4680]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sanstoragecsi-operator-bundle not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 22:49:25 crc restorecon[4680]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sanstoragecsi-operator-bundle/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 22:49:25 crc restorecon[4680]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/silicom-sts-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 22:49:25 crc restorecon[4680]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/silicom-sts-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 22:49:25 crc restorecon[4680]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/smilecdr-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 22:49:25 crc restorecon[4680]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/smilecdr-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 22:49:25 crc restorecon[4680]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sriov-fec not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 22:49:25 crc restorecon[4680]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sriov-fec/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 22:49:25 crc restorecon[4680]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/stackable-commons-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 22:49:25 crc restorecon[4680]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/stackable-commons-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 22:49:25 crc restorecon[4680]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/stackable-zookeeper-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 22:49:25 crc restorecon[4680]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/stackable-zookeeper-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 22:49:25 crc restorecon[4680]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/t8c-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 22:49:25 crc restorecon[4680]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/t8c-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 22:49:25 crc restorecon[4680]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/t8c-tsc-client-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 22:49:25 crc restorecon[4680]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/t8c-tsc-client-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 22:49:25 crc restorecon[4680]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tawon-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 22:49:25 crc restorecon[4680]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tawon-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 22:49:25 crc restorecon[4680]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tigera-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 22:49:25 crc restorecon[4680]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tigera-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 22:49:25 crc restorecon[4680]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/timemachine-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 22:49:25 crc restorecon[4680]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/timemachine-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 22:49:25 crc restorecon[4680]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/vault-secrets-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 22:49:25 crc restorecon[4680]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/vault-secrets-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 22:49:25 crc restorecon[4680]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/vcp-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 22:49:25 crc restorecon[4680]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/vcp-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 22:49:25 crc restorecon[4680]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/webotx-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 22:49:25 crc restorecon[4680]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/webotx-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 22:49:25 crc restorecon[4680]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/xcrypt-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 22:49:25 crc restorecon[4680]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/xcrypt-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 22:49:25 crc restorecon[4680]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/zabbix-operator-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 22:49:25 crc restorecon[4680]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/zabbix-operator-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 22:49:25 crc restorecon[4680]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/cache not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 22:49:25 crc restorecon[4680]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 22:49:25 crc restorecon[4680]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 22:49:25 crc restorecon[4680]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/00000-1.psg not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 22:49:25 crc restorecon[4680]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/00000-1.psg.pmt not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 22:49:25 crc restorecon[4680]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/db.pmt not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 22:49:25 crc restorecon[4680]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/index.pmt not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 22:49:25 crc restorecon[4680]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/main.pix not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 22:49:25 crc restorecon[4680]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/overflow.pix not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 22:49:25 crc restorecon[4680]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/digest not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 22:49:25 crc restorecon[4680]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/utilities not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 22:49:25 crc restorecon[4680]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/utilities/copy-content not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 22:49:25 crc restorecon[4680]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 22:49:25 crc restorecon[4680]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/containers/extract-utilities/63709497 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 22:49:25 crc restorecon[4680]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/containers/extract-utilities/d966b7fd not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 22:49:25 crc restorecon[4680]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/containers/extract-utilities/f5773757 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 22:49:25 crc restorecon[4680]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/containers/extract-content/81c9edb9 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 22:49:25 crc restorecon[4680]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/containers/extract-content/57bf57ee not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 22:49:25 crc restorecon[4680]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/containers/extract-content/86f5e6aa not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 22:49:25 crc restorecon[4680]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/containers/registry-server/0aabe31d not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 22:49:25 crc restorecon[4680]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/containers/registry-server/d2af85c2 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 22:49:25 crc restorecon[4680]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/containers/registry-server/09d157d9 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 22:49:25 crc restorecon[4680]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 22:49:25 crc restorecon[4680]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/cache not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 22:49:25 crc restorecon[4680]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 22:49:25 crc restorecon[4680]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 22:49:25 crc restorecon[4680]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/00000-1.psg not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 22:49:25 crc restorecon[4680]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/00000-1.psg.pmt not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 22:49:25 crc restorecon[4680]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/db.pmt not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 22:49:25 crc restorecon[4680]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/index.pmt not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 22:49:25 crc restorecon[4680]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/main.pix not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 22:49:25 crc restorecon[4680]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/overflow.pix not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 22:49:25 crc restorecon[4680]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/digest not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 22:49:25 crc restorecon[4680]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 22:49:25 crc restorecon[4680]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/3scale-community-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 22:49:25 crc restorecon[4680]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/3scale-community-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 22:49:25 crc restorecon[4680]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-acm-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 22:49:25 crc restorecon[4680]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-acm-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 22:49:25 crc restorecon[4680]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-acmpca-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 22:49:25 crc restorecon[4680]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-acmpca-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 22:49:25 crc restorecon[4680]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-apigateway-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 22:49:25 crc restorecon[4680]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-apigateway-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 22:49:25 crc restorecon[4680]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-apigatewayv2-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 22:49:25 crc restorecon[4680]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-apigatewayv2-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 22:49:25 crc restorecon[4680]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-applicationautoscaling-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 22:49:25 crc restorecon[4680]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-applicationautoscaling-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 22:49:25 crc restorecon[4680]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-athena-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 22:49:25 crc restorecon[4680]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-athena-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 22:49:25 crc restorecon[4680]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-cloudfront-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 22:49:25 crc restorecon[4680]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-cloudfront-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 22:49:25 crc restorecon[4680]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-cloudtrail-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 22:49:25 crc restorecon[4680]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-cloudtrail-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 22:49:25 crc restorecon[4680]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-cloudwatch-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 22:49:25 crc restorecon[4680]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-cloudwatch-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 22:49:25 crc restorecon[4680]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-cloudwatchlogs-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 22:49:25 crc restorecon[4680]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-cloudwatchlogs-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 22:49:25 crc restorecon[4680]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-documentdb-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 22:49:25 crc restorecon[4680]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-documentdb-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 22:49:25 crc restorecon[4680]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-dynamodb-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 22:49:25 crc restorecon[4680]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-dynamodb-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 22:49:25 crc restorecon[4680]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-ec2-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 22:49:25 crc restorecon[4680]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-ec2-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 22:49:25 crc restorecon[4680]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-ecr-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 22:49:25 crc restorecon[4680]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-ecr-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 22:49:25 crc restorecon[4680]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-ecs-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 22:49:25 crc restorecon[4680]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-ecs-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 22:49:25 crc restorecon[4680]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-efs-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 22:49:25 crc restorecon[4680]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-efs-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 22:49:25 crc restorecon[4680]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-eks-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 22:49:25 crc restorecon[4680]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-eks-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 22:49:25 crc restorecon[4680]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-elasticache-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 22:49:25 crc restorecon[4680]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-elasticache-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 22:49:25 crc restorecon[4680]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-elbv2-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 22:49:25 crc restorecon[4680]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-elbv2-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 22:49:25 crc restorecon[4680]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-emrcontainers-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 22:49:25 crc restorecon[4680]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-emrcontainers-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 22:49:25 crc restorecon[4680]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-eventbridge-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 22:49:25 crc restorecon[4680]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-eventbridge-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 22:49:25 crc restorecon[4680]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-iam-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 22:49:25 crc restorecon[4680]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-iam-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 22:49:25 crc restorecon[4680]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-kafka-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 22:49:25 crc restorecon[4680]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-kafka-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 22:49:25 crc restorecon[4680]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-keyspaces-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 22:49:25 crc restorecon[4680]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-keyspaces-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 22:49:25 crc restorecon[4680]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-kinesis-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 22:49:25 crc restorecon[4680]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-kinesis-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 22:49:25 crc restorecon[4680]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-kms-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 22:49:25 crc restorecon[4680]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-kms-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 22:49:25 crc restorecon[4680]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-lambda-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 22:49:25 crc restorecon[4680]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-lambda-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 22:49:25 crc restorecon[4680]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-memorydb-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 22:49:25 crc restorecon[4680]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-memorydb-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 22:49:25 crc restorecon[4680]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-mq-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 22:49:25 crc restorecon[4680]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-mq-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 22:49:25 crc restorecon[4680]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-networkfirewall-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 22:49:25 crc restorecon[4680]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-networkfirewall-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 22:49:25 crc restorecon[4680]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-opensearchservice-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 22:49:25 crc restorecon[4680]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-opensearchservice-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 22:49:25 crc restorecon[4680]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-organizations-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 22:49:25 crc restorecon[4680]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-organizations-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 22:49:25 crc restorecon[4680]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-pipes-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 22:49:25 crc restorecon[4680]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-pipes-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 22:49:25 crc restorecon[4680]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-prometheusservice-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 22:49:25 crc restorecon[4680]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-prometheusservice-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 22:49:25 crc restorecon[4680]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-rds-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 22:49:25 crc restorecon[4680]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-rds-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 22:49:25 crc restorecon[4680]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-recyclebin-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 22:49:25 crc restorecon[4680]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-recyclebin-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 22:49:25 crc restorecon[4680]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-route53-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 22:49:25 crc restorecon[4680]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-route53-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 22:49:25 crc restorecon[4680]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-route53resolver-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 22:49:25 crc restorecon[4680]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-route53resolver-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 22:49:25 crc restorecon[4680]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-s3-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 22:49:25 crc restorecon[4680]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-s3-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 22:49:25 crc restorecon[4680]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-sagemaker-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 22:49:25 crc restorecon[4680]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-sagemaker-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 22:49:25 crc restorecon[4680]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-secretsmanager-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 22:49:25 crc restorecon[4680]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-secretsmanager-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 22:49:25 crc restorecon[4680]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-ses-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 22:49:25 crc restorecon[4680]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-ses-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 22:49:25 crc restorecon[4680]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-sfn-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 22:49:25 crc restorecon[4680]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-sfn-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 22:49:25 crc restorecon[4680]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-sns-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 22:49:25 crc restorecon[4680]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-sns-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 22:49:25 crc restorecon[4680]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-sqs-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 22:49:25 crc restorecon[4680]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-sqs-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 22:49:25 crc restorecon[4680]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-ssm-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 22:49:25 crc restorecon[4680]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-ssm-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 22:49:25 crc restorecon[4680]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-wafv2-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 22:49:25 crc restorecon[4680]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-wafv2-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 22:49:25 crc restorecon[4680]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aerospike-kubernetes-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 22:49:25 crc restorecon[4680]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aerospike-kubernetes-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 22:49:25 crc restorecon[4680]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/airflow-helm-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 22:49:25 crc restorecon[4680]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/airflow-helm-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 22:49:25 crc restorecon[4680]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/alloydb-omni-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 22:49:25 crc restorecon[4680]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/alloydb-omni-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 22:49:25 crc restorecon[4680]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/alvearie-imaging-ingestion not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 22:49:25 crc restorecon[4680]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/alvearie-imaging-ingestion/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 22:49:25 crc restorecon[4680]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/amd-gpu-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 22:49:25 crc restorecon[4680]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/amd-gpu-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 22:49:25 crc restorecon[4680]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/analytics-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 22:49:25 crc restorecon[4680]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/analytics-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 22:49:25 crc restorecon[4680]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/annotationlab not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 22:49:25 crc restorecon[4680]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/annotationlab/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 22:49:25 crc restorecon[4680]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apicast-community-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 22:49:25 crc restorecon[4680]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apicast-community-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 22:49:25 crc restorecon[4680]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apicurio-api-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 22:49:25 crc restorecon[4680]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apicurio-api-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 22:49:25 crc restorecon[4680]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apicurio-registry not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 22:49:25 crc restorecon[4680]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apicurio-registry/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 22:49:25 crc restorecon[4680]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apicurito not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 22:49:25 crc restorecon[4680]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apicurito/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 22:49:25 crc restorecon[4680]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apimatic-kubernetes-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 22:49:25 crc restorecon[4680]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apimatic-kubernetes-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 22:49:25 crc restorecon[4680]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/application-services-metering-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 22:49:25 crc restorecon[4680]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/application-services-metering-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 22:49:25 crc restorecon[4680]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aqua not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 22:49:25 crc restorecon[4680]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aqua/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 22:49:25 crc restorecon[4680]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/argocd-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 22:49:25 crc restorecon[4680]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/argocd-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 22:49:25 crc restorecon[4680]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/assisted-service-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 22:49:25 crc restorecon[4680]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/assisted-service-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 22:49:25 crc restorecon[4680]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/authorino-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 22:49:25 crc restorecon[4680]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/authorino-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 22:49:25 crc restorecon[4680]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/automotive-infra not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 22:49:25 crc restorecon[4680]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/automotive-infra/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 22:49:25 crc restorecon[4680]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aws-efs-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 22:49:25 crc restorecon[4680]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aws-efs-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 22:49:25 crc restorecon[4680]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/awss3-operator-registry not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 22:49:25 crc restorecon[4680]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/awss3-operator-registry/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 22:49:25 crc restorecon[4680]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/azure-service-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 22:49:25 crc restorecon[4680]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/azure-service-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 22:49:25 crc restorecon[4680]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/beegfs-csi-driver-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 22:49:25 crc restorecon[4680]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/beegfs-csi-driver-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 22:49:25 crc restorecon[4680]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/bpfman-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 22:49:25 crc restorecon[4680]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/bpfman-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 22:49:25 crc restorecon[4680]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/camel-k not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 22:49:25 crc restorecon[4680]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/camel-k/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 22:49:25 crc restorecon[4680]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/camel-karavan-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 22:49:25 crc restorecon[4680]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/camel-karavan-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 22:49:25 crc restorecon[4680]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cass-operator-community not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 22:49:25 crc restorecon[4680]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cass-operator-community/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 22:49:25 crc restorecon[4680]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cert-manager not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 22:49:25 crc restorecon[4680]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cert-manager/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 22:49:25 crc restorecon[4680]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cert-utils-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 22:49:25 crc restorecon[4680]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cert-utils-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 22:49:25 crc restorecon[4680]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cluster-aas-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 22:49:25 crc restorecon[4680]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cluster-aas-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 22:49:25 crc restorecon[4680]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cluster-impairment-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 22:49:25 crc restorecon[4680]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cluster-impairment-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 22:49:25 crc restorecon[4680]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cluster-manager not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 22:49:25 crc restorecon[4680]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cluster-manager/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 22:49:25 crc restorecon[4680]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cockroachdb not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 22:49:25 crc restorecon[4680]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cockroachdb/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 22:49:25 crc restorecon[4680]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/codeflare-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 22:49:25 crc restorecon[4680]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/codeflare-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 22:49:25 crc restorecon[4680]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/community-kubevirt-hyperconverged not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 22:49:25 crc restorecon[4680]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/community-kubevirt-hyperconverged/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 22:49:25 crc restorecon[4680]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/community-trivy-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 22:49:25 crc restorecon[4680]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/community-trivy-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 22:49:25 crc restorecon[4680]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/community-windows-machine-config-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 22:49:25 crc restorecon[4680]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/community-windows-machine-config-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 22:49:25 crc restorecon[4680]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/customized-user-remediation not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 22:49:25 crc restorecon[4680]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/customized-user-remediation/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 22:49:25 crc restorecon[4680]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cxl-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 22:49:25 crc restorecon[4680]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cxl-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 22:49:25 crc restorecon[4680]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dapr-kubernetes-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 22:49:25 crc restorecon[4680]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dapr-kubernetes-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 22:49:25 crc restorecon[4680]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/datadog-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 22:49:25 crc restorecon[4680]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/datadog-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 22:49:25 crc restorecon[4680]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/datatrucker-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 22:49:25 crc restorecon[4680]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/datatrucker-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 22:49:25 crc restorecon[4680]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dbaas-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 22:49:25 crc restorecon[4680]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dbaas-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 22:49:25 crc restorecon[4680]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/debezium-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 22:49:25 crc restorecon[4680]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/debezium-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 22:49:25 crc restorecon[4680]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dell-csm-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 22:49:25 crc restorecon[4680]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dell-csm-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 22:49:25 crc restorecon[4680]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/deployment-validation-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 22:49:25 crc restorecon[4680]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/deployment-validation-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 22:49:25 crc restorecon[4680]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/devopsinabox not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 22:49:25 crc restorecon[4680]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/devopsinabox/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 22:49:25 crc restorecon[4680]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dns-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 22:49:25 crc restorecon[4680]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dns-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 22:49:25 crc restorecon[4680]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dynatrace-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 22:49:25 crc restorecon[4680]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dynatrace-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 22:49:25 crc restorecon[4680]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/eclipse-amlen-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 22:49:25 crc restorecon[4680]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/eclipse-amlen-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 22:49:25 crc restorecon[4680]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/eclipse-che not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 22:49:25 crc restorecon[4680]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/eclipse-che/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 22:49:25 crc restorecon[4680]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ecr-secret-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 22:49:25 crc restorecon[4680]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ecr-secret-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 22:49:25 crc restorecon[4680]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/edp-keycloak-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 22:49:25 crc restorecon[4680]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/edp-keycloak-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 22:49:25 crc restorecon[4680]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/eginnovations-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 22:49:25 crc restorecon[4680]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/eginnovations-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 22:49:25 crc restorecon[4680]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/egressip-ipam-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 22:49:25 crc restorecon[4680]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/egressip-ipam-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 22:49:25 crc restorecon[4680]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ember-csi-community-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 22:49:25 crc restorecon[4680]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ember-csi-community-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 22:49:25 crc restorecon[4680]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/etcd not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 22:49:25 crc restorecon[4680]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/etcd/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 22:49:25 crc restorecon[4680]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/eventing-kogito not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 22:49:25 crc restorecon[4680]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/eventing-kogito/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 22:49:25 crc restorecon[4680]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/external-secrets-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 22:49:25 crc restorecon[4680]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/external-secrets-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 22:49:25 crc restorecon[4680]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/falcon-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 22:49:25 crc restorecon[4680]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/falcon-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 22:49:25 crc restorecon[4680]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fence-agents-remediation not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 22:49:25 crc restorecon[4680]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fence-agents-remediation/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 22:49:25 crc restorecon[4680]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/flink-kubernetes-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 22:49:25 crc restorecon[4680]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/flink-kubernetes-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 22:49:25 crc restorecon[4680]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/flux not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 22:49:25 crc restorecon[4680]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/flux/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 22:49:25 crc restorecon[4680]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/k8gb not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 22:49:25 crc restorecon[4680]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/k8gb/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 22:49:25 crc restorecon[4680]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fossul-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 22:49:25 crc restorecon[4680]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fossul-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 22:49:25 crc restorecon[4680]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/github-arc-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 22:49:25 crc restorecon[4680]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/github-arc-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 22:49:25 crc restorecon[4680]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/gitops-primer not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 22:49:25 crc restorecon[4680]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/gitops-primer/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 22:49:25 crc restorecon[4680]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/gitwebhook-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 22:49:25 crc restorecon[4680]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/gitwebhook-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 22:49:25 crc restorecon[4680]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/global-load-balancer-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 22:49:25 crc restorecon[4680]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/global-load-balancer-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 22:49:25 crc restorecon[4680]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/grafana-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 22:49:25 crc restorecon[4680]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/grafana-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 22:49:25 crc restorecon[4680]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/group-sync-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 22:49:25 crc restorecon[4680]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/group-sync-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 22:49:25 crc restorecon[4680]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hawtio-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 22:49:25 crc restorecon[4680]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hawtio-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 22:49:25 crc restorecon[4680]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hazelcast-platform-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 22:49:25 crc restorecon[4680]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hazelcast-platform-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 22:49:25 crc restorecon[4680]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hedvig-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 22:49:25 crc restorecon[4680]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hedvig-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 22:49:25 crc restorecon[4680]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hive-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 22:49:25 crc restorecon[4680]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hive-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 22:49:25 crc restorecon[4680]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/horreum-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 22:49:25 crc restorecon[4680]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/horreum-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 22:49:25 crc restorecon[4680]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hyperfoil-bundle not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 22:49:25 crc restorecon[4680]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hyperfoil-bundle/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 22:49:25 crc restorecon[4680]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-block-csi-operator-community not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 22:49:25 crc restorecon[4680]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-block-csi-operator-community/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 22:49:25 crc restorecon[4680]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-security-verify-access-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 22:49:25 crc restorecon[4680]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-security-verify-access-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 22:49:25 crc restorecon[4680]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-spectrum-scale-csi-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 22:49:25 crc restorecon[4680]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-spectrum-scale-csi-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 22:49:25 crc restorecon[4680]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibmcloud-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 22:49:25 crc restorecon[4680]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibmcloud-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 22:49:25 crc restorecon[4680]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/infinispan not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 22:49:25 crc restorecon[4680]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/infinispan/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 22:49:25 crc restorecon[4680]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/integrity-shield-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 22:49:25 crc restorecon[4680]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/integrity-shield-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 22:49:25 crc restorecon[4680]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ipfs-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 22:49:25 crc restorecon[4680]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ipfs-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 22:49:25 crc restorecon[4680]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/istio-workspace-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 22:49:25 crc restorecon[4680]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/istio-workspace-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 22:49:25 crc restorecon[4680]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/jaeger not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 22:49:25 crc restorecon[4680]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/jaeger/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 22:49:25 crc restorecon[4680]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kaoto-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 22:49:25 crc restorecon[4680]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kaoto-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 22:49:25 crc restorecon[4680]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keda not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 22:49:25 crc restorecon[4680]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keda/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 22:49:25 crc restorecon[4680]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keepalived-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 22:49:25 crc restorecon[4680]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keepalived-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 22:49:25 crc restorecon[4680]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keycloak-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 22:49:25 crc restorecon[4680]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keycloak-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 22:49:25 crc restorecon[4680]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keycloak-permissions-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 22:49:25 crc restorecon[4680]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keycloak-permissions-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 22:49:25 crc restorecon[4680]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/klusterlet not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 22:49:25 crc restorecon[4680]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/klusterlet/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 22:49:25 crc restorecon[4680]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kogito-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 22:49:25 crc restorecon[4680]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kogito-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 22:49:25 crc restorecon[4680]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/koku-metrics-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 22:49:25 crc restorecon[4680]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/koku-metrics-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 22:49:25 crc restorecon[4680]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/konveyor-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 22:49:25 crc restorecon[4680]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/konveyor-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 22:49:25 crc restorecon[4680]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/korrel8r not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 22:49:25 crc restorecon[4680]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/korrel8r/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 22:49:25 crc restorecon[4680]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kuadrant-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 22:49:25 crc restorecon[4680]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kuadrant-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 22:49:25 crc restorecon[4680]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kube-green not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 22:49:25 crc restorecon[4680]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kube-green/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 22:49:25 crc restorecon[4680]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubecost not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 22:49:25 crc restorecon[4680]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubecost/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 22:49:25 crc restorecon[4680]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubernetes-imagepuller-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 22:49:25 crc restorecon[4680]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubernetes-imagepuller-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 22:49:25 crc restorecon[4680]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubeturbo not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 22:49:25 crc restorecon[4680]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubeturbo/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 22:49:25 crc restorecon[4680]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/l5-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 22:49:25 crc restorecon[4680]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/l5-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 22:49:25 crc restorecon[4680]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/layer7-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 22:49:25 crc restorecon[4680]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/layer7-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 22:49:25 crc restorecon[4680]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/lbconfig-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 22:49:25 crc restorecon[4680]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/lbconfig-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 22:49:25 crc restorecon[4680]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/lib-bucket-provisioner not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 22:49:25 crc restorecon[4680]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/lib-bucket-provisioner/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 22:49:25 crc restorecon[4680]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/limitador-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 22:49:25 crc restorecon[4680]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/limitador-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 22:49:25 crc restorecon[4680]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/logging-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 22:49:25 crc restorecon[4680]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/logging-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 22:49:25 crc restorecon[4680]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/loki-helm-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 22:49:25 crc restorecon[4680]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/loki-helm-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 22:49:25 crc restorecon[4680]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/loki-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 22:49:25 crc restorecon[4680]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/loki-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 22:49:25 crc restorecon[4680]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/machine-deletion-remediation not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 22:49:25 crc restorecon[4680]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/machine-deletion-remediation/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 22:49:25 crc restorecon[4680]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mariadb-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 22:49:25 crc restorecon[4680]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mariadb-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 22:49:25 crc restorecon[4680]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/marin3r not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 22:49:25 crc restorecon[4680]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/marin3r/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 22:49:25 crc restorecon[4680]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mercury-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 22:49:25 crc restorecon[4680]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mercury-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 22:49:25 crc restorecon[4680]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/microcks not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 22:49:25 crc restorecon[4680]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/microcks/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 22:49:25 crc restorecon[4680]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mongodb-atlas-kubernetes not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 22:49:25 crc restorecon[4680]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mongodb-atlas-kubernetes/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 22:49:25 crc restorecon[4680]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mongodb-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 22:49:25 crc restorecon[4680]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mongodb-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 22:49:25 crc restorecon[4680]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/move2kube-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 22:49:25 crc restorecon[4680]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/move2kube-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 22:49:25 crc restorecon[4680]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/multi-nic-cni-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 22:49:25 crc restorecon[4680]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/multi-nic-cni-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 22:49:25 crc restorecon[4680]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/multicluster-global-hub-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 22:49:25 crc restorecon[4680]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/multicluster-global-hub-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 22:49:25 crc restorecon[4680]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/multicluster-operators-subscription not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 22:49:25 crc restorecon[4680]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/multicluster-operators-subscription/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 22:49:25 crc restorecon[4680]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/must-gather-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 22:49:25 crc restorecon[4680]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/must-gather-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 22:49:25 crc restorecon[4680]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/namespace-configuration-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 22:49:25 crc restorecon[4680]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/namespace-configuration-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 22:49:25 crc restorecon[4680]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ncn-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 22:49:25 crc restorecon[4680]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ncn-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 22:49:25 crc restorecon[4680]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ndmspc-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 22:49:25 crc restorecon[4680]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ndmspc-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 22:49:25 crc restorecon[4680]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/netobserv-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 22:49:25 crc restorecon[4680]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/netobserv-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 22:49:25 crc restorecon[4680]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/neuvector-community-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 22:49:25 crc restorecon[4680]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/neuvector-community-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 22:49:25 crc restorecon[4680]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nexus-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 22:49:25 crc restorecon[4680]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nexus-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 22:49:25 crc restorecon[4680]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nexus-operator-m88i not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 22:49:25 crc restorecon[4680]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nexus-operator-m88i/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 22:49:25 crc restorecon[4680]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nfs-provisioner-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 22:49:25 crc restorecon[4680]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nfs-provisioner-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 22:49:25 crc restorecon[4680]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nlp-server not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 22:49:25 crc restorecon[4680]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nlp-server/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 22:49:25 crc restorecon[4680]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/node-discovery-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 22:49:25 crc restorecon[4680]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/node-discovery-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 22:49:25 crc restorecon[4680]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/node-healthcheck-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 22:49:25 crc restorecon[4680]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/node-healthcheck-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 22:49:25 crc restorecon[4680]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/node-maintenance-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 22:49:25 crc restorecon[4680]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/node-maintenance-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 22:49:25 crc restorecon[4680]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nsm-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 22:49:25 crc restorecon[4680]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nsm-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 22:49:25 crc restorecon[4680]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/oadp-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 22:49:25 crc restorecon[4680]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/oadp-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 22:49:25 crc restorecon[4680]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/observability-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 22:49:25 crc restorecon[4680]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/observability-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 22:49:25 crc restorecon[4680]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/oci-ccm-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 22:49:25 crc restorecon[4680]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/oci-ccm-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 22:49:25 crc restorecon[4680]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ocm-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 22:49:25 crc restorecon[4680]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ocm-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 22:49:25 crc restorecon[4680]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odoo-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 22:49:25 crc restorecon[4680]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odoo-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 22:49:25 crc restorecon[4680]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/opendatahub-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 22:49:25 crc restorecon[4680]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/opendatahub-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 22:49:25 crc restorecon[4680]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openebs not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 22:49:25 crc restorecon[4680]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openebs/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 22:49:25 crc restorecon[4680]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-nfd-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 22:49:25 crc restorecon[4680]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-nfd-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 22:49:25 crc restorecon[4680]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-node-upgrade-mutex-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 22:49:25 crc restorecon[4680]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-node-upgrade-mutex-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 22:49:25 crc restorecon[4680]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-qiskit-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 22:49:25 crc restorecon[4680]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-qiskit-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 22:49:25 crc restorecon[4680]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/opentelemetry-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 22:49:25 crc restorecon[4680]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/opentelemetry-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 22:49:25 crc restorecon[4680]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/patch-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 22:49:25 crc restorecon[4680]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/patch-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 22:49:25 crc restorecon[4680]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/patterns-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 22:49:25 crc restorecon[4680]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/patterns-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 22:49:25 crc restorecon[4680]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pcc-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 22:49:25 crc restorecon[4680]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pcc-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 22:49:25 crc restorecon[4680]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pelorus-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 22:49:25 crc restorecon[4680]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pelorus-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 22:49:25 crc restorecon[4680]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/percona-xtradb-cluster-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 22:49:25 crc restorecon[4680]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/percona-xtradb-cluster-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 22:49:25 crc restorecon[4680]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/portworx-essentials not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 22:49:25 crc restorecon[4680]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/portworx-essentials/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 22:49:25 crc restorecon[4680]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/postgresql not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 22:49:25 crc restorecon[4680]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/postgresql/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 22:49:25 crc restorecon[4680]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/proactive-node-scaling-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 22:49:25 crc restorecon[4680]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/proactive-node-scaling-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 22:49:25 crc restorecon[4680]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/project-quay not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 22:49:25 crc restorecon[4680]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/project-quay/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 22:49:25 crc restorecon[4680]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/prometheus not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 22:49:25 crc restorecon[4680]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/prometheus/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 22:49:25 crc restorecon[4680]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/prometheus-exporter-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 22:49:25 crc restorecon[4680]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/prometheus-exporter-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 22:49:25 crc restorecon[4680]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/prometurbo not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 22:49:25 crc restorecon[4680]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/prometurbo/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 22:49:25 crc restorecon[4680]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pubsubplus-eventbroker-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 22:49:25 crc restorecon[4680]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pubsubplus-eventbroker-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 22:49:25 crc restorecon[4680]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pulp-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 22:49:25 crc restorecon[4680]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pulp-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 22:49:25 crc restorecon[4680]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rabbitmq-cluster-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 22:49:25 crc restorecon[4680]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rabbitmq-cluster-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 22:49:25 crc restorecon[4680]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rabbitmq-messaging-topology-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 22:49:25 crc restorecon[4680]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rabbitmq-messaging-topology-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 22:49:25 crc restorecon[4680]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/redis-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 22:49:25 crc restorecon[4680]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/redis-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 22:49:25 crc restorecon[4680]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/reportportal-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 22:49:25 crc restorecon[4680]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/reportportal-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 22:49:25 crc restorecon[4680]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/resource-locker-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 22:49:25 crc restorecon[4680]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/resource-locker-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 22:49:25 crc restorecon[4680]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhoas-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 22:49:25 crc restorecon[4680]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhoas-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 22:49:25 crc restorecon[4680]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ripsaw not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 22:49:25 crc restorecon[4680]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ripsaw/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 22:49:25 crc restorecon[4680]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sailoperator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 22:49:25 crc restorecon[4680]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sailoperator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 22:49:25 crc restorecon[4680]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sap-commerce-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 22:49:25 crc restorecon[4680]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sap-commerce-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 22:49:25 crc restorecon[4680]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sap-data-intelligence-observer-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 22:49:25 crc restorecon[4680]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sap-data-intelligence-observer-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 22:49:25 crc restorecon[4680]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sap-hana-express-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 22:49:25 crc restorecon[4680]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sap-hana-express-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 22:49:25 crc restorecon[4680]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/seldon-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 22:49:25 crc restorecon[4680]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/seldon-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 22:49:25 crc restorecon[4680]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/self-node-remediation not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 22:49:25 crc restorecon[4680]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/self-node-remediation/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 22:49:25 crc restorecon[4680]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/service-binding-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 22:49:25 crc restorecon[4680]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/service-binding-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 22:49:25 crc restorecon[4680]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/shipwright-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 22:49:25 crc restorecon[4680]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/shipwright-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 22:49:25 crc restorecon[4680]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sigstore-helm-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 22:49:25 crc restorecon[4680]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sigstore-helm-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 22:49:25 crc restorecon[4680]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/silicom-sts-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 22:49:25 crc restorecon[4680]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/silicom-sts-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 22:49:25 crc restorecon[4680]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/skupper-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 22:49:25 crc restorecon[4680]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/skupper-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 22:49:25 crc restorecon[4680]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/snapscheduler not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 22:49:25 crc restorecon[4680]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/snapscheduler/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 22:49:25 crc restorecon[4680]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/snyk-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 22:49:25 crc restorecon[4680]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/snyk-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 22:49:25 crc restorecon[4680]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/socmmd not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 22:49:25 crc restorecon[4680]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/socmmd/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 22:49:25 crc restorecon[4680]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sonar-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 22:49:25 crc restorecon[4680]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sonar-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 22:49:25 crc restorecon[4680]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sosivio not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 22:49:25 crc restorecon[4680]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sosivio/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 22:49:25 crc restorecon[4680]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sonataflow-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 22:49:25 crc restorecon[4680]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sonataflow-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 22:49:25 crc restorecon[4680]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sosreport-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 22:49:25 crc restorecon[4680]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sosreport-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 22:49:25 crc restorecon[4680]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/spark-helm-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 22:49:25 crc restorecon[4680]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/spark-helm-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 22:49:25 crc restorecon[4680]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/special-resource-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 22:49:25 crc restorecon[4680]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/special-resource-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 22:49:25 crc restorecon[4680]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/stolostron not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 22:49:25 crc restorecon[4680]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/stolostron/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 22:49:25 crc restorecon[4680]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/stolostron-engine not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 22:49:25 crc restorecon[4680]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/stolostron-engine/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 22:49:25 crc restorecon[4680]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/strimzi-kafka-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 22:49:25 crc restorecon[4680]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/strimzi-kafka-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 22:49:25 crc restorecon[4680]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/syndesis not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 22:49:25 crc restorecon[4680]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/syndesis/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 22:49:25 crc restorecon[4680]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/t8c not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 22:49:25 crc restorecon[4680]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/t8c/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 22:49:25 crc restorecon[4680]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tagger not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 22:49:25 crc restorecon[4680]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tagger/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 22:49:25 crc restorecon[4680]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tempo-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 22:49:25 crc restorecon[4680]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tempo-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 22:49:25 crc restorecon[4680]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tf-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 22:49:25 crc restorecon[4680]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tf-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 22:49:25 crc restorecon[4680]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tidb-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 22:49:25 crc restorecon[4680]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tidb-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 22:49:25 crc restorecon[4680]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/trident-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 22:49:25 crc restorecon[4680]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/trident-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 22:49:25 crc restorecon[4680]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/trustify-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 22:49:25 crc restorecon[4680]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/trustify-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 22:49:25 crc restorecon[4680]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ucs-ci-solutions-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 22:49:25 crc restorecon[4680]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ucs-ci-solutions-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 22:49:25 crc restorecon[4680]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/universal-crossplane not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 22:49:25 crc restorecon[4680]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/universal-crossplane/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 22:49:25 crc restorecon[4680]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/varnish-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 22:49:25 crc restorecon[4680]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/varnish-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 22:49:25 crc restorecon[4680]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/vault-config-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 22:49:25 crc restorecon[4680]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/vault-config-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 22:49:25 crc restorecon[4680]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/verticadb-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 22:49:25 crc restorecon[4680]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/verticadb-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 22:49:25 crc restorecon[4680]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/volume-expander-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 22:49:25 crc restorecon[4680]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/volume-expander-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 22:49:25 crc restorecon[4680]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/wandb-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 22:49:25 crc restorecon[4680]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/wandb-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 22:49:25 crc restorecon[4680]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/windup-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 22:49:25 crc restorecon[4680]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/windup-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 22:49:25 crc restorecon[4680]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/yaks not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 22:49:25 crc restorecon[4680]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/yaks/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 22:49:25 crc restorecon[4680]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/utilities not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 22:49:25 crc restorecon[4680]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/utilities/copy-content not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 22:49:25 crc restorecon[4680]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 22:49:25 crc restorecon[4680]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/containers/extract-utilities/c0fe7256 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 22:49:25 crc restorecon[4680]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/containers/extract-utilities/c30319e4 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 22:49:25 crc restorecon[4680]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/containers/extract-utilities/e6b1dd45 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 22:49:25 crc restorecon[4680]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/containers/extract-content/2bb643f0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 22:49:25 crc restorecon[4680]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/containers/extract-content/920de426 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 22:49:25 crc restorecon[4680]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/containers/extract-content/70fa1e87 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 22:49:25 crc restorecon[4680]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/containers/registry-server/a1c12a2f not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 22:49:25 crc restorecon[4680]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/containers/registry-server/9442e6c7 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 22:49:25 crc restorecon[4680]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/containers/registry-server/5b45ec72 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 22:49:25 crc restorecon[4680]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 22:49:25 crc restorecon[4680]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 22:49:25 crc restorecon[4680]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/abot-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 22:49:25 crc restorecon[4680]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/abot-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 22:49:25 crc restorecon[4680]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aerospike-kubernetes-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 22:49:25 crc restorecon[4680]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aerospike-kubernetes-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 22:49:25 crc restorecon[4680]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aikit-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 22:49:25 crc restorecon[4680]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aikit-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 22:49:25 crc restorecon[4680]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anzo-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 22:49:25 crc restorecon[4680]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anzo-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 22:49:25 crc restorecon[4680]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anzograph-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 22:49:25 crc restorecon[4680]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anzograph-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 22:49:25 crc restorecon[4680]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anzounstructured-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 22:49:25 crc restorecon[4680]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anzounstructured-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 22:49:25 crc restorecon[4680]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cloudbees-ci-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 22:49:25 crc restorecon[4680]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cloudbees-ci-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 22:49:25 crc restorecon[4680]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cockroachdb-certified-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 22:49:25 crc restorecon[4680]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cockroachdb-certified-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 22:49:25 crc restorecon[4680]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/crunchy-postgres-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 22:49:25 crc restorecon[4680]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/crunchy-postgres-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 22:49:25 crc restorecon[4680]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/datadog-operator-certified-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 22:49:25 crc restorecon[4680]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/datadog-operator-certified-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 22:49:25 crc restorecon[4680]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dynatrace-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 22:49:25 crc restorecon[4680]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dynatrace-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 22:49:25 crc restorecon[4680]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/entando-k8s-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 22:49:25 crc restorecon[4680]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/entando-k8s-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 22:49:25 crc restorecon[4680]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/flux not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 22:49:25 crc restorecon[4680]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/flux/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 22:49:25 crc restorecon[4680]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/instana-agent-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 22:49:25 crc restorecon[4680]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/instana-agent-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 22:49:25 crc restorecon[4680]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/iomesh-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 22:49:25 crc restorecon[4680]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/iomesh-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 22:49:25 crc restorecon[4680]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/joget-dx-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 22:49:25 crc restorecon[4680]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/joget-dx-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 22:49:25 crc restorecon[4680]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/joget-dx8-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 22:49:25 crc restorecon[4680]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/joget-dx8-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 22:49:25 crc restorecon[4680]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/k10-kasten-operator-paygo-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 22:49:25 crc restorecon[4680]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/k10-kasten-operator-paygo-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 22:49:25 crc restorecon[4680]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/k10-kasten-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 22:49:25 crc restorecon[4680]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/k10-kasten-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 22:49:25 crc restorecon[4680]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/k10-kasten-operator-term-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 22:49:25 crc restorecon[4680]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/k10-kasten-operator-term-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 22:49:25 crc restorecon[4680]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubemq-operator-marketplace-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 22:49:25 crc restorecon[4680]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubemq-operator-marketplace-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 22:49:25 crc restorecon[4680]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubeturbo-certified-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 22:49:25 crc restorecon[4680]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubeturbo-certified-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 22:49:25 crc restorecon[4680]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/linstor-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 22:49:25 crc restorecon[4680]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/linstor-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 22:49:25 crc restorecon[4680]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/marketplace-games-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 22:49:25 crc restorecon[4680]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/marketplace-games-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 22:49:25 crc restorecon[4680]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/model-builder-for-vision-certified-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 22:49:25 crc restorecon[4680]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/model-builder-for-vision-certified-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 22:49:25 crc restorecon[4680]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/neuvector-certified-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 22:49:25 crc restorecon[4680]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/neuvector-certified-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 22:49:25 crc restorecon[4680]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ovms-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 22:49:25 crc restorecon[4680]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ovms-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 22:49:25 crc restorecon[4680]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pachyderm-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 22:49:25 crc restorecon[4680]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pachyderm-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 22:49:25 crc restorecon[4680]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/redis-enterprise-operator-cert-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 22:49:25 crc restorecon[4680]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/redis-enterprise-operator-cert-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 22:49:25 crc restorecon[4680]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/seldon-deploy-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 22:49:25 crc restorecon[4680]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/seldon-deploy-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 22:49:25 crc restorecon[4680]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/starburst-enterprise-helm-operator-paygo-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 22:49:25 crc restorecon[4680]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/starburst-enterprise-helm-operator-paygo-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 22:49:25 crc restorecon[4680]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/starburst-enterprise-helm-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 22:49:25 crc restorecon[4680]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/starburst-enterprise-helm-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 22:49:25 crc restorecon[4680]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/t8c-certified-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 22:49:25 crc restorecon[4680]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/t8c-certified-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 22:49:25 crc restorecon[4680]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/timemachine-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 22:49:25 crc restorecon[4680]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/timemachine-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 22:49:25 crc restorecon[4680]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/vfunction-server-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 22:49:25 crc restorecon[4680]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/vfunction-server-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 22:49:25 crc restorecon[4680]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/xcrypt-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 22:49:25 crc restorecon[4680]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/xcrypt-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 22:49:25 crc restorecon[4680]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/yugabyte-platform-operator-bundle-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 22:49:25 crc restorecon[4680]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/yugabyte-platform-operator-bundle-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 22:49:25 crc restorecon[4680]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/zabbix-operator-certified-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 22:49:25 crc restorecon[4680]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/zabbix-operator-certified-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 22:49:25 crc restorecon[4680]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/cache not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 22:49:25 crc restorecon[4680]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 22:49:25 crc restorecon[4680]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 22:49:25 crc restorecon[4680]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/00000-1.psg not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 22:49:25 crc restorecon[4680]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/00000-1.psg.pmt not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 22:49:25 crc restorecon[4680]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/db.pmt not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 22:49:25 crc restorecon[4680]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/index.pmt not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 22:49:25 crc restorecon[4680]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/main.pix not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 22:49:25 crc restorecon[4680]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/overflow.pix not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 22:49:25 crc restorecon[4680]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/digest not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 22:49:25 crc restorecon[4680]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/utilities not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 22:49:25 crc restorecon[4680]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/utilities/copy-content not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 22:49:25 crc restorecon[4680]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 22:49:25 crc restorecon[4680]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/containers/extract-utilities/3c9f3a59 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 22:49:25 crc restorecon[4680]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/containers/extract-utilities/1091c11b not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 22:49:25 crc restorecon[4680]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/containers/extract-utilities/9a6821c6 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 22:49:25 crc restorecon[4680]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/containers/extract-content/ec0c35e2 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 22:49:25 crc restorecon[4680]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/containers/extract-content/517f37e7 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 22:49:25 crc restorecon[4680]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/containers/extract-content/6214fe78 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 22:49:25 crc restorecon[4680]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/containers/registry-server/ba189c8b not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 22:49:25 crc restorecon[4680]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/containers/registry-server/351e4f31 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 22:49:25 crc restorecon[4680]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/containers/registry-server/c0f219ff not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 22:49:25 crc restorecon[4680]: /var/lib/kubelet/pods/3dcd261975c3d6b9a6ad6367fd4facd3/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c247,c522 Dec 10 22:49:25 crc restorecon[4680]: /var/lib/kubelet/pods/3dcd261975c3d6b9a6ad6367fd4facd3/containers/wait-for-host-port/8069f607 not reset as customized by admin to system_u:object_r:container_file_t:s0:c378,c723 Dec 10 22:49:25 crc restorecon[4680]: /var/lib/kubelet/pods/3dcd261975c3d6b9a6ad6367fd4facd3/containers/wait-for-host-port/559c3d82 not reset as customized by admin to system_u:object_r:container_file_t:s0:c133,c223 Dec 10 22:49:25 crc restorecon[4680]: /var/lib/kubelet/pods/3dcd261975c3d6b9a6ad6367fd4facd3/containers/wait-for-host-port/605ad488 not reset as customized by admin to system_u:object_r:container_file_t:s0:c247,c522 Dec 10 22:49:25 crc restorecon[4680]: /var/lib/kubelet/pods/3dcd261975c3d6b9a6ad6367fd4facd3/containers/kube-scheduler/148df488 not reset as customized by admin to system_u:object_r:container_file_t:s0:c378,c723 Dec 10 22:49:25 crc restorecon[4680]: /var/lib/kubelet/pods/3dcd261975c3d6b9a6ad6367fd4facd3/containers/kube-scheduler/3bf6dcb4 not reset as customized by admin to system_u:object_r:container_file_t:s0:c133,c223 Dec 10 22:49:25 crc restorecon[4680]: /var/lib/kubelet/pods/3dcd261975c3d6b9a6ad6367fd4facd3/containers/kube-scheduler/022a2feb not reset as customized by admin to system_u:object_r:container_file_t:s0:c247,c522 Dec 10 22:49:25 crc restorecon[4680]: /var/lib/kubelet/pods/3dcd261975c3d6b9a6ad6367fd4facd3/containers/kube-scheduler-cert-syncer/938c3924 not reset as customized by admin to system_u:object_r:container_file_t:s0:c378,c723 Dec 10 22:49:25 crc restorecon[4680]: /var/lib/kubelet/pods/3dcd261975c3d6b9a6ad6367fd4facd3/containers/kube-scheduler-cert-syncer/729fe23e not reset as customized by admin to system_u:object_r:container_file_t:s0:c133,c223 Dec 10 22:49:25 crc restorecon[4680]: /var/lib/kubelet/pods/3dcd261975c3d6b9a6ad6367fd4facd3/containers/kube-scheduler-cert-syncer/1fd5cbd4 not reset as customized by admin to system_u:object_r:container_file_t:s0:c247,c522 Dec 10 22:49:25 crc restorecon[4680]: /var/lib/kubelet/pods/3dcd261975c3d6b9a6ad6367fd4facd3/containers/kube-scheduler-recovery-controller/a96697e1 not reset as customized by admin to system_u:object_r:container_file_t:s0:c378,c723 Dec 10 22:49:25 crc restorecon[4680]: /var/lib/kubelet/pods/3dcd261975c3d6b9a6ad6367fd4facd3/containers/kube-scheduler-recovery-controller/e155ddca not reset as customized by admin to system_u:object_r:container_file_t:s0:c133,c223 Dec 10 22:49:25 crc restorecon[4680]: /var/lib/kubelet/pods/3dcd261975c3d6b9a6ad6367fd4facd3/containers/kube-scheduler-recovery-controller/10dd0e0f not reset as customized by admin to system_u:object_r:container_file_t:s0:c247,c522 Dec 10 22:49:25 crc restorecon[4680]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-trusted-ca-bundle not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Dec 10 22:49:25 crc restorecon[4680]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-trusted-ca-bundle/..2025_02_24_06_09_35.3018472960 not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Dec 10 22:49:25 crc restorecon[4680]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-trusted-ca-bundle/..2025_02_24_06_09_35.3018472960/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Dec 10 22:49:25 crc restorecon[4680]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-trusted-ca-bundle/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Dec 10 22:49:25 crc restorecon[4680]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-trusted-ca-bundle/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Dec 10 22:49:25 crc restorecon[4680]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/audit-policies not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Dec 10 22:49:25 crc restorecon[4680]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/audit-policies/..2025_02_24_06_09_35.4262376737 not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Dec 10 22:49:25 crc restorecon[4680]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/audit-policies/..2025_02_24_06_09_35.4262376737/audit.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Dec 10 22:49:25 crc restorecon[4680]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/audit-policies/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Dec 10 22:49:25 crc restorecon[4680]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/audit-policies/audit.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Dec 10 22:49:25 crc restorecon[4680]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-cliconfig not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Dec 10 22:49:25 crc restorecon[4680]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-cliconfig/..2025_02_24_06_09_35.2630275752 not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Dec 10 22:49:25 crc restorecon[4680]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-cliconfig/..2025_02_24_06_09_35.2630275752/v4-0-config-system-cliconfig not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Dec 10 22:49:25 crc restorecon[4680]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-cliconfig/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Dec 10 22:49:25 crc restorecon[4680]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-cliconfig/v4-0-config-system-cliconfig not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Dec 10 22:49:25 crc restorecon[4680]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-service-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Dec 10 22:49:25 crc restorecon[4680]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-service-ca/..2025_02_24_06_09_35.2376963788 not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Dec 10 22:49:25 crc restorecon[4680]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-service-ca/..2025_02_24_06_09_35.2376963788/service-ca.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Dec 10 22:49:25 crc restorecon[4680]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-service-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Dec 10 22:49:25 crc restorecon[4680]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-service-ca/service-ca.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Dec 10 22:49:25 crc restorecon[4680]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Dec 10 22:49:25 crc restorecon[4680]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/containers/oauth-openshift/6f2c8392 not reset as customized by admin to system_u:object_r:container_file_t:s0:c267,c588 Dec 10 22:49:25 crc restorecon[4680]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/containers/oauth-openshift/bd241ad9 not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Dec 10 22:49:25 crc restorecon[4680]: /var/lib/kubelet/plugins not reset as customized by admin to system_u:object_r:container_file_t:s0 Dec 10 22:49:25 crc restorecon[4680]: /var/lib/kubelet/plugins/csi-hostpath not reset as customized by admin to system_u:object_r:container_file_t:s0 Dec 10 22:49:25 crc restorecon[4680]: /var/lib/kubelet/plugins/csi-hostpath/csi.sock not reset as customized by admin to system_u:object_r:container_file_t:s0 Dec 10 22:49:25 crc restorecon[4680]: /var/lib/kubelet/plugins/kubernetes.io not reset as customized by admin to system_u:object_r:container_file_t:s0 Dec 10 22:49:25 crc restorecon[4680]: /var/lib/kubelet/plugins/kubernetes.io/csi not reset as customized by admin to system_u:object_r:container_file_t:s0 Dec 10 22:49:25 crc restorecon[4680]: /var/lib/kubelet/plugins/kubernetes.io/csi/kubevirt.io.hostpath-provisioner not reset as customized by admin to system_u:object_r:container_file_t:s0 Dec 10 22:49:25 crc restorecon[4680]: /var/lib/kubelet/plugins/kubernetes.io/csi/kubevirt.io.hostpath-provisioner/1f4776af88835e41c12b831b4c9fed40233456d14189815a54dbe7f892fc1983 not reset as customized by admin to system_u:object_r:container_file_t:s0 Dec 10 22:49:25 crc restorecon[4680]: /var/lib/kubelet/plugins/kubernetes.io/csi/kubevirt.io.hostpath-provisioner/1f4776af88835e41c12b831b4c9fed40233456d14189815a54dbe7f892fc1983/globalmount not reset as customized by admin to system_u:object_r:container_file_t:s0 Dec 10 22:49:25 crc restorecon[4680]: /var/lib/kubelet/plugins/kubernetes.io/csi/kubevirt.io.hostpath-provisioner/1f4776af88835e41c12b831b4c9fed40233456d14189815a54dbe7f892fc1983/vol_data.json not reset as customized by admin to system_u:object_r:container_file_t:s0 Dec 10 22:49:25 crc restorecon[4680]: /var/lib/kubelet/plugins_registry not reset as customized by admin to system_u:object_r:container_file_t:s0 Dec 10 22:49:25 crc restorecon[4680]: Relabeled /var/usrlocal/bin/kubenswrapper from system_u:object_r:bin_t:s0 to system_u:object_r:kubelet_exec_t:s0 Dec 10 22:49:25 crc kubenswrapper[4791]: Flag --container-runtime-endpoint has been deprecated, This parameter should be set via the config file specified by the Kubelet's --config flag. See https://kubernetes.io/docs/tasks/administer-cluster/kubelet-config-file/ for more information. Dec 10 22:49:25 crc kubenswrapper[4791]: Flag --minimum-container-ttl-duration has been deprecated, Use --eviction-hard or --eviction-soft instead. Will be removed in a future version. Dec 10 22:49:25 crc kubenswrapper[4791]: Flag --volume-plugin-dir has been deprecated, This parameter should be set via the config file specified by the Kubelet's --config flag. See https://kubernetes.io/docs/tasks/administer-cluster/kubelet-config-file/ for more information. Dec 10 22:49:25 crc kubenswrapper[4791]: Flag --register-with-taints has been deprecated, This parameter should be set via the config file specified by the Kubelet's --config flag. See https://kubernetes.io/docs/tasks/administer-cluster/kubelet-config-file/ for more information. Dec 10 22:49:25 crc kubenswrapper[4791]: Flag --pod-infra-container-image has been deprecated, will be removed in a future release. Image garbage collector will get sandbox image information from CRI. Dec 10 22:49:25 crc kubenswrapper[4791]: Flag --system-reserved has been deprecated, This parameter should be set via the config file specified by the Kubelet's --config flag. See https://kubernetes.io/docs/tasks/administer-cluster/kubelet-config-file/ for more information. Dec 10 22:49:25 crc kubenswrapper[4791]: I1210 22:49:25.704859 4791 server.go:211] "--pod-infra-container-image will not be pruned by the image garbage collector in kubelet and should also be set in the remote runtime" Dec 10 22:49:25 crc kubenswrapper[4791]: W1210 22:49:25.708714 4791 feature_gate.go:330] unrecognized feature gate: Example Dec 10 22:49:25 crc kubenswrapper[4791]: W1210 22:49:25.708735 4791 feature_gate.go:330] unrecognized feature gate: IngressControllerLBSubnetsAWS Dec 10 22:49:25 crc kubenswrapper[4791]: W1210 22:49:25.708742 4791 feature_gate.go:330] unrecognized feature gate: InsightsConfigAPI Dec 10 22:49:25 crc kubenswrapper[4791]: W1210 22:49:25.708747 4791 feature_gate.go:330] unrecognized feature gate: InsightsRuntimeExtractor Dec 10 22:49:25 crc kubenswrapper[4791]: W1210 22:49:25.708751 4791 feature_gate.go:330] unrecognized feature gate: AutomatedEtcdBackup Dec 10 22:49:25 crc kubenswrapper[4791]: W1210 22:49:25.708756 4791 feature_gate.go:330] unrecognized feature gate: NetworkLiveMigration Dec 10 22:49:25 crc kubenswrapper[4791]: W1210 22:49:25.708760 4791 feature_gate.go:330] unrecognized feature gate: NetworkDiagnosticsConfig Dec 10 22:49:25 crc kubenswrapper[4791]: W1210 22:49:25.708765 4791 feature_gate.go:330] unrecognized feature gate: RouteAdvertisements Dec 10 22:49:25 crc kubenswrapper[4791]: W1210 22:49:25.708769 4791 feature_gate.go:330] unrecognized feature gate: VSphereControlPlaneMachineSet Dec 10 22:49:25 crc kubenswrapper[4791]: W1210 22:49:25.708774 4791 feature_gate.go:330] unrecognized feature gate: AzureWorkloadIdentity Dec 10 22:49:25 crc kubenswrapper[4791]: W1210 22:49:25.708778 4791 feature_gate.go:330] unrecognized feature gate: InsightsConfig Dec 10 22:49:25 crc kubenswrapper[4791]: W1210 22:49:25.708782 4791 feature_gate.go:330] unrecognized feature gate: NutanixMultiSubnets Dec 10 22:49:25 crc kubenswrapper[4791]: W1210 22:49:25.708787 4791 feature_gate.go:330] unrecognized feature gate: PlatformOperators Dec 10 22:49:25 crc kubenswrapper[4791]: W1210 22:49:25.708793 4791 feature_gate.go:353] Setting GA feature gate CloudDualStackNodeIPs=true. It will be removed in a future release. Dec 10 22:49:25 crc kubenswrapper[4791]: W1210 22:49:25.708799 4791 feature_gate.go:330] unrecognized feature gate: ConsolePluginContentSecurityPolicy Dec 10 22:49:25 crc kubenswrapper[4791]: W1210 22:49:25.708804 4791 feature_gate.go:330] unrecognized feature gate: HardwareSpeed Dec 10 22:49:25 crc kubenswrapper[4791]: W1210 22:49:25.708809 4791 feature_gate.go:330] unrecognized feature gate: AdditionalRoutingCapabilities Dec 10 22:49:25 crc kubenswrapper[4791]: W1210 22:49:25.708814 4791 feature_gate.go:330] unrecognized feature gate: MachineAPIOperatorDisableMachineHealthCheckController Dec 10 22:49:25 crc kubenswrapper[4791]: W1210 22:49:25.708820 4791 feature_gate.go:330] unrecognized feature gate: AdminNetworkPolicy Dec 10 22:49:25 crc kubenswrapper[4791]: W1210 22:49:25.708824 4791 feature_gate.go:330] unrecognized feature gate: MetricsCollectionProfiles Dec 10 22:49:25 crc kubenswrapper[4791]: W1210 22:49:25.708836 4791 feature_gate.go:330] unrecognized feature gate: MachineConfigNodes Dec 10 22:49:25 crc kubenswrapper[4791]: W1210 22:49:25.708841 4791 feature_gate.go:330] unrecognized feature gate: GatewayAPI Dec 10 22:49:25 crc kubenswrapper[4791]: W1210 22:49:25.708846 4791 feature_gate.go:330] unrecognized feature gate: SigstoreImageVerification Dec 10 22:49:25 crc kubenswrapper[4791]: W1210 22:49:25.708852 4791 feature_gate.go:353] Setting GA feature gate DisableKubeletCloudCredentialProviders=true. It will be removed in a future release. Dec 10 22:49:25 crc kubenswrapper[4791]: W1210 22:49:25.708857 4791 feature_gate.go:330] unrecognized feature gate: MultiArchInstallGCP Dec 10 22:49:25 crc kubenswrapper[4791]: W1210 22:49:25.708863 4791 feature_gate.go:330] unrecognized feature gate: CSIDriverSharedResource Dec 10 22:49:25 crc kubenswrapper[4791]: W1210 22:49:25.708867 4791 feature_gate.go:330] unrecognized feature gate: BuildCSIVolumes Dec 10 22:49:25 crc kubenswrapper[4791]: W1210 22:49:25.708872 4791 feature_gate.go:330] unrecognized feature gate: ManagedBootImages Dec 10 22:49:25 crc kubenswrapper[4791]: W1210 22:49:25.708876 4791 feature_gate.go:330] unrecognized feature gate: IngressControllerDynamicConfigurationManager Dec 10 22:49:25 crc kubenswrapper[4791]: W1210 22:49:25.708889 4791 feature_gate.go:353] Setting GA feature gate ValidatingAdmissionPolicy=true. It will be removed in a future release. Dec 10 22:49:25 crc kubenswrapper[4791]: W1210 22:49:25.708895 4791 feature_gate.go:330] unrecognized feature gate: NetworkSegmentation Dec 10 22:49:25 crc kubenswrapper[4791]: W1210 22:49:25.708900 4791 feature_gate.go:330] unrecognized feature gate: InsightsOnDemandDataGather Dec 10 22:49:25 crc kubenswrapper[4791]: W1210 22:49:25.708905 4791 feature_gate.go:330] unrecognized feature gate: NewOLM Dec 10 22:49:25 crc kubenswrapper[4791]: W1210 22:49:25.708909 4791 feature_gate.go:330] unrecognized feature gate: OVNObservability Dec 10 22:49:25 crc kubenswrapper[4791]: W1210 22:49:25.708914 4791 feature_gate.go:330] unrecognized feature gate: AWSClusterHostedDNS Dec 10 22:49:25 crc kubenswrapper[4791]: W1210 22:49:25.708918 4791 feature_gate.go:330] unrecognized feature gate: OpenShiftPodSecurityAdmission Dec 10 22:49:25 crc kubenswrapper[4791]: W1210 22:49:25.708923 4791 feature_gate.go:330] unrecognized feature gate: ManagedBootImagesAWS Dec 10 22:49:25 crc kubenswrapper[4791]: W1210 22:49:25.708927 4791 feature_gate.go:330] unrecognized feature gate: VSphereStaticIPs Dec 10 22:49:25 crc kubenswrapper[4791]: W1210 22:49:25.708932 4791 feature_gate.go:330] unrecognized feature gate: BareMetalLoadBalancer Dec 10 22:49:25 crc kubenswrapper[4791]: W1210 22:49:25.708936 4791 feature_gate.go:330] unrecognized feature gate: MachineAPIMigration Dec 10 22:49:25 crc kubenswrapper[4791]: W1210 22:49:25.708940 4791 feature_gate.go:330] unrecognized feature gate: NodeDisruptionPolicy Dec 10 22:49:25 crc kubenswrapper[4791]: W1210 22:49:25.708945 4791 feature_gate.go:330] unrecognized feature gate: ClusterMonitoringConfig Dec 10 22:49:25 crc kubenswrapper[4791]: W1210 22:49:25.708949 4791 feature_gate.go:330] unrecognized feature gate: PersistentIPsForVirtualization Dec 10 22:49:25 crc kubenswrapper[4791]: W1210 22:49:25.708955 4791 feature_gate.go:351] Setting deprecated feature gate KMSv1=true. It will be removed in a future release. Dec 10 22:49:25 crc kubenswrapper[4791]: W1210 22:49:25.708961 4791 feature_gate.go:330] unrecognized feature gate: GCPClusterHostedDNS Dec 10 22:49:25 crc kubenswrapper[4791]: W1210 22:49:25.708965 4791 feature_gate.go:330] unrecognized feature gate: ChunkSizeMiB Dec 10 22:49:25 crc kubenswrapper[4791]: W1210 22:49:25.708970 4791 feature_gate.go:330] unrecognized feature gate: MultiArchInstallAzure Dec 10 22:49:25 crc kubenswrapper[4791]: W1210 22:49:25.708974 4791 feature_gate.go:330] unrecognized feature gate: DNSNameResolver Dec 10 22:49:25 crc kubenswrapper[4791]: W1210 22:49:25.708979 4791 feature_gate.go:330] unrecognized feature gate: VSphereMultiNetworks Dec 10 22:49:25 crc kubenswrapper[4791]: W1210 22:49:25.708983 4791 feature_gate.go:330] unrecognized feature gate: ClusterAPIInstall Dec 10 22:49:25 crc kubenswrapper[4791]: W1210 22:49:25.708987 4791 feature_gate.go:330] unrecognized feature gate: MixedCPUsAllocation Dec 10 22:49:25 crc kubenswrapper[4791]: W1210 22:49:25.708991 4791 feature_gate.go:330] unrecognized feature gate: ExternalOIDC Dec 10 22:49:25 crc kubenswrapper[4791]: W1210 22:49:25.708996 4791 feature_gate.go:330] unrecognized feature gate: VolumeGroupSnapshot Dec 10 22:49:25 crc kubenswrapper[4791]: W1210 22:49:25.709000 4791 feature_gate.go:330] unrecognized feature gate: SetEIPForNLBIngressController Dec 10 22:49:25 crc kubenswrapper[4791]: W1210 22:49:25.709005 4791 feature_gate.go:330] unrecognized feature gate: PinnedImages Dec 10 22:49:25 crc kubenswrapper[4791]: W1210 22:49:25.709010 4791 feature_gate.go:330] unrecognized feature gate: ClusterAPIInstallIBMCloud Dec 10 22:49:25 crc kubenswrapper[4791]: W1210 22:49:25.709017 4791 feature_gate.go:330] unrecognized feature gate: OnClusterBuild Dec 10 22:49:25 crc kubenswrapper[4791]: W1210 22:49:25.709022 4791 feature_gate.go:330] unrecognized feature gate: ImageStreamImportMode Dec 10 22:49:25 crc kubenswrapper[4791]: W1210 22:49:25.709026 4791 feature_gate.go:330] unrecognized feature gate: GCPLabelsTags Dec 10 22:49:25 crc kubenswrapper[4791]: W1210 22:49:25.709031 4791 feature_gate.go:330] unrecognized feature gate: VSphereDriverConfiguration Dec 10 22:49:25 crc kubenswrapper[4791]: W1210 22:49:25.709036 4791 feature_gate.go:330] unrecognized feature gate: MultiArchInstallAWS Dec 10 22:49:25 crc kubenswrapper[4791]: W1210 22:49:25.709040 4791 feature_gate.go:330] unrecognized feature gate: SignatureStores Dec 10 22:49:25 crc kubenswrapper[4791]: W1210 22:49:25.709045 4791 feature_gate.go:330] unrecognized feature gate: PrivateHostedZoneAWS Dec 10 22:49:25 crc kubenswrapper[4791]: W1210 22:49:25.709049 4791 feature_gate.go:330] unrecognized feature gate: AWSEFSDriverVolumeMetrics Dec 10 22:49:25 crc kubenswrapper[4791]: W1210 22:49:25.709054 4791 feature_gate.go:330] unrecognized feature gate: AlibabaPlatform Dec 10 22:49:25 crc kubenswrapper[4791]: W1210 22:49:25.709058 4791 feature_gate.go:330] unrecognized feature gate: BootcNodeManagement Dec 10 22:49:25 crc kubenswrapper[4791]: W1210 22:49:25.709062 4791 feature_gate.go:330] unrecognized feature gate: UpgradeStatus Dec 10 22:49:25 crc kubenswrapper[4791]: W1210 22:49:25.709066 4791 feature_gate.go:330] unrecognized feature gate: MinimumKubeletVersion Dec 10 22:49:25 crc kubenswrapper[4791]: W1210 22:49:25.709071 4791 feature_gate.go:330] unrecognized feature gate: EtcdBackendQuota Dec 10 22:49:25 crc kubenswrapper[4791]: W1210 22:49:25.709075 4791 feature_gate.go:330] unrecognized feature gate: VSphereMultiVCenters Dec 10 22:49:25 crc kubenswrapper[4791]: W1210 22:49:25.709080 4791 feature_gate.go:330] unrecognized feature gate: MachineAPIProviderOpenStack Dec 10 22:49:25 crc kubenswrapper[4791]: I1210 22:49:25.709183 4791 flags.go:64] FLAG: --address="0.0.0.0" Dec 10 22:49:25 crc kubenswrapper[4791]: I1210 22:49:25.709193 4791 flags.go:64] FLAG: --allowed-unsafe-sysctls="[]" Dec 10 22:49:25 crc kubenswrapper[4791]: I1210 22:49:25.709201 4791 flags.go:64] FLAG: --anonymous-auth="true" Dec 10 22:49:25 crc kubenswrapper[4791]: I1210 22:49:25.709208 4791 flags.go:64] FLAG: --application-metrics-count-limit="100" Dec 10 22:49:25 crc kubenswrapper[4791]: I1210 22:49:25.709215 4791 flags.go:64] FLAG: --authentication-token-webhook="false" Dec 10 22:49:25 crc kubenswrapper[4791]: I1210 22:49:25.709221 4791 flags.go:64] FLAG: --authentication-token-webhook-cache-ttl="2m0s" Dec 10 22:49:25 crc kubenswrapper[4791]: I1210 22:49:25.709229 4791 flags.go:64] FLAG: --authorization-mode="AlwaysAllow" Dec 10 22:49:25 crc kubenswrapper[4791]: I1210 22:49:25.709236 4791 flags.go:64] FLAG: --authorization-webhook-cache-authorized-ttl="5m0s" Dec 10 22:49:25 crc kubenswrapper[4791]: I1210 22:49:25.709242 4791 flags.go:64] FLAG: --authorization-webhook-cache-unauthorized-ttl="30s" Dec 10 22:49:25 crc kubenswrapper[4791]: I1210 22:49:25.709248 4791 flags.go:64] FLAG: --boot-id-file="/proc/sys/kernel/random/boot_id" Dec 10 22:49:25 crc kubenswrapper[4791]: I1210 22:49:25.709254 4791 flags.go:64] FLAG: --bootstrap-kubeconfig="/etc/kubernetes/kubeconfig" Dec 10 22:49:25 crc kubenswrapper[4791]: I1210 22:49:25.709259 4791 flags.go:64] FLAG: --cert-dir="/var/lib/kubelet/pki" Dec 10 22:49:25 crc kubenswrapper[4791]: I1210 22:49:25.709265 4791 flags.go:64] FLAG: --cgroup-driver="cgroupfs" Dec 10 22:49:25 crc kubenswrapper[4791]: I1210 22:49:25.709271 4791 flags.go:64] FLAG: --cgroup-root="" Dec 10 22:49:25 crc kubenswrapper[4791]: I1210 22:49:25.709276 4791 flags.go:64] FLAG: --cgroups-per-qos="true" Dec 10 22:49:25 crc kubenswrapper[4791]: I1210 22:49:25.709283 4791 flags.go:64] FLAG: --client-ca-file="" Dec 10 22:49:25 crc kubenswrapper[4791]: I1210 22:49:25.709288 4791 flags.go:64] FLAG: --cloud-config="" Dec 10 22:49:25 crc kubenswrapper[4791]: I1210 22:49:25.709293 4791 flags.go:64] FLAG: --cloud-provider="" Dec 10 22:49:25 crc kubenswrapper[4791]: I1210 22:49:25.709299 4791 flags.go:64] FLAG: --cluster-dns="[]" Dec 10 22:49:25 crc kubenswrapper[4791]: I1210 22:49:25.709306 4791 flags.go:64] FLAG: --cluster-domain="" Dec 10 22:49:25 crc kubenswrapper[4791]: I1210 22:49:25.709311 4791 flags.go:64] FLAG: --config="/etc/kubernetes/kubelet.conf" Dec 10 22:49:25 crc kubenswrapper[4791]: I1210 22:49:25.709317 4791 flags.go:64] FLAG: --config-dir="" Dec 10 22:49:25 crc kubenswrapper[4791]: I1210 22:49:25.709322 4791 flags.go:64] FLAG: --container-hints="/etc/cadvisor/container_hints.json" Dec 10 22:49:25 crc kubenswrapper[4791]: I1210 22:49:25.709328 4791 flags.go:64] FLAG: --container-log-max-files="5" Dec 10 22:49:25 crc kubenswrapper[4791]: I1210 22:49:25.709384 4791 flags.go:64] FLAG: --container-log-max-size="10Mi" Dec 10 22:49:25 crc kubenswrapper[4791]: I1210 22:49:25.709390 4791 flags.go:64] FLAG: --container-runtime-endpoint="/var/run/crio/crio.sock" Dec 10 22:49:25 crc kubenswrapper[4791]: I1210 22:49:25.709395 4791 flags.go:64] FLAG: --containerd="/run/containerd/containerd.sock" Dec 10 22:49:25 crc kubenswrapper[4791]: I1210 22:49:25.709401 4791 flags.go:64] FLAG: --containerd-namespace="k8s.io" Dec 10 22:49:25 crc kubenswrapper[4791]: I1210 22:49:25.709406 4791 flags.go:64] FLAG: --contention-profiling="false" Dec 10 22:49:25 crc kubenswrapper[4791]: I1210 22:49:25.709412 4791 flags.go:64] FLAG: --cpu-cfs-quota="true" Dec 10 22:49:25 crc kubenswrapper[4791]: I1210 22:49:25.709417 4791 flags.go:64] FLAG: --cpu-cfs-quota-period="100ms" Dec 10 22:49:25 crc kubenswrapper[4791]: I1210 22:49:25.709422 4791 flags.go:64] FLAG: --cpu-manager-policy="none" Dec 10 22:49:25 crc kubenswrapper[4791]: I1210 22:49:25.709427 4791 flags.go:64] FLAG: --cpu-manager-policy-options="" Dec 10 22:49:25 crc kubenswrapper[4791]: I1210 22:49:25.709433 4791 flags.go:64] FLAG: --cpu-manager-reconcile-period="10s" Dec 10 22:49:25 crc kubenswrapper[4791]: I1210 22:49:25.709438 4791 flags.go:64] FLAG: --enable-controller-attach-detach="true" Dec 10 22:49:25 crc kubenswrapper[4791]: I1210 22:49:25.709444 4791 flags.go:64] FLAG: --enable-debugging-handlers="true" Dec 10 22:49:25 crc kubenswrapper[4791]: I1210 22:49:25.709449 4791 flags.go:64] FLAG: --enable-load-reader="false" Dec 10 22:49:25 crc kubenswrapper[4791]: I1210 22:49:25.709454 4791 flags.go:64] FLAG: --enable-server="true" Dec 10 22:49:25 crc kubenswrapper[4791]: I1210 22:49:25.709459 4791 flags.go:64] FLAG: --enforce-node-allocatable="[pods]" Dec 10 22:49:25 crc kubenswrapper[4791]: I1210 22:49:25.709468 4791 flags.go:64] FLAG: --event-burst="100" Dec 10 22:49:25 crc kubenswrapper[4791]: I1210 22:49:25.709474 4791 flags.go:64] FLAG: --event-qps="50" Dec 10 22:49:25 crc kubenswrapper[4791]: I1210 22:49:25.709479 4791 flags.go:64] FLAG: --event-storage-age-limit="default=0" Dec 10 22:49:25 crc kubenswrapper[4791]: I1210 22:49:25.709484 4791 flags.go:64] FLAG: --event-storage-event-limit="default=0" Dec 10 22:49:25 crc kubenswrapper[4791]: I1210 22:49:25.709489 4791 flags.go:64] FLAG: --eviction-hard="" Dec 10 22:49:25 crc kubenswrapper[4791]: I1210 22:49:25.709496 4791 flags.go:64] FLAG: --eviction-max-pod-grace-period="0" Dec 10 22:49:25 crc kubenswrapper[4791]: I1210 22:49:25.709502 4791 flags.go:64] FLAG: --eviction-minimum-reclaim="" Dec 10 22:49:25 crc kubenswrapper[4791]: I1210 22:49:25.709507 4791 flags.go:64] FLAG: --eviction-pressure-transition-period="5m0s" Dec 10 22:49:25 crc kubenswrapper[4791]: I1210 22:49:25.709513 4791 flags.go:64] FLAG: --eviction-soft="" Dec 10 22:49:25 crc kubenswrapper[4791]: I1210 22:49:25.709519 4791 flags.go:64] FLAG: --eviction-soft-grace-period="" Dec 10 22:49:25 crc kubenswrapper[4791]: I1210 22:49:25.709524 4791 flags.go:64] FLAG: --exit-on-lock-contention="false" Dec 10 22:49:25 crc kubenswrapper[4791]: I1210 22:49:25.709529 4791 flags.go:64] FLAG: --experimental-allocatable-ignore-eviction="false" Dec 10 22:49:25 crc kubenswrapper[4791]: I1210 22:49:25.709535 4791 flags.go:64] FLAG: --experimental-mounter-path="" Dec 10 22:49:25 crc kubenswrapper[4791]: I1210 22:49:25.709540 4791 flags.go:64] FLAG: --fail-cgroupv1="false" Dec 10 22:49:25 crc kubenswrapper[4791]: I1210 22:49:25.709545 4791 flags.go:64] FLAG: --fail-swap-on="true" Dec 10 22:49:25 crc kubenswrapper[4791]: I1210 22:49:25.709550 4791 flags.go:64] FLAG: --feature-gates="" Dec 10 22:49:25 crc kubenswrapper[4791]: I1210 22:49:25.709557 4791 flags.go:64] FLAG: --file-check-frequency="20s" Dec 10 22:49:25 crc kubenswrapper[4791]: I1210 22:49:25.709562 4791 flags.go:64] FLAG: --global-housekeeping-interval="1m0s" Dec 10 22:49:25 crc kubenswrapper[4791]: I1210 22:49:25.709569 4791 flags.go:64] FLAG: --hairpin-mode="promiscuous-bridge" Dec 10 22:49:25 crc kubenswrapper[4791]: I1210 22:49:25.709574 4791 flags.go:64] FLAG: --healthz-bind-address="127.0.0.1" Dec 10 22:49:25 crc kubenswrapper[4791]: I1210 22:49:25.709580 4791 flags.go:64] FLAG: --healthz-port="10248" Dec 10 22:49:25 crc kubenswrapper[4791]: I1210 22:49:25.709585 4791 flags.go:64] FLAG: --help="false" Dec 10 22:49:25 crc kubenswrapper[4791]: I1210 22:49:25.709591 4791 flags.go:64] FLAG: --hostname-override="" Dec 10 22:49:25 crc kubenswrapper[4791]: I1210 22:49:25.709596 4791 flags.go:64] FLAG: --housekeeping-interval="10s" Dec 10 22:49:25 crc kubenswrapper[4791]: I1210 22:49:25.709602 4791 flags.go:64] FLAG: --http-check-frequency="20s" Dec 10 22:49:25 crc kubenswrapper[4791]: I1210 22:49:25.709607 4791 flags.go:64] FLAG: --image-credential-provider-bin-dir="" Dec 10 22:49:25 crc kubenswrapper[4791]: I1210 22:49:25.709612 4791 flags.go:64] FLAG: --image-credential-provider-config="" Dec 10 22:49:25 crc kubenswrapper[4791]: I1210 22:49:25.709617 4791 flags.go:64] FLAG: --image-gc-high-threshold="85" Dec 10 22:49:25 crc kubenswrapper[4791]: I1210 22:49:25.709622 4791 flags.go:64] FLAG: --image-gc-low-threshold="80" Dec 10 22:49:25 crc kubenswrapper[4791]: I1210 22:49:25.709627 4791 flags.go:64] FLAG: --image-service-endpoint="" Dec 10 22:49:25 crc kubenswrapper[4791]: I1210 22:49:25.709632 4791 flags.go:64] FLAG: --kernel-memcg-notification="false" Dec 10 22:49:25 crc kubenswrapper[4791]: I1210 22:49:25.709637 4791 flags.go:64] FLAG: --kube-api-burst="100" Dec 10 22:49:25 crc kubenswrapper[4791]: I1210 22:49:25.709643 4791 flags.go:64] FLAG: --kube-api-content-type="application/vnd.kubernetes.protobuf" Dec 10 22:49:25 crc kubenswrapper[4791]: I1210 22:49:25.709649 4791 flags.go:64] FLAG: --kube-api-qps="50" Dec 10 22:49:25 crc kubenswrapper[4791]: I1210 22:49:25.709654 4791 flags.go:64] FLAG: --kube-reserved="" Dec 10 22:49:25 crc kubenswrapper[4791]: I1210 22:49:25.709659 4791 flags.go:64] FLAG: --kube-reserved-cgroup="" Dec 10 22:49:25 crc kubenswrapper[4791]: I1210 22:49:25.709664 4791 flags.go:64] FLAG: --kubeconfig="/var/lib/kubelet/kubeconfig" Dec 10 22:49:25 crc kubenswrapper[4791]: I1210 22:49:25.709670 4791 flags.go:64] FLAG: --kubelet-cgroups="" Dec 10 22:49:25 crc kubenswrapper[4791]: I1210 22:49:25.709675 4791 flags.go:64] FLAG: --local-storage-capacity-isolation="true" Dec 10 22:49:25 crc kubenswrapper[4791]: I1210 22:49:25.709680 4791 flags.go:64] FLAG: --lock-file="" Dec 10 22:49:25 crc kubenswrapper[4791]: I1210 22:49:25.709686 4791 flags.go:64] FLAG: --log-cadvisor-usage="false" Dec 10 22:49:25 crc kubenswrapper[4791]: I1210 22:49:25.709691 4791 flags.go:64] FLAG: --log-flush-frequency="5s" Dec 10 22:49:25 crc kubenswrapper[4791]: I1210 22:49:25.709696 4791 flags.go:64] FLAG: --log-json-info-buffer-size="0" Dec 10 22:49:25 crc kubenswrapper[4791]: I1210 22:49:25.709705 4791 flags.go:64] FLAG: --log-json-split-stream="false" Dec 10 22:49:25 crc kubenswrapper[4791]: I1210 22:49:25.709710 4791 flags.go:64] FLAG: --log-text-info-buffer-size="0" Dec 10 22:49:25 crc kubenswrapper[4791]: I1210 22:49:25.709715 4791 flags.go:64] FLAG: --log-text-split-stream="false" Dec 10 22:49:25 crc kubenswrapper[4791]: I1210 22:49:25.709720 4791 flags.go:64] FLAG: --logging-format="text" Dec 10 22:49:25 crc kubenswrapper[4791]: I1210 22:49:25.709726 4791 flags.go:64] FLAG: --machine-id-file="/etc/machine-id,/var/lib/dbus/machine-id" Dec 10 22:49:25 crc kubenswrapper[4791]: I1210 22:49:25.709732 4791 flags.go:64] FLAG: --make-iptables-util-chains="true" Dec 10 22:49:25 crc kubenswrapper[4791]: I1210 22:49:25.709737 4791 flags.go:64] FLAG: --manifest-url="" Dec 10 22:49:25 crc kubenswrapper[4791]: I1210 22:49:25.709742 4791 flags.go:64] FLAG: --manifest-url-header="" Dec 10 22:49:25 crc kubenswrapper[4791]: I1210 22:49:25.709749 4791 flags.go:64] FLAG: --max-housekeeping-interval="15s" Dec 10 22:49:25 crc kubenswrapper[4791]: I1210 22:49:25.709755 4791 flags.go:64] FLAG: --max-open-files="1000000" Dec 10 22:49:25 crc kubenswrapper[4791]: I1210 22:49:25.709761 4791 flags.go:64] FLAG: --max-pods="110" Dec 10 22:49:25 crc kubenswrapper[4791]: I1210 22:49:25.709767 4791 flags.go:64] FLAG: --maximum-dead-containers="-1" Dec 10 22:49:25 crc kubenswrapper[4791]: I1210 22:49:25.709772 4791 flags.go:64] FLAG: --maximum-dead-containers-per-container="1" Dec 10 22:49:25 crc kubenswrapper[4791]: I1210 22:49:25.709778 4791 flags.go:64] FLAG: --memory-manager-policy="None" Dec 10 22:49:25 crc kubenswrapper[4791]: I1210 22:49:25.709783 4791 flags.go:64] FLAG: --minimum-container-ttl-duration="6m0s" Dec 10 22:49:25 crc kubenswrapper[4791]: I1210 22:49:25.709788 4791 flags.go:64] FLAG: --minimum-image-ttl-duration="2m0s" Dec 10 22:49:25 crc kubenswrapper[4791]: I1210 22:49:25.709793 4791 flags.go:64] FLAG: --node-ip="192.168.126.11" Dec 10 22:49:25 crc kubenswrapper[4791]: I1210 22:49:25.709798 4791 flags.go:64] FLAG: --node-labels="node-role.kubernetes.io/control-plane=,node-role.kubernetes.io/master=,node.openshift.io/os_id=rhcos" Dec 10 22:49:25 crc kubenswrapper[4791]: I1210 22:49:25.709811 4791 flags.go:64] FLAG: --node-status-max-images="50" Dec 10 22:49:25 crc kubenswrapper[4791]: I1210 22:49:25.709816 4791 flags.go:64] FLAG: --node-status-update-frequency="10s" Dec 10 22:49:25 crc kubenswrapper[4791]: I1210 22:49:25.709821 4791 flags.go:64] FLAG: --oom-score-adj="-999" Dec 10 22:49:25 crc kubenswrapper[4791]: I1210 22:49:25.709827 4791 flags.go:64] FLAG: --pod-cidr="" Dec 10 22:49:25 crc kubenswrapper[4791]: I1210 22:49:25.709833 4791 flags.go:64] FLAG: --pod-infra-container-image="quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:33549946e22a9ffa738fd94b1345f90921bc8f92fa6137784cb33c77ad806f9d" Dec 10 22:49:25 crc kubenswrapper[4791]: I1210 22:49:25.709843 4791 flags.go:64] FLAG: --pod-manifest-path="" Dec 10 22:49:25 crc kubenswrapper[4791]: I1210 22:49:25.709848 4791 flags.go:64] FLAG: --pod-max-pids="-1" Dec 10 22:49:25 crc kubenswrapper[4791]: I1210 22:49:25.709853 4791 flags.go:64] FLAG: --pods-per-core="0" Dec 10 22:49:25 crc kubenswrapper[4791]: I1210 22:49:25.709859 4791 flags.go:64] FLAG: --port="10250" Dec 10 22:49:25 crc kubenswrapper[4791]: I1210 22:49:25.709864 4791 flags.go:64] FLAG: --protect-kernel-defaults="false" Dec 10 22:49:25 crc kubenswrapper[4791]: I1210 22:49:25.709869 4791 flags.go:64] FLAG: --provider-id="" Dec 10 22:49:25 crc kubenswrapper[4791]: I1210 22:49:25.709875 4791 flags.go:64] FLAG: --qos-reserved="" Dec 10 22:49:25 crc kubenswrapper[4791]: I1210 22:49:25.709881 4791 flags.go:64] FLAG: --read-only-port="10255" Dec 10 22:49:25 crc kubenswrapper[4791]: I1210 22:49:25.709887 4791 flags.go:64] FLAG: --register-node="true" Dec 10 22:49:25 crc kubenswrapper[4791]: I1210 22:49:25.709893 4791 flags.go:64] FLAG: --register-schedulable="true" Dec 10 22:49:25 crc kubenswrapper[4791]: I1210 22:49:25.709898 4791 flags.go:64] FLAG: --register-with-taints="node-role.kubernetes.io/master=:NoSchedule" Dec 10 22:49:25 crc kubenswrapper[4791]: I1210 22:49:25.709907 4791 flags.go:64] FLAG: --registry-burst="10" Dec 10 22:49:25 crc kubenswrapper[4791]: I1210 22:49:25.709913 4791 flags.go:64] FLAG: --registry-qps="5" Dec 10 22:49:25 crc kubenswrapper[4791]: I1210 22:49:25.709918 4791 flags.go:64] FLAG: --reserved-cpus="" Dec 10 22:49:25 crc kubenswrapper[4791]: I1210 22:49:25.709923 4791 flags.go:64] FLAG: --reserved-memory="" Dec 10 22:49:25 crc kubenswrapper[4791]: I1210 22:49:25.709930 4791 flags.go:64] FLAG: --resolv-conf="/etc/resolv.conf" Dec 10 22:49:25 crc kubenswrapper[4791]: I1210 22:49:25.709936 4791 flags.go:64] FLAG: --root-dir="/var/lib/kubelet" Dec 10 22:49:25 crc kubenswrapper[4791]: I1210 22:49:25.709942 4791 flags.go:64] FLAG: --rotate-certificates="false" Dec 10 22:49:25 crc kubenswrapper[4791]: I1210 22:49:25.709947 4791 flags.go:64] FLAG: --rotate-server-certificates="false" Dec 10 22:49:25 crc kubenswrapper[4791]: I1210 22:49:25.709952 4791 flags.go:64] FLAG: --runonce="false" Dec 10 22:49:25 crc kubenswrapper[4791]: I1210 22:49:25.709957 4791 flags.go:64] FLAG: --runtime-cgroups="/system.slice/crio.service" Dec 10 22:49:25 crc kubenswrapper[4791]: I1210 22:49:25.709962 4791 flags.go:64] FLAG: --runtime-request-timeout="2m0s" Dec 10 22:49:25 crc kubenswrapper[4791]: I1210 22:49:25.709968 4791 flags.go:64] FLAG: --seccomp-default="false" Dec 10 22:49:25 crc kubenswrapper[4791]: I1210 22:49:25.709973 4791 flags.go:64] FLAG: --serialize-image-pulls="true" Dec 10 22:49:25 crc kubenswrapper[4791]: I1210 22:49:25.709979 4791 flags.go:64] FLAG: --storage-driver-buffer-duration="1m0s" Dec 10 22:49:25 crc kubenswrapper[4791]: I1210 22:49:25.709984 4791 flags.go:64] FLAG: --storage-driver-db="cadvisor" Dec 10 22:49:25 crc kubenswrapper[4791]: I1210 22:49:25.709989 4791 flags.go:64] FLAG: --storage-driver-host="localhost:8086" Dec 10 22:49:25 crc kubenswrapper[4791]: I1210 22:49:25.709994 4791 flags.go:64] FLAG: --storage-driver-password="root" Dec 10 22:49:25 crc kubenswrapper[4791]: I1210 22:49:25.710000 4791 flags.go:64] FLAG: --storage-driver-secure="false" Dec 10 22:49:25 crc kubenswrapper[4791]: I1210 22:49:25.710005 4791 flags.go:64] FLAG: --storage-driver-table="stats" Dec 10 22:49:25 crc kubenswrapper[4791]: I1210 22:49:25.710011 4791 flags.go:64] FLAG: --storage-driver-user="root" Dec 10 22:49:25 crc kubenswrapper[4791]: I1210 22:49:25.710016 4791 flags.go:64] FLAG: --streaming-connection-idle-timeout="4h0m0s" Dec 10 22:49:25 crc kubenswrapper[4791]: I1210 22:49:25.710021 4791 flags.go:64] FLAG: --sync-frequency="1m0s" Dec 10 22:49:25 crc kubenswrapper[4791]: I1210 22:49:25.710026 4791 flags.go:64] FLAG: --system-cgroups="" Dec 10 22:49:25 crc kubenswrapper[4791]: I1210 22:49:25.710031 4791 flags.go:64] FLAG: --system-reserved="cpu=200m,ephemeral-storage=350Mi,memory=350Mi" Dec 10 22:49:25 crc kubenswrapper[4791]: I1210 22:49:25.710039 4791 flags.go:64] FLAG: --system-reserved-cgroup="" Dec 10 22:49:25 crc kubenswrapper[4791]: I1210 22:49:25.710044 4791 flags.go:64] FLAG: --tls-cert-file="" Dec 10 22:49:25 crc kubenswrapper[4791]: I1210 22:49:25.710049 4791 flags.go:64] FLAG: --tls-cipher-suites="[]" Dec 10 22:49:25 crc kubenswrapper[4791]: I1210 22:49:25.710064 4791 flags.go:64] FLAG: --tls-min-version="" Dec 10 22:49:25 crc kubenswrapper[4791]: I1210 22:49:25.710069 4791 flags.go:64] FLAG: --tls-private-key-file="" Dec 10 22:49:25 crc kubenswrapper[4791]: I1210 22:49:25.710074 4791 flags.go:64] FLAG: --topology-manager-policy="none" Dec 10 22:49:25 crc kubenswrapper[4791]: I1210 22:49:25.710079 4791 flags.go:64] FLAG: --topology-manager-policy-options="" Dec 10 22:49:25 crc kubenswrapper[4791]: I1210 22:49:25.710084 4791 flags.go:64] FLAG: --topology-manager-scope="container" Dec 10 22:49:25 crc kubenswrapper[4791]: I1210 22:49:25.710089 4791 flags.go:64] FLAG: --v="2" Dec 10 22:49:25 crc kubenswrapper[4791]: I1210 22:49:25.710103 4791 flags.go:64] FLAG: --version="false" Dec 10 22:49:25 crc kubenswrapper[4791]: I1210 22:49:25.710110 4791 flags.go:64] FLAG: --vmodule="" Dec 10 22:49:25 crc kubenswrapper[4791]: I1210 22:49:25.710116 4791 flags.go:64] FLAG: --volume-plugin-dir="/etc/kubernetes/kubelet-plugins/volume/exec" Dec 10 22:49:25 crc kubenswrapper[4791]: I1210 22:49:25.710122 4791 flags.go:64] FLAG: --volume-stats-agg-period="1m0s" Dec 10 22:49:25 crc kubenswrapper[4791]: W1210 22:49:25.710235 4791 feature_gate.go:330] unrecognized feature gate: ClusterAPIInstall Dec 10 22:49:25 crc kubenswrapper[4791]: W1210 22:49:25.710241 4791 feature_gate.go:330] unrecognized feature gate: AlibabaPlatform Dec 10 22:49:25 crc kubenswrapper[4791]: W1210 22:49:25.710246 4791 feature_gate.go:330] unrecognized feature gate: ManagedBootImages Dec 10 22:49:25 crc kubenswrapper[4791]: W1210 22:49:25.710251 4791 feature_gate.go:330] unrecognized feature gate: PersistentIPsForVirtualization Dec 10 22:49:25 crc kubenswrapper[4791]: W1210 22:49:25.710256 4791 feature_gate.go:330] unrecognized feature gate: VolumeGroupSnapshot Dec 10 22:49:25 crc kubenswrapper[4791]: W1210 22:49:25.710260 4791 feature_gate.go:330] unrecognized feature gate: RouteAdvertisements Dec 10 22:49:25 crc kubenswrapper[4791]: W1210 22:49:25.710265 4791 feature_gate.go:330] unrecognized feature gate: OpenShiftPodSecurityAdmission Dec 10 22:49:25 crc kubenswrapper[4791]: W1210 22:49:25.710270 4791 feature_gate.go:330] unrecognized feature gate: NetworkLiveMigration Dec 10 22:49:25 crc kubenswrapper[4791]: W1210 22:49:25.710274 4791 feature_gate.go:330] unrecognized feature gate: MetricsCollectionProfiles Dec 10 22:49:25 crc kubenswrapper[4791]: W1210 22:49:25.710278 4791 feature_gate.go:330] unrecognized feature gate: NutanixMultiSubnets Dec 10 22:49:25 crc kubenswrapper[4791]: W1210 22:49:25.710283 4791 feature_gate.go:330] unrecognized feature gate: OnClusterBuild Dec 10 22:49:25 crc kubenswrapper[4791]: W1210 22:49:25.710288 4791 feature_gate.go:330] unrecognized feature gate: AWSClusterHostedDNS Dec 10 22:49:25 crc kubenswrapper[4791]: W1210 22:49:25.710293 4791 feature_gate.go:330] unrecognized feature gate: SetEIPForNLBIngressController Dec 10 22:49:25 crc kubenswrapper[4791]: W1210 22:49:25.710298 4791 feature_gate.go:330] unrecognized feature gate: HardwareSpeed Dec 10 22:49:25 crc kubenswrapper[4791]: W1210 22:49:25.710306 4791 feature_gate.go:330] unrecognized feature gate: VSphereMultiVCenters Dec 10 22:49:25 crc kubenswrapper[4791]: W1210 22:49:25.710310 4791 feature_gate.go:330] unrecognized feature gate: SignatureStores Dec 10 22:49:25 crc kubenswrapper[4791]: W1210 22:49:25.710314 4791 feature_gate.go:330] unrecognized feature gate: Example Dec 10 22:49:25 crc kubenswrapper[4791]: W1210 22:49:25.710319 4791 feature_gate.go:330] unrecognized feature gate: NetworkDiagnosticsConfig Dec 10 22:49:25 crc kubenswrapper[4791]: W1210 22:49:25.710324 4791 feature_gate.go:330] unrecognized feature gate: ClusterMonitoringConfig Dec 10 22:49:25 crc kubenswrapper[4791]: W1210 22:49:25.710328 4791 feature_gate.go:330] unrecognized feature gate: AWSEFSDriverVolumeMetrics Dec 10 22:49:25 crc kubenswrapper[4791]: W1210 22:49:25.710332 4791 feature_gate.go:330] unrecognized feature gate: InsightsRuntimeExtractor Dec 10 22:49:25 crc kubenswrapper[4791]: W1210 22:49:25.710352 4791 feature_gate.go:330] unrecognized feature gate: OVNObservability Dec 10 22:49:25 crc kubenswrapper[4791]: W1210 22:49:25.710388 4791 feature_gate.go:330] unrecognized feature gate: InsightsConfigAPI Dec 10 22:49:25 crc kubenswrapper[4791]: W1210 22:49:25.710393 4791 feature_gate.go:330] unrecognized feature gate: AdminNetworkPolicy Dec 10 22:49:25 crc kubenswrapper[4791]: W1210 22:49:25.710398 4791 feature_gate.go:330] unrecognized feature gate: ManagedBootImagesAWS Dec 10 22:49:25 crc kubenswrapper[4791]: W1210 22:49:25.710402 4791 feature_gate.go:330] unrecognized feature gate: NodeDisruptionPolicy Dec 10 22:49:25 crc kubenswrapper[4791]: W1210 22:49:25.710407 4791 feature_gate.go:330] unrecognized feature gate: MachineAPIMigration Dec 10 22:49:25 crc kubenswrapper[4791]: W1210 22:49:25.710412 4791 feature_gate.go:330] unrecognized feature gate: InsightsOnDemandDataGather Dec 10 22:49:25 crc kubenswrapper[4791]: W1210 22:49:25.710416 4791 feature_gate.go:330] unrecognized feature gate: NetworkSegmentation Dec 10 22:49:25 crc kubenswrapper[4791]: W1210 22:49:25.710422 4791 feature_gate.go:351] Setting deprecated feature gate KMSv1=true. It will be removed in a future release. Dec 10 22:49:25 crc kubenswrapper[4791]: W1210 22:49:25.710427 4791 feature_gate.go:330] unrecognized feature gate: GatewayAPI Dec 10 22:49:25 crc kubenswrapper[4791]: W1210 22:49:25.710432 4791 feature_gate.go:330] unrecognized feature gate: ImageStreamImportMode Dec 10 22:49:25 crc kubenswrapper[4791]: W1210 22:49:25.710436 4791 feature_gate.go:330] unrecognized feature gate: IngressControllerDynamicConfigurationManager Dec 10 22:49:25 crc kubenswrapper[4791]: W1210 22:49:25.710441 4791 feature_gate.go:330] unrecognized feature gate: MultiArchInstallAzure Dec 10 22:49:25 crc kubenswrapper[4791]: W1210 22:49:25.710446 4791 feature_gate.go:330] unrecognized feature gate: VSphereDriverConfiguration Dec 10 22:49:25 crc kubenswrapper[4791]: W1210 22:49:25.710450 4791 feature_gate.go:330] unrecognized feature gate: BareMetalLoadBalancer Dec 10 22:49:25 crc kubenswrapper[4791]: W1210 22:49:25.710455 4791 feature_gate.go:330] unrecognized feature gate: DNSNameResolver Dec 10 22:49:25 crc kubenswrapper[4791]: W1210 22:49:25.710459 4791 feature_gate.go:330] unrecognized feature gate: MachineConfigNodes Dec 10 22:49:25 crc kubenswrapper[4791]: W1210 22:49:25.710463 4791 feature_gate.go:330] unrecognized feature gate: SigstoreImageVerification Dec 10 22:49:25 crc kubenswrapper[4791]: W1210 22:49:25.710468 4791 feature_gate.go:330] unrecognized feature gate: PinnedImages Dec 10 22:49:25 crc kubenswrapper[4791]: W1210 22:49:25.710472 4791 feature_gate.go:330] unrecognized feature gate: UpgradeStatus Dec 10 22:49:25 crc kubenswrapper[4791]: W1210 22:49:25.710476 4791 feature_gate.go:330] unrecognized feature gate: ClusterAPIInstallIBMCloud Dec 10 22:49:25 crc kubenswrapper[4791]: W1210 22:49:25.710481 4791 feature_gate.go:330] unrecognized feature gate: MultiArchInstallAWS Dec 10 22:49:25 crc kubenswrapper[4791]: W1210 22:49:25.710486 4791 feature_gate.go:330] unrecognized feature gate: AutomatedEtcdBackup Dec 10 22:49:25 crc kubenswrapper[4791]: W1210 22:49:25.710490 4791 feature_gate.go:330] unrecognized feature gate: ExternalOIDC Dec 10 22:49:25 crc kubenswrapper[4791]: W1210 22:49:25.710495 4791 feature_gate.go:330] unrecognized feature gate: VSphereControlPlaneMachineSet Dec 10 22:49:25 crc kubenswrapper[4791]: W1210 22:49:25.710502 4791 feature_gate.go:330] unrecognized feature gate: EtcdBackendQuota Dec 10 22:49:25 crc kubenswrapper[4791]: W1210 22:49:25.710506 4791 feature_gate.go:330] unrecognized feature gate: ConsolePluginContentSecurityPolicy Dec 10 22:49:25 crc kubenswrapper[4791]: W1210 22:49:25.710511 4791 feature_gate.go:330] unrecognized feature gate: MachineAPIProviderOpenStack Dec 10 22:49:25 crc kubenswrapper[4791]: W1210 22:49:25.710516 4791 feature_gate.go:330] unrecognized feature gate: GCPLabelsTags Dec 10 22:49:25 crc kubenswrapper[4791]: W1210 22:49:25.710520 4791 feature_gate.go:330] unrecognized feature gate: CSIDriverSharedResource Dec 10 22:49:25 crc kubenswrapper[4791]: W1210 22:49:25.710525 4791 feature_gate.go:330] unrecognized feature gate: MixedCPUsAllocation Dec 10 22:49:25 crc kubenswrapper[4791]: W1210 22:49:25.710530 4791 feature_gate.go:330] unrecognized feature gate: AzureWorkloadIdentity Dec 10 22:49:25 crc kubenswrapper[4791]: W1210 22:49:25.710534 4791 feature_gate.go:330] unrecognized feature gate: GCPClusterHostedDNS Dec 10 22:49:25 crc kubenswrapper[4791]: W1210 22:49:25.710541 4791 feature_gate.go:330] unrecognized feature gate: MachineAPIOperatorDisableMachineHealthCheckController Dec 10 22:49:25 crc kubenswrapper[4791]: W1210 22:49:25.710546 4791 feature_gate.go:330] unrecognized feature gate: MinimumKubeletVersion Dec 10 22:49:25 crc kubenswrapper[4791]: W1210 22:49:25.710550 4791 feature_gate.go:330] unrecognized feature gate: BootcNodeManagement Dec 10 22:49:25 crc kubenswrapper[4791]: W1210 22:49:25.710556 4791 feature_gate.go:353] Setting GA feature gate ValidatingAdmissionPolicy=true. It will be removed in a future release. Dec 10 22:49:25 crc kubenswrapper[4791]: W1210 22:49:25.710562 4791 feature_gate.go:330] unrecognized feature gate: VSphereStaticIPs Dec 10 22:49:25 crc kubenswrapper[4791]: W1210 22:49:25.710567 4791 feature_gate.go:353] Setting GA feature gate DisableKubeletCloudCredentialProviders=true. It will be removed in a future release. Dec 10 22:49:25 crc kubenswrapper[4791]: W1210 22:49:25.710573 4791 feature_gate.go:330] unrecognized feature gate: VSphereMultiNetworks Dec 10 22:49:25 crc kubenswrapper[4791]: W1210 22:49:25.710579 4791 feature_gate.go:330] unrecognized feature gate: InsightsConfig Dec 10 22:49:25 crc kubenswrapper[4791]: W1210 22:49:25.710584 4791 feature_gate.go:330] unrecognized feature gate: PlatformOperators Dec 10 22:49:25 crc kubenswrapper[4791]: W1210 22:49:25.710588 4791 feature_gate.go:330] unrecognized feature gate: NewOLM Dec 10 22:49:25 crc kubenswrapper[4791]: W1210 22:49:25.710594 4791 feature_gate.go:330] unrecognized feature gate: MultiArchInstallGCP Dec 10 22:49:25 crc kubenswrapper[4791]: W1210 22:49:25.710598 4791 feature_gate.go:330] unrecognized feature gate: IngressControllerLBSubnetsAWS Dec 10 22:49:25 crc kubenswrapper[4791]: W1210 22:49:25.710603 4791 feature_gate.go:330] unrecognized feature gate: BuildCSIVolumes Dec 10 22:49:25 crc kubenswrapper[4791]: W1210 22:49:25.710609 4791 feature_gate.go:353] Setting GA feature gate CloudDualStackNodeIPs=true. It will be removed in a future release. Dec 10 22:49:25 crc kubenswrapper[4791]: W1210 22:49:25.710615 4791 feature_gate.go:330] unrecognized feature gate: AdditionalRoutingCapabilities Dec 10 22:49:25 crc kubenswrapper[4791]: W1210 22:49:25.710620 4791 feature_gate.go:330] unrecognized feature gate: ChunkSizeMiB Dec 10 22:49:25 crc kubenswrapper[4791]: W1210 22:49:25.710624 4791 feature_gate.go:330] unrecognized feature gate: PrivateHostedZoneAWS Dec 10 22:49:25 crc kubenswrapper[4791]: I1210 22:49:25.710799 4791 feature_gate.go:386] feature gates: {map[CloudDualStackNodeIPs:true DisableKubeletCloudCredentialProviders:true DynamicResourceAllocation:false EventedPLEG:false KMSv1:true MaxUnavailableStatefulSet:false NodeSwap:false ProcMountType:false RouteExternalCertificate:false ServiceAccountTokenNodeBinding:false TranslateStreamCloseWebsocketRequests:false UserNamespacesPodSecurityStandards:false UserNamespacesSupport:false ValidatingAdmissionPolicy:true VolumeAttributesClass:false]} Dec 10 22:49:25 crc kubenswrapper[4791]: I1210 22:49:25.720159 4791 server.go:491] "Kubelet version" kubeletVersion="v1.31.5" Dec 10 22:49:25 crc kubenswrapper[4791]: I1210 22:49:25.720206 4791 server.go:493] "Golang settings" GOGC="" GOMAXPROCS="" GOTRACEBACK="" Dec 10 22:49:25 crc kubenswrapper[4791]: W1210 22:49:25.720285 4791 feature_gate.go:330] unrecognized feature gate: HardwareSpeed Dec 10 22:49:25 crc kubenswrapper[4791]: W1210 22:49:25.720295 4791 feature_gate.go:330] unrecognized feature gate: MachineAPIOperatorDisableMachineHealthCheckController Dec 10 22:49:25 crc kubenswrapper[4791]: W1210 22:49:25.720299 4791 feature_gate.go:330] unrecognized feature gate: ManagedBootImages Dec 10 22:49:25 crc kubenswrapper[4791]: W1210 22:49:25.720304 4791 feature_gate.go:330] unrecognized feature gate: IngressControllerDynamicConfigurationManager Dec 10 22:49:25 crc kubenswrapper[4791]: W1210 22:49:25.720308 4791 feature_gate.go:330] unrecognized feature gate: MachineAPIMigration Dec 10 22:49:25 crc kubenswrapper[4791]: W1210 22:49:25.720312 4791 feature_gate.go:330] unrecognized feature gate: VSphereMultiVCenters Dec 10 22:49:25 crc kubenswrapper[4791]: W1210 22:49:25.720316 4791 feature_gate.go:330] unrecognized feature gate: SignatureStores Dec 10 22:49:25 crc kubenswrapper[4791]: W1210 22:49:25.720320 4791 feature_gate.go:330] unrecognized feature gate: MachineConfigNodes Dec 10 22:49:25 crc kubenswrapper[4791]: W1210 22:49:25.720324 4791 feature_gate.go:330] unrecognized feature gate: NetworkSegmentation Dec 10 22:49:25 crc kubenswrapper[4791]: W1210 22:49:25.720328 4791 feature_gate.go:330] unrecognized feature gate: PinnedImages Dec 10 22:49:25 crc kubenswrapper[4791]: W1210 22:49:25.720332 4791 feature_gate.go:330] unrecognized feature gate: VSphereControlPlaneMachineSet Dec 10 22:49:25 crc kubenswrapper[4791]: W1210 22:49:25.720335 4791 feature_gate.go:330] unrecognized feature gate: InsightsConfig Dec 10 22:49:25 crc kubenswrapper[4791]: W1210 22:49:25.720366 4791 feature_gate.go:330] unrecognized feature gate: ClusterAPIInstall Dec 10 22:49:25 crc kubenswrapper[4791]: W1210 22:49:25.720371 4791 feature_gate.go:330] unrecognized feature gate: VolumeGroupSnapshot Dec 10 22:49:25 crc kubenswrapper[4791]: W1210 22:49:25.720375 4791 feature_gate.go:330] unrecognized feature gate: BareMetalLoadBalancer Dec 10 22:49:25 crc kubenswrapper[4791]: W1210 22:49:25.720380 4791 feature_gate.go:330] unrecognized feature gate: AWSEFSDriverVolumeMetrics Dec 10 22:49:25 crc kubenswrapper[4791]: W1210 22:49:25.720385 4791 feature_gate.go:330] unrecognized feature gate: OpenShiftPodSecurityAdmission Dec 10 22:49:25 crc kubenswrapper[4791]: W1210 22:49:25.720390 4791 feature_gate.go:330] unrecognized feature gate: Example Dec 10 22:49:25 crc kubenswrapper[4791]: W1210 22:49:25.720397 4791 feature_gate.go:353] Setting GA feature gate CloudDualStackNodeIPs=true. It will be removed in a future release. Dec 10 22:49:25 crc kubenswrapper[4791]: W1210 22:49:25.720403 4791 feature_gate.go:330] unrecognized feature gate: OnClusterBuild Dec 10 22:49:25 crc kubenswrapper[4791]: W1210 22:49:25.720408 4791 feature_gate.go:330] unrecognized feature gate: ImageStreamImportMode Dec 10 22:49:25 crc kubenswrapper[4791]: W1210 22:49:25.720413 4791 feature_gate.go:330] unrecognized feature gate: GCPLabelsTags Dec 10 22:49:25 crc kubenswrapper[4791]: W1210 22:49:25.720417 4791 feature_gate.go:330] unrecognized feature gate: GatewayAPI Dec 10 22:49:25 crc kubenswrapper[4791]: W1210 22:49:25.720422 4791 feature_gate.go:330] unrecognized feature gate: InsightsOnDemandDataGather Dec 10 22:49:25 crc kubenswrapper[4791]: W1210 22:49:25.720426 4791 feature_gate.go:330] unrecognized feature gate: NewOLM Dec 10 22:49:25 crc kubenswrapper[4791]: W1210 22:49:25.720430 4791 feature_gate.go:330] unrecognized feature gate: MultiArchInstallAzure Dec 10 22:49:25 crc kubenswrapper[4791]: W1210 22:49:25.720434 4791 feature_gate.go:330] unrecognized feature gate: ExternalOIDC Dec 10 22:49:25 crc kubenswrapper[4791]: W1210 22:49:25.720438 4791 feature_gate.go:330] unrecognized feature gate: ClusterAPIInstallIBMCloud Dec 10 22:49:25 crc kubenswrapper[4791]: W1210 22:49:25.720442 4791 feature_gate.go:330] unrecognized feature gate: ChunkSizeMiB Dec 10 22:49:25 crc kubenswrapper[4791]: W1210 22:49:25.720445 4791 feature_gate.go:330] unrecognized feature gate: MachineAPIProviderOpenStack Dec 10 22:49:25 crc kubenswrapper[4791]: W1210 22:49:25.720449 4791 feature_gate.go:330] unrecognized feature gate: InsightsRuntimeExtractor Dec 10 22:49:25 crc kubenswrapper[4791]: W1210 22:49:25.720453 4791 feature_gate.go:330] unrecognized feature gate: SigstoreImageVerification Dec 10 22:49:25 crc kubenswrapper[4791]: W1210 22:49:25.720457 4791 feature_gate.go:353] Setting GA feature gate ValidatingAdmissionPolicy=true. It will be removed in a future release. Dec 10 22:49:25 crc kubenswrapper[4791]: W1210 22:49:25.720464 4791 feature_gate.go:330] unrecognized feature gate: EtcdBackendQuota Dec 10 22:49:25 crc kubenswrapper[4791]: W1210 22:49:25.720467 4791 feature_gate.go:330] unrecognized feature gate: NetworkLiveMigration Dec 10 22:49:25 crc kubenswrapper[4791]: W1210 22:49:25.720471 4791 feature_gate.go:330] unrecognized feature gate: RouteAdvertisements Dec 10 22:49:25 crc kubenswrapper[4791]: W1210 22:49:25.720475 4791 feature_gate.go:330] unrecognized feature gate: AdminNetworkPolicy Dec 10 22:49:25 crc kubenswrapper[4791]: W1210 22:49:25.720479 4791 feature_gate.go:330] unrecognized feature gate: MetricsCollectionProfiles Dec 10 22:49:25 crc kubenswrapper[4791]: W1210 22:49:25.720484 4791 feature_gate.go:353] Setting GA feature gate DisableKubeletCloudCredentialProviders=true. It will be removed in a future release. Dec 10 22:49:25 crc kubenswrapper[4791]: W1210 22:49:25.720489 4791 feature_gate.go:330] unrecognized feature gate: BootcNodeManagement Dec 10 22:49:25 crc kubenswrapper[4791]: W1210 22:49:25.720493 4791 feature_gate.go:330] unrecognized feature gate: MinimumKubeletVersion Dec 10 22:49:25 crc kubenswrapper[4791]: W1210 22:49:25.720497 4791 feature_gate.go:330] unrecognized feature gate: CSIDriverSharedResource Dec 10 22:49:25 crc kubenswrapper[4791]: W1210 22:49:25.720501 4791 feature_gate.go:330] unrecognized feature gate: SetEIPForNLBIngressController Dec 10 22:49:25 crc kubenswrapper[4791]: W1210 22:49:25.720505 4791 feature_gate.go:330] unrecognized feature gate: IngressControllerLBSubnetsAWS Dec 10 22:49:25 crc kubenswrapper[4791]: W1210 22:49:25.720508 4791 feature_gate.go:330] unrecognized feature gate: InsightsConfigAPI Dec 10 22:49:25 crc kubenswrapper[4791]: W1210 22:49:25.720513 4791 feature_gate.go:330] unrecognized feature gate: AdditionalRoutingCapabilities Dec 10 22:49:25 crc kubenswrapper[4791]: W1210 22:49:25.720517 4791 feature_gate.go:330] unrecognized feature gate: VSphereMultiNetworks Dec 10 22:49:25 crc kubenswrapper[4791]: W1210 22:49:25.720520 4791 feature_gate.go:330] unrecognized feature gate: MixedCPUsAllocation Dec 10 22:49:25 crc kubenswrapper[4791]: W1210 22:49:25.720524 4791 feature_gate.go:330] unrecognized feature gate: ClusterMonitoringConfig Dec 10 22:49:25 crc kubenswrapper[4791]: W1210 22:49:25.720528 4791 feature_gate.go:330] unrecognized feature gate: AlibabaPlatform Dec 10 22:49:25 crc kubenswrapper[4791]: W1210 22:49:25.720532 4791 feature_gate.go:330] unrecognized feature gate: PrivateHostedZoneAWS Dec 10 22:49:25 crc kubenswrapper[4791]: W1210 22:49:25.720536 4791 feature_gate.go:330] unrecognized feature gate: GCPClusterHostedDNS Dec 10 22:49:25 crc kubenswrapper[4791]: W1210 22:49:25.720540 4791 feature_gate.go:330] unrecognized feature gate: UpgradeStatus Dec 10 22:49:25 crc kubenswrapper[4791]: W1210 22:49:25.720544 4791 feature_gate.go:330] unrecognized feature gate: PlatformOperators Dec 10 22:49:25 crc kubenswrapper[4791]: W1210 22:49:25.720548 4791 feature_gate.go:330] unrecognized feature gate: AutomatedEtcdBackup Dec 10 22:49:25 crc kubenswrapper[4791]: W1210 22:49:25.720551 4791 feature_gate.go:330] unrecognized feature gate: NetworkDiagnosticsConfig Dec 10 22:49:25 crc kubenswrapper[4791]: W1210 22:49:25.720554 4791 feature_gate.go:330] unrecognized feature gate: MultiArchInstallGCP Dec 10 22:49:25 crc kubenswrapper[4791]: W1210 22:49:25.720558 4791 feature_gate.go:330] unrecognized feature gate: BuildCSIVolumes Dec 10 22:49:25 crc kubenswrapper[4791]: W1210 22:49:25.720562 4791 feature_gate.go:330] unrecognized feature gate: VSphereStaticIPs Dec 10 22:49:25 crc kubenswrapper[4791]: W1210 22:49:25.720565 4791 feature_gate.go:330] unrecognized feature gate: DNSNameResolver Dec 10 22:49:25 crc kubenswrapper[4791]: W1210 22:49:25.720568 4791 feature_gate.go:330] unrecognized feature gate: MultiArchInstallAWS Dec 10 22:49:25 crc kubenswrapper[4791]: W1210 22:49:25.720572 4791 feature_gate.go:330] unrecognized feature gate: ConsolePluginContentSecurityPolicy Dec 10 22:49:25 crc kubenswrapper[4791]: W1210 22:49:25.720575 4791 feature_gate.go:330] unrecognized feature gate: VSphereDriverConfiguration Dec 10 22:49:25 crc kubenswrapper[4791]: W1210 22:49:25.720579 4791 feature_gate.go:330] unrecognized feature gate: AzureWorkloadIdentity Dec 10 22:49:25 crc kubenswrapper[4791]: W1210 22:49:25.720582 4791 feature_gate.go:330] unrecognized feature gate: AWSClusterHostedDNS Dec 10 22:49:25 crc kubenswrapper[4791]: W1210 22:49:25.720586 4791 feature_gate.go:330] unrecognized feature gate: ManagedBootImagesAWS Dec 10 22:49:25 crc kubenswrapper[4791]: W1210 22:49:25.720589 4791 feature_gate.go:330] unrecognized feature gate: NodeDisruptionPolicy Dec 10 22:49:25 crc kubenswrapper[4791]: W1210 22:49:25.720593 4791 feature_gate.go:330] unrecognized feature gate: PersistentIPsForVirtualization Dec 10 22:49:25 crc kubenswrapper[4791]: W1210 22:49:25.720597 4791 feature_gate.go:351] Setting deprecated feature gate KMSv1=true. It will be removed in a future release. Dec 10 22:49:25 crc kubenswrapper[4791]: W1210 22:49:25.720602 4791 feature_gate.go:330] unrecognized feature gate: NutanixMultiSubnets Dec 10 22:49:25 crc kubenswrapper[4791]: W1210 22:49:25.720606 4791 feature_gate.go:330] unrecognized feature gate: OVNObservability Dec 10 22:49:25 crc kubenswrapper[4791]: I1210 22:49:25.720612 4791 feature_gate.go:386] feature gates: {map[CloudDualStackNodeIPs:true DisableKubeletCloudCredentialProviders:true DynamicResourceAllocation:false EventedPLEG:false KMSv1:true MaxUnavailableStatefulSet:false NodeSwap:false ProcMountType:false RouteExternalCertificate:false ServiceAccountTokenNodeBinding:false TranslateStreamCloseWebsocketRequests:false UserNamespacesPodSecurityStandards:false UserNamespacesSupport:false ValidatingAdmissionPolicy:true VolumeAttributesClass:false]} Dec 10 22:49:25 crc kubenswrapper[4791]: W1210 22:49:25.720715 4791 feature_gate.go:330] unrecognized feature gate: MixedCPUsAllocation Dec 10 22:49:25 crc kubenswrapper[4791]: W1210 22:49:25.720721 4791 feature_gate.go:330] unrecognized feature gate: VSphereStaticIPs Dec 10 22:49:25 crc kubenswrapper[4791]: W1210 22:49:25.720725 4791 feature_gate.go:330] unrecognized feature gate: VSphereMultiNetworks Dec 10 22:49:25 crc kubenswrapper[4791]: W1210 22:49:25.720730 4791 feature_gate.go:353] Setting GA feature gate CloudDualStackNodeIPs=true. It will be removed in a future release. Dec 10 22:49:25 crc kubenswrapper[4791]: W1210 22:49:25.720736 4791 feature_gate.go:330] unrecognized feature gate: NetworkDiagnosticsConfig Dec 10 22:49:25 crc kubenswrapper[4791]: W1210 22:49:25.720740 4791 feature_gate.go:330] unrecognized feature gate: MinimumKubeletVersion Dec 10 22:49:25 crc kubenswrapper[4791]: W1210 22:49:25.720744 4791 feature_gate.go:330] unrecognized feature gate: VSphereMultiVCenters Dec 10 22:49:25 crc kubenswrapper[4791]: W1210 22:49:25.720748 4791 feature_gate.go:330] unrecognized feature gate: AWSClusterHostedDNS Dec 10 22:49:25 crc kubenswrapper[4791]: W1210 22:49:25.720752 4791 feature_gate.go:330] unrecognized feature gate: CSIDriverSharedResource Dec 10 22:49:25 crc kubenswrapper[4791]: W1210 22:49:25.720757 4791 feature_gate.go:353] Setting GA feature gate DisableKubeletCloudCredentialProviders=true. It will be removed in a future release. Dec 10 22:49:25 crc kubenswrapper[4791]: W1210 22:49:25.720762 4791 feature_gate.go:330] unrecognized feature gate: MultiArchInstallAWS Dec 10 22:49:25 crc kubenswrapper[4791]: W1210 22:49:25.720766 4791 feature_gate.go:330] unrecognized feature gate: OnClusterBuild Dec 10 22:49:25 crc kubenswrapper[4791]: W1210 22:49:25.720770 4791 feature_gate.go:330] unrecognized feature gate: UpgradeStatus Dec 10 22:49:25 crc kubenswrapper[4791]: W1210 22:49:25.720774 4791 feature_gate.go:330] unrecognized feature gate: ClusterMonitoringConfig Dec 10 22:49:25 crc kubenswrapper[4791]: W1210 22:49:25.720777 4791 feature_gate.go:330] unrecognized feature gate: BareMetalLoadBalancer Dec 10 22:49:25 crc kubenswrapper[4791]: W1210 22:49:25.720781 4791 feature_gate.go:330] unrecognized feature gate: SigstoreImageVerification Dec 10 22:49:25 crc kubenswrapper[4791]: W1210 22:49:25.720785 4791 feature_gate.go:330] unrecognized feature gate: NodeDisruptionPolicy Dec 10 22:49:25 crc kubenswrapper[4791]: W1210 22:49:25.720789 4791 feature_gate.go:330] unrecognized feature gate: ClusterAPIInstallIBMCloud Dec 10 22:49:25 crc kubenswrapper[4791]: W1210 22:49:25.720792 4791 feature_gate.go:330] unrecognized feature gate: DNSNameResolver Dec 10 22:49:25 crc kubenswrapper[4791]: W1210 22:49:25.720796 4791 feature_gate.go:330] unrecognized feature gate: AWSEFSDriverVolumeMetrics Dec 10 22:49:25 crc kubenswrapper[4791]: W1210 22:49:25.720799 4791 feature_gate.go:330] unrecognized feature gate: InsightsRuntimeExtractor Dec 10 22:49:25 crc kubenswrapper[4791]: W1210 22:49:25.720802 4791 feature_gate.go:330] unrecognized feature gate: GCPLabelsTags Dec 10 22:49:25 crc kubenswrapper[4791]: W1210 22:49:25.720806 4791 feature_gate.go:330] unrecognized feature gate: MetricsCollectionProfiles Dec 10 22:49:25 crc kubenswrapper[4791]: W1210 22:49:25.720810 4791 feature_gate.go:330] unrecognized feature gate: IngressControllerDynamicConfigurationManager Dec 10 22:49:25 crc kubenswrapper[4791]: W1210 22:49:25.720813 4791 feature_gate.go:330] unrecognized feature gate: MultiArchInstallAzure Dec 10 22:49:25 crc kubenswrapper[4791]: W1210 22:49:25.720816 4791 feature_gate.go:330] unrecognized feature gate: AdditionalRoutingCapabilities Dec 10 22:49:25 crc kubenswrapper[4791]: W1210 22:49:25.720820 4791 feature_gate.go:330] unrecognized feature gate: GatewayAPI Dec 10 22:49:25 crc kubenswrapper[4791]: W1210 22:49:25.720823 4791 feature_gate.go:330] unrecognized feature gate: Example Dec 10 22:49:25 crc kubenswrapper[4791]: W1210 22:49:25.720826 4791 feature_gate.go:330] unrecognized feature gate: MachineAPIProviderOpenStack Dec 10 22:49:25 crc kubenswrapper[4791]: W1210 22:49:25.720830 4791 feature_gate.go:330] unrecognized feature gate: MachineConfigNodes Dec 10 22:49:25 crc kubenswrapper[4791]: W1210 22:49:25.720833 4791 feature_gate.go:330] unrecognized feature gate: InsightsConfig Dec 10 22:49:25 crc kubenswrapper[4791]: W1210 22:49:25.720837 4791 feature_gate.go:330] unrecognized feature gate: HardwareSpeed Dec 10 22:49:25 crc kubenswrapper[4791]: W1210 22:49:25.720840 4791 feature_gate.go:330] unrecognized feature gate: OVNObservability Dec 10 22:49:25 crc kubenswrapper[4791]: W1210 22:49:25.720844 4791 feature_gate.go:330] unrecognized feature gate: BootcNodeManagement Dec 10 22:49:25 crc kubenswrapper[4791]: W1210 22:49:25.720848 4791 feature_gate.go:330] unrecognized feature gate: SignatureStores Dec 10 22:49:25 crc kubenswrapper[4791]: W1210 22:49:25.720851 4791 feature_gate.go:330] unrecognized feature gate: MultiArchInstallGCP Dec 10 22:49:25 crc kubenswrapper[4791]: W1210 22:49:25.720855 4791 feature_gate.go:330] unrecognized feature gate: PinnedImages Dec 10 22:49:25 crc kubenswrapper[4791]: W1210 22:49:25.720858 4791 feature_gate.go:330] unrecognized feature gate: OpenShiftPodSecurityAdmission Dec 10 22:49:25 crc kubenswrapper[4791]: W1210 22:49:25.720862 4791 feature_gate.go:330] unrecognized feature gate: ExternalOIDC Dec 10 22:49:25 crc kubenswrapper[4791]: W1210 22:49:25.720866 4791 feature_gate.go:330] unrecognized feature gate: NetworkLiveMigration Dec 10 22:49:25 crc kubenswrapper[4791]: W1210 22:49:25.720869 4791 feature_gate.go:330] unrecognized feature gate: MachineAPIOperatorDisableMachineHealthCheckController Dec 10 22:49:25 crc kubenswrapper[4791]: W1210 22:49:25.720874 4791 feature_gate.go:330] unrecognized feature gate: VolumeGroupSnapshot Dec 10 22:49:25 crc kubenswrapper[4791]: W1210 22:49:25.720878 4791 feature_gate.go:330] unrecognized feature gate: MachineAPIMigration Dec 10 22:49:25 crc kubenswrapper[4791]: W1210 22:49:25.720882 4791 feature_gate.go:330] unrecognized feature gate: ChunkSizeMiB Dec 10 22:49:25 crc kubenswrapper[4791]: W1210 22:49:25.720886 4791 feature_gate.go:330] unrecognized feature gate: IngressControllerLBSubnetsAWS Dec 10 22:49:25 crc kubenswrapper[4791]: W1210 22:49:25.720890 4791 feature_gate.go:330] unrecognized feature gate: PersistentIPsForVirtualization Dec 10 22:49:25 crc kubenswrapper[4791]: W1210 22:49:25.720894 4791 feature_gate.go:330] unrecognized feature gate: RouteAdvertisements Dec 10 22:49:25 crc kubenswrapper[4791]: W1210 22:49:25.720899 4791 feature_gate.go:330] unrecognized feature gate: InsightsConfigAPI Dec 10 22:49:25 crc kubenswrapper[4791]: W1210 22:49:25.720907 4791 feature_gate.go:330] unrecognized feature gate: InsightsOnDemandDataGather Dec 10 22:49:25 crc kubenswrapper[4791]: W1210 22:49:25.720915 4791 feature_gate.go:330] unrecognized feature gate: PlatformOperators Dec 10 22:49:25 crc kubenswrapper[4791]: W1210 22:49:25.720921 4791 feature_gate.go:330] unrecognized feature gate: ManagedBootImages Dec 10 22:49:25 crc kubenswrapper[4791]: W1210 22:49:25.720925 4791 feature_gate.go:330] unrecognized feature gate: ImageStreamImportMode Dec 10 22:49:25 crc kubenswrapper[4791]: W1210 22:49:25.720932 4791 feature_gate.go:330] unrecognized feature gate: AzureWorkloadIdentity Dec 10 22:49:25 crc kubenswrapper[4791]: W1210 22:49:25.720937 4791 feature_gate.go:330] unrecognized feature gate: AdminNetworkPolicy Dec 10 22:49:25 crc kubenswrapper[4791]: W1210 22:49:25.720941 4791 feature_gate.go:330] unrecognized feature gate: NetworkSegmentation Dec 10 22:49:25 crc kubenswrapper[4791]: W1210 22:49:25.720946 4791 feature_gate.go:330] unrecognized feature gate: VSphereControlPlaneMachineSet Dec 10 22:49:25 crc kubenswrapper[4791]: W1210 22:49:25.720952 4791 feature_gate.go:353] Setting GA feature gate ValidatingAdmissionPolicy=true. It will be removed in a future release. Dec 10 22:49:25 crc kubenswrapper[4791]: W1210 22:49:25.720957 4791 feature_gate.go:330] unrecognized feature gate: VSphereDriverConfiguration Dec 10 22:49:25 crc kubenswrapper[4791]: W1210 22:49:25.720962 4791 feature_gate.go:330] unrecognized feature gate: SetEIPForNLBIngressController Dec 10 22:49:25 crc kubenswrapper[4791]: W1210 22:49:25.720968 4791 feature_gate.go:330] unrecognized feature gate: GCPClusterHostedDNS Dec 10 22:49:25 crc kubenswrapper[4791]: W1210 22:49:25.720974 4791 feature_gate.go:330] unrecognized feature gate: AutomatedEtcdBackup Dec 10 22:49:25 crc kubenswrapper[4791]: W1210 22:49:25.720980 4791 feature_gate.go:351] Setting deprecated feature gate KMSv1=true. It will be removed in a future release. Dec 10 22:49:25 crc kubenswrapper[4791]: W1210 22:49:25.720985 4791 feature_gate.go:330] unrecognized feature gate: NutanixMultiSubnets Dec 10 22:49:25 crc kubenswrapper[4791]: W1210 22:49:25.720989 4791 feature_gate.go:330] unrecognized feature gate: ConsolePluginContentSecurityPolicy Dec 10 22:49:25 crc kubenswrapper[4791]: W1210 22:49:25.720992 4791 feature_gate.go:330] unrecognized feature gate: NewOLM Dec 10 22:49:25 crc kubenswrapper[4791]: W1210 22:49:25.720997 4791 feature_gate.go:330] unrecognized feature gate: EtcdBackendQuota Dec 10 22:49:25 crc kubenswrapper[4791]: W1210 22:49:25.721001 4791 feature_gate.go:330] unrecognized feature gate: AlibabaPlatform Dec 10 22:49:25 crc kubenswrapper[4791]: W1210 22:49:25.721005 4791 feature_gate.go:330] unrecognized feature gate: ClusterAPIInstall Dec 10 22:49:25 crc kubenswrapper[4791]: W1210 22:49:25.721008 4791 feature_gate.go:330] unrecognized feature gate: ManagedBootImagesAWS Dec 10 22:49:25 crc kubenswrapper[4791]: W1210 22:49:25.721011 4791 feature_gate.go:330] unrecognized feature gate: BuildCSIVolumes Dec 10 22:49:25 crc kubenswrapper[4791]: W1210 22:49:25.721016 4791 feature_gate.go:330] unrecognized feature gate: PrivateHostedZoneAWS Dec 10 22:49:25 crc kubenswrapper[4791]: I1210 22:49:25.721023 4791 feature_gate.go:386] feature gates: {map[CloudDualStackNodeIPs:true DisableKubeletCloudCredentialProviders:true DynamicResourceAllocation:false EventedPLEG:false KMSv1:true MaxUnavailableStatefulSet:false NodeSwap:false ProcMountType:false RouteExternalCertificate:false ServiceAccountTokenNodeBinding:false TranslateStreamCloseWebsocketRequests:false UserNamespacesPodSecurityStandards:false UserNamespacesSupport:false ValidatingAdmissionPolicy:true VolumeAttributesClass:false]} Dec 10 22:49:25 crc kubenswrapper[4791]: I1210 22:49:25.722256 4791 server.go:940] "Client rotation is on, will bootstrap in background" Dec 10 22:49:25 crc kubenswrapper[4791]: I1210 22:49:25.726102 4791 bootstrap.go:85] "Current kubeconfig file contents are still valid, no bootstrap necessary" Dec 10 22:49:25 crc kubenswrapper[4791]: I1210 22:49:25.726223 4791 certificate_store.go:130] Loading cert/key pair from "/var/lib/kubelet/pki/kubelet-client-current.pem". Dec 10 22:49:25 crc kubenswrapper[4791]: I1210 22:49:25.726783 4791 server.go:997] "Starting client certificate rotation" Dec 10 22:49:25 crc kubenswrapper[4791]: I1210 22:49:25.726817 4791 certificate_manager.go:356] kubernetes.io/kube-apiserver-client-kubelet: Certificate rotation is enabled Dec 10 22:49:25 crc kubenswrapper[4791]: I1210 22:49:25.727087 4791 certificate_manager.go:356] kubernetes.io/kube-apiserver-client-kubelet: Certificate expiration is 2026-02-24 05:52:08 +0000 UTC, rotation deadline is 2025-12-06 20:26:54.452650951 +0000 UTC Dec 10 22:49:25 crc kubenswrapper[4791]: I1210 22:49:25.727256 4791 certificate_manager.go:356] kubernetes.io/kube-apiserver-client-kubelet: Rotating certificates Dec 10 22:49:25 crc kubenswrapper[4791]: I1210 22:49:25.731942 4791 dynamic_cafile_content.go:123] "Loaded a new CA Bundle and Verifier" name="client-ca-bundle::/etc/kubernetes/kubelet-ca.crt" Dec 10 22:49:25 crc kubenswrapper[4791]: I1210 22:49:25.733761 4791 dynamic_cafile_content.go:161] "Starting controller" name="client-ca-bundle::/etc/kubernetes/kubelet-ca.crt" Dec 10 22:49:25 crc kubenswrapper[4791]: E1210 22:49:25.733870 4791 certificate_manager.go:562] "Unhandled Error" err="kubernetes.io/kube-apiserver-client-kubelet: Failed while requesting a signed certificate from the control plane: cannot create certificate signing request: Post \"https://api-int.crc.testing:6443/apis/certificates.k8s.io/v1/certificatesigningrequests\": dial tcp 38.102.83.30:6443: connect: connection refused" logger="UnhandledError" Dec 10 22:49:25 crc kubenswrapper[4791]: I1210 22:49:25.742457 4791 log.go:25] "Validated CRI v1 runtime API" Dec 10 22:49:25 crc kubenswrapper[4791]: I1210 22:49:25.780304 4791 log.go:25] "Validated CRI v1 image API" Dec 10 22:49:25 crc kubenswrapper[4791]: I1210 22:49:25.782206 4791 server.go:1437] "Using cgroup driver setting received from the CRI runtime" cgroupDriver="systemd" Dec 10 22:49:25 crc kubenswrapper[4791]: I1210 22:49:25.785433 4791 fs.go:133] Filesystem UUIDs: map[0b076daa-c26a-46d2-b3a6-72a8dbc6e257:/dev/vda4 2025-12-10-22-45-00-00:/dev/sr0 7B77-95E7:/dev/vda2 de0497b0-db1b-465a-b278-03db02455c71:/dev/vda3] Dec 10 22:49:25 crc kubenswrapper[4791]: I1210 22:49:25.785516 4791 fs.go:134] Filesystem partitions: map[/dev/shm:{mountpoint:/dev/shm major:0 minor:22 fsType:tmpfs blockSize:0} /dev/vda3:{mountpoint:/boot major:252 minor:3 fsType:ext4 blockSize:0} /dev/vda4:{mountpoint:/var major:252 minor:4 fsType:xfs blockSize:0} /run:{mountpoint:/run major:0 minor:24 fsType:tmpfs blockSize:0} /run/user/1000:{mountpoint:/run/user/1000 major:0 minor:42 fsType:tmpfs blockSize:0} /tmp:{mountpoint:/tmp major:0 minor:30 fsType:tmpfs blockSize:0} /var/lib/etcd:{mountpoint:/var/lib/etcd major:0 minor:41 fsType:tmpfs blockSize:0}] Dec 10 22:49:25 crc kubenswrapper[4791]: I1210 22:49:25.814137 4791 manager.go:217] Machine: {Timestamp:2025-12-10 22:49:25.812595988 +0000 UTC m=+0.242213631 CPUVendorID:AuthenticAMD NumCores:12 NumPhysicalCores:1 NumSockets:12 CpuFrequency:2800000 MemoryCapacity:33654128640 SwapCapacity:0 MemoryByType:map[] NVMInfo:{MemoryModeCapacity:0 AppDirectModeCapacity:0 AvgPowerBudget:0} HugePages:[{PageSize:1048576 NumPages:0} {PageSize:2048 NumPages:0}] MachineID:21801e6708c44f15b81395eb736a7cec SystemUUID:442c0b68-8dce-47e0-bd68-d5210d7e0493 BootID:4fab3af6-f657-4dfc-8ef2-12b7c978c94f Filesystems:[{Device:/run/user/1000 DeviceMajor:0 DeviceMinor:42 Capacity:3365412864 Type:vfs Inodes:821634 HasInodes:true} {Device:/var/lib/etcd DeviceMajor:0 DeviceMinor:41 Capacity:1073741824 Type:vfs Inodes:4108170 HasInodes:true} {Device:/dev/shm DeviceMajor:0 DeviceMinor:22 Capacity:16827064320 Type:vfs Inodes:4108170 HasInodes:true} {Device:/run DeviceMajor:0 DeviceMinor:24 Capacity:6730825728 Type:vfs Inodes:819200 HasInodes:true} {Device:/dev/vda4 DeviceMajor:252 DeviceMinor:4 Capacity:85292941312 Type:vfs Inodes:41679680 HasInodes:true} {Device:/tmp DeviceMajor:0 DeviceMinor:30 Capacity:16827064320 Type:vfs Inodes:1048576 HasInodes:true} {Device:/dev/vda3 DeviceMajor:252 DeviceMinor:3 Capacity:366869504 Type:vfs Inodes:98304 HasInodes:true}] DiskMap:map[252:0:{Name:vda Major:252 Minor:0 Size:214748364800 Scheduler:none}] NetworkDevices:[{Name:br-ex MacAddress:fa:16:3e:e2:1c:eb Speed:0 Mtu:1500} {Name:br-int MacAddress:d6:39:55:2e:22:71 Speed:0 Mtu:1400} {Name:ens3 MacAddress:fa:16:3e:e2:1c:eb Speed:-1 Mtu:1500} {Name:ens7 MacAddress:fa:16:3e:f5:e8:fb Speed:-1 Mtu:1500} {Name:ens7.20 MacAddress:52:54:00:81:ce:73 Speed:-1 Mtu:1496} {Name:ens7.21 MacAddress:52:54:00:30:bb:b6 Speed:-1 Mtu:1496} {Name:ens7.22 MacAddress:52:54:00:98:e2:22 Speed:-1 Mtu:1496} {Name:eth10 MacAddress:1a:6c:bd:82:db:26 Speed:0 Mtu:1500} {Name:ovn-k8s-mp0 MacAddress:0a:58:0a:d9:00:02 Speed:0 Mtu:1400} {Name:ovs-system MacAddress:96:46:27:59:b3:dd Speed:0 Mtu:1500}] Topology:[{Id:0 Memory:33654128640 HugePages:[{PageSize:1048576 NumPages:0} {PageSize:2048 NumPages:0}] Cores:[{Id:0 Threads:[0] Caches:[{Id:0 Size:32768 Type:Data Level:1} {Id:0 Size:32768 Type:Instruction Level:1} {Id:0 Size:524288 Type:Unified Level:2}] UncoreCaches:[{Id:0 Size:16777216 Type:Unified Level:3}] SocketID:0 BookID: DrawerID:} {Id:0 Threads:[1] Caches:[{Id:1 Size:32768 Type:Data Level:1} {Id:1 Size:32768 Type:Instruction Level:1} {Id:1 Size:524288 Type:Unified Level:2}] UncoreCaches:[{Id:1 Size:16777216 Type:Unified Level:3}] SocketID:1 BookID: DrawerID:} {Id:0 Threads:[10] Caches:[{Id:10 Size:32768 Type:Data Level:1} {Id:10 Size:32768 Type:Instruction Level:1} {Id:10 Size:524288 Type:Unified Level:2}] UncoreCaches:[{Id:10 Size:16777216 Type:Unified Level:3}] SocketID:10 BookID: DrawerID:} {Id:0 Threads:[11] Caches:[{Id:11 Size:32768 Type:Data Level:1} {Id:11 Size:32768 Type:Instruction Level:1} {Id:11 Size:524288 Type:Unified Level:2}] UncoreCaches:[{Id:11 Size:16777216 Type:Unified Level:3}] SocketID:11 BookID: DrawerID:} {Id:0 Threads:[2] Caches:[{Id:2 Size:32768 Type:Data Level:1} {Id:2 Size:32768 Type:Instruction Level:1} {Id:2 Size:524288 Type:Unified Level:2}] UncoreCaches:[{Id:2 Size:16777216 Type:Unified Level:3}] SocketID:2 BookID: DrawerID:} {Id:0 Threads:[3] Caches:[{Id:3 Size:32768 Type:Data Level:1} {Id:3 Size:32768 Type:Instruction Level:1} {Id:3 Size:524288 Type:Unified Level:2}] UncoreCaches:[{Id:3 Size:16777216 Type:Unified Level:3}] SocketID:3 BookID: DrawerID:} {Id:0 Threads:[4] Caches:[{Id:4 Size:32768 Type:Data Level:1} {Id:4 Size:32768 Type:Instruction Level:1} {Id:4 Size:524288 Type:Unified Level:2}] UncoreCaches:[{Id:4 Size:16777216 Type:Unified Level:3}] SocketID:4 BookID: DrawerID:} {Id:0 Threads:[5] Caches:[{Id:5 Size:32768 Type:Data Level:1} {Id:5 Size:32768 Type:Instruction Level:1} {Id:5 Size:524288 Type:Unified Level:2}] UncoreCaches:[{Id:5 Size:16777216 Type:Unified Level:3}] SocketID:5 BookID: DrawerID:} {Id:0 Threads:[6] Caches:[{Id:6 Size:32768 Type:Data Level:1} {Id:6 Size:32768 Type:Instruction Level:1} {Id:6 Size:524288 Type:Unified Level:2}] UncoreCaches:[{Id:6 Size:16777216 Type:Unified Level:3}] SocketID:6 BookID: DrawerID:} {Id:0 Threads:[7] Caches:[{Id:7 Size:32768 Type:Data Level:1} {Id:7 Size:32768 Type:Instruction Level:1} {Id:7 Size:524288 Type:Unified Level:2}] UncoreCaches:[{Id:7 Size:16777216 Type:Unified Level:3}] SocketID:7 BookID: DrawerID:} {Id:0 Threads:[8] Caches:[{Id:8 Size:32768 Type:Data Level:1} {Id:8 Size:32768 Type:Instruction Level:1} {Id:8 Size:524288 Type:Unified Level:2}] UncoreCaches:[{Id:8 Size:16777216 Type:Unified Level:3}] SocketID:8 BookID: DrawerID:} {Id:0 Threads:[9] Caches:[{Id:9 Size:32768 Type:Data Level:1} {Id:9 Size:32768 Type:Instruction Level:1} {Id:9 Size:524288 Type:Unified Level:2}] UncoreCaches:[{Id:9 Size:16777216 Type:Unified Level:3}] SocketID:9 BookID: DrawerID:}] Caches:[] Distances:[10]}] CloudProvider:Unknown InstanceType:Unknown InstanceID:None} Dec 10 22:49:25 crc kubenswrapper[4791]: I1210 22:49:25.814417 4791 manager_no_libpfm.go:29] cAdvisor is build without cgo and/or libpfm support. Perf event counters are not available. Dec 10 22:49:25 crc kubenswrapper[4791]: I1210 22:49:25.814551 4791 manager.go:233] Version: {KernelVersion:5.14.0-427.50.2.el9_4.x86_64 ContainerOsVersion:Red Hat Enterprise Linux CoreOS 418.94.202502100215-0 DockerVersion: DockerAPIVersion: CadvisorVersion: CadvisorRevision:} Dec 10 22:49:25 crc kubenswrapper[4791]: I1210 22:49:25.815141 4791 swap_util.go:113] "Swap is on" /proc/swaps contents="Filename\t\t\t\tType\t\tSize\t\tUsed\t\tPriority" Dec 10 22:49:25 crc kubenswrapper[4791]: I1210 22:49:25.815359 4791 container_manager_linux.go:267] "Container manager verified user specified cgroup-root exists" cgroupRoot=[] Dec 10 22:49:25 crc kubenswrapper[4791]: I1210 22:49:25.815397 4791 container_manager_linux.go:272] "Creating Container Manager object based on Node Config" nodeConfig={"NodeName":"crc","RuntimeCgroupsName":"/system.slice/crio.service","SystemCgroupsName":"/system.slice","KubeletCgroupsName":"","KubeletOOMScoreAdj":-999,"ContainerRuntime":"","CgroupsPerQOS":true,"CgroupRoot":"/","CgroupDriver":"systemd","KubeletRootDir":"/var/lib/kubelet","ProtectKernelDefaults":true,"KubeReservedCgroupName":"","SystemReservedCgroupName":"","ReservedSystemCPUs":{},"EnforceNodeAllocatable":{"pods":{}},"KubeReserved":null,"SystemReserved":{"cpu":"200m","ephemeral-storage":"350Mi","memory":"350Mi"},"HardEvictionThresholds":[{"Signal":"nodefs.inodesFree","Operator":"LessThan","Value":{"Quantity":null,"Percentage":0.05},"GracePeriod":0,"MinReclaim":null},{"Signal":"imagefs.available","Operator":"LessThan","Value":{"Quantity":null,"Percentage":0.15},"GracePeriod":0,"MinReclaim":null},{"Signal":"imagefs.inodesFree","Operator":"LessThan","Value":{"Quantity":null,"Percentage":0.05},"GracePeriod":0,"MinReclaim":null},{"Signal":"memory.available","Operator":"LessThan","Value":{"Quantity":"100Mi","Percentage":0},"GracePeriod":0,"MinReclaim":null},{"Signal":"nodefs.available","Operator":"LessThan","Value":{"Quantity":null,"Percentage":0.1},"GracePeriod":0,"MinReclaim":null}],"QOSReserved":{},"CPUManagerPolicy":"none","CPUManagerPolicyOptions":null,"TopologyManagerScope":"container","CPUManagerReconcilePeriod":10000000000,"ExperimentalMemoryManagerPolicy":"None","ExperimentalMemoryManagerReservedMemory":null,"PodPidsLimit":4096,"EnforceCPULimits":true,"CPUCFSQuotaPeriod":100000000,"TopologyManagerPolicy":"none","TopologyManagerPolicyOptions":null,"CgroupVersion":2} Dec 10 22:49:25 crc kubenswrapper[4791]: I1210 22:49:25.815707 4791 topology_manager.go:138] "Creating topology manager with none policy" Dec 10 22:49:25 crc kubenswrapper[4791]: I1210 22:49:25.815725 4791 container_manager_linux.go:303] "Creating device plugin manager" Dec 10 22:49:25 crc kubenswrapper[4791]: I1210 22:49:25.816008 4791 manager.go:142] "Creating Device Plugin manager" path="/var/lib/kubelet/device-plugins/kubelet.sock" Dec 10 22:49:25 crc kubenswrapper[4791]: I1210 22:49:25.816071 4791 server.go:66] "Creating device plugin registration server" version="v1beta1" socket="/var/lib/kubelet/device-plugins/kubelet.sock" Dec 10 22:49:25 crc kubenswrapper[4791]: I1210 22:49:25.816294 4791 state_mem.go:36] "Initialized new in-memory state store" Dec 10 22:49:25 crc kubenswrapper[4791]: I1210 22:49:25.816810 4791 server.go:1245] "Using root directory" path="/var/lib/kubelet" Dec 10 22:49:25 crc kubenswrapper[4791]: I1210 22:49:25.817735 4791 kubelet.go:418] "Attempting to sync node with API server" Dec 10 22:49:25 crc kubenswrapper[4791]: I1210 22:49:25.817776 4791 kubelet.go:313] "Adding static pod path" path="/etc/kubernetes/manifests" Dec 10 22:49:25 crc kubenswrapper[4791]: I1210 22:49:25.817825 4791 file.go:69] "Watching path" path="/etc/kubernetes/manifests" Dec 10 22:49:25 crc kubenswrapper[4791]: I1210 22:49:25.817852 4791 kubelet.go:324] "Adding apiserver pod source" Dec 10 22:49:25 crc kubenswrapper[4791]: I1210 22:49:25.817876 4791 apiserver.go:42] "Waiting for node sync before watching apiserver pods" Dec 10 22:49:25 crc kubenswrapper[4791]: I1210 22:49:25.820083 4791 kuberuntime_manager.go:262] "Container runtime initialized" containerRuntime="cri-o" version="1.31.5-4.rhaos4.18.gitdad78d5.el9" apiVersion="v1" Dec 10 22:49:25 crc kubenswrapper[4791]: I1210 22:49:25.820469 4791 certificate_store.go:130] Loading cert/key pair from "/var/lib/kubelet/pki/kubelet-server-current.pem". Dec 10 22:49:25 crc kubenswrapper[4791]: W1210 22:49:25.820475 4791 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.Service: Get "https://api-int.crc.testing:6443/api/v1/services?fieldSelector=spec.clusterIP%21%3DNone&limit=500&resourceVersion=0": dial tcp 38.102.83.30:6443: connect: connection refused Dec 10 22:49:25 crc kubenswrapper[4791]: E1210 22:49:25.820538 4791 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.Service: failed to list *v1.Service: Get \"https://api-int.crc.testing:6443/api/v1/services?fieldSelector=spec.clusterIP%21%3DNone&limit=500&resourceVersion=0\": dial tcp 38.102.83.30:6443: connect: connection refused" logger="UnhandledError" Dec 10 22:49:25 crc kubenswrapper[4791]: W1210 22:49:25.820524 4791 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.Node: Get "https://api-int.crc.testing:6443/api/v1/nodes?fieldSelector=metadata.name%3Dcrc&limit=500&resourceVersion=0": dial tcp 38.102.83.30:6443: connect: connection refused Dec 10 22:49:25 crc kubenswrapper[4791]: E1210 22:49:25.820613 4791 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.Node: failed to list *v1.Node: Get \"https://api-int.crc.testing:6443/api/v1/nodes?fieldSelector=metadata.name%3Dcrc&limit=500&resourceVersion=0\": dial tcp 38.102.83.30:6443: connect: connection refused" logger="UnhandledError" Dec 10 22:49:25 crc kubenswrapper[4791]: I1210 22:49:25.821368 4791 kubelet.go:854] "Not starting ClusterTrustBundle informer because we are in static kubelet mode" Dec 10 22:49:25 crc kubenswrapper[4791]: I1210 22:49:25.821956 4791 plugins.go:603] "Loaded volume plugin" pluginName="kubernetes.io/portworx-volume" Dec 10 22:49:25 crc kubenswrapper[4791]: I1210 22:49:25.821976 4791 plugins.go:603] "Loaded volume plugin" pluginName="kubernetes.io/empty-dir" Dec 10 22:49:25 crc kubenswrapper[4791]: I1210 22:49:25.821984 4791 plugins.go:603] "Loaded volume plugin" pluginName="kubernetes.io/git-repo" Dec 10 22:49:25 crc kubenswrapper[4791]: I1210 22:49:25.821991 4791 plugins.go:603] "Loaded volume plugin" pluginName="kubernetes.io/host-path" Dec 10 22:49:25 crc kubenswrapper[4791]: I1210 22:49:25.822001 4791 plugins.go:603] "Loaded volume plugin" pluginName="kubernetes.io/nfs" Dec 10 22:49:25 crc kubenswrapper[4791]: I1210 22:49:25.822008 4791 plugins.go:603] "Loaded volume plugin" pluginName="kubernetes.io/secret" Dec 10 22:49:25 crc kubenswrapper[4791]: I1210 22:49:25.822016 4791 plugins.go:603] "Loaded volume plugin" pluginName="kubernetes.io/iscsi" Dec 10 22:49:25 crc kubenswrapper[4791]: I1210 22:49:25.822032 4791 plugins.go:603] "Loaded volume plugin" pluginName="kubernetes.io/downward-api" Dec 10 22:49:25 crc kubenswrapper[4791]: I1210 22:49:25.822045 4791 plugins.go:603] "Loaded volume plugin" pluginName="kubernetes.io/fc" Dec 10 22:49:25 crc kubenswrapper[4791]: I1210 22:49:25.822053 4791 plugins.go:603] "Loaded volume plugin" pluginName="kubernetes.io/configmap" Dec 10 22:49:25 crc kubenswrapper[4791]: I1210 22:49:25.822078 4791 plugins.go:603] "Loaded volume plugin" pluginName="kubernetes.io/projected" Dec 10 22:49:25 crc kubenswrapper[4791]: I1210 22:49:25.822085 4791 plugins.go:603] "Loaded volume plugin" pluginName="kubernetes.io/local-volume" Dec 10 22:49:25 crc kubenswrapper[4791]: I1210 22:49:25.822289 4791 plugins.go:603] "Loaded volume plugin" pluginName="kubernetes.io/csi" Dec 10 22:49:25 crc kubenswrapper[4791]: I1210 22:49:25.822827 4791 server.go:1280] "Started kubelet" Dec 10 22:49:25 crc kubenswrapper[4791]: I1210 22:49:25.823265 4791 ratelimit.go:55] "Setting rate limiting for endpoint" service="podresources" qps=100 burstTokens=10 Dec 10 22:49:25 crc kubenswrapper[4791]: I1210 22:49:25.823261 4791 server.go:163] "Starting to listen" address="0.0.0.0" port=10250 Dec 10 22:49:25 crc kubenswrapper[4791]: I1210 22:49:25.824030 4791 server.go:236] "Starting to serve the podresources API" endpoint="unix:/var/lib/kubelet/pod-resources/kubelet.sock" Dec 10 22:49:25 crc kubenswrapper[4791]: I1210 22:49:25.824271 4791 csi_plugin.go:884] Failed to contact API server when waiting for CSINode publishing: Get "https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csinodes/crc?resourceVersion=0": dial tcp 38.102.83.30:6443: connect: connection refused Dec 10 22:49:25 crc systemd[1]: Started Kubernetes Kubelet. Dec 10 22:49:25 crc kubenswrapper[4791]: I1210 22:49:25.825836 4791 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate rotation is enabled Dec 10 22:49:25 crc kubenswrapper[4791]: I1210 22:49:25.825894 4791 fs_resource_analyzer.go:67] "Starting FS ResourceAnalyzer" Dec 10 22:49:25 crc kubenswrapper[4791]: I1210 22:49:25.826012 4791 server.go:460] "Adding debug handlers to kubelet server" Dec 10 22:49:25 crc kubenswrapper[4791]: E1210 22:49:25.826365 4791 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Dec 10 22:49:25 crc kubenswrapper[4791]: I1210 22:49:25.826303 4791 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2025-11-07 18:07:07.484445052 +0000 UTC Dec 10 22:49:25 crc kubenswrapper[4791]: I1210 22:49:25.826437 4791 volume_manager.go:287] "The desired_state_of_world populator starts" Dec 10 22:49:25 crc kubenswrapper[4791]: I1210 22:49:25.826502 4791 volume_manager.go:289] "Starting Kubelet Volume Manager" Dec 10 22:49:25 crc kubenswrapper[4791]: I1210 22:49:25.826578 4791 desired_state_of_world_populator.go:146] "Desired state populator starts to run" Dec 10 22:49:25 crc kubenswrapper[4791]: E1210 22:49:25.826624 4791 controller.go:145] "Failed to ensure lease exists, will retry" err="Get \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": dial tcp 38.102.83.30:6443: connect: connection refused" interval="200ms" Dec 10 22:49:25 crc kubenswrapper[4791]: W1210 22:49:25.827943 4791 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.CSIDriver: Get "https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csidrivers?limit=500&resourceVersion=0": dial tcp 38.102.83.30:6443: connect: connection refused Dec 10 22:49:25 crc kubenswrapper[4791]: E1210 22:49:25.828043 4791 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.CSIDriver: failed to list *v1.CSIDriver: Get \"https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csidrivers?limit=500&resourceVersion=0\": dial tcp 38.102.83.30:6443: connect: connection refused" logger="UnhandledError" Dec 10 22:49:25 crc kubenswrapper[4791]: E1210 22:49:25.826185 4791 event.go:368] "Unable to write event (may retry after sleeping)" err="Post \"https://api-int.crc.testing:6443/api/v1/namespaces/default/events\": dial tcp 38.102.83.30:6443: connect: connection refused" event="&Event{ObjectMeta:{crc.187ffc4829de3e7d default 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Node,Namespace:,Name:crc,UID:crc,APIVersion:,ResourceVersion:,FieldPath:,},Reason:Starting,Message:Starting kubelet.,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2025-12-10 22:49:25.822799485 +0000 UTC m=+0.252417098,LastTimestamp:2025-12-10 22:49:25.822799485 +0000 UTC m=+0.252417098,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Dec 10 22:49:25 crc kubenswrapper[4791]: I1210 22:49:25.830598 4791 factory.go:153] Registering CRI-O factory Dec 10 22:49:25 crc kubenswrapper[4791]: I1210 22:49:25.830658 4791 factory.go:221] Registration of the crio container factory successfully Dec 10 22:49:25 crc kubenswrapper[4791]: I1210 22:49:25.830795 4791 factory.go:219] Registration of the containerd container factory failed: unable to create containerd client: containerd: cannot unix dial containerd api service: dial unix /run/containerd/containerd.sock: connect: no such file or directory Dec 10 22:49:25 crc kubenswrapper[4791]: I1210 22:49:25.830811 4791 factory.go:55] Registering systemd factory Dec 10 22:49:25 crc kubenswrapper[4791]: I1210 22:49:25.830819 4791 factory.go:221] Registration of the systemd container factory successfully Dec 10 22:49:25 crc kubenswrapper[4791]: I1210 22:49:25.830840 4791 factory.go:103] Registering Raw factory Dec 10 22:49:25 crc kubenswrapper[4791]: I1210 22:49:25.830855 4791 manager.go:1196] Started watching for new ooms in manager Dec 10 22:49:25 crc kubenswrapper[4791]: I1210 22:49:25.832066 4791 manager.go:319] Starting recovery of all containers Dec 10 22:49:25 crc kubenswrapper[4791]: I1210 22:49:25.843989 4791 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="31d8b7a1-420e-4252-a5b7-eebe8a111292" volumeName="kubernetes.io/projected/31d8b7a1-420e-4252-a5b7-eebe8a111292-kube-api-access-zgdk5" seLinuxMountContext="" Dec 10 22:49:25 crc kubenswrapper[4791]: I1210 22:49:25.844070 4791 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" volumeName="kubernetes.io/projected/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-kube-api-access-ngvvp" seLinuxMountContext="" Dec 10 22:49:25 crc kubenswrapper[4791]: I1210 22:49:25.844100 4791 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" volumeName="kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-idp-0-file-data" seLinuxMountContext="" Dec 10 22:49:25 crc kubenswrapper[4791]: I1210 22:49:25.844123 4791 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="87cf06ed-a83f-41a7-828d-70653580a8cb" volumeName="kubernetes.io/configmap/87cf06ed-a83f-41a7-828d-70653580a8cb-config-volume" seLinuxMountContext="" Dec 10 22:49:25 crc kubenswrapper[4791]: I1210 22:49:25.844149 4791 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="efdd0498-1daa-4136-9a4a-3b948c2293fc" volumeName="kubernetes.io/secret/efdd0498-1daa-4136-9a4a-3b948c2293fc-webhook-certs" seLinuxMountContext="" Dec 10 22:49:25 crc kubenswrapper[4791]: I1210 22:49:25.844173 4791 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="308be0ea-9f5f-4b29-aeb1-5abd31a0b17b" volumeName="kubernetes.io/projected/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-kube-api-access-6ccd8" seLinuxMountContext="" Dec 10 22:49:25 crc kubenswrapper[4791]: I1210 22:49:25.844198 4791 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" volumeName="kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-service-ca" seLinuxMountContext="" Dec 10 22:49:25 crc kubenswrapper[4791]: I1210 22:49:25.844220 4791 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="8f668bae-612b-4b75-9490-919e737c6a3b" volumeName="kubernetes.io/configmap/8f668bae-612b-4b75-9490-919e737c6a3b-registry-certificates" seLinuxMountContext="" Dec 10 22:49:25 crc kubenswrapper[4791]: I1210 22:49:25.844251 4791 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="96b93a3a-6083-4aea-8eab-fe1aa8245ad9" volumeName="kubernetes.io/secret/96b93a3a-6083-4aea-8eab-fe1aa8245ad9-metrics-tls" seLinuxMountContext="" Dec 10 22:49:25 crc kubenswrapper[4791]: I1210 22:49:25.844276 4791 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="b6312bbd-5731-4ea0-a20f-81d5a57df44a" volumeName="kubernetes.io/secret/b6312bbd-5731-4ea0-a20f-81d5a57df44a-profile-collector-cert" seLinuxMountContext="" Dec 10 22:49:25 crc kubenswrapper[4791]: I1210 22:49:25.844300 4791 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1bf7eb37-55a3-4c65-b768-a94c82151e69" volumeName="kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-audit" seLinuxMountContext="" Dec 10 22:49:25 crc kubenswrapper[4791]: I1210 22:49:25.844328 4791 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1bf7eb37-55a3-4c65-b768-a94c82151e69" volumeName="kubernetes.io/secret/1bf7eb37-55a3-4c65-b768-a94c82151e69-encryption-config" seLinuxMountContext="" Dec 10 22:49:25 crc kubenswrapper[4791]: I1210 22:49:25.844392 4791 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="3ab1a177-2de0-46d9-b765-d0d0649bb42e" volumeName="kubernetes.io/projected/3ab1a177-2de0-46d9-b765-d0d0649bb42e-kube-api-access-4d4hj" seLinuxMountContext="" Dec 10 22:49:25 crc kubenswrapper[4791]: I1210 22:49:25.844423 4791 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="57a731c4-ef35-47a8-b875-bfb08a7f8011" volumeName="kubernetes.io/empty-dir/57a731c4-ef35-47a8-b875-bfb08a7f8011-catalog-content" seLinuxMountContext="" Dec 10 22:49:25 crc kubenswrapper[4791]: I1210 22:49:25.844447 4791 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="87cf06ed-a83f-41a7-828d-70653580a8cb" volumeName="kubernetes.io/secret/87cf06ed-a83f-41a7-828d-70653580a8cb-metrics-tls" seLinuxMountContext="" Dec 10 22:49:25 crc kubenswrapper[4791]: I1210 22:49:25.844473 4791 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="b11524ee-3fca-4b1b-9cdf-6da289fdbc7d" volumeName="kubernetes.io/projected/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d-kube-api-access-x4zgh" seLinuxMountContext="" Dec 10 22:49:25 crc kubenswrapper[4791]: I1210 22:49:25.844496 4791 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="b6cd30de-2eeb-49a2-ab40-9167f4560ff5" volumeName="kubernetes.io/projected/b6cd30de-2eeb-49a2-ab40-9167f4560ff5-kube-api-access-pj782" seLinuxMountContext="" Dec 10 22:49:25 crc kubenswrapper[4791]: I1210 22:49:25.844519 4791 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1bf7eb37-55a3-4c65-b768-a94c82151e69" volumeName="kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-image-import-ca" seLinuxMountContext="" Dec 10 22:49:25 crc kubenswrapper[4791]: I1210 22:49:25.844552 4791 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1bf7eb37-55a3-4c65-b768-a94c82151e69" volumeName="kubernetes.io/secret/1bf7eb37-55a3-4c65-b768-a94c82151e69-etcd-client" seLinuxMountContext="" Dec 10 22:49:25 crc kubenswrapper[4791]: I1210 22:49:25.844576 4791 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="8f668bae-612b-4b75-9490-919e737c6a3b" volumeName="kubernetes.io/secret/8f668bae-612b-4b75-9490-919e737c6a3b-installation-pull-secrets" seLinuxMountContext="" Dec 10 22:49:25 crc kubenswrapper[4791]: I1210 22:49:25.844598 4791 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="96b93a3a-6083-4aea-8eab-fe1aa8245ad9" volumeName="kubernetes.io/projected/96b93a3a-6083-4aea-8eab-fe1aa8245ad9-kube-api-access-nzwt7" seLinuxMountContext="" Dec 10 22:49:25 crc kubenswrapper[4791]: I1210 22:49:25.844621 4791 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="a31745f5-9847-4afe-82a5-3161cc66ca93" volumeName="kubernetes.io/projected/a31745f5-9847-4afe-82a5-3161cc66ca93-bound-sa-token" seLinuxMountContext="" Dec 10 22:49:25 crc kubenswrapper[4791]: I1210 22:49:25.844643 4791 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="c03ee662-fb2f-4fc4-a2c1-af487c19d254" volumeName="kubernetes.io/configmap/c03ee662-fb2f-4fc4-a2c1-af487c19d254-service-ca-bundle" seLinuxMountContext="" Dec 10 22:49:25 crc kubenswrapper[4791]: I1210 22:49:25.844668 4791 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="0b78653f-4ff9-4508-8672-245ed9b561e3" volumeName="kubernetes.io/configmap/0b78653f-4ff9-4508-8672-245ed9b561e3-service-ca" seLinuxMountContext="" Dec 10 22:49:25 crc kubenswrapper[4791]: I1210 22:49:25.844692 4791 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1bf7eb37-55a3-4c65-b768-a94c82151e69" volumeName="kubernetes.io/projected/1bf7eb37-55a3-4c65-b768-a94c82151e69-kube-api-access-sb6h7" seLinuxMountContext="" Dec 10 22:49:25 crc kubenswrapper[4791]: I1210 22:49:25.844717 4791 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5441d097-087c-4d9a-baa8-b210afa90fc9" volumeName="kubernetes.io/configmap/5441d097-087c-4d9a-baa8-b210afa90fc9-client-ca" seLinuxMountContext="" Dec 10 22:49:25 crc kubenswrapper[4791]: I1210 22:49:25.844745 4791 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5b88f790-22fa-440e-b583-365168c0b23d" volumeName="kubernetes.io/projected/5b88f790-22fa-440e-b583-365168c0b23d-kube-api-access-jkwtn" seLinuxMountContext="" Dec 10 22:49:25 crc kubenswrapper[4791]: I1210 22:49:25.844769 4791 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5fe579f8-e8a6-4643-bce5-a661393c4dde" volumeName="kubernetes.io/secret/5fe579f8-e8a6-4643-bce5-a661393c4dde-certs" seLinuxMountContext="" Dec 10 22:49:25 crc kubenswrapper[4791]: I1210 22:49:25.844799 4791 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="09ae3b1a-e8e7-4524-b54b-61eab6f9239a" volumeName="kubernetes.io/configmap/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-audit-policies" seLinuxMountContext="" Dec 10 22:49:25 crc kubenswrapper[4791]: I1210 22:49:25.844822 4791 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="22c825df-677d-4ca6-82db-3454ed06e783" volumeName="kubernetes.io/configmap/22c825df-677d-4ca6-82db-3454ed06e783-auth-proxy-config" seLinuxMountContext="" Dec 10 22:49:25 crc kubenswrapper[4791]: I1210 22:49:25.844896 4791 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="b6cd30de-2eeb-49a2-ab40-9167f4560ff5" volumeName="kubernetes.io/configmap/b6cd30de-2eeb-49a2-ab40-9167f4560ff5-marketplace-trusted-ca" seLinuxMountContext="" Dec 10 22:49:25 crc kubenswrapper[4791]: I1210 22:49:25.844927 4791 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="d75a4c96-2883-4a0b-bab2-0fab2b6c0b49" volumeName="kubernetes.io/configmap/d75a4c96-2883-4a0b-bab2-0fab2b6c0b49-iptables-alerter-script" seLinuxMountContext="" Dec 10 22:49:25 crc kubenswrapper[4791]: I1210 22:49:25.844953 4791 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="fda69060-fa79-4696-b1a6-7980f124bf7c" volumeName="kubernetes.io/secret/fda69060-fa79-4696-b1a6-7980f124bf7c-proxy-tls" seLinuxMountContext="" Dec 10 22:49:25 crc kubenswrapper[4791]: I1210 22:49:25.844976 4791 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="efdd0498-1daa-4136-9a4a-3b948c2293fc" volumeName="kubernetes.io/projected/efdd0498-1daa-4136-9a4a-3b948c2293fc-kube-api-access-fqsjt" seLinuxMountContext="" Dec 10 22:49:25 crc kubenswrapper[4791]: I1210 22:49:25.845001 4791 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1d611f23-29be-4491-8495-bee1670e935f" volumeName="kubernetes.io/projected/1d611f23-29be-4491-8495-bee1670e935f-kube-api-access-bf2bz" seLinuxMountContext="" Dec 10 22:49:25 crc kubenswrapper[4791]: I1210 22:49:25.845025 4791 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="43509403-f426-496e-be36-56cef71462f5" volumeName="kubernetes.io/projected/43509403-f426-496e-be36-56cef71462f5-kube-api-access-qg5z5" seLinuxMountContext="" Dec 10 22:49:25 crc kubenswrapper[4791]: I1210 22:49:25.845051 4791 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5b88f790-22fa-440e-b583-365168c0b23d" volumeName="kubernetes.io/secret/5b88f790-22fa-440e-b583-365168c0b23d-metrics-certs" seLinuxMountContext="" Dec 10 22:49:25 crc kubenswrapper[4791]: I1210 22:49:25.845077 4791 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6731426b-95fe-49ff-bb5f-40441049fde2" volumeName="kubernetes.io/secret/6731426b-95fe-49ff-bb5f-40441049fde2-control-plane-machine-set-operator-tls" seLinuxMountContext="" Dec 10 22:49:25 crc kubenswrapper[4791]: I1210 22:49:25.845102 4791 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="7583ce53-e0fe-4a16-9e4d-50516596a136" volumeName="kubernetes.io/configmap/7583ce53-e0fe-4a16-9e4d-50516596a136-client-ca" seLinuxMountContext="" Dec 10 22:49:25 crc kubenswrapper[4791]: I1210 22:49:25.845126 4791 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="c03ee662-fb2f-4fc4-a2c1-af487c19d254" volumeName="kubernetes.io/projected/c03ee662-fb2f-4fc4-a2c1-af487c19d254-kube-api-access-v47cf" seLinuxMountContext="" Dec 10 22:49:25 crc kubenswrapper[4791]: I1210 22:49:25.845151 4791 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="d75a4c96-2883-4a0b-bab2-0fab2b6c0b49" volumeName="kubernetes.io/projected/d75a4c96-2883-4a0b-bab2-0fab2b6c0b49-kube-api-access-rczfb" seLinuxMountContext="" Dec 10 22:49:25 crc kubenswrapper[4791]: I1210 22:49:25.845174 4791 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="e7e6199b-1264-4501-8953-767f51328d08" volumeName="kubernetes.io/secret/e7e6199b-1264-4501-8953-767f51328d08-serving-cert" seLinuxMountContext="" Dec 10 22:49:25 crc kubenswrapper[4791]: I1210 22:49:25.845198 4791 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="09ae3b1a-e8e7-4524-b54b-61eab6f9239a" volumeName="kubernetes.io/configmap/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-trusted-ca-bundle" seLinuxMountContext="" Dec 10 22:49:25 crc kubenswrapper[4791]: I1210 22:49:25.845220 4791 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6509e943-70c6-444c-bc41-48a544e36fbd" volumeName="kubernetes.io/configmap/6509e943-70c6-444c-bc41-48a544e36fbd-trusted-ca-bundle" seLinuxMountContext="" Dec 10 22:49:25 crc kubenswrapper[4791]: I1210 22:49:25.845243 4791 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="9d4552c7-cd75-42dd-8880-30dd377c49a4" volumeName="kubernetes.io/configmap/9d4552c7-cd75-42dd-8880-30dd377c49a4-trusted-ca" seLinuxMountContext="" Dec 10 22:49:25 crc kubenswrapper[4791]: I1210 22:49:25.845269 4791 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="b6cd30de-2eeb-49a2-ab40-9167f4560ff5" volumeName="kubernetes.io/secret/b6cd30de-2eeb-49a2-ab40-9167f4560ff5-marketplace-operator-metrics" seLinuxMountContext="" Dec 10 22:49:25 crc kubenswrapper[4791]: I1210 22:49:25.845292 4791 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="bc5039c0-ea34-426b-a2b7-fbbc87b49a6d" volumeName="kubernetes.io/secret/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d-serving-cert" seLinuxMountContext="" Dec 10 22:49:25 crc kubenswrapper[4791]: I1210 22:49:25.845331 4791 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="ef543e1b-8068-4ea3-b32a-61027b32e95d" volumeName="kubernetes.io/configmap/ef543e1b-8068-4ea3-b32a-61027b32e95d-env-overrides" seLinuxMountContext="" Dec 10 22:49:25 crc kubenswrapper[4791]: I1210 22:49:25.845393 4791 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="20b0d48f-5fd6-431c-a545-e3c800c7b866" volumeName="kubernetes.io/projected/20b0d48f-5fd6-431c-a545-e3c800c7b866-kube-api-access-w9rds" seLinuxMountContext="" Dec 10 22:49:25 crc kubenswrapper[4791]: I1210 22:49:25.845419 4791 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="44663579-783b-4372-86d6-acf235a62d72" volumeName="kubernetes.io/projected/44663579-783b-4372-86d6-acf235a62d72-kube-api-access-vt5rc" seLinuxMountContext="" Dec 10 22:49:25 crc kubenswrapper[4791]: I1210 22:49:25.845445 4791 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" volumeName="kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-template-provider-selection" seLinuxMountContext="" Dec 10 22:49:25 crc kubenswrapper[4791]: I1210 22:49:25.845471 4791 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6ea678ab-3438-413e-bfe3-290ae7725660" volumeName="kubernetes.io/projected/6ea678ab-3438-413e-bfe3-290ae7725660-kube-api-access-htfz6" seLinuxMountContext="" Dec 10 22:49:25 crc kubenswrapper[4791]: I1210 22:49:25.845505 4791 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="8cea82b4-6893-4ddc-af9f-1bb5ae425c5b" volumeName="kubernetes.io/configmap/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b-config" seLinuxMountContext="" Dec 10 22:49:25 crc kubenswrapper[4791]: I1210 22:49:25.845576 4791 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6ea678ab-3438-413e-bfe3-290ae7725660" volumeName="kubernetes.io/secret/6ea678ab-3438-413e-bfe3-290ae7725660-ovn-node-metrics-cert" seLinuxMountContext="" Dec 10 22:49:25 crc kubenswrapper[4791]: I1210 22:49:25.845605 4791 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="7539238d-5fe0-46ed-884e-1c3b566537ec" volumeName="kubernetes.io/configmap/7539238d-5fe0-46ed-884e-1c3b566537ec-config" seLinuxMountContext="" Dec 10 22:49:25 crc kubenswrapper[4791]: I1210 22:49:25.845632 4791 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="925f1c65-6136-48ba-85aa-3a3b50560753" volumeName="kubernetes.io/secret/925f1c65-6136-48ba-85aa-3a3b50560753-ovn-control-plane-metrics-cert" seLinuxMountContext="" Dec 10 22:49:25 crc kubenswrapper[4791]: I1210 22:49:25.845657 4791 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="0b574797-001e-440a-8f4e-c0be86edad0f" volumeName="kubernetes.io/secret/0b574797-001e-440a-8f4e-c0be86edad0f-proxy-tls" seLinuxMountContext="" Dec 10 22:49:25 crc kubenswrapper[4791]: I1210 22:49:25.845683 4791 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="31d8b7a1-420e-4252-a5b7-eebe8a111292" volumeName="kubernetes.io/secret/31d8b7a1-420e-4252-a5b7-eebe8a111292-proxy-tls" seLinuxMountContext="" Dec 10 22:49:25 crc kubenswrapper[4791]: I1210 22:49:25.845711 4791 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="3b6479f0-333b-4a96-9adf-2099afdc2447" volumeName="kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr" seLinuxMountContext="" Dec 10 22:49:25 crc kubenswrapper[4791]: I1210 22:49:25.845737 4791 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="43509403-f426-496e-be36-56cef71462f5" volumeName="kubernetes.io/secret/43509403-f426-496e-be36-56cef71462f5-console-serving-cert" seLinuxMountContext="" Dec 10 22:49:25 crc kubenswrapper[4791]: I1210 22:49:25.845764 4791 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="57a731c4-ef35-47a8-b875-bfb08a7f8011" volumeName="kubernetes.io/projected/57a731c4-ef35-47a8-b875-bfb08a7f8011-kube-api-access-cfbct" seLinuxMountContext="" Dec 10 22:49:25 crc kubenswrapper[4791]: I1210 22:49:25.845788 4791 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="a31745f5-9847-4afe-82a5-3161cc66ca93" volumeName="kubernetes.io/secret/a31745f5-9847-4afe-82a5-3161cc66ca93-metrics-tls" seLinuxMountContext="" Dec 10 22:49:25 crc kubenswrapper[4791]: I1210 22:49:25.845813 4791 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" volumeName="kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-cliconfig" seLinuxMountContext="" Dec 10 22:49:25 crc kubenswrapper[4791]: I1210 22:49:25.845839 4791 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" volumeName="kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-template-error" seLinuxMountContext="" Dec 10 22:49:25 crc kubenswrapper[4791]: I1210 22:49:25.845864 4791 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="4bb40260-dbaa-4fb0-84df-5e680505d512" volumeName="kubernetes.io/projected/4bb40260-dbaa-4fb0-84df-5e680505d512-kube-api-access-2w9zh" seLinuxMountContext="" Dec 10 22:49:25 crc kubenswrapper[4791]: I1210 22:49:25.845887 4791 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6731426b-95fe-49ff-bb5f-40441049fde2" volumeName="kubernetes.io/projected/6731426b-95fe-49ff-bb5f-40441049fde2-kube-api-access-x7zkh" seLinuxMountContext="" Dec 10 22:49:25 crc kubenswrapper[4791]: I1210 22:49:25.845910 4791 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="7539238d-5fe0-46ed-884e-1c3b566537ec" volumeName="kubernetes.io/secret/7539238d-5fe0-46ed-884e-1c3b566537ec-serving-cert" seLinuxMountContext="" Dec 10 22:49:25 crc kubenswrapper[4791]: I1210 22:49:25.845934 4791 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6ea678ab-3438-413e-bfe3-290ae7725660" volumeName="kubernetes.io/configmap/6ea678ab-3438-413e-bfe3-290ae7725660-env-overrides" seLinuxMountContext="" Dec 10 22:49:25 crc kubenswrapper[4791]: I1210 22:49:25.845960 4791 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6ea678ab-3438-413e-bfe3-290ae7725660" volumeName="kubernetes.io/configmap/6ea678ab-3438-413e-bfe3-290ae7725660-ovnkube-script-lib" seLinuxMountContext="" Dec 10 22:49:25 crc kubenswrapper[4791]: I1210 22:49:25.845983 4791 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="8f668bae-612b-4b75-9490-919e737c6a3b" volumeName="kubernetes.io/projected/8f668bae-612b-4b75-9490-919e737c6a3b-registry-tls" seLinuxMountContext="" Dec 10 22:49:25 crc kubenswrapper[4791]: I1210 22:49:25.846007 4791 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="09efc573-dbb6-4249-bd59-9b87aba8dd28" volumeName="kubernetes.io/configmap/09efc573-dbb6-4249-bd59-9b87aba8dd28-etcd-ca" seLinuxMountContext="" Dec 10 22:49:25 crc kubenswrapper[4791]: I1210 22:49:25.846032 4791 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1bf7eb37-55a3-4c65-b768-a94c82151e69" volumeName="kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-etcd-serving-ca" seLinuxMountContext="" Dec 10 22:49:25 crc kubenswrapper[4791]: I1210 22:49:25.846055 4791 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1bf7eb37-55a3-4c65-b768-a94c82151e69" volumeName="kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-config" seLinuxMountContext="" Dec 10 22:49:25 crc kubenswrapper[4791]: I1210 22:49:25.846078 4791 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1d611f23-29be-4491-8495-bee1670e935f" volumeName="kubernetes.io/empty-dir/1d611f23-29be-4491-8495-bee1670e935f-utilities" seLinuxMountContext="" Dec 10 22:49:25 crc kubenswrapper[4791]: I1210 22:49:25.846111 4791 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="496e6271-fb68-4057-954e-a0d97a4afa3f" volumeName="kubernetes.io/projected/496e6271-fb68-4057-954e-a0d97a4afa3f-kube-api-access" seLinuxMountContext="" Dec 10 22:49:25 crc kubenswrapper[4791]: I1210 22:49:25.846136 4791 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="8f668bae-612b-4b75-9490-919e737c6a3b" volumeName="kubernetes.io/projected/8f668bae-612b-4b75-9490-919e737c6a3b-kube-api-access-kfwg7" seLinuxMountContext="" Dec 10 22:49:25 crc kubenswrapper[4791]: I1210 22:49:25.846161 4791 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" volumeName="kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-audit-policies" seLinuxMountContext="" Dec 10 22:49:25 crc kubenswrapper[4791]: I1210 22:49:25.846186 4791 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6509e943-70c6-444c-bc41-48a544e36fbd" volumeName="kubernetes.io/projected/6509e943-70c6-444c-bc41-48a544e36fbd-kube-api-access-6g6sz" seLinuxMountContext="" Dec 10 22:49:25 crc kubenswrapper[4791]: I1210 22:49:25.846211 4791 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="09ae3b1a-e8e7-4524-b54b-61eab6f9239a" volumeName="kubernetes.io/configmap/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-etcd-serving-ca" seLinuxMountContext="" Dec 10 22:49:25 crc kubenswrapper[4791]: I1210 22:49:25.846240 4791 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="09efc573-dbb6-4249-bd59-9b87aba8dd28" volumeName="kubernetes.io/configmap/09efc573-dbb6-4249-bd59-9b87aba8dd28-config" seLinuxMountContext="" Dec 10 22:49:25 crc kubenswrapper[4791]: I1210 22:49:25.846264 4791 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="09efc573-dbb6-4249-bd59-9b87aba8dd28" volumeName="kubernetes.io/projected/09efc573-dbb6-4249-bd59-9b87aba8dd28-kube-api-access-8tdtz" seLinuxMountContext="" Dec 10 22:49:25 crc kubenswrapper[4791]: I1210 22:49:25.846289 4791 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="0b78653f-4ff9-4508-8672-245ed9b561e3" volumeName="kubernetes.io/projected/0b78653f-4ff9-4508-8672-245ed9b561e3-kube-api-access" seLinuxMountContext="" Dec 10 22:49:25 crc kubenswrapper[4791]: I1210 22:49:25.846312 4791 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="37a5e44f-9a88-4405-be8a-b645485e7312" volumeName="kubernetes.io/projected/37a5e44f-9a88-4405-be8a-b645485e7312-kube-api-access-rdwmf" seLinuxMountContext="" Dec 10 22:49:25 crc kubenswrapper[4791]: I1210 22:49:25.846336 4791 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="09ae3b1a-e8e7-4524-b54b-61eab6f9239a" volumeName="kubernetes.io/projected/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-kube-api-access-zkvpv" seLinuxMountContext="" Dec 10 22:49:25 crc kubenswrapper[4791]: I1210 22:49:25.846399 4791 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6509e943-70c6-444c-bc41-48a544e36fbd" volumeName="kubernetes.io/secret/6509e943-70c6-444c-bc41-48a544e36fbd-serving-cert" seLinuxMountContext="" Dec 10 22:49:25 crc kubenswrapper[4791]: I1210 22:49:25.846424 4791 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="ef543e1b-8068-4ea3-b32a-61027b32e95d" volumeName="kubernetes.io/secret/ef543e1b-8068-4ea3-b32a-61027b32e95d-webhook-cert" seLinuxMountContext="" Dec 10 22:49:25 crc kubenswrapper[4791]: I1210 22:49:25.846449 4791 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="09efc573-dbb6-4249-bd59-9b87aba8dd28" volumeName="kubernetes.io/secret/09efc573-dbb6-4249-bd59-9b87aba8dd28-serving-cert" seLinuxMountContext="" Dec 10 22:49:25 crc kubenswrapper[4791]: I1210 22:49:25.846474 4791 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1d611f23-29be-4491-8495-bee1670e935f" volumeName="kubernetes.io/empty-dir/1d611f23-29be-4491-8495-bee1670e935f-catalog-content" seLinuxMountContext="" Dec 10 22:49:25 crc kubenswrapper[4791]: I1210 22:49:25.846501 4791 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="57a731c4-ef35-47a8-b875-bfb08a7f8011" volumeName="kubernetes.io/empty-dir/57a731c4-ef35-47a8-b875-bfb08a7f8011-utilities" seLinuxMountContext="" Dec 10 22:49:25 crc kubenswrapper[4791]: I1210 22:49:25.846528 4791 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="8cea82b4-6893-4ddc-af9f-1bb5ae425c5b" volumeName="kubernetes.io/secret/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b-serving-cert" seLinuxMountContext="" Dec 10 22:49:25 crc kubenswrapper[4791]: I1210 22:49:25.846551 4791 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="bf126b07-da06-4140-9a57-dfd54fc6b486" volumeName="kubernetes.io/secret/bf126b07-da06-4140-9a57-dfd54fc6b486-image-registry-operator-tls" seLinuxMountContext="" Dec 10 22:49:25 crc kubenswrapper[4791]: I1210 22:49:25.846577 4791 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="7539238d-5fe0-46ed-884e-1c3b566537ec" volumeName="kubernetes.io/projected/7539238d-5fe0-46ed-884e-1c3b566537ec-kube-api-access-tk88c" seLinuxMountContext="" Dec 10 22:49:25 crc kubenswrapper[4791]: I1210 22:49:25.846602 4791 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="7583ce53-e0fe-4a16-9e4d-50516596a136" volumeName="kubernetes.io/projected/7583ce53-e0fe-4a16-9e4d-50516596a136-kube-api-access-xcphl" seLinuxMountContext="" Dec 10 22:49:25 crc kubenswrapper[4791]: I1210 22:49:25.846625 4791 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="8f668bae-612b-4b75-9490-919e737c6a3b" volumeName="kubernetes.io/projected/8f668bae-612b-4b75-9490-919e737c6a3b-bound-sa-token" seLinuxMountContext="" Dec 10 22:49:25 crc kubenswrapper[4791]: I1210 22:49:25.846649 4791 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1bf7eb37-55a3-4c65-b768-a94c82151e69" volumeName="kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-trusted-ca-bundle" seLinuxMountContext="" Dec 10 22:49:25 crc kubenswrapper[4791]: I1210 22:49:25.846675 4791 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="20b0d48f-5fd6-431c-a545-e3c800c7b866" volumeName="kubernetes.io/secret/20b0d48f-5fd6-431c-a545-e3c800c7b866-cert" seLinuxMountContext="" Dec 10 22:49:25 crc kubenswrapper[4791]: I1210 22:49:25.846700 4791 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5441d097-087c-4d9a-baa8-b210afa90fc9" volumeName="kubernetes.io/configmap/5441d097-087c-4d9a-baa8-b210afa90fc9-config" seLinuxMountContext="" Dec 10 22:49:25 crc kubenswrapper[4791]: I1210 22:49:25.846724 4791 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6402fda4-df10-493c-b4e5-d0569419652d" volumeName="kubernetes.io/secret/6402fda4-df10-493c-b4e5-d0569419652d-machine-api-operator-tls" seLinuxMountContext="" Dec 10 22:49:25 crc kubenswrapper[4791]: I1210 22:49:25.846749 4791 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6509e943-70c6-444c-bc41-48a544e36fbd" volumeName="kubernetes.io/configmap/6509e943-70c6-444c-bc41-48a544e36fbd-service-ca-bundle" seLinuxMountContext="" Dec 10 22:49:25 crc kubenswrapper[4791]: I1210 22:49:25.846773 4791 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="925f1c65-6136-48ba-85aa-3a3b50560753" volumeName="kubernetes.io/projected/925f1c65-6136-48ba-85aa-3a3b50560753-kube-api-access-s4n52" seLinuxMountContext="" Dec 10 22:49:25 crc kubenswrapper[4791]: I1210 22:49:25.846991 4791 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="bf126b07-da06-4140-9a57-dfd54fc6b486" volumeName="kubernetes.io/projected/bf126b07-da06-4140-9a57-dfd54fc6b486-bound-sa-token" seLinuxMountContext="" Dec 10 22:49:25 crc kubenswrapper[4791]: I1210 22:49:25.847018 4791 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="c03ee662-fb2f-4fc4-a2c1-af487c19d254" volumeName="kubernetes.io/secret/c03ee662-fb2f-4fc4-a2c1-af487c19d254-default-certificate" seLinuxMountContext="" Dec 10 22:49:25 crc kubenswrapper[4791]: I1210 22:49:25.847041 4791 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="c03ee662-fb2f-4fc4-a2c1-af487c19d254" volumeName="kubernetes.io/secret/c03ee662-fb2f-4fc4-a2c1-af487c19d254-metrics-certs" seLinuxMountContext="" Dec 10 22:49:25 crc kubenswrapper[4791]: I1210 22:49:25.847066 4791 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="bd23aa5c-e532-4e53-bccf-e79f130c5ae8" volumeName="kubernetes.io/projected/bd23aa5c-e532-4e53-bccf-e79f130c5ae8-kube-api-access-jhbk2" seLinuxMountContext="" Dec 10 22:49:25 crc kubenswrapper[4791]: I1210 22:49:25.847100 4791 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="25e176fe-21b4-4974-b1ed-c8b94f112a7f" volumeName="kubernetes.io/projected/25e176fe-21b4-4974-b1ed-c8b94f112a7f-kube-api-access-d4lsv" seLinuxMountContext="" Dec 10 22:49:25 crc kubenswrapper[4791]: I1210 22:49:25.847129 4791 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="308be0ea-9f5f-4b29-aeb1-5abd31a0b17b" volumeName="kubernetes.io/empty-dir/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-tmpfs" seLinuxMountContext="" Dec 10 22:49:25 crc kubenswrapper[4791]: I1210 22:49:25.847157 4791 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="3cb93b32-e0ae-4377-b9c8-fdb9842c6d59" volumeName="kubernetes.io/configmap/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59-serviceca" seLinuxMountContext="" Dec 10 22:49:25 crc kubenswrapper[4791]: I1210 22:49:25.847184 4791 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="496e6271-fb68-4057-954e-a0d97a4afa3f" volumeName="kubernetes.io/configmap/496e6271-fb68-4057-954e-a0d97a4afa3f-config" seLinuxMountContext="" Dec 10 22:49:25 crc kubenswrapper[4791]: I1210 22:49:25.847212 4791 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="925f1c65-6136-48ba-85aa-3a3b50560753" volumeName="kubernetes.io/configmap/925f1c65-6136-48ba-85aa-3a3b50560753-ovnkube-config" seLinuxMountContext="" Dec 10 22:49:25 crc kubenswrapper[4791]: I1210 22:49:25.847244 4791 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="8f668bae-612b-4b75-9490-919e737c6a3b" volumeName="kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" seLinuxMountContext="" Dec 10 22:49:25 crc kubenswrapper[4791]: I1210 22:49:25.848218 4791 reconstruct.go:144] "Volume is marked device as uncertain and added into the actual state" volumeName="kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" deviceMountPath="/var/lib/kubelet/plugins/kubernetes.io/csi/kubevirt.io.hostpath-provisioner/1f4776af88835e41c12b831b4c9fed40233456d14189815a54dbe7f892fc1983/globalmount" Dec 10 22:49:25 crc kubenswrapper[4791]: I1210 22:49:25.848276 4791 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="9d4552c7-cd75-42dd-8880-30dd377c49a4" volumeName="kubernetes.io/secret/9d4552c7-cd75-42dd-8880-30dd377c49a4-serving-cert" seLinuxMountContext="" Dec 10 22:49:25 crc kubenswrapper[4791]: I1210 22:49:25.848309 4791 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="a0128f3a-b052-44ed-a84e-c4c8aaf17c13" volumeName="kubernetes.io/projected/a0128f3a-b052-44ed-a84e-c4c8aaf17c13-kube-api-access-gf66m" seLinuxMountContext="" Dec 10 22:49:25 crc kubenswrapper[4791]: I1210 22:49:25.848397 4791 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="bc5039c0-ea34-426b-a2b7-fbbc87b49a6d" volumeName="kubernetes.io/empty-dir/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d-available-featuregates" seLinuxMountContext="" Dec 10 22:49:25 crc kubenswrapper[4791]: I1210 22:49:25.848441 4791 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="87cf06ed-a83f-41a7-828d-70653580a8cb" volumeName="kubernetes.io/projected/87cf06ed-a83f-41a7-828d-70653580a8cb-kube-api-access-d6qdx" seLinuxMountContext="" Dec 10 22:49:25 crc kubenswrapper[4791]: I1210 22:49:25.848468 4791 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="f88749ec-7931-4ee7-b3fc-1ec5e11f92e9" volumeName="kubernetes.io/projected/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9-kube-api-access-dbsvg" seLinuxMountContext="" Dec 10 22:49:25 crc kubenswrapper[4791]: I1210 22:49:25.848493 4791 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="fda69060-fa79-4696-b1a6-7980f124bf7c" volumeName="kubernetes.io/configmap/fda69060-fa79-4696-b1a6-7980f124bf7c-mcd-auth-proxy-config" seLinuxMountContext="" Dec 10 22:49:25 crc kubenswrapper[4791]: I1210 22:49:25.848520 4791 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1bf7eb37-55a3-4c65-b768-a94c82151e69" volumeName="kubernetes.io/secret/1bf7eb37-55a3-4c65-b768-a94c82151e69-serving-cert" seLinuxMountContext="" Dec 10 22:49:25 crc kubenswrapper[4791]: I1210 22:49:25.848546 4791 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="31d8b7a1-420e-4252-a5b7-eebe8a111292" volumeName="kubernetes.io/configmap/31d8b7a1-420e-4252-a5b7-eebe8a111292-auth-proxy-config" seLinuxMountContext="" Dec 10 22:49:25 crc kubenswrapper[4791]: I1210 22:49:25.848570 4791 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="37a5e44f-9a88-4405-be8a-b645485e7312" volumeName="kubernetes.io/secret/37a5e44f-9a88-4405-be8a-b645485e7312-metrics-tls" seLinuxMountContext="" Dec 10 22:49:25 crc kubenswrapper[4791]: I1210 22:49:25.848597 4791 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" volumeName="kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-template-login" seLinuxMountContext="" Dec 10 22:49:25 crc kubenswrapper[4791]: I1210 22:49:25.848629 4791 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5441d097-087c-4d9a-baa8-b210afa90fc9" volumeName="kubernetes.io/projected/5441d097-087c-4d9a-baa8-b210afa90fc9-kube-api-access-2d4wz" seLinuxMountContext="" Dec 10 22:49:25 crc kubenswrapper[4791]: I1210 22:49:25.848654 4791 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="308be0ea-9f5f-4b29-aeb1-5abd31a0b17b" volumeName="kubernetes.io/secret/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-webhook-cert" seLinuxMountContext="" Dec 10 22:49:25 crc kubenswrapper[4791]: I1210 22:49:25.848680 4791 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="4bb40260-dbaa-4fb0-84df-5e680505d512" volumeName="kubernetes.io/configmap/4bb40260-dbaa-4fb0-84df-5e680505d512-cni-binary-copy" seLinuxMountContext="" Dec 10 22:49:25 crc kubenswrapper[4791]: I1210 22:49:25.848705 4791 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="9d751cbb-f2e2-430d-9754-c882a5e924a5" volumeName="kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl" seLinuxMountContext="" Dec 10 22:49:25 crc kubenswrapper[4791]: I1210 22:49:25.848732 4791 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="b6312bbd-5731-4ea0-a20f-81d5a57df44a" volumeName="kubernetes.io/secret/b6312bbd-5731-4ea0-a20f-81d5a57df44a-srv-cert" seLinuxMountContext="" Dec 10 22:49:25 crc kubenswrapper[4791]: I1210 22:49:25.848758 4791 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="31d8b7a1-420e-4252-a5b7-eebe8a111292" volumeName="kubernetes.io/configmap/31d8b7a1-420e-4252-a5b7-eebe8a111292-images" seLinuxMountContext="" Dec 10 22:49:25 crc kubenswrapper[4791]: I1210 22:49:25.848787 4791 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="3cb93b32-e0ae-4377-b9c8-fdb9842c6d59" volumeName="kubernetes.io/projected/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59-kube-api-access-wxkg8" seLinuxMountContext="" Dec 10 22:49:25 crc kubenswrapper[4791]: I1210 22:49:25.848814 4791 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="496e6271-fb68-4057-954e-a0d97a4afa3f" volumeName="kubernetes.io/secret/496e6271-fb68-4057-954e-a0d97a4afa3f-serving-cert" seLinuxMountContext="" Dec 10 22:49:25 crc kubenswrapper[4791]: I1210 22:49:25.848841 4791 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="01ab3dd5-8196-46d0-ad33-122e2ca51def" volumeName="kubernetes.io/configmap/01ab3dd5-8196-46d0-ad33-122e2ca51def-config" seLinuxMountContext="" Dec 10 22:49:25 crc kubenswrapper[4791]: I1210 22:49:25.848868 4791 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="01ab3dd5-8196-46d0-ad33-122e2ca51def" volumeName="kubernetes.io/projected/01ab3dd5-8196-46d0-ad33-122e2ca51def-kube-api-access-w7l8j" seLinuxMountContext="" Dec 10 22:49:25 crc kubenswrapper[4791]: I1210 22:49:25.848896 4791 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="0b574797-001e-440a-8f4e-c0be86edad0f" volumeName="kubernetes.io/projected/0b574797-001e-440a-8f4e-c0be86edad0f-kube-api-access-lzf88" seLinuxMountContext="" Dec 10 22:49:25 crc kubenswrapper[4791]: I1210 22:49:25.848921 4791 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1386a44e-36a2-460c-96d0-0359d2b6f0f5" volumeName="kubernetes.io/projected/1386a44e-36a2-460c-96d0-0359d2b6f0f5-kube-api-access" seLinuxMountContext="" Dec 10 22:49:25 crc kubenswrapper[4791]: I1210 22:49:25.848947 4791 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c" volumeName="kubernetes.io/secret/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c-serving-cert" seLinuxMountContext="" Dec 10 22:49:25 crc kubenswrapper[4791]: I1210 22:49:25.848972 4791 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="4bb40260-dbaa-4fb0-84df-5e680505d512" volumeName="kubernetes.io/configmap/4bb40260-dbaa-4fb0-84df-5e680505d512-multus-daemon-config" seLinuxMountContext="" Dec 10 22:49:25 crc kubenswrapper[4791]: I1210 22:49:25.848997 4791 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="7583ce53-e0fe-4a16-9e4d-50516596a136" volumeName="kubernetes.io/configmap/7583ce53-e0fe-4a16-9e4d-50516596a136-proxy-ca-bundles" seLinuxMountContext="" Dec 10 22:49:25 crc kubenswrapper[4791]: I1210 22:49:25.849025 4791 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="a0128f3a-b052-44ed-a84e-c4c8aaf17c13" volumeName="kubernetes.io/secret/a0128f3a-b052-44ed-a84e-c4c8aaf17c13-samples-operator-tls" seLinuxMountContext="" Dec 10 22:49:25 crc kubenswrapper[4791]: I1210 22:49:25.849050 4791 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="e7e6199b-1264-4501-8953-767f51328d08" volumeName="kubernetes.io/projected/e7e6199b-1264-4501-8953-767f51328d08-kube-api-access" seLinuxMountContext="" Dec 10 22:49:25 crc kubenswrapper[4791]: I1210 22:49:25.849075 4791 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" volumeName="kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-trusted-ca-bundle" seLinuxMountContext="" Dec 10 22:49:25 crc kubenswrapper[4791]: I1210 22:49:25.849100 4791 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5441d097-087c-4d9a-baa8-b210afa90fc9" volumeName="kubernetes.io/secret/5441d097-087c-4d9a-baa8-b210afa90fc9-serving-cert" seLinuxMountContext="" Dec 10 22:49:25 crc kubenswrapper[4791]: I1210 22:49:25.849125 4791 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5fe579f8-e8a6-4643-bce5-a661393c4dde" volumeName="kubernetes.io/secret/5fe579f8-e8a6-4643-bce5-a661393c4dde-node-bootstrap-token" seLinuxMountContext="" Dec 10 22:49:25 crc kubenswrapper[4791]: I1210 22:49:25.849150 4791 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="f88749ec-7931-4ee7-b3fc-1ec5e11f92e9" volumeName="kubernetes.io/secret/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9-srv-cert" seLinuxMountContext="" Dec 10 22:49:25 crc kubenswrapper[4791]: I1210 22:49:25.849175 4791 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="e7e6199b-1264-4501-8953-767f51328d08" volumeName="kubernetes.io/configmap/e7e6199b-1264-4501-8953-767f51328d08-config" seLinuxMountContext="" Dec 10 22:49:25 crc kubenswrapper[4791]: I1210 22:49:25.849205 4791 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="ef543e1b-8068-4ea3-b32a-61027b32e95d" volumeName="kubernetes.io/configmap/ef543e1b-8068-4ea3-b32a-61027b32e95d-ovnkube-identity-cm" seLinuxMountContext="" Dec 10 22:49:25 crc kubenswrapper[4791]: I1210 22:49:25.849232 4791 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="09ae3b1a-e8e7-4524-b54b-61eab6f9239a" volumeName="kubernetes.io/secret/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-encryption-config" seLinuxMountContext="" Dec 10 22:49:25 crc kubenswrapper[4791]: I1210 22:49:25.849257 4791 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c" volumeName="kubernetes.io/projected/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c-kube-api-access-qs4fp" seLinuxMountContext="" Dec 10 22:49:25 crc kubenswrapper[4791]: I1210 22:49:25.849283 4791 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="7bb08738-c794-4ee8-9972-3a62ca171029" volumeName="kubernetes.io/projected/7bb08738-c794-4ee8-9972-3a62ca171029-kube-api-access-279lb" seLinuxMountContext="" Dec 10 22:49:25 crc kubenswrapper[4791]: I1210 22:49:25.849311 4791 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="8f668bae-612b-4b75-9490-919e737c6a3b" volumeName="kubernetes.io/empty-dir/8f668bae-612b-4b75-9490-919e737c6a3b-ca-trust-extracted" seLinuxMountContext="" Dec 10 22:49:25 crc kubenswrapper[4791]: I1210 22:49:25.849370 4791 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d" volumeName="kubernetes.io/projected/cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d-kube-api-access-x2m85" seLinuxMountContext="" Dec 10 22:49:25 crc kubenswrapper[4791]: I1210 22:49:25.849401 4791 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="0b78653f-4ff9-4508-8672-245ed9b561e3" volumeName="kubernetes.io/secret/0b78653f-4ff9-4508-8672-245ed9b561e3-serving-cert" seLinuxMountContext="" Dec 10 22:49:25 crc kubenswrapper[4791]: I1210 22:49:25.849429 4791 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" volumeName="kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-ocp-branding-template" seLinuxMountContext="" Dec 10 22:49:25 crc kubenswrapper[4791]: I1210 22:49:25.849456 4791 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5225d0e4-402f-4861-b410-819f433b1803" volumeName="kubernetes.io/empty-dir/5225d0e4-402f-4861-b410-819f433b1803-catalog-content" seLinuxMountContext="" Dec 10 22:49:25 crc kubenswrapper[4791]: I1210 22:49:25.849483 4791 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5225d0e4-402f-4861-b410-819f433b1803" volumeName="kubernetes.io/empty-dir/5225d0e4-402f-4861-b410-819f433b1803-utilities" seLinuxMountContext="" Dec 10 22:49:25 crc kubenswrapper[4791]: I1210 22:49:25.849508 4791 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="7bb08738-c794-4ee8-9972-3a62ca171029" volumeName="kubernetes.io/configmap/7bb08738-c794-4ee8-9972-3a62ca171029-cni-sysctl-allowlist" seLinuxMountContext="" Dec 10 22:49:25 crc kubenswrapper[4791]: I1210 22:49:25.850471 4791 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" volumeName="kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-serving-cert" seLinuxMountContext="" Dec 10 22:49:25 crc kubenswrapper[4791]: I1210 22:49:25.850539 4791 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6402fda4-df10-493c-b4e5-d0569419652d" volumeName="kubernetes.io/configmap/6402fda4-df10-493c-b4e5-d0569419652d-images" seLinuxMountContext="" Dec 10 22:49:25 crc kubenswrapper[4791]: I1210 22:49:25.850554 4791 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="925f1c65-6136-48ba-85aa-3a3b50560753" volumeName="kubernetes.io/configmap/925f1c65-6136-48ba-85aa-3a3b50560753-env-overrides" seLinuxMountContext="" Dec 10 22:49:25 crc kubenswrapper[4791]: I1210 22:49:25.850567 4791 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="bc5039c0-ea34-426b-a2b7-fbbc87b49a6d" volumeName="kubernetes.io/projected/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d-kube-api-access-mnrrd" seLinuxMountContext="" Dec 10 22:49:25 crc kubenswrapper[4791]: I1210 22:49:25.850581 4791 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="ef543e1b-8068-4ea3-b32a-61027b32e95d" volumeName="kubernetes.io/projected/ef543e1b-8068-4ea3-b32a-61027b32e95d-kube-api-access-s2kz5" seLinuxMountContext="" Dec 10 22:49:25 crc kubenswrapper[4791]: I1210 22:49:25.850597 4791 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="01ab3dd5-8196-46d0-ad33-122e2ca51def" volumeName="kubernetes.io/secret/01ab3dd5-8196-46d0-ad33-122e2ca51def-serving-cert" seLinuxMountContext="" Dec 10 22:49:25 crc kubenswrapper[4791]: I1210 22:49:25.850625 4791 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6402fda4-df10-493c-b4e5-d0569419652d" volumeName="kubernetes.io/projected/6402fda4-df10-493c-b4e5-d0569419652d-kube-api-access-mg5zb" seLinuxMountContext="" Dec 10 22:49:25 crc kubenswrapper[4791]: I1210 22:49:25.850640 4791 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49ef4625-1d3a-4a9f-b595-c2433d32326d" volumeName="kubernetes.io/projected/49ef4625-1d3a-4a9f-b595-c2433d32326d-kube-api-access-pjr6v" seLinuxMountContext="" Dec 10 22:49:25 crc kubenswrapper[4791]: I1210 22:49:25.850654 4791 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5fe579f8-e8a6-4643-bce5-a661393c4dde" volumeName="kubernetes.io/projected/5fe579f8-e8a6-4643-bce5-a661393c4dde-kube-api-access-fcqwp" seLinuxMountContext="" Dec 10 22:49:25 crc kubenswrapper[4791]: I1210 22:49:25.850667 4791 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6402fda4-df10-493c-b4e5-d0569419652d" volumeName="kubernetes.io/configmap/6402fda4-df10-493c-b4e5-d0569419652d-config" seLinuxMountContext="" Dec 10 22:49:25 crc kubenswrapper[4791]: I1210 22:49:25.850682 4791 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="22c825df-677d-4ca6-82db-3454ed06e783" volumeName="kubernetes.io/configmap/22c825df-677d-4ca6-82db-3454ed06e783-config" seLinuxMountContext="" Dec 10 22:49:25 crc kubenswrapper[4791]: I1210 22:49:25.850695 4791 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="22c825df-677d-4ca6-82db-3454ed06e783" volumeName="kubernetes.io/projected/22c825df-677d-4ca6-82db-3454ed06e783-kube-api-access-7c4vf" seLinuxMountContext="" Dec 10 22:49:25 crc kubenswrapper[4791]: I1210 22:49:25.850710 4791 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="25e176fe-21b4-4974-b1ed-c8b94f112a7f" volumeName="kubernetes.io/configmap/25e176fe-21b4-4974-b1ed-c8b94f112a7f-signing-cabundle" seLinuxMountContext="" Dec 10 22:49:25 crc kubenswrapper[4791]: I1210 22:49:25.850726 4791 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="308be0ea-9f5f-4b29-aeb1-5abd31a0b17b" volumeName="kubernetes.io/secret/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-apiservice-cert" seLinuxMountContext="" Dec 10 22:49:25 crc kubenswrapper[4791]: I1210 22:49:25.850739 4791 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="43509403-f426-496e-be36-56cef71462f5" volumeName="kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-service-ca" seLinuxMountContext="" Dec 10 22:49:25 crc kubenswrapper[4791]: I1210 22:49:25.850752 4791 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6509e943-70c6-444c-bc41-48a544e36fbd" volumeName="kubernetes.io/configmap/6509e943-70c6-444c-bc41-48a544e36fbd-config" seLinuxMountContext="" Dec 10 22:49:25 crc kubenswrapper[4791]: I1210 22:49:25.850764 4791 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="7583ce53-e0fe-4a16-9e4d-50516596a136" volumeName="kubernetes.io/configmap/7583ce53-e0fe-4a16-9e4d-50516596a136-config" seLinuxMountContext="" Dec 10 22:49:25 crc kubenswrapper[4791]: I1210 22:49:25.850776 4791 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="3ab1a177-2de0-46d9-b765-d0d0649bb42e" volumeName="kubernetes.io/secret/3ab1a177-2de0-46d9-b765-d0d0649bb42e-package-server-manager-serving-cert" seLinuxMountContext="" Dec 10 22:49:25 crc kubenswrapper[4791]: I1210 22:49:25.850788 4791 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" volumeName="kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert" seLinuxMountContext="" Dec 10 22:49:25 crc kubenswrapper[4791]: I1210 22:49:25.850802 4791 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="b6312bbd-5731-4ea0-a20f-81d5a57df44a" volumeName="kubernetes.io/projected/b6312bbd-5731-4ea0-a20f-81d5a57df44a-kube-api-access-249nr" seLinuxMountContext="" Dec 10 22:49:25 crc kubenswrapper[4791]: I1210 22:49:25.850814 4791 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="bf126b07-da06-4140-9a57-dfd54fc6b486" volumeName="kubernetes.io/projected/bf126b07-da06-4140-9a57-dfd54fc6b486-kube-api-access-rnphk" seLinuxMountContext="" Dec 10 22:49:25 crc kubenswrapper[4791]: I1210 22:49:25.850830 4791 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="c03ee662-fb2f-4fc4-a2c1-af487c19d254" volumeName="kubernetes.io/secret/c03ee662-fb2f-4fc4-a2c1-af487c19d254-stats-auth" seLinuxMountContext="" Dec 10 22:49:25 crc kubenswrapper[4791]: I1210 22:49:25.850841 4791 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c" volumeName="kubernetes.io/configmap/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c-config" seLinuxMountContext="" Dec 10 22:49:25 crc kubenswrapper[4791]: I1210 22:49:25.850853 4791 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="22c825df-677d-4ca6-82db-3454ed06e783" volumeName="kubernetes.io/secret/22c825df-677d-4ca6-82db-3454ed06e783-machine-approver-tls" seLinuxMountContext="" Dec 10 22:49:25 crc kubenswrapper[4791]: I1210 22:49:25.850866 4791 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" volumeName="kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-router-certs" seLinuxMountContext="" Dec 10 22:49:25 crc kubenswrapper[4791]: I1210 22:49:25.850880 4791 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="8f668bae-612b-4b75-9490-919e737c6a3b" volumeName="kubernetes.io/configmap/8f668bae-612b-4b75-9490-919e737c6a3b-trusted-ca" seLinuxMountContext="" Dec 10 22:49:25 crc kubenswrapper[4791]: I1210 22:49:25.850893 4791 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="a31745f5-9847-4afe-82a5-3161cc66ca93" volumeName="kubernetes.io/configmap/a31745f5-9847-4afe-82a5-3161cc66ca93-trusted-ca" seLinuxMountContext="" Dec 10 22:49:25 crc kubenswrapper[4791]: I1210 22:49:25.850907 4791 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5225d0e4-402f-4861-b410-819f433b1803" volumeName="kubernetes.io/projected/5225d0e4-402f-4861-b410-819f433b1803-kube-api-access-9xfj7" seLinuxMountContext="" Dec 10 22:49:25 crc kubenswrapper[4791]: I1210 22:49:25.850922 4791 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="7583ce53-e0fe-4a16-9e4d-50516596a136" volumeName="kubernetes.io/secret/7583ce53-e0fe-4a16-9e4d-50516596a136-serving-cert" seLinuxMountContext="" Dec 10 22:49:25 crc kubenswrapper[4791]: I1210 22:49:25.850934 4791 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="7bb08738-c794-4ee8-9972-3a62ca171029" volumeName="kubernetes.io/configmap/7bb08738-c794-4ee8-9972-3a62ca171029-cni-binary-copy" seLinuxMountContext="" Dec 10 22:49:25 crc kubenswrapper[4791]: I1210 22:49:25.850946 4791 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="09efc573-dbb6-4249-bd59-9b87aba8dd28" volumeName="kubernetes.io/secret/09efc573-dbb6-4249-bd59-9b87aba8dd28-etcd-client" seLinuxMountContext="" Dec 10 22:49:25 crc kubenswrapper[4791]: I1210 22:49:25.850960 4791 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1386a44e-36a2-460c-96d0-0359d2b6f0f5" volumeName="kubernetes.io/configmap/1386a44e-36a2-460c-96d0-0359d2b6f0f5-config" seLinuxMountContext="" Dec 10 22:49:25 crc kubenswrapper[4791]: I1210 22:49:25.850973 4791 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1386a44e-36a2-460c-96d0-0359d2b6f0f5" volumeName="kubernetes.io/secret/1386a44e-36a2-460c-96d0-0359d2b6f0f5-serving-cert" seLinuxMountContext="" Dec 10 22:49:25 crc kubenswrapper[4791]: I1210 22:49:25.850987 4791 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="43509403-f426-496e-be36-56cef71462f5" volumeName="kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-oauth-serving-cert" seLinuxMountContext="" Dec 10 22:49:25 crc kubenswrapper[4791]: I1210 22:49:25.850998 4791 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" volumeName="kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-session" seLinuxMountContext="" Dec 10 22:49:25 crc kubenswrapper[4791]: I1210 22:49:25.851010 4791 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="fda69060-fa79-4696-b1a6-7980f124bf7c" volumeName="kubernetes.io/projected/fda69060-fa79-4696-b1a6-7980f124bf7c-kube-api-access-xcgwh" seLinuxMountContext="" Dec 10 22:49:25 crc kubenswrapper[4791]: I1210 22:49:25.851022 4791 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="8cea82b4-6893-4ddc-af9f-1bb5ae425c5b" volumeName="kubernetes.io/projected/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b-kube-api-access-w4xd4" seLinuxMountContext="" Dec 10 22:49:25 crc kubenswrapper[4791]: I1210 22:49:25.851035 4791 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="9d4552c7-cd75-42dd-8880-30dd377c49a4" volumeName="kubernetes.io/configmap/9d4552c7-cd75-42dd-8880-30dd377c49a4-config" seLinuxMountContext="" Dec 10 22:49:25 crc kubenswrapper[4791]: I1210 22:49:25.851047 4791 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="a31745f5-9847-4afe-82a5-3161cc66ca93" volumeName="kubernetes.io/projected/a31745f5-9847-4afe-82a5-3161cc66ca93-kube-api-access-lz9wn" seLinuxMountContext="" Dec 10 22:49:25 crc kubenswrapper[4791]: I1210 22:49:25.851058 4791 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="b11524ee-3fca-4b1b-9cdf-6da289fdbc7d" volumeName="kubernetes.io/empty-dir/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d-utilities" seLinuxMountContext="" Dec 10 22:49:25 crc kubenswrapper[4791]: I1210 22:49:25.851071 4791 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="f88749ec-7931-4ee7-b3fc-1ec5e11f92e9" volumeName="kubernetes.io/secret/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9-profile-collector-cert" seLinuxMountContext="" Dec 10 22:49:25 crc kubenswrapper[4791]: I1210 22:49:25.851083 4791 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="bf126b07-da06-4140-9a57-dfd54fc6b486" volumeName="kubernetes.io/configmap/bf126b07-da06-4140-9a57-dfd54fc6b486-trusted-ca" seLinuxMountContext="" Dec 10 22:49:25 crc kubenswrapper[4791]: I1210 22:49:25.851105 4791 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="09ae3b1a-e8e7-4524-b54b-61eab6f9239a" volumeName="kubernetes.io/secret/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-etcd-client" seLinuxMountContext="" Dec 10 22:49:25 crc kubenswrapper[4791]: I1210 22:49:25.851116 4791 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="43509403-f426-496e-be36-56cef71462f5" volumeName="kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-trusted-ca-bundle" seLinuxMountContext="" Dec 10 22:49:25 crc kubenswrapper[4791]: I1210 22:49:25.851129 4791 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="43509403-f426-496e-be36-56cef71462f5" volumeName="kubernetes.io/secret/43509403-f426-496e-be36-56cef71462f5-console-oauth-config" seLinuxMountContext="" Dec 10 22:49:25 crc kubenswrapper[4791]: I1210 22:49:25.851144 4791 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" volumeName="kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf" seLinuxMountContext="" Dec 10 22:49:25 crc kubenswrapper[4791]: I1210 22:49:25.851157 4791 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="9d4552c7-cd75-42dd-8880-30dd377c49a4" volumeName="kubernetes.io/projected/9d4552c7-cd75-42dd-8880-30dd377c49a4-kube-api-access-pcxfs" seLinuxMountContext="" Dec 10 22:49:25 crc kubenswrapper[4791]: I1210 22:49:25.851169 4791 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6ea678ab-3438-413e-bfe3-290ae7725660" volumeName="kubernetes.io/configmap/6ea678ab-3438-413e-bfe3-290ae7725660-ovnkube-config" seLinuxMountContext="" Dec 10 22:49:25 crc kubenswrapper[4791]: I1210 22:49:25.851181 4791 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="b11524ee-3fca-4b1b-9cdf-6da289fdbc7d" volumeName="kubernetes.io/empty-dir/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d-catalog-content" seLinuxMountContext="" Dec 10 22:49:25 crc kubenswrapper[4791]: I1210 22:49:25.851194 4791 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="09ae3b1a-e8e7-4524-b54b-61eab6f9239a" volumeName="kubernetes.io/secret/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-serving-cert" seLinuxMountContext="" Dec 10 22:49:25 crc kubenswrapper[4791]: I1210 22:49:25.851206 4791 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="09efc573-dbb6-4249-bd59-9b87aba8dd28" volumeName="kubernetes.io/configmap/09efc573-dbb6-4249-bd59-9b87aba8dd28-etcd-service-ca" seLinuxMountContext="" Dec 10 22:49:25 crc kubenswrapper[4791]: I1210 22:49:25.851220 4791 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="0b574797-001e-440a-8f4e-c0be86edad0f" volumeName="kubernetes.io/configmap/0b574797-001e-440a-8f4e-c0be86edad0f-mcc-auth-proxy-config" seLinuxMountContext="" Dec 10 22:49:25 crc kubenswrapper[4791]: I1210 22:49:25.851231 4791 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="25e176fe-21b4-4974-b1ed-c8b94f112a7f" volumeName="kubernetes.io/secret/25e176fe-21b4-4974-b1ed-c8b94f112a7f-signing-key" seLinuxMountContext="" Dec 10 22:49:25 crc kubenswrapper[4791]: I1210 22:49:25.851244 4791 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="43509403-f426-496e-be36-56cef71462f5" volumeName="kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-console-config" seLinuxMountContext="" Dec 10 22:49:25 crc kubenswrapper[4791]: I1210 22:49:25.851256 4791 reconstruct.go:97] "Volume reconstruction finished" Dec 10 22:49:25 crc kubenswrapper[4791]: I1210 22:49:25.851268 4791 reconciler.go:26] "Reconciler: start to sync state" Dec 10 22:49:25 crc kubenswrapper[4791]: I1210 22:49:25.863020 4791 manager.go:324] Recovery completed Dec 10 22:49:25 crc kubenswrapper[4791]: I1210 22:49:25.875271 4791 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 10 22:49:25 crc kubenswrapper[4791]: I1210 22:49:25.879781 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 22:49:25 crc kubenswrapper[4791]: I1210 22:49:25.879821 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 22:49:25 crc kubenswrapper[4791]: I1210 22:49:25.879830 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 22:49:25 crc kubenswrapper[4791]: I1210 22:49:25.880358 4791 cpu_manager.go:225] "Starting CPU manager" policy="none" Dec 10 22:49:25 crc kubenswrapper[4791]: I1210 22:49:25.880375 4791 cpu_manager.go:226] "Reconciling" reconcilePeriod="10s" Dec 10 22:49:25 crc kubenswrapper[4791]: I1210 22:49:25.880394 4791 state_mem.go:36] "Initialized new in-memory state store" Dec 10 22:49:25 crc kubenswrapper[4791]: I1210 22:49:25.880744 4791 kubelet_network_linux.go:50] "Initialized iptables rules." protocol="IPv4" Dec 10 22:49:25 crc kubenswrapper[4791]: I1210 22:49:25.883422 4791 kubelet_network_linux.go:50] "Initialized iptables rules." protocol="IPv6" Dec 10 22:49:25 crc kubenswrapper[4791]: I1210 22:49:25.883477 4791 status_manager.go:217] "Starting to sync pod status with apiserver" Dec 10 22:49:25 crc kubenswrapper[4791]: I1210 22:49:25.883506 4791 kubelet.go:2335] "Starting kubelet main sync loop" Dec 10 22:49:25 crc kubenswrapper[4791]: E1210 22:49:25.883579 4791 kubelet.go:2359] "Skipping pod synchronization" err="[container runtime status check may not have completed yet, PLEG is not healthy: pleg has yet to be successful]" Dec 10 22:49:25 crc kubenswrapper[4791]: W1210 22:49:25.886978 4791 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.RuntimeClass: Get "https://api-int.crc.testing:6443/apis/node.k8s.io/v1/runtimeclasses?limit=500&resourceVersion=0": dial tcp 38.102.83.30:6443: connect: connection refused Dec 10 22:49:25 crc kubenswrapper[4791]: E1210 22:49:25.887061 4791 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.RuntimeClass: failed to list *v1.RuntimeClass: Get \"https://api-int.crc.testing:6443/apis/node.k8s.io/v1/runtimeclasses?limit=500&resourceVersion=0\": dial tcp 38.102.83.30:6443: connect: connection refused" logger="UnhandledError" Dec 10 22:49:25 crc kubenswrapper[4791]: I1210 22:49:25.888508 4791 policy_none.go:49] "None policy: Start" Dec 10 22:49:25 crc kubenswrapper[4791]: I1210 22:49:25.889191 4791 memory_manager.go:170] "Starting memorymanager" policy="None" Dec 10 22:49:25 crc kubenswrapper[4791]: I1210 22:49:25.889221 4791 state_mem.go:35] "Initializing new in-memory state store" Dec 10 22:49:25 crc kubenswrapper[4791]: E1210 22:49:25.926780 4791 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Dec 10 22:49:25 crc kubenswrapper[4791]: I1210 22:49:25.941798 4791 manager.go:334] "Starting Device Plugin manager" Dec 10 22:49:25 crc kubenswrapper[4791]: I1210 22:49:25.941879 4791 manager.go:513] "Failed to read data from checkpoint" checkpoint="kubelet_internal_checkpoint" err="checkpoint is not found" Dec 10 22:49:25 crc kubenswrapper[4791]: I1210 22:49:25.941896 4791 server.go:79] "Starting device plugin registration server" Dec 10 22:49:25 crc kubenswrapper[4791]: I1210 22:49:25.942562 4791 eviction_manager.go:189] "Eviction manager: starting control loop" Dec 10 22:49:25 crc kubenswrapper[4791]: I1210 22:49:25.942599 4791 container_log_manager.go:189] "Initializing container log rotate workers" workers=1 monitorPeriod="10s" Dec 10 22:49:25 crc kubenswrapper[4791]: I1210 22:49:25.942893 4791 plugin_watcher.go:51] "Plugin Watcher Start" path="/var/lib/kubelet/plugins_registry" Dec 10 22:49:25 crc kubenswrapper[4791]: I1210 22:49:25.942991 4791 plugin_manager.go:116] "The desired_state_of_world populator (plugin watcher) starts" Dec 10 22:49:25 crc kubenswrapper[4791]: I1210 22:49:25.943000 4791 plugin_manager.go:118] "Starting Kubelet Plugin Manager" Dec 10 22:49:25 crc kubenswrapper[4791]: E1210 22:49:25.949237 4791 eviction_manager.go:285] "Eviction manager: failed to get summary stats" err="failed to get node info: node \"crc\" not found" Dec 10 22:49:25 crc kubenswrapper[4791]: I1210 22:49:25.984256 4791 kubelet.go:2421] "SyncLoop ADD" source="file" pods=["openshift-kube-apiserver/kube-apiserver-crc","openshift-kube-controller-manager/kube-controller-manager-crc","openshift-kube-scheduler/openshift-kube-scheduler-crc","openshift-machine-config-operator/kube-rbac-proxy-crio-crc","openshift-etcd/etcd-crc"] Dec 10 22:49:25 crc kubenswrapper[4791]: I1210 22:49:25.984461 4791 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 10 22:49:25 crc kubenswrapper[4791]: I1210 22:49:25.985704 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 22:49:25 crc kubenswrapper[4791]: I1210 22:49:25.985799 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 22:49:25 crc kubenswrapper[4791]: I1210 22:49:25.985847 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 22:49:25 crc kubenswrapper[4791]: I1210 22:49:25.986097 4791 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 10 22:49:25 crc kubenswrapper[4791]: I1210 22:49:25.986446 4791 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/kube-apiserver-crc" Dec 10 22:49:25 crc kubenswrapper[4791]: I1210 22:49:25.986487 4791 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 10 22:49:25 crc kubenswrapper[4791]: I1210 22:49:25.987504 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 22:49:25 crc kubenswrapper[4791]: I1210 22:49:25.987556 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 22:49:25 crc kubenswrapper[4791]: I1210 22:49:25.987578 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 22:49:25 crc kubenswrapper[4791]: I1210 22:49:25.988216 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 22:49:25 crc kubenswrapper[4791]: I1210 22:49:25.988249 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 22:49:25 crc kubenswrapper[4791]: I1210 22:49:25.988261 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 22:49:25 crc kubenswrapper[4791]: I1210 22:49:25.988364 4791 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 10 22:49:25 crc kubenswrapper[4791]: I1210 22:49:25.988581 4791 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Dec 10 22:49:25 crc kubenswrapper[4791]: I1210 22:49:25.988648 4791 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 10 22:49:25 crc kubenswrapper[4791]: I1210 22:49:25.989142 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 22:49:25 crc kubenswrapper[4791]: I1210 22:49:25.989191 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 22:49:25 crc kubenswrapper[4791]: I1210 22:49:25.989222 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 22:49:25 crc kubenswrapper[4791]: I1210 22:49:25.989297 4791 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 10 22:49:25 crc kubenswrapper[4791]: I1210 22:49:25.989506 4791 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" Dec 10 22:49:25 crc kubenswrapper[4791]: I1210 22:49:25.989596 4791 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 10 22:49:25 crc kubenswrapper[4791]: I1210 22:49:25.989800 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 22:49:25 crc kubenswrapper[4791]: I1210 22:49:25.989861 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 22:49:25 crc kubenswrapper[4791]: I1210 22:49:25.989882 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 22:49:25 crc kubenswrapper[4791]: I1210 22:49:25.990224 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 22:49:25 crc kubenswrapper[4791]: I1210 22:49:25.990249 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 22:49:25 crc kubenswrapper[4791]: I1210 22:49:25.990259 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 22:49:25 crc kubenswrapper[4791]: I1210 22:49:25.990366 4791 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 10 22:49:25 crc kubenswrapper[4791]: I1210 22:49:25.990691 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 22:49:25 crc kubenswrapper[4791]: I1210 22:49:25.990717 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 22:49:25 crc kubenswrapper[4791]: I1210 22:49:25.990731 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 22:49:25 crc kubenswrapper[4791]: I1210 22:49:25.990821 4791 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" Dec 10 22:49:25 crc kubenswrapper[4791]: I1210 22:49:25.990875 4791 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 10 22:49:25 crc kubenswrapper[4791]: I1210 22:49:25.991113 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 22:49:25 crc kubenswrapper[4791]: I1210 22:49:25.991140 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 22:49:25 crc kubenswrapper[4791]: I1210 22:49:25.991148 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 22:49:25 crc kubenswrapper[4791]: I1210 22:49:25.991385 4791 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-etcd/etcd-crc" Dec 10 22:49:25 crc kubenswrapper[4791]: I1210 22:49:25.991409 4791 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 10 22:49:25 crc kubenswrapper[4791]: I1210 22:49:25.992012 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 22:49:25 crc kubenswrapper[4791]: I1210 22:49:25.992058 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 22:49:25 crc kubenswrapper[4791]: I1210 22:49:25.992072 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 22:49:25 crc kubenswrapper[4791]: I1210 22:49:25.992060 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 22:49:25 crc kubenswrapper[4791]: I1210 22:49:25.992122 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 22:49:25 crc kubenswrapper[4791]: I1210 22:49:25.992150 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 22:49:26 crc kubenswrapper[4791]: E1210 22:49:26.027483 4791 controller.go:145] "Failed to ensure lease exists, will retry" err="Get \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": dial tcp 38.102.83.30:6443: connect: connection refused" interval="400ms" Dec 10 22:49:26 crc kubenswrapper[4791]: I1210 22:49:26.042808 4791 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 10 22:49:26 crc kubenswrapper[4791]: I1210 22:49:26.044407 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 22:49:26 crc kubenswrapper[4791]: I1210 22:49:26.044448 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 22:49:26 crc kubenswrapper[4791]: I1210 22:49:26.044461 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 22:49:26 crc kubenswrapper[4791]: I1210 22:49:26.044490 4791 kubelet_node_status.go:76] "Attempting to register node" node="crc" Dec 10 22:49:26 crc kubenswrapper[4791]: E1210 22:49:26.045062 4791 kubelet_node_status.go:99] "Unable to register node with API server" err="Post \"https://api-int.crc.testing:6443/api/v1/nodes\": dial tcp 38.102.83.30:6443: connect: connection refused" node="crc" Dec 10 22:49:26 crc kubenswrapper[4791]: I1210 22:49:26.055497 4791 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-cert-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Dec 10 22:49:26 crc kubenswrapper[4791]: I1210 22:49:26.055584 4791 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/f614b9022728cf315e60c057852e563e-resource-dir\") pod \"kube-controller-manager-crc\" (UID: \"f614b9022728cf315e60c057852e563e\") " pod="openshift-kube-controller-manager/kube-controller-manager-crc" Dec 10 22:49:26 crc kubenswrapper[4791]: I1210 22:49:26.055635 4791 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/3dcd261975c3d6b9a6ad6367fd4facd3-cert-dir\") pod \"openshift-kube-scheduler-crc\" (UID: \"3dcd261975c3d6b9a6ad6367fd4facd3\") " pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" Dec 10 22:49:26 crc kubenswrapper[4791]: I1210 22:49:26.055672 4791 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-kube\" (UniqueName: \"kubernetes.io/host-path/d1b160f5dda77d281dd8e69ec8d817f9-etc-kube\") pod \"kube-rbac-proxy-crio-crc\" (UID: \"d1b160f5dda77d281dd8e69ec8d817f9\") " pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" Dec 10 22:49:26 crc kubenswrapper[4791]: I1210 22:49:26.055704 4791 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-resource-dir\") pod \"kube-apiserver-crc\" (UID: \"f4b27818a5e8e43d0dc095d08835c792\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Dec 10 22:49:26 crc kubenswrapper[4791]: I1210 22:49:26.055742 4791 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/f614b9022728cf315e60c057852e563e-cert-dir\") pod \"kube-controller-manager-crc\" (UID: \"f614b9022728cf315e60c057852e563e\") " pod="openshift-kube-controller-manager/kube-controller-manager-crc" Dec 10 22:49:26 crc kubenswrapper[4791]: I1210 22:49:26.055775 4791 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-lib-kubelet\" (UniqueName: \"kubernetes.io/host-path/d1b160f5dda77d281dd8e69ec8d817f9-var-lib-kubelet\") pod \"kube-rbac-proxy-crio-crc\" (UID: \"d1b160f5dda77d281dd8e69ec8d817f9\") " pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" Dec 10 22:49:26 crc kubenswrapper[4791]: I1210 22:49:26.055808 4791 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/3dcd261975c3d6b9a6ad6367fd4facd3-resource-dir\") pod \"openshift-kube-scheduler-crc\" (UID: \"3dcd261975c3d6b9a6ad6367fd4facd3\") " pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" Dec 10 22:49:26 crc kubenswrapper[4791]: I1210 22:49:26.055856 4791 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"data-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-data-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Dec 10 22:49:26 crc kubenswrapper[4791]: I1210 22:49:26.055901 4791 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"usr-local-bin\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-usr-local-bin\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Dec 10 22:49:26 crc kubenswrapper[4791]: I1210 22:49:26.055916 4791 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"log-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-log-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Dec 10 22:49:26 crc kubenswrapper[4791]: I1210 22:49:26.055934 4791 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-audit-dir\") pod \"kube-apiserver-crc\" (UID: \"f4b27818a5e8e43d0dc095d08835c792\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Dec 10 22:49:26 crc kubenswrapper[4791]: I1210 22:49:26.055964 4791 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"static-pod-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-static-pod-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Dec 10 22:49:26 crc kubenswrapper[4791]: I1210 22:49:26.056041 4791 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-resource-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Dec 10 22:49:26 crc kubenswrapper[4791]: I1210 22:49:26.056067 4791 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-cert-dir\") pod \"kube-apiserver-crc\" (UID: \"f4b27818a5e8e43d0dc095d08835c792\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Dec 10 22:49:26 crc kubenswrapper[4791]: I1210 22:49:26.157639 4791 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-cert-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Dec 10 22:49:26 crc kubenswrapper[4791]: I1210 22:49:26.158169 4791 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/f614b9022728cf315e60c057852e563e-resource-dir\") pod \"kube-controller-manager-crc\" (UID: \"f614b9022728cf315e60c057852e563e\") " pod="openshift-kube-controller-manager/kube-controller-manager-crc" Dec 10 22:49:26 crc kubenswrapper[4791]: I1210 22:49:26.158367 4791 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/3dcd261975c3d6b9a6ad6367fd4facd3-cert-dir\") pod \"openshift-kube-scheduler-crc\" (UID: \"3dcd261975c3d6b9a6ad6367fd4facd3\") " pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" Dec 10 22:49:26 crc kubenswrapper[4791]: I1210 22:49:26.158525 4791 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-kube\" (UniqueName: \"kubernetes.io/host-path/d1b160f5dda77d281dd8e69ec8d817f9-etc-kube\") pod \"kube-rbac-proxy-crio-crc\" (UID: \"d1b160f5dda77d281dd8e69ec8d817f9\") " pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" Dec 10 22:49:26 crc kubenswrapper[4791]: I1210 22:49:26.158663 4791 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-resource-dir\") pod \"kube-apiserver-crc\" (UID: \"f4b27818a5e8e43d0dc095d08835c792\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Dec 10 22:49:26 crc kubenswrapper[4791]: I1210 22:49:26.158867 4791 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-resource-dir\") pod \"kube-apiserver-crc\" (UID: \"f4b27818a5e8e43d0dc095d08835c792\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Dec 10 22:49:26 crc kubenswrapper[4791]: I1210 22:49:26.158528 4791 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/3dcd261975c3d6b9a6ad6367fd4facd3-cert-dir\") pod \"openshift-kube-scheduler-crc\" (UID: \"3dcd261975c3d6b9a6ad6367fd4facd3\") " pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" Dec 10 22:49:26 crc kubenswrapper[4791]: I1210 22:49:26.158567 4791 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-kube\" (UniqueName: \"kubernetes.io/host-path/d1b160f5dda77d281dd8e69ec8d817f9-etc-kube\") pod \"kube-rbac-proxy-crio-crc\" (UID: \"d1b160f5dda77d281dd8e69ec8d817f9\") " pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" Dec 10 22:49:26 crc kubenswrapper[4791]: I1210 22:49:26.157998 4791 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-cert-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Dec 10 22:49:26 crc kubenswrapper[4791]: I1210 22:49:26.158880 4791 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/f614b9022728cf315e60c057852e563e-cert-dir\") pod \"kube-controller-manager-crc\" (UID: \"f614b9022728cf315e60c057852e563e\") " pod="openshift-kube-controller-manager/kube-controller-manager-crc" Dec 10 22:49:26 crc kubenswrapper[4791]: I1210 22:49:26.159074 4791 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-lib-kubelet\" (UniqueName: \"kubernetes.io/host-path/d1b160f5dda77d281dd8e69ec8d817f9-var-lib-kubelet\") pod \"kube-rbac-proxy-crio-crc\" (UID: \"d1b160f5dda77d281dd8e69ec8d817f9\") " pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" Dec 10 22:49:26 crc kubenswrapper[4791]: I1210 22:49:26.158267 4791 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/f614b9022728cf315e60c057852e563e-resource-dir\") pod \"kube-controller-manager-crc\" (UID: \"f614b9022728cf315e60c057852e563e\") " pod="openshift-kube-controller-manager/kube-controller-manager-crc" Dec 10 22:49:26 crc kubenswrapper[4791]: I1210 22:49:26.159128 4791 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/3dcd261975c3d6b9a6ad6367fd4facd3-resource-dir\") pod \"openshift-kube-scheduler-crc\" (UID: \"3dcd261975c3d6b9a6ad6367fd4facd3\") " pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" Dec 10 22:49:26 crc kubenswrapper[4791]: I1210 22:49:26.158928 4791 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/f614b9022728cf315e60c057852e563e-cert-dir\") pod \"kube-controller-manager-crc\" (UID: \"f614b9022728cf315e60c057852e563e\") " pod="openshift-kube-controller-manager/kube-controller-manager-crc" Dec 10 22:49:26 crc kubenswrapper[4791]: I1210 22:49:26.159100 4791 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/3dcd261975c3d6b9a6ad6367fd4facd3-resource-dir\") pod \"openshift-kube-scheduler-crc\" (UID: \"3dcd261975c3d6b9a6ad6367fd4facd3\") " pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" Dec 10 22:49:26 crc kubenswrapper[4791]: I1210 22:49:26.159142 4791 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-lib-kubelet\" (UniqueName: \"kubernetes.io/host-path/d1b160f5dda77d281dd8e69ec8d817f9-var-lib-kubelet\") pod \"kube-rbac-proxy-crio-crc\" (UID: \"d1b160f5dda77d281dd8e69ec8d817f9\") " pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" Dec 10 22:49:26 crc kubenswrapper[4791]: I1210 22:49:26.159208 4791 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"data-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-data-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Dec 10 22:49:26 crc kubenswrapper[4791]: I1210 22:49:26.159179 4791 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"data-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-data-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Dec 10 22:49:26 crc kubenswrapper[4791]: I1210 22:49:26.159288 4791 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"usr-local-bin\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-usr-local-bin\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Dec 10 22:49:26 crc kubenswrapper[4791]: I1210 22:49:26.159331 4791 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"log-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-log-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Dec 10 22:49:26 crc kubenswrapper[4791]: I1210 22:49:26.159378 4791 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-audit-dir\") pod \"kube-apiserver-crc\" (UID: \"f4b27818a5e8e43d0dc095d08835c792\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Dec 10 22:49:26 crc kubenswrapper[4791]: I1210 22:49:26.159402 4791 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"static-pod-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-static-pod-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Dec 10 22:49:26 crc kubenswrapper[4791]: I1210 22:49:26.159416 4791 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"usr-local-bin\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-usr-local-bin\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Dec 10 22:49:26 crc kubenswrapper[4791]: I1210 22:49:26.159441 4791 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-resource-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Dec 10 22:49:26 crc kubenswrapper[4791]: I1210 22:49:26.159504 4791 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-audit-dir\") pod \"kube-apiserver-crc\" (UID: \"f4b27818a5e8e43d0dc095d08835c792\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Dec 10 22:49:26 crc kubenswrapper[4791]: I1210 22:49:26.159495 4791 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-cert-dir\") pod \"kube-apiserver-crc\" (UID: \"f4b27818a5e8e43d0dc095d08835c792\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Dec 10 22:49:26 crc kubenswrapper[4791]: I1210 22:49:26.159543 4791 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-cert-dir\") pod \"kube-apiserver-crc\" (UID: \"f4b27818a5e8e43d0dc095d08835c792\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Dec 10 22:49:26 crc kubenswrapper[4791]: I1210 22:49:26.159590 4791 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"static-pod-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-static-pod-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Dec 10 22:49:26 crc kubenswrapper[4791]: I1210 22:49:26.159604 4791 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-resource-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Dec 10 22:49:26 crc kubenswrapper[4791]: I1210 22:49:26.159577 4791 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"log-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-log-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Dec 10 22:49:26 crc kubenswrapper[4791]: I1210 22:49:26.246180 4791 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 10 22:49:26 crc kubenswrapper[4791]: I1210 22:49:26.247866 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 22:49:26 crc kubenswrapper[4791]: I1210 22:49:26.247925 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 22:49:26 crc kubenswrapper[4791]: I1210 22:49:26.247971 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 22:49:26 crc kubenswrapper[4791]: I1210 22:49:26.248016 4791 kubelet_node_status.go:76] "Attempting to register node" node="crc" Dec 10 22:49:26 crc kubenswrapper[4791]: E1210 22:49:26.248779 4791 kubelet_node_status.go:99] "Unable to register node with API server" err="Post \"https://api-int.crc.testing:6443/api/v1/nodes\": dial tcp 38.102.83.30:6443: connect: connection refused" node="crc" Dec 10 22:49:26 crc kubenswrapper[4791]: I1210 22:49:26.325602 4791 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/kube-apiserver-crc" Dec 10 22:49:26 crc kubenswrapper[4791]: I1210 22:49:26.340668 4791 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Dec 10 22:49:26 crc kubenswrapper[4791]: I1210 22:49:26.363271 4791 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" Dec 10 22:49:26 crc kubenswrapper[4791]: W1210 22:49:26.368715 4791 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podf4b27818a5e8e43d0dc095d08835c792.slice/crio-932145d8a3a61f561379d353180c8cfa0b1b7d158d90c0318fe6addb06438a9d WatchSource:0}: Error finding container 932145d8a3a61f561379d353180c8cfa0b1b7d158d90c0318fe6addb06438a9d: Status 404 returned error can't find the container with id 932145d8a3a61f561379d353180c8cfa0b1b7d158d90c0318fe6addb06438a9d Dec 10 22:49:26 crc kubenswrapper[4791]: W1210 22:49:26.372433 4791 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podf614b9022728cf315e60c057852e563e.slice/crio-1d8ef5cb073e4b4a2597d166721d4c955335dd230724b2c7ee66348f1af03995 WatchSource:0}: Error finding container 1d8ef5cb073e4b4a2597d166721d4c955335dd230724b2c7ee66348f1af03995: Status 404 returned error can't find the container with id 1d8ef5cb073e4b4a2597d166721d4c955335dd230724b2c7ee66348f1af03995 Dec 10 22:49:26 crc kubenswrapper[4791]: I1210 22:49:26.373507 4791 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-etcd/etcd-crc" Dec 10 22:49:26 crc kubenswrapper[4791]: I1210 22:49:26.379529 4791 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" Dec 10 22:49:26 crc kubenswrapper[4791]: W1210 22:49:26.382937 4791 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod3dcd261975c3d6b9a6ad6367fd4facd3.slice/crio-ea9c24a0dc00ee0618818b27298b47db081acf54e93f3157c14fb396e1375eb2 WatchSource:0}: Error finding container ea9c24a0dc00ee0618818b27298b47db081acf54e93f3157c14fb396e1375eb2: Status 404 returned error can't find the container with id ea9c24a0dc00ee0618818b27298b47db081acf54e93f3157c14fb396e1375eb2 Dec 10 22:49:26 crc kubenswrapper[4791]: W1210 22:49:26.395317 4791 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod2139d3e2895fc6797b9c76a1b4c9886d.slice/crio-98e2b71dc016f9979130bb00b7f2501de1e234456500a807e2bfdf34279dd9eb WatchSource:0}: Error finding container 98e2b71dc016f9979130bb00b7f2501de1e234456500a807e2bfdf34279dd9eb: Status 404 returned error can't find the container with id 98e2b71dc016f9979130bb00b7f2501de1e234456500a807e2bfdf34279dd9eb Dec 10 22:49:26 crc kubenswrapper[4791]: W1210 22:49:26.404644 4791 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podd1b160f5dda77d281dd8e69ec8d817f9.slice/crio-03c453a0254fae7937d338db20b3f8d0e47aa30fd1947c9bbe10e140e2faa5f5 WatchSource:0}: Error finding container 03c453a0254fae7937d338db20b3f8d0e47aa30fd1947c9bbe10e140e2faa5f5: Status 404 returned error can't find the container with id 03c453a0254fae7937d338db20b3f8d0e47aa30fd1947c9bbe10e140e2faa5f5 Dec 10 22:49:26 crc kubenswrapper[4791]: E1210 22:49:26.428835 4791 controller.go:145] "Failed to ensure lease exists, will retry" err="Get \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": dial tcp 38.102.83.30:6443: connect: connection refused" interval="800ms" Dec 10 22:49:26 crc kubenswrapper[4791]: W1210 22:49:26.643925 4791 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.Node: Get "https://api-int.crc.testing:6443/api/v1/nodes?fieldSelector=metadata.name%3Dcrc&limit=500&resourceVersion=0": dial tcp 38.102.83.30:6443: connect: connection refused Dec 10 22:49:26 crc kubenswrapper[4791]: E1210 22:49:26.644063 4791 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.Node: failed to list *v1.Node: Get \"https://api-int.crc.testing:6443/api/v1/nodes?fieldSelector=metadata.name%3Dcrc&limit=500&resourceVersion=0\": dial tcp 38.102.83.30:6443: connect: connection refused" logger="UnhandledError" Dec 10 22:49:26 crc kubenswrapper[4791]: I1210 22:49:26.649486 4791 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 10 22:49:26 crc kubenswrapper[4791]: I1210 22:49:26.651107 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 22:49:26 crc kubenswrapper[4791]: I1210 22:49:26.651158 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 22:49:26 crc kubenswrapper[4791]: I1210 22:49:26.651173 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 22:49:26 crc kubenswrapper[4791]: I1210 22:49:26.651217 4791 kubelet_node_status.go:76] "Attempting to register node" node="crc" Dec 10 22:49:26 crc kubenswrapper[4791]: E1210 22:49:26.651908 4791 kubelet_node_status.go:99] "Unable to register node with API server" err="Post \"https://api-int.crc.testing:6443/api/v1/nodes\": dial tcp 38.102.83.30:6443: connect: connection refused" node="crc" Dec 10 22:49:26 crc kubenswrapper[4791]: I1210 22:49:26.825922 4791 csi_plugin.go:884] Failed to contact API server when waiting for CSINode publishing: Get "https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csinodes/crc?resourceVersion=0": dial tcp 38.102.83.30:6443: connect: connection refused Dec 10 22:49:26 crc kubenswrapper[4791]: I1210 22:49:26.827031 4791 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2025-12-09 08:53:57.069787714 +0000 UTC Dec 10 22:49:26 crc kubenswrapper[4791]: I1210 22:49:26.888084 4791 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" event={"ID":"d1b160f5dda77d281dd8e69ec8d817f9","Type":"ContainerStarted","Data":"03c453a0254fae7937d338db20b3f8d0e47aa30fd1947c9bbe10e140e2faa5f5"} Dec 10 22:49:26 crc kubenswrapper[4791]: I1210 22:49:26.889150 4791 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-etcd/etcd-crc" event={"ID":"2139d3e2895fc6797b9c76a1b4c9886d","Type":"ContainerStarted","Data":"98e2b71dc016f9979130bb00b7f2501de1e234456500a807e2bfdf34279dd9eb"} Dec 10 22:49:26 crc kubenswrapper[4791]: I1210 22:49:26.890489 4791 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" event={"ID":"3dcd261975c3d6b9a6ad6367fd4facd3","Type":"ContainerStarted","Data":"ea9c24a0dc00ee0618818b27298b47db081acf54e93f3157c14fb396e1375eb2"} Dec 10 22:49:26 crc kubenswrapper[4791]: I1210 22:49:26.891535 4791 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" event={"ID":"f614b9022728cf315e60c057852e563e","Type":"ContainerStarted","Data":"1d8ef5cb073e4b4a2597d166721d4c955335dd230724b2c7ee66348f1af03995"} Dec 10 22:49:26 crc kubenswrapper[4791]: I1210 22:49:26.892695 4791 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"f4b27818a5e8e43d0dc095d08835c792","Type":"ContainerStarted","Data":"932145d8a3a61f561379d353180c8cfa0b1b7d158d90c0318fe6addb06438a9d"} Dec 10 22:49:27 crc kubenswrapper[4791]: W1210 22:49:27.197266 4791 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.CSIDriver: Get "https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csidrivers?limit=500&resourceVersion=0": dial tcp 38.102.83.30:6443: connect: connection refused Dec 10 22:49:27 crc kubenswrapper[4791]: E1210 22:49:27.197372 4791 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.CSIDriver: failed to list *v1.CSIDriver: Get \"https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csidrivers?limit=500&resourceVersion=0\": dial tcp 38.102.83.30:6443: connect: connection refused" logger="UnhandledError" Dec 10 22:49:27 crc kubenswrapper[4791]: E1210 22:49:27.230516 4791 controller.go:145] "Failed to ensure lease exists, will retry" err="Get \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": dial tcp 38.102.83.30:6443: connect: connection refused" interval="1.6s" Dec 10 22:49:27 crc kubenswrapper[4791]: W1210 22:49:27.256975 4791 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.Service: Get "https://api-int.crc.testing:6443/api/v1/services?fieldSelector=spec.clusterIP%21%3DNone&limit=500&resourceVersion=0": dial tcp 38.102.83.30:6443: connect: connection refused Dec 10 22:49:27 crc kubenswrapper[4791]: E1210 22:49:27.257132 4791 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.Service: failed to list *v1.Service: Get \"https://api-int.crc.testing:6443/api/v1/services?fieldSelector=spec.clusterIP%21%3DNone&limit=500&resourceVersion=0\": dial tcp 38.102.83.30:6443: connect: connection refused" logger="UnhandledError" Dec 10 22:49:27 crc kubenswrapper[4791]: W1210 22:49:27.317381 4791 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.RuntimeClass: Get "https://api-int.crc.testing:6443/apis/node.k8s.io/v1/runtimeclasses?limit=500&resourceVersion=0": dial tcp 38.102.83.30:6443: connect: connection refused Dec 10 22:49:27 crc kubenswrapper[4791]: E1210 22:49:27.317503 4791 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.RuntimeClass: failed to list *v1.RuntimeClass: Get \"https://api-int.crc.testing:6443/apis/node.k8s.io/v1/runtimeclasses?limit=500&resourceVersion=0\": dial tcp 38.102.83.30:6443: connect: connection refused" logger="UnhandledError" Dec 10 22:49:27 crc kubenswrapper[4791]: I1210 22:49:27.452703 4791 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 10 22:49:27 crc kubenswrapper[4791]: I1210 22:49:27.455031 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 22:49:27 crc kubenswrapper[4791]: I1210 22:49:27.455086 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 22:49:27 crc kubenswrapper[4791]: I1210 22:49:27.455099 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 22:49:27 crc kubenswrapper[4791]: I1210 22:49:27.455128 4791 kubelet_node_status.go:76] "Attempting to register node" node="crc" Dec 10 22:49:27 crc kubenswrapper[4791]: E1210 22:49:27.455875 4791 kubelet_node_status.go:99] "Unable to register node with API server" err="Post \"https://api-int.crc.testing:6443/api/v1/nodes\": dial tcp 38.102.83.30:6443: connect: connection refused" node="crc" Dec 10 22:49:27 crc kubenswrapper[4791]: I1210 22:49:27.826115 4791 csi_plugin.go:884] Failed to contact API server when waiting for CSINode publishing: Get "https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csinodes/crc?resourceVersion=0": dial tcp 38.102.83.30:6443: connect: connection refused Dec 10 22:49:27 crc kubenswrapper[4791]: I1210 22:49:27.827137 4791 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2025-12-04 00:15:22.963293715 +0000 UTC Dec 10 22:49:27 crc kubenswrapper[4791]: I1210 22:49:27.897744 4791 generic.go:334] "Generic (PLEG): container finished" podID="f4b27818a5e8e43d0dc095d08835c792" containerID="c5856d6b26c19fda13b5c9ac2892a7e2f9c6f79feb8173ecfd9b33b673327d67" exitCode=0 Dec 10 22:49:27 crc kubenswrapper[4791]: I1210 22:49:27.897817 4791 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"f4b27818a5e8e43d0dc095d08835c792","Type":"ContainerDied","Data":"c5856d6b26c19fda13b5c9ac2892a7e2f9c6f79feb8173ecfd9b33b673327d67"} Dec 10 22:49:27 crc kubenswrapper[4791]: I1210 22:49:27.897878 4791 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 10 22:49:27 crc kubenswrapper[4791]: I1210 22:49:27.899189 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 22:49:27 crc kubenswrapper[4791]: I1210 22:49:27.899219 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 22:49:27 crc kubenswrapper[4791]: I1210 22:49:27.899229 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 22:49:27 crc kubenswrapper[4791]: I1210 22:49:27.900356 4791 generic.go:334] "Generic (PLEG): container finished" podID="d1b160f5dda77d281dd8e69ec8d817f9" containerID="8c489105d0dfb302732d4f5bd2fc9d000967a071802735b778d49bdbc428b264" exitCode=0 Dec 10 22:49:27 crc kubenswrapper[4791]: I1210 22:49:27.900428 4791 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" event={"ID":"d1b160f5dda77d281dd8e69ec8d817f9","Type":"ContainerDied","Data":"8c489105d0dfb302732d4f5bd2fc9d000967a071802735b778d49bdbc428b264"} Dec 10 22:49:27 crc kubenswrapper[4791]: I1210 22:49:27.900650 4791 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 10 22:49:27 crc kubenswrapper[4791]: I1210 22:49:27.901191 4791 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 10 22:49:27 crc kubenswrapper[4791]: I1210 22:49:27.902382 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 22:49:27 crc kubenswrapper[4791]: I1210 22:49:27.902422 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 22:49:27 crc kubenswrapper[4791]: I1210 22:49:27.902437 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 22:49:27 crc kubenswrapper[4791]: I1210 22:49:27.902477 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 22:49:27 crc kubenswrapper[4791]: I1210 22:49:27.902522 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 22:49:27 crc kubenswrapper[4791]: I1210 22:49:27.902556 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 22:49:27 crc kubenswrapper[4791]: I1210 22:49:27.902567 4791 generic.go:334] "Generic (PLEG): container finished" podID="2139d3e2895fc6797b9c76a1b4c9886d" containerID="0f65c864d885319ce2a08d61c30be24b3ffb7d23ea30c82ed7faf039cb2c006d" exitCode=0 Dec 10 22:49:27 crc kubenswrapper[4791]: I1210 22:49:27.902621 4791 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-etcd/etcd-crc" event={"ID":"2139d3e2895fc6797b9c76a1b4c9886d","Type":"ContainerDied","Data":"0f65c864d885319ce2a08d61c30be24b3ffb7d23ea30c82ed7faf039cb2c006d"} Dec 10 22:49:27 crc kubenswrapper[4791]: I1210 22:49:27.902760 4791 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 10 22:49:27 crc kubenswrapper[4791]: I1210 22:49:27.904508 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 22:49:27 crc kubenswrapper[4791]: I1210 22:49:27.904568 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 22:49:27 crc kubenswrapper[4791]: I1210 22:49:27.904597 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 22:49:27 crc kubenswrapper[4791]: I1210 22:49:27.905068 4791 generic.go:334] "Generic (PLEG): container finished" podID="3dcd261975c3d6b9a6ad6367fd4facd3" containerID="9b7a5502803eaad9f76f5ffa91b29fbb2657402065a03a4d6433c096aac7264f" exitCode=0 Dec 10 22:49:27 crc kubenswrapper[4791]: I1210 22:49:27.905832 4791 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 10 22:49:27 crc kubenswrapper[4791]: I1210 22:49:27.906008 4791 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" event={"ID":"3dcd261975c3d6b9a6ad6367fd4facd3","Type":"ContainerDied","Data":"9b7a5502803eaad9f76f5ffa91b29fbb2657402065a03a4d6433c096aac7264f"} Dec 10 22:49:27 crc kubenswrapper[4791]: I1210 22:49:27.909371 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 22:49:27 crc kubenswrapper[4791]: I1210 22:49:27.909403 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 22:49:27 crc kubenswrapper[4791]: I1210 22:49:27.909422 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 22:49:27 crc kubenswrapper[4791]: I1210 22:49:27.921664 4791 certificate_manager.go:356] kubernetes.io/kube-apiserver-client-kubelet: Rotating certificates Dec 10 22:49:27 crc kubenswrapper[4791]: E1210 22:49:27.923332 4791 certificate_manager.go:562] "Unhandled Error" err="kubernetes.io/kube-apiserver-client-kubelet: Failed while requesting a signed certificate from the control plane: cannot create certificate signing request: Post \"https://api-int.crc.testing:6443/apis/certificates.k8s.io/v1/certificatesigningrequests\": dial tcp 38.102.83.30:6443: connect: connection refused" logger="UnhandledError" Dec 10 22:49:27 crc kubenswrapper[4791]: I1210 22:49:27.950106 4791 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" event={"ID":"f614b9022728cf315e60c057852e563e","Type":"ContainerStarted","Data":"224d73713312376a58ec6cc2bc8fdb6ffa44c4749ae0e7691d5fe27b9afca09f"} Dec 10 22:49:27 crc kubenswrapper[4791]: I1210 22:49:27.950364 4791 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" event={"ID":"f614b9022728cf315e60c057852e563e","Type":"ContainerStarted","Data":"372eeaffd40992c6f06efc4df6dfcb44b31de64ee8256c21508cdd7c74ecb96e"} Dec 10 22:49:28 crc kubenswrapper[4791]: W1210 22:49:28.404201 4791 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.Node: Get "https://api-int.crc.testing:6443/api/v1/nodes?fieldSelector=metadata.name%3Dcrc&limit=500&resourceVersion=0": dial tcp 38.102.83.30:6443: connect: connection refused Dec 10 22:49:28 crc kubenswrapper[4791]: E1210 22:49:28.404322 4791 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.Node: failed to list *v1.Node: Get \"https://api-int.crc.testing:6443/api/v1/nodes?fieldSelector=metadata.name%3Dcrc&limit=500&resourceVersion=0\": dial tcp 38.102.83.30:6443: connect: connection refused" logger="UnhandledError" Dec 10 22:49:28 crc kubenswrapper[4791]: I1210 22:49:28.825645 4791 csi_plugin.go:884] Failed to contact API server when waiting for CSINode publishing: Get "https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csinodes/crc?resourceVersion=0": dial tcp 38.102.83.30:6443: connect: connection refused Dec 10 22:49:28 crc kubenswrapper[4791]: I1210 22:49:28.827838 4791 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2025-11-13 08:59:49.697247549 +0000 UTC Dec 10 22:49:28 crc kubenswrapper[4791]: E1210 22:49:28.831669 4791 controller.go:145] "Failed to ensure lease exists, will retry" err="Get \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": dial tcp 38.102.83.30:6443: connect: connection refused" interval="3.2s" Dec 10 22:49:28 crc kubenswrapper[4791]: I1210 22:49:28.958188 4791 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" event={"ID":"f614b9022728cf315e60c057852e563e","Type":"ContainerStarted","Data":"76b6013b062cac5b0b4d12a94f38ce7b43f9c7ea216063455d88a956ac04d26e"} Dec 10 22:49:29 crc kubenswrapper[4791]: I1210 22:49:29.056933 4791 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 10 22:49:29 crc kubenswrapper[4791]: I1210 22:49:29.059157 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 22:49:29 crc kubenswrapper[4791]: I1210 22:49:29.059228 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 22:49:29 crc kubenswrapper[4791]: I1210 22:49:29.059252 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 22:49:29 crc kubenswrapper[4791]: I1210 22:49:29.059286 4791 kubelet_node_status.go:76] "Attempting to register node" node="crc" Dec 10 22:49:29 crc kubenswrapper[4791]: E1210 22:49:29.060056 4791 kubelet_node_status.go:99] "Unable to register node with API server" err="Post \"https://api-int.crc.testing:6443/api/v1/nodes\": dial tcp 38.102.83.30:6443: connect: connection refused" node="crc" Dec 10 22:49:29 crc kubenswrapper[4791]: W1210 22:49:29.122132 4791 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.CSIDriver: Get "https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csidrivers?limit=500&resourceVersion=0": dial tcp 38.102.83.30:6443: connect: connection refused Dec 10 22:49:29 crc kubenswrapper[4791]: E1210 22:49:29.122273 4791 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.CSIDriver: failed to list *v1.CSIDriver: Get \"https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csidrivers?limit=500&resourceVersion=0\": dial tcp 38.102.83.30:6443: connect: connection refused" logger="UnhandledError" Dec 10 22:49:29 crc kubenswrapper[4791]: E1210 22:49:29.501132 4791 event.go:368] "Unable to write event (may retry after sleeping)" err="Post \"https://api-int.crc.testing:6443/api/v1/namespaces/default/events\": dial tcp 38.102.83.30:6443: connect: connection refused" event="&Event{ObjectMeta:{crc.187ffc4829de3e7d default 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Node,Namespace:,Name:crc,UID:crc,APIVersion:,ResourceVersion:,FieldPath:,},Reason:Starting,Message:Starting kubelet.,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2025-12-10 22:49:25.822799485 +0000 UTC m=+0.252417098,LastTimestamp:2025-12-10 22:49:25.822799485 +0000 UTC m=+0.252417098,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Dec 10 22:49:29 crc kubenswrapper[4791]: W1210 22:49:29.722655 4791 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.RuntimeClass: Get "https://api-int.crc.testing:6443/apis/node.k8s.io/v1/runtimeclasses?limit=500&resourceVersion=0": dial tcp 38.102.83.30:6443: connect: connection refused Dec 10 22:49:29 crc kubenswrapper[4791]: E1210 22:49:29.722782 4791 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.RuntimeClass: failed to list *v1.RuntimeClass: Get \"https://api-int.crc.testing:6443/apis/node.k8s.io/v1/runtimeclasses?limit=500&resourceVersion=0\": dial tcp 38.102.83.30:6443: connect: connection refused" logger="UnhandledError" Dec 10 22:49:29 crc kubenswrapper[4791]: I1210 22:49:29.825082 4791 csi_plugin.go:884] Failed to contact API server when waiting for CSINode publishing: Get "https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csinodes/crc?resourceVersion=0": dial tcp 38.102.83.30:6443: connect: connection refused Dec 10 22:49:29 crc kubenswrapper[4791]: I1210 22:49:29.828507 4791 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2025-12-18 02:28:26.049861734 +0000 UTC Dec 10 22:49:29 crc kubenswrapper[4791]: I1210 22:49:29.828560 4791 certificate_manager.go:356] kubernetes.io/kubelet-serving: Waiting 171h38m56.221306965s for next certificate rotation Dec 10 22:49:29 crc kubenswrapper[4791]: I1210 22:49:29.964614 4791 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"f4b27818a5e8e43d0dc095d08835c792","Type":"ContainerStarted","Data":"acafe12186df99d1059a31260de0eafce35ff7771bcb8f5c6753b72665d6f60b"} Dec 10 22:49:29 crc kubenswrapper[4791]: I1210 22:49:29.967669 4791 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" event={"ID":"d1b160f5dda77d281dd8e69ec8d817f9","Type":"ContainerStarted","Data":"511bd056768a15ea0f650a2b1a6e87e3c71dfc75696222a54f7376227ad480c6"} Dec 10 22:49:29 crc kubenswrapper[4791]: I1210 22:49:29.970384 4791 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-etcd/etcd-crc" event={"ID":"2139d3e2895fc6797b9c76a1b4c9886d","Type":"ContainerStarted","Data":"e31a76d1d5b2812ba83e24e38268e25ec8bed98fc3da9adcd06bda5822b6ace1"} Dec 10 22:49:29 crc kubenswrapper[4791]: I1210 22:49:29.973280 4791 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" event={"ID":"3dcd261975c3d6b9a6ad6367fd4facd3","Type":"ContainerStarted","Data":"cd4158bc59b30f0f1716bf43c1ae7c9916a5b7d7164d074987fc1f19ba49b778"} Dec 10 22:49:30 crc kubenswrapper[4791]: W1210 22:49:30.081702 4791 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.Service: Get "https://api-int.crc.testing:6443/api/v1/services?fieldSelector=spec.clusterIP%21%3DNone&limit=500&resourceVersion=0": dial tcp 38.102.83.30:6443: connect: connection refused Dec 10 22:49:30 crc kubenswrapper[4791]: E1210 22:49:30.081845 4791 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.Service: failed to list *v1.Service: Get \"https://api-int.crc.testing:6443/api/v1/services?fieldSelector=spec.clusterIP%21%3DNone&limit=500&resourceVersion=0\": dial tcp 38.102.83.30:6443: connect: connection refused" logger="UnhandledError" Dec 10 22:49:30 crc kubenswrapper[4791]: I1210 22:49:30.825737 4791 csi_plugin.go:884] Failed to contact API server when waiting for CSINode publishing: Get "https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csinodes/crc?resourceVersion=0": dial tcp 38.102.83.30:6443: connect: connection refused Dec 10 22:49:30 crc kubenswrapper[4791]: I1210 22:49:30.977797 4791 generic.go:334] "Generic (PLEG): container finished" podID="2139d3e2895fc6797b9c76a1b4c9886d" containerID="e31a76d1d5b2812ba83e24e38268e25ec8bed98fc3da9adcd06bda5822b6ace1" exitCode=0 Dec 10 22:49:30 crc kubenswrapper[4791]: I1210 22:49:30.977882 4791 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-etcd/etcd-crc" event={"ID":"2139d3e2895fc6797b9c76a1b4c9886d","Type":"ContainerDied","Data":"e31a76d1d5b2812ba83e24e38268e25ec8bed98fc3da9adcd06bda5822b6ace1"} Dec 10 22:49:30 crc kubenswrapper[4791]: I1210 22:49:30.977940 4791 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 10 22:49:30 crc kubenswrapper[4791]: I1210 22:49:30.979162 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 22:49:30 crc kubenswrapper[4791]: I1210 22:49:30.979211 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 22:49:30 crc kubenswrapper[4791]: I1210 22:49:30.979237 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 22:49:30 crc kubenswrapper[4791]: I1210 22:49:30.981526 4791 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 10 22:49:30 crc kubenswrapper[4791]: I1210 22:49:30.981560 4791 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" event={"ID":"f614b9022728cf315e60c057852e563e","Type":"ContainerStarted","Data":"78e950e122e761d950da1a3c60c06476e53920e8795f6c61db22c01e22d5136e"} Dec 10 22:49:30 crc kubenswrapper[4791]: I1210 22:49:30.982499 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 22:49:30 crc kubenswrapper[4791]: I1210 22:49:30.982538 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 22:49:30 crc kubenswrapper[4791]: I1210 22:49:30.982551 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 22:49:31 crc kubenswrapper[4791]: I1210 22:49:31.825837 4791 csi_plugin.go:884] Failed to contact API server when waiting for CSINode publishing: Get "https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csinodes/crc?resourceVersion=0": dial tcp 38.102.83.30:6443: connect: connection refused Dec 10 22:49:31 crc kubenswrapper[4791]: I1210 22:49:31.986077 4791 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"f4b27818a5e8e43d0dc095d08835c792","Type":"ContainerStarted","Data":"2e6e324ec6f4f7bb69fdfab4ea3ac1aa55e480dd8e2a8984b0fe5b4c9f66636e"} Dec 10 22:49:31 crc kubenswrapper[4791]: I1210 22:49:31.988043 4791 generic.go:334] "Generic (PLEG): container finished" podID="2139d3e2895fc6797b9c76a1b4c9886d" containerID="b5b5eceae562f3986b543d41de1e977efb85a91d84c98e09159ca6d092eda691" exitCode=0 Dec 10 22:49:31 crc kubenswrapper[4791]: I1210 22:49:31.988085 4791 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-etcd/etcd-crc" event={"ID":"2139d3e2895fc6797b9c76a1b4c9886d","Type":"ContainerDied","Data":"b5b5eceae562f3986b543d41de1e977efb85a91d84c98e09159ca6d092eda691"} Dec 10 22:49:31 crc kubenswrapper[4791]: I1210 22:49:31.990766 4791 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" event={"ID":"3dcd261975c3d6b9a6ad6367fd4facd3","Type":"ContainerStarted","Data":"656065fd3cbdc6a25dab5d06c7baa00af16d8f62fc5da687efe9c6cc91945de0"} Dec 10 22:49:31 crc kubenswrapper[4791]: I1210 22:49:31.990842 4791 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 10 22:49:31 crc kubenswrapper[4791]: I1210 22:49:31.991704 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 22:49:31 crc kubenswrapper[4791]: I1210 22:49:31.991742 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 22:49:31 crc kubenswrapper[4791]: I1210 22:49:31.991759 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 22:49:32 crc kubenswrapper[4791]: E1210 22:49:32.033114 4791 controller.go:145] "Failed to ensure lease exists, will retry" err="Get \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": dial tcp 38.102.83.30:6443: connect: connection refused" interval="6.4s" Dec 10 22:49:32 crc kubenswrapper[4791]: I1210 22:49:32.260684 4791 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 10 22:49:32 crc kubenswrapper[4791]: I1210 22:49:32.262616 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 22:49:32 crc kubenswrapper[4791]: I1210 22:49:32.262672 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 22:49:32 crc kubenswrapper[4791]: I1210 22:49:32.262689 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 22:49:32 crc kubenswrapper[4791]: I1210 22:49:32.262724 4791 kubelet_node_status.go:76] "Attempting to register node" node="crc" Dec 10 22:49:32 crc kubenswrapper[4791]: E1210 22:49:32.263447 4791 kubelet_node_status.go:99] "Unable to register node with API server" err="Post \"https://api-int.crc.testing:6443/api/v1/nodes\": dial tcp 38.102.83.30:6443: connect: connection refused" node="crc" Dec 10 22:49:32 crc kubenswrapper[4791]: I1210 22:49:32.306549 4791 certificate_manager.go:356] kubernetes.io/kube-apiserver-client-kubelet: Rotating certificates Dec 10 22:49:32 crc kubenswrapper[4791]: E1210 22:49:32.307660 4791 certificate_manager.go:562] "Unhandled Error" err="kubernetes.io/kube-apiserver-client-kubelet: Failed while requesting a signed certificate from the control plane: cannot create certificate signing request: Post \"https://api-int.crc.testing:6443/apis/certificates.k8s.io/v1/certificatesigningrequests\": dial tcp 38.102.83.30:6443: connect: connection refused" logger="UnhandledError" Dec 10 22:49:32 crc kubenswrapper[4791]: I1210 22:49:32.825445 4791 csi_plugin.go:884] Failed to contact API server when waiting for CSINode publishing: Get "https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csinodes/crc?resourceVersion=0": dial tcp 38.102.83.30:6443: connect: connection refused Dec 10 22:49:32 crc kubenswrapper[4791]: I1210 22:49:32.997246 4791 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" event={"ID":"3dcd261975c3d6b9a6ad6367fd4facd3","Type":"ContainerStarted","Data":"90ed7fff816b1772ba9ecd95587a59cbed7c30ac302fff7e7c5de934c912bcf4"} Dec 10 22:49:32 crc kubenswrapper[4791]: I1210 22:49:32.997475 4791 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 10 22:49:32 crc kubenswrapper[4791]: I1210 22:49:32.998701 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 22:49:32 crc kubenswrapper[4791]: I1210 22:49:32.998745 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 22:49:32 crc kubenswrapper[4791]: I1210 22:49:32.998760 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 22:49:33 crc kubenswrapper[4791]: I1210 22:49:33.002183 4791 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"f4b27818a5e8e43d0dc095d08835c792","Type":"ContainerStarted","Data":"0c4d036ab269f62561fe7ef1586012cb3a16e3583a08b9488a0616ec00c8e0ae"} Dec 10 22:49:33 crc kubenswrapper[4791]: I1210 22:49:33.002222 4791 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"f4b27818a5e8e43d0dc095d08835c792","Type":"ContainerStarted","Data":"ea4c346507862a8c12756a3061d6e25e788a8d79f4621b62f81f4322185c82a8"} Dec 10 22:49:33 crc kubenswrapper[4791]: I1210 22:49:33.002242 4791 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 10 22:49:33 crc kubenswrapper[4791]: I1210 22:49:33.003838 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 22:49:33 crc kubenswrapper[4791]: I1210 22:49:33.003918 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 22:49:33 crc kubenswrapper[4791]: I1210 22:49:33.003937 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 22:49:33 crc kubenswrapper[4791]: W1210 22:49:33.409425 4791 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.Node: Get "https://api-int.crc.testing:6443/api/v1/nodes?fieldSelector=metadata.name%3Dcrc&limit=500&resourceVersion=0": dial tcp 38.102.83.30:6443: connect: connection refused Dec 10 22:49:33 crc kubenswrapper[4791]: E1210 22:49:33.409565 4791 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.Node: failed to list *v1.Node: Get \"https://api-int.crc.testing:6443/api/v1/nodes?fieldSelector=metadata.name%3Dcrc&limit=500&resourceVersion=0\": dial tcp 38.102.83.30:6443: connect: connection refused" logger="UnhandledError" Dec 10 22:49:33 crc kubenswrapper[4791]: W1210 22:49:33.475933 4791 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.Service: Get "https://api-int.crc.testing:6443/api/v1/services?fieldSelector=spec.clusterIP%21%3DNone&limit=500&resourceVersion=0": dial tcp 38.102.83.30:6443: connect: connection refused Dec 10 22:49:33 crc kubenswrapper[4791]: E1210 22:49:33.476052 4791 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.Service: failed to list *v1.Service: Get \"https://api-int.crc.testing:6443/api/v1/services?fieldSelector=spec.clusterIP%21%3DNone&limit=500&resourceVersion=0\": dial tcp 38.102.83.30:6443: connect: connection refused" logger="UnhandledError" Dec 10 22:49:33 crc kubenswrapper[4791]: I1210 22:49:33.825235 4791 csi_plugin.go:884] Failed to contact API server when waiting for CSINode publishing: Get "https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csinodes/crc?resourceVersion=0": dial tcp 38.102.83.30:6443: connect: connection refused Dec 10 22:49:34 crc kubenswrapper[4791]: I1210 22:49:34.007701 4791 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"f4b27818a5e8e43d0dc095d08835c792","Type":"ContainerStarted","Data":"d6aa7f8c5d908d90008f2df49860ac8a3480136da567e3d85f86135d779204f9"} Dec 10 22:49:34 crc kubenswrapper[4791]: I1210 22:49:34.007767 4791 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 10 22:49:34 crc kubenswrapper[4791]: I1210 22:49:34.013220 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 22:49:34 crc kubenswrapper[4791]: I1210 22:49:34.013280 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 22:49:34 crc kubenswrapper[4791]: I1210 22:49:34.013293 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 22:49:34 crc kubenswrapper[4791]: I1210 22:49:34.014776 4791 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-etcd/etcd-crc" event={"ID":"2139d3e2895fc6797b9c76a1b4c9886d","Type":"ContainerStarted","Data":"61db531940fa6f2effdae5a005b4c91bd2698bd5d1bff9588c171a65ad550eca"} Dec 10 22:49:34 crc kubenswrapper[4791]: I1210 22:49:34.014868 4791 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 10 22:49:34 crc kubenswrapper[4791]: I1210 22:49:34.014940 4791 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" Dec 10 22:49:34 crc kubenswrapper[4791]: I1210 22:49:34.015766 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 22:49:34 crc kubenswrapper[4791]: I1210 22:49:34.015799 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 22:49:34 crc kubenswrapper[4791]: I1210 22:49:34.015810 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 22:49:34 crc kubenswrapper[4791]: W1210 22:49:34.264945 4791 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.RuntimeClass: Get "https://api-int.crc.testing:6443/apis/node.k8s.io/v1/runtimeclasses?limit=500&resourceVersion=0": dial tcp 38.102.83.30:6443: connect: connection refused Dec 10 22:49:34 crc kubenswrapper[4791]: E1210 22:49:34.265387 4791 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.RuntimeClass: failed to list *v1.RuntimeClass: Get \"https://api-int.crc.testing:6443/apis/node.k8s.io/v1/runtimeclasses?limit=500&resourceVersion=0\": dial tcp 38.102.83.30:6443: connect: connection refused" logger="UnhandledError" Dec 10 22:49:35 crc kubenswrapper[4791]: I1210 22:49:35.022807 4791 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-etcd/etcd-crc" event={"ID":"2139d3e2895fc6797b9c76a1b4c9886d","Type":"ContainerStarted","Data":"5aa8ffa518f3a68b60b275f8df4851b2564aa4984b61efdc9470b73072133e43"} Dec 10 22:49:35 crc kubenswrapper[4791]: I1210 22:49:35.022877 4791 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-etcd/etcd-crc" event={"ID":"2139d3e2895fc6797b9c76a1b4c9886d","Type":"ContainerStarted","Data":"ba54f95b1e00843b854db9a10fe6897592159ccacf30ed0f76002093a13eedd3"} Dec 10 22:49:35 crc kubenswrapper[4791]: I1210 22:49:35.022878 4791 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 10 22:49:35 crc kubenswrapper[4791]: I1210 22:49:35.023004 4791 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 10 22:49:35 crc kubenswrapper[4791]: I1210 22:49:35.023067 4791 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-kube-apiserver/kube-apiserver-crc" Dec 10 22:49:35 crc kubenswrapper[4791]: I1210 22:49:35.024271 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 22:49:35 crc kubenswrapper[4791]: I1210 22:49:35.024311 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 22:49:35 crc kubenswrapper[4791]: I1210 22:49:35.024318 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 22:49:35 crc kubenswrapper[4791]: I1210 22:49:35.024332 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 22:49:35 crc kubenswrapper[4791]: I1210 22:49:35.024383 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 22:49:35 crc kubenswrapper[4791]: I1210 22:49:35.024341 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 22:49:35 crc kubenswrapper[4791]: I1210 22:49:35.506140 4791 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-kube-apiserver/kube-apiserver-crc" Dec 10 22:49:35 crc kubenswrapper[4791]: I1210 22:49:35.603155 4791 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-kube-apiserver/kube-apiserver-crc" Dec 10 22:49:35 crc kubenswrapper[4791]: E1210 22:49:35.949403 4791 eviction_manager.go:285] "Eviction manager: failed to get summary stats" err="failed to get node info: node \"crc\" not found" Dec 10 22:49:36 crc kubenswrapper[4791]: I1210 22:49:36.030224 4791 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 10 22:49:36 crc kubenswrapper[4791]: I1210 22:49:36.030378 4791 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-etcd/etcd-crc" event={"ID":"2139d3e2895fc6797b9c76a1b4c9886d","Type":"ContainerStarted","Data":"82556824d63ce6e46df14c45689ff1f81ebf729abe902061815c040b323898e9"} Dec 10 22:49:36 crc kubenswrapper[4791]: I1210 22:49:36.030442 4791 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-etcd/etcd-crc" event={"ID":"2139d3e2895fc6797b9c76a1b4c9886d","Type":"ContainerStarted","Data":"16f989503edc78d9451638e2cf6cf4b19a2cf2b8809b10b237541cb829303091"} Dec 10 22:49:36 crc kubenswrapper[4791]: I1210 22:49:36.030551 4791 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 10 22:49:36 crc kubenswrapper[4791]: I1210 22:49:36.031923 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 22:49:36 crc kubenswrapper[4791]: I1210 22:49:36.031970 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 22:49:36 crc kubenswrapper[4791]: I1210 22:49:36.031988 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 22:49:36 crc kubenswrapper[4791]: I1210 22:49:36.032193 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 22:49:36 crc kubenswrapper[4791]: I1210 22:49:36.032258 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 22:49:36 crc kubenswrapper[4791]: I1210 22:49:36.032278 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 22:49:36 crc kubenswrapper[4791]: I1210 22:49:36.661627 4791 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-etcd/etcd-crc" Dec 10 22:49:36 crc kubenswrapper[4791]: I1210 22:49:36.726129 4791 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Dec 10 22:49:36 crc kubenswrapper[4791]: I1210 22:49:36.726497 4791 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 10 22:49:36 crc kubenswrapper[4791]: I1210 22:49:36.728092 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 22:49:36 crc kubenswrapper[4791]: I1210 22:49:36.728172 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 22:49:36 crc kubenswrapper[4791]: I1210 22:49:36.728191 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 22:49:37 crc kubenswrapper[4791]: I1210 22:49:37.033052 4791 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 10 22:49:37 crc kubenswrapper[4791]: I1210 22:49:37.033055 4791 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 10 22:49:37 crc kubenswrapper[4791]: I1210 22:49:37.034675 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 22:49:37 crc kubenswrapper[4791]: I1210 22:49:37.034728 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 22:49:37 crc kubenswrapper[4791]: I1210 22:49:37.034746 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 22:49:37 crc kubenswrapper[4791]: I1210 22:49:37.035458 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 22:49:37 crc kubenswrapper[4791]: I1210 22:49:37.035527 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 22:49:37 crc kubenswrapper[4791]: I1210 22:49:37.035546 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 22:49:38 crc kubenswrapper[4791]: I1210 22:49:38.036201 4791 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 10 22:49:38 crc kubenswrapper[4791]: I1210 22:49:38.037906 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 22:49:38 crc kubenswrapper[4791]: I1210 22:49:38.037969 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 22:49:38 crc kubenswrapper[4791]: I1210 22:49:38.037993 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 22:49:38 crc kubenswrapper[4791]: I1210 22:49:38.663847 4791 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 10 22:49:38 crc kubenswrapper[4791]: I1210 22:49:38.665604 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 22:49:38 crc kubenswrapper[4791]: I1210 22:49:38.665668 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 22:49:38 crc kubenswrapper[4791]: I1210 22:49:38.665687 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 22:49:38 crc kubenswrapper[4791]: I1210 22:49:38.665727 4791 kubelet_node_status.go:76] "Attempting to register node" node="crc" Dec 10 22:49:40 crc kubenswrapper[4791]: I1210 22:49:40.652211 4791 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Dec 10 22:49:40 crc kubenswrapper[4791]: I1210 22:49:40.652620 4791 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 10 22:49:40 crc kubenswrapper[4791]: I1210 22:49:40.654156 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 22:49:40 crc kubenswrapper[4791]: I1210 22:49:40.654216 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 22:49:40 crc kubenswrapper[4791]: I1210 22:49:40.654237 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 22:49:41 crc kubenswrapper[4791]: I1210 22:49:41.074683 4791 certificate_manager.go:356] kubernetes.io/kube-apiserver-client-kubelet: Rotating certificates Dec 10 22:49:41 crc kubenswrapper[4791]: I1210 22:49:41.362256 4791 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Dec 10 22:49:41 crc kubenswrapper[4791]: I1210 22:49:41.362574 4791 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 10 22:49:41 crc kubenswrapper[4791]: I1210 22:49:41.364414 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 22:49:41 crc kubenswrapper[4791]: I1210 22:49:41.364475 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 22:49:41 crc kubenswrapper[4791]: I1210 22:49:41.364494 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 22:49:41 crc kubenswrapper[4791]: I1210 22:49:41.563625 4791 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Dec 10 22:49:41 crc kubenswrapper[4791]: I1210 22:49:41.572521 4791 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Dec 10 22:49:42 crc kubenswrapper[4791]: I1210 22:49:42.047913 4791 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 10 22:49:42 crc kubenswrapper[4791]: I1210 22:49:42.049776 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 22:49:42 crc kubenswrapper[4791]: I1210 22:49:42.049819 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 22:49:42 crc kubenswrapper[4791]: I1210 22:49:42.049831 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 22:49:42 crc kubenswrapper[4791]: I1210 22:49:42.055817 4791 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Dec 10 22:49:42 crc kubenswrapper[4791]: I1210 22:49:42.713289 4791 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-etcd/etcd-crc" Dec 10 22:49:42 crc kubenswrapper[4791]: I1210 22:49:42.713660 4791 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 10 22:49:42 crc kubenswrapper[4791]: I1210 22:49:42.715393 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 22:49:42 crc kubenswrapper[4791]: I1210 22:49:42.715450 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 22:49:42 crc kubenswrapper[4791]: I1210 22:49:42.715472 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 22:49:43 crc kubenswrapper[4791]: I1210 22:49:43.051368 4791 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 10 22:49:43 crc kubenswrapper[4791]: I1210 22:49:43.052589 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 22:49:43 crc kubenswrapper[4791]: I1210 22:49:43.052642 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 22:49:43 crc kubenswrapper[4791]: I1210 22:49:43.052659 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 22:49:43 crc kubenswrapper[4791]: I1210 22:49:43.652227 4791 patch_prober.go:28] interesting pod/kube-controller-manager-crc container/cluster-policy-controller namespace/openshift-kube-controller-manager: Startup probe status=failure output="Get \"https://192.168.126.11:10357/healthz\": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers)" start-of-body= Dec 10 22:49:43 crc kubenswrapper[4791]: I1210 22:49:43.652453 4791 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-kube-controller-manager/kube-controller-manager-crc" podUID="f614b9022728cf315e60c057852e563e" containerName="cluster-policy-controller" probeResult="failure" output="Get \"https://192.168.126.11:10357/healthz\": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers)" Dec 10 22:49:44 crc kubenswrapper[4791]: I1210 22:49:44.056393 4791 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 10 22:49:44 crc kubenswrapper[4791]: I1210 22:49:44.058425 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 22:49:44 crc kubenswrapper[4791]: I1210 22:49:44.058454 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 22:49:44 crc kubenswrapper[4791]: I1210 22:49:44.058466 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 22:49:44 crc kubenswrapper[4791]: I1210 22:49:44.786680 4791 patch_prober.go:28] interesting pod/kube-apiserver-crc container/kube-apiserver-check-endpoints namespace/openshift-kube-apiserver: Liveness probe status=failure output="Get \"https://192.168.126.11:17697/healthz\": read tcp 192.168.126.11:42876->192.168.126.11:17697: read: connection reset by peer" start-of-body= Dec 10 22:49:44 crc kubenswrapper[4791]: I1210 22:49:44.786771 4791 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-check-endpoints" probeResult="failure" output="Get \"https://192.168.126.11:17697/healthz\": read tcp 192.168.126.11:42876->192.168.126.11:17697: read: connection reset by peer" Dec 10 22:49:44 crc kubenswrapper[4791]: I1210 22:49:44.825176 4791 csi_plugin.go:884] Failed to contact API server when waiting for CSINode publishing: Get "https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csinodes/crc?resourceVersion=0": net/http: TLS handshake timeout Dec 10 22:49:44 crc kubenswrapper[4791]: I1210 22:49:44.872559 4791 patch_prober.go:28] interesting pod/kube-apiserver-crc container/kube-apiserver namespace/openshift-kube-apiserver: Startup probe status=failure output="HTTP probe failed with statuscode: 403" start-of-body={"kind":"Status","apiVersion":"v1","metadata":{},"status":"Failure","message":"forbidden: User \"system:anonymous\" cannot get path \"/livez\"","reason":"Forbidden","details":{},"code":403} Dec 10 22:49:44 crc kubenswrapper[4791]: I1210 22:49:44.872655 4791 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver" probeResult="failure" output="HTTP probe failed with statuscode: 403" Dec 10 22:49:44 crc kubenswrapper[4791]: I1210 22:49:44.878047 4791 patch_prober.go:28] interesting pod/kube-apiserver-crc container/kube-apiserver namespace/openshift-kube-apiserver: Startup probe status=failure output="HTTP probe failed with statuscode: 403" start-of-body={"kind":"Status","apiVersion":"v1","metadata":{},"status":"Failure","message":"forbidden: User \"system:anonymous\" cannot get path \"/livez\": RBAC: [clusterrole.rbac.authorization.k8s.io \"system:public-info-viewer\" not found, clusterrole.rbac.authorization.k8s.io \"system:openshift:public-info-viewer\" not found]","reason":"Forbidden","details":{},"code":403} Dec 10 22:49:44 crc kubenswrapper[4791]: I1210 22:49:44.878136 4791 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver" probeResult="failure" output="HTTP probe failed with statuscode: 403" Dec 10 22:49:45 crc kubenswrapper[4791]: I1210 22:49:45.061138 4791 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-kube-apiserver_kube-apiserver-crc_f4b27818a5e8e43d0dc095d08835c792/kube-apiserver-check-endpoints/0.log" Dec 10 22:49:45 crc kubenswrapper[4791]: I1210 22:49:45.062772 4791 generic.go:334] "Generic (PLEG): container finished" podID="f4b27818a5e8e43d0dc095d08835c792" containerID="d6aa7f8c5d908d90008f2df49860ac8a3480136da567e3d85f86135d779204f9" exitCode=255 Dec 10 22:49:45 crc kubenswrapper[4791]: I1210 22:49:45.062825 4791 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"f4b27818a5e8e43d0dc095d08835c792","Type":"ContainerDied","Data":"d6aa7f8c5d908d90008f2df49860ac8a3480136da567e3d85f86135d779204f9"} Dec 10 22:49:45 crc kubenswrapper[4791]: I1210 22:49:45.063100 4791 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 10 22:49:45 crc kubenswrapper[4791]: I1210 22:49:45.064166 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 22:49:45 crc kubenswrapper[4791]: I1210 22:49:45.064201 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 22:49:45 crc kubenswrapper[4791]: I1210 22:49:45.064214 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 22:49:45 crc kubenswrapper[4791]: I1210 22:49:45.064786 4791 scope.go:117] "RemoveContainer" containerID="d6aa7f8c5d908d90008f2df49860ac8a3480136da567e3d85f86135d779204f9" Dec 10 22:49:45 crc kubenswrapper[4791]: I1210 22:49:45.618005 4791 patch_prober.go:28] interesting pod/kube-apiserver-crc container/kube-apiserver namespace/openshift-kube-apiserver: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[+]ping ok Dec 10 22:49:45 crc kubenswrapper[4791]: [+]log ok Dec 10 22:49:45 crc kubenswrapper[4791]: [+]etcd ok Dec 10 22:49:45 crc kubenswrapper[4791]: [+]poststarthook/openshift.io-oauth-apiserver-reachable ok Dec 10 22:49:45 crc kubenswrapper[4791]: [+]poststarthook/start-apiserver-admission-initializer ok Dec 10 22:49:45 crc kubenswrapper[4791]: [+]poststarthook/quota.openshift.io-clusterquotamapping ok Dec 10 22:49:45 crc kubenswrapper[4791]: [+]poststarthook/openshift.io-api-request-count-filter ok Dec 10 22:49:45 crc kubenswrapper[4791]: [+]poststarthook/openshift.io-startkubeinformers ok Dec 10 22:49:45 crc kubenswrapper[4791]: [+]poststarthook/openshift.io-openshift-apiserver-reachable ok Dec 10 22:49:45 crc kubenswrapper[4791]: [+]poststarthook/generic-apiserver-start-informers ok Dec 10 22:49:45 crc kubenswrapper[4791]: [+]poststarthook/priority-and-fairness-config-consumer ok Dec 10 22:49:45 crc kubenswrapper[4791]: [+]poststarthook/priority-and-fairness-filter ok Dec 10 22:49:45 crc kubenswrapper[4791]: [+]poststarthook/storage-object-count-tracker-hook ok Dec 10 22:49:45 crc kubenswrapper[4791]: [+]poststarthook/start-apiextensions-informers ok Dec 10 22:49:45 crc kubenswrapper[4791]: [+]poststarthook/start-apiextensions-controllers ok Dec 10 22:49:45 crc kubenswrapper[4791]: [+]poststarthook/crd-informer-synced ok Dec 10 22:49:45 crc kubenswrapper[4791]: [+]poststarthook/start-system-namespaces-controller ok Dec 10 22:49:45 crc kubenswrapper[4791]: [+]poststarthook/start-cluster-authentication-info-controller ok Dec 10 22:49:45 crc kubenswrapper[4791]: [+]poststarthook/start-kube-apiserver-identity-lease-controller ok Dec 10 22:49:45 crc kubenswrapper[4791]: [+]poststarthook/start-kube-apiserver-identity-lease-garbage-collector ok Dec 10 22:49:45 crc kubenswrapper[4791]: [+]poststarthook/start-legacy-token-tracking-controller ok Dec 10 22:49:45 crc kubenswrapper[4791]: [+]poststarthook/start-service-ip-repair-controllers ok Dec 10 22:49:45 crc kubenswrapper[4791]: [-]poststarthook/rbac/bootstrap-roles failed: reason withheld Dec 10 22:49:45 crc kubenswrapper[4791]: [-]poststarthook/scheduling/bootstrap-system-priority-classes failed: reason withheld Dec 10 22:49:45 crc kubenswrapper[4791]: [+]poststarthook/priority-and-fairness-config-producer ok Dec 10 22:49:45 crc kubenswrapper[4791]: [+]poststarthook/bootstrap-controller ok Dec 10 22:49:45 crc kubenswrapper[4791]: [+]poststarthook/aggregator-reload-proxy-client-cert ok Dec 10 22:49:45 crc kubenswrapper[4791]: [+]poststarthook/start-kube-aggregator-informers ok Dec 10 22:49:45 crc kubenswrapper[4791]: [+]poststarthook/apiservice-status-local-available-controller ok Dec 10 22:49:45 crc kubenswrapper[4791]: [+]poststarthook/apiservice-status-remote-available-controller ok Dec 10 22:49:45 crc kubenswrapper[4791]: [+]poststarthook/apiservice-registration-controller ok Dec 10 22:49:45 crc kubenswrapper[4791]: [+]poststarthook/apiservice-wait-for-first-sync ok Dec 10 22:49:45 crc kubenswrapper[4791]: [+]poststarthook/apiservice-discovery-controller ok Dec 10 22:49:45 crc kubenswrapper[4791]: [+]poststarthook/kube-apiserver-autoregistration ok Dec 10 22:49:45 crc kubenswrapper[4791]: [+]autoregister-completion ok Dec 10 22:49:45 crc kubenswrapper[4791]: [+]poststarthook/apiservice-openapi-controller ok Dec 10 22:49:45 crc kubenswrapper[4791]: [+]poststarthook/apiservice-openapiv3-controller ok Dec 10 22:49:45 crc kubenswrapper[4791]: livez check failed Dec 10 22:49:45 crc kubenswrapper[4791]: I1210 22:49:45.618099 4791 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver" probeResult="failure" output="HTTP probe failed with statuscode: 500" Dec 10 22:49:45 crc kubenswrapper[4791]: E1210 22:49:45.950492 4791 eviction_manager.go:285] "Eviction manager: failed to get summary stats" err="failed to get node info: node \"crc\" not found" Dec 10 22:49:46 crc kubenswrapper[4791]: I1210 22:49:46.067761 4791 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-kube-apiserver_kube-apiserver-crc_f4b27818a5e8e43d0dc095d08835c792/kube-apiserver-check-endpoints/0.log" Dec 10 22:49:46 crc kubenswrapper[4791]: I1210 22:49:46.069559 4791 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"f4b27818a5e8e43d0dc095d08835c792","Type":"ContainerStarted","Data":"1b2e11ff2bf8789fa1d0d23ae648b3e6ef653106e5bc4bbb5dd6447f7cd6f05d"} Dec 10 22:49:46 crc kubenswrapper[4791]: I1210 22:49:46.069687 4791 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 10 22:49:46 crc kubenswrapper[4791]: I1210 22:49:46.070449 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 22:49:46 crc kubenswrapper[4791]: I1210 22:49:46.070481 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 22:49:46 crc kubenswrapper[4791]: I1210 22:49:46.070495 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 22:49:48 crc kubenswrapper[4791]: I1210 22:49:48.173500 4791 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-kube-apiserver/kube-apiserver-crc" Dec 10 22:49:48 crc kubenswrapper[4791]: I1210 22:49:48.173688 4791 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 10 22:49:48 crc kubenswrapper[4791]: I1210 22:49:48.175636 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 22:49:48 crc kubenswrapper[4791]: I1210 22:49:48.175700 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 22:49:48 crc kubenswrapper[4791]: I1210 22:49:48.175716 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 22:49:49 crc kubenswrapper[4791]: E1210 22:49:49.850461 4791 controller.go:145] "Failed to ensure lease exists, will retry" err="Get \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": context deadline exceeded" interval="7s" Dec 10 22:49:49 crc kubenswrapper[4791]: I1210 22:49:49.854073 4791 trace.go:236] Trace[1242645008]: "Reflector ListAndWatch" name:k8s.io/client-go/informers/factory.go:160 (10-Dec-2025 22:49:34.930) (total time: 14923ms): Dec 10 22:49:49 crc kubenswrapper[4791]: Trace[1242645008]: ---"Objects listed" error: 14923ms (22:49:49.853) Dec 10 22:49:49 crc kubenswrapper[4791]: Trace[1242645008]: [14.923999745s] [14.923999745s] END Dec 10 22:49:49 crc kubenswrapper[4791]: I1210 22:49:49.854115 4791 reflector.go:368] Caches populated for *v1.CSIDriver from k8s.io/client-go/informers/factory.go:160 Dec 10 22:49:49 crc kubenswrapper[4791]: I1210 22:49:49.854662 4791 reflector.go:368] Caches populated for *v1.Node from k8s.io/client-go/informers/factory.go:160 Dec 10 22:49:49 crc kubenswrapper[4791]: I1210 22:49:49.854686 4791 reflector.go:368] Caches populated for *v1.RuntimeClass from k8s.io/client-go/informers/factory.go:160 Dec 10 22:49:49 crc kubenswrapper[4791]: I1210 22:49:49.855593 4791 reflector.go:368] Caches populated for *v1.Service from k8s.io/client-go/informers/factory.go:160 Dec 10 22:49:49 crc kubenswrapper[4791]: E1210 22:49:49.860380 4791 kubelet_node_status.go:99] "Unable to register node with API server" err="nodes \"crc\" is forbidden: autoscaling.openshift.io/ManagedNode infra config cache not synchronized" node="crc" Dec 10 22:49:49 crc kubenswrapper[4791]: I1210 22:49:49.877497 4791 reflector.go:368] Caches populated for *v1.CertificateSigningRequest from k8s.io/client-go/tools/watch/informerwatcher.go:146 Dec 10 22:49:49 crc kubenswrapper[4791]: I1210 22:49:49.886425 4791 reconstruct.go:205] "DevicePaths of reconstructed volumes updated" Dec 10 22:49:49 crc kubenswrapper[4791]: I1210 22:49:49.931399 4791 csr.go:261] certificate signing request csr-nwdhq is approved, waiting to be issued Dec 10 22:49:49 crc kubenswrapper[4791]: I1210 22:49:49.945893 4791 csr.go:257] certificate signing request csr-nwdhq is issued Dec 10 22:49:50 crc kubenswrapper[4791]: I1210 22:49:50.607627 4791 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-kube-apiserver/kube-apiserver-crc" Dec 10 22:49:50 crc kubenswrapper[4791]: I1210 22:49:50.613559 4791 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-kube-apiserver/kube-apiserver-crc" Dec 10 22:49:50 crc kubenswrapper[4791]: I1210 22:49:50.664350 4791 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Dec 10 22:49:50 crc kubenswrapper[4791]: I1210 22:49:50.671761 4791 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Dec 10 22:49:50 crc kubenswrapper[4791]: I1210 22:49:50.835817 4791 apiserver.go:52] "Watching apiserver" Dec 10 22:49:50 crc kubenswrapper[4791]: I1210 22:49:50.838567 4791 reflector.go:368] Caches populated for *v1.Pod from pkg/kubelet/config/apiserver.go:66 Dec 10 22:49:50 crc kubenswrapper[4791]: I1210 22:49:50.839026 4791 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-dns/node-resolver-tchxk","openshift-kube-apiserver/kube-apiserver-crc","openshift-network-diagnostics/network-check-source-55646444c4-trplf","openshift-network-operator/network-operator-58b4c7f79c-55gtf","openshift-network-operator/iptables-alerter-4ln5h","openshift-kube-controller-manager/kube-controller-manager-crc","openshift-network-console/networking-console-plugin-85b44fc459-gdk6g","openshift-network-diagnostics/network-check-target-xd92c","openshift-network-node-identity/network-node-identity-vrzqb"] Dec 10 22:49:50 crc kubenswrapper[4791]: I1210 22:49:50.839406 4791 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" Dec 10 22:49:50 crc kubenswrapper[4791]: I1210 22:49:50.839681 4791 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 10 22:49:50 crc kubenswrapper[4791]: I1210 22:49:50.839717 4791 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-node-identity/network-node-identity-vrzqb" Dec 10 22:49:50 crc kubenswrapper[4791]: E1210 22:49:50.839744 4791 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 10 22:49:50 crc kubenswrapper[4791]: I1210 22:49:50.839791 4791 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 10 22:49:50 crc kubenswrapper[4791]: I1210 22:49:50.839915 4791 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 10 22:49:50 crc kubenswrapper[4791]: E1210 22:49:50.839922 4791 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 10 22:49:50 crc kubenswrapper[4791]: I1210 22:49:50.839951 4791 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-operator/iptables-alerter-4ln5h" Dec 10 22:49:50 crc kubenswrapper[4791]: E1210 22:49:50.840059 4791 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 10 22:49:50 crc kubenswrapper[4791]: I1210 22:49:50.840442 4791 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-dns/node-resolver-tchxk" Dec 10 22:49:50 crc kubenswrapper[4791]: I1210 22:49:50.842125 4791 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-network-operator"/"metrics-tls" Dec 10 22:49:50 crc kubenswrapper[4791]: I1210 22:49:50.843135 4791 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-node-identity"/"env-overrides" Dec 10 22:49:50 crc kubenswrapper[4791]: I1210 22:49:50.843138 4791 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-node-identity"/"ovnkube-identity-cm" Dec 10 22:49:50 crc kubenswrapper[4791]: I1210 22:49:50.843420 4791 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-node-identity"/"openshift-service-ca.crt" Dec 10 22:49:50 crc kubenswrapper[4791]: I1210 22:49:50.843500 4791 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-operator"/"openshift-service-ca.crt" Dec 10 22:49:50 crc kubenswrapper[4791]: I1210 22:49:50.843782 4791 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-operator"/"iptables-alerter-script" Dec 10 22:49:50 crc kubenswrapper[4791]: I1210 22:49:50.843873 4791 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-dns"/"openshift-service-ca.crt" Dec 10 22:49:50 crc kubenswrapper[4791]: I1210 22:49:50.843955 4791 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-operator"/"kube-root-ca.crt" Dec 10 22:49:50 crc kubenswrapper[4791]: I1210 22:49:50.844064 4791 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-dns"/"node-resolver-dockercfg-kz9s7" Dec 10 22:49:50 crc kubenswrapper[4791]: I1210 22:49:50.844160 4791 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-node-identity"/"kube-root-ca.crt" Dec 10 22:49:50 crc kubenswrapper[4791]: I1210 22:49:50.844297 4791 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-network-node-identity"/"network-node-identity-cert" Dec 10 22:49:50 crc kubenswrapper[4791]: I1210 22:49:50.844642 4791 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-dns"/"kube-root-ca.crt" Dec 10 22:49:50 crc kubenswrapper[4791]: I1210 22:49:50.862061 4791 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-10T22:49:50Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-operator]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-10T22:49:50Z\\\",\\\"message\\\":\\\"containers with unready status: [network-operator]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":false,\\\"restartCount\\\":5,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Dec 10 22:49:50 crc kubenswrapper[4791]: I1210 22:49:50.880298 4791 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-10T22:49:50Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-10T22:49:50Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Dec 10 22:49:50 crc kubenswrapper[4791]: I1210 22:49:50.890189 4791 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-10T22:49:50Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-10T22:49:50Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Dec 10 22:49:50 crc kubenswrapper[4791]: I1210 22:49:50.897800 4791 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-tchxk" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"c0ab74e0-42a9-4f42-ac80-16030cf4ffff\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T22:49:50Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T22:49:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T22:49:50Z\\\",\\\"message\\\":\\\"containers with unready status: [dns-node-resolver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T22:49:50Z\\\",\\\"message\\\":\\\"containers with unready status: [dns-node-resolver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zn2qt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-10T22:49:50Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-tchxk\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Dec 10 22:49:50 crc kubenswrapper[4791]: I1210 22:49:50.904823 4791 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-image-registry/node-ca-vczjq"] Dec 10 22:49:50 crc kubenswrapper[4791]: I1210 22:49:50.905119 4791 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-image-registry/node-ca-vczjq" Dec 10 22:49:50 crc kubenswrapper[4791]: I1210 22:49:50.906897 4791 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-image-registry"/"kube-root-ca.crt" Dec 10 22:49:50 crc kubenswrapper[4791]: I1210 22:49:50.907112 4791 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-image-registry"/"image-registry-certificates" Dec 10 22:49:50 crc kubenswrapper[4791]: I1210 22:49:50.908080 4791 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-image-registry"/"openshift-service-ca.crt" Dec 10 22:49:50 crc kubenswrapper[4791]: I1210 22:49:50.908504 4791 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-image-registry"/"node-ca-dockercfg-4777p" Dec 10 22:49:50 crc kubenswrapper[4791]: I1210 22:49:50.910403 4791 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-10T22:49:50Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-operator]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-10T22:49:50Z\\\",\\\"message\\\":\\\"containers with unready status: [network-operator]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":false,\\\"restartCount\\\":5,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Dec 10 22:49:50 crc kubenswrapper[4791]: I1210 22:49:50.922619 4791 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"698bbe9a-fbe5-436d-a835-5e80df77ba24\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T22:49:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T22:49:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T22:49:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T22:49:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T22:49:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://224d73713312376a58ec6cc2bc8fdb6ffa44c4749ae0e7691d5fe27b9afca09f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T22:49:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://372eeaffd40992c6f06efc4df6dfcb44b31de64ee8256c21508cdd7c74ecb96e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T22:49:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://76b6013b062cac5b0b4d12a94f38ce7b43f9c7ea216063455d88a956ac04d26e\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T22:49:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://78e950e122e761d950da1a3c60c06476e53920e8795f6c61db22c01e22d5136e\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T22:49:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-10T22:49:25Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Dec 10 22:49:50 crc kubenswrapper[4791]: I1210 22:49:50.927746 4791 desired_state_of_world_populator.go:154] "Finished populating initial desired state of world" Dec 10 22:49:50 crc kubenswrapper[4791]: I1210 22:49:50.931915 4791 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-10T22:49:50Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [webhook approver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-10T22:49:50Z\\\",\\\"message\\\":\\\"containers with unready status: [webhook approver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":false,\\\"restartCount\\\":6,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Dec 10 22:49:50 crc kubenswrapper[4791]: I1210 22:49:50.941645 4791 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-10T22:49:50Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-10T22:49:50Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Dec 10 22:49:50 crc kubenswrapper[4791]: I1210 22:49:50.947728 4791 certificate_manager.go:356] kubernetes.io/kube-apiserver-client-kubelet: Certificate expiration is 2026-12-10 22:44:49 +0000 UTC, rotation deadline is 2026-08-24 23:40:57.019399412 +0000 UTC Dec 10 22:49:50 crc kubenswrapper[4791]: I1210 22:49:50.947795 4791 certificate_manager.go:356] kubernetes.io/kube-apiserver-client-kubelet: Waiting 6168h51m6.071607713s for next certificate rotation Dec 10 22:49:50 crc kubenswrapper[4791]: I1210 22:49:50.950137 4791 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-10T22:49:50Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-10T22:49:50Z\\\",\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Dec 10 22:49:50 crc kubenswrapper[4791]: I1210 22:49:50.959259 4791 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f8349fea-c49b-49a1-b5ee-032c5dffe021\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T22:49:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T22:49:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T22:49:25Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T22:49:25Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T22:49:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://acafe12186df99d1059a31260de0eafce35ff7771bcb8f5c6753b72665d6f60b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T22:49:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ea4c346507862a8c12756a3061d6e25e788a8d79f4621b62f81f4322185c82a8\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T22:49:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://2e6e324ec6f4f7bb69fdfab4ea3ac1aa55e480dd8e2a8984b0fe5b4c9f66636e\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T22:49:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://1b2e11ff2bf8789fa1d0d23ae648b3e6ef653106e5bc4bbb5dd6447f7cd6f05d\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d6aa7f8c5d908d90008f2df49860ac8a3480136da567e3d85f86135d779204f9\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-12-10T22:49:44Z\\\",\\\"message\\\":\\\"W1210 22:49:33.845733 1 cmd.go:257] Using insecure, self-signed certificates\\\\nI1210 22:49:33.846108 1 crypto.go:601] Generating new CA for check-endpoints-signer@1765406973 cert, and key in /tmp/serving-cert-2585701456/serving-signer.crt, /tmp/serving-cert-2585701456/serving-signer.key\\\\nI1210 22:49:34.529748 1 observer_polling.go:159] Starting file observer\\\\nW1210 22:49:34.532225 1 builder.go:272] unable to get owner reference (falling back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": dial tcp [::1]:6443: connect: connection refused\\\\nI1210 22:49:34.532599 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1210 22:49:34.533977 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-2585701456/tls.crt::/tmp/serving-cert-2585701456/tls.key\\\\\\\"\\\\nF1210 22:49:44.774710 1 cmd.go:182] error initializing delegating authentication: unable to load configmap based request-header-client-ca-file: Get \\\\\\\"https://localhost:6443/api/v1/namespaces/kube-system/configmaps/extension-apiserver-authentication\\\\\\\": net/http: TLS handshake timeout\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-10T22:49:33Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T22:49:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://0c4d036ab269f62561fe7ef1586012cb3a16e3583a08b9488a0616ec00c8e0ae\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T22:49:32Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c5856d6b26c19fda13b5c9ac2892a7e2f9c6f79feb8173ecfd9b33b673327d67\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c5856d6b26c19fda13b5c9ac2892a7e2f9c6f79feb8173ecfd9b33b673327d67\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-10T22:49:27Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-10T22:49:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-10T22:49:25Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Dec 10 22:49:50 crc kubenswrapper[4791]: I1210 22:49:50.968203 4791 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"698bbe9a-fbe5-436d-a835-5e80df77ba24\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T22:49:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T22:49:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T22:49:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T22:49:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T22:49:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://224d73713312376a58ec6cc2bc8fdb6ffa44c4749ae0e7691d5fe27b9afca09f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T22:49:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://372eeaffd40992c6f06efc4df6dfcb44b31de64ee8256c21508cdd7c74ecb96e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T22:49:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://76b6013b062cac5b0b4d12a94f38ce7b43f9c7ea216063455d88a956ac04d26e\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T22:49:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://78e950e122e761d950da1a3c60c06476e53920e8795f6c61db22c01e22d5136e\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T22:49:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-10T22:49:25Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Dec 10 22:49:50 crc kubenswrapper[4791]: I1210 22:49:50.979304 4791 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-10T22:49:50Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [webhook approver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-10T22:49:50Z\\\",\\\"message\\\":\\\"containers with unready status: [webhook approver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":false,\\\"restartCount\\\":6,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Dec 10 22:49:50 crc kubenswrapper[4791]: I1210 22:49:50.989554 4791 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-10T22:49:50Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-10T22:49:50Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Dec 10 22:49:50 crc kubenswrapper[4791]: I1210 22:49:50.991497 4791 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-ngvvp\" (UniqueName: \"kubernetes.io/projected/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-kube-api-access-ngvvp\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") " Dec 10 22:49:50 crc kubenswrapper[4791]: I1210 22:49:50.991542 4791 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-x7zkh\" (UniqueName: \"kubernetes.io/projected/6731426b-95fe-49ff-bb5f-40441049fde2-kube-api-access-x7zkh\") pod \"6731426b-95fe-49ff-bb5f-40441049fde2\" (UID: \"6731426b-95fe-49ff-bb5f-40441049fde2\") " Dec 10 22:49:50 crc kubenswrapper[4791]: I1210 22:49:50.991567 4791 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/0b78653f-4ff9-4508-8672-245ed9b561e3-kube-api-access\") pod \"0b78653f-4ff9-4508-8672-245ed9b561e3\" (UID: \"0b78653f-4ff9-4508-8672-245ed9b561e3\") " Dec 10 22:49:50 crc kubenswrapper[4791]: I1210 22:49:50.991596 4791 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-cliconfig\" (UniqueName: \"kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-cliconfig\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") " Dec 10 22:49:50 crc kubenswrapper[4791]: I1210 22:49:50.991619 4791 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/a31745f5-9847-4afe-82a5-3161cc66ca93-bound-sa-token\") pod \"a31745f5-9847-4afe-82a5-3161cc66ca93\" (UID: \"a31745f5-9847-4afe-82a5-3161cc66ca93\") " Dec 10 22:49:50 crc kubenswrapper[4791]: I1210 22:49:50.991643 4791 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-trusted-ca-bundle\") pod \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\" (UID: \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\") " Dec 10 22:49:50 crc kubenswrapper[4791]: I1210 22:49:50.991667 4791 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"machine-api-operator-tls\" (UniqueName: \"kubernetes.io/secret/6402fda4-df10-493c-b4e5-d0569419652d-machine-api-operator-tls\") pod \"6402fda4-df10-493c-b4e5-d0569419652d\" (UID: \"6402fda4-df10-493c-b4e5-d0569419652d\") " Dec 10 22:49:50 crc kubenswrapper[4791]: I1210 22:49:50.991689 4791 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"package-server-manager-serving-cert\" (UniqueName: \"kubernetes.io/secret/3ab1a177-2de0-46d9-b765-d0d0649bb42e-package-server-manager-serving-cert\") pod \"3ab1a177-2de0-46d9-b765-d0d0649bb42e\" (UID: \"3ab1a177-2de0-46d9-b765-d0d0649bb42e\") " Dec 10 22:49:50 crc kubenswrapper[4791]: I1210 22:49:50.991715 4791 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-9xfj7\" (UniqueName: \"kubernetes.io/projected/5225d0e4-402f-4861-b410-819f433b1803-kube-api-access-9xfj7\") pod \"5225d0e4-402f-4861-b410-819f433b1803\" (UID: \"5225d0e4-402f-4861-b410-819f433b1803\") " Dec 10 22:49:50 crc kubenswrapper[4791]: I1210 22:49:50.991735 4791 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/bf126b07-da06-4140-9a57-dfd54fc6b486-trusted-ca\") pod \"bf126b07-da06-4140-9a57-dfd54fc6b486\" (UID: \"bf126b07-da06-4140-9a57-dfd54fc6b486\") " Dec 10 22:49:50 crc kubenswrapper[4791]: I1210 22:49:50.991755 4791 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etcd-serving-ca\" (UniqueName: \"kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-etcd-serving-ca\") pod \"1bf7eb37-55a3-4c65-b768-a94c82151e69\" (UID: \"1bf7eb37-55a3-4c65-b768-a94c82151e69\") " Dec 10 22:49:50 crc kubenswrapper[4791]: I1210 22:49:50.991777 4791 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-router-certs\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-router-certs\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") " Dec 10 22:49:50 crc kubenswrapper[4791]: I1210 22:49:50.991799 4791 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-trusted-ca-bundle\") pod \"43509403-f426-496e-be36-56cef71462f5\" (UID: \"43509403-f426-496e-be36-56cef71462f5\") " Dec 10 22:49:50 crc kubenswrapper[4791]: I1210 22:49:50.991822 4791 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-bf2bz\" (UniqueName: \"kubernetes.io/projected/1d611f23-29be-4491-8495-bee1670e935f-kube-api-access-bf2bz\") pod \"1d611f23-29be-4491-8495-bee1670e935f\" (UID: \"1d611f23-29be-4491-8495-bee1670e935f\") " Dec 10 22:49:50 crc kubenswrapper[4791]: I1210 22:49:50.991845 4791 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/bf126b07-da06-4140-9a57-dfd54fc6b486-bound-sa-token\") pod \"bf126b07-da06-4140-9a57-dfd54fc6b486\" (UID: \"bf126b07-da06-4140-9a57-dfd54fc6b486\") " Dec 10 22:49:50 crc kubenswrapper[4791]: I1210 22:49:50.991869 4791 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-w7l8j\" (UniqueName: \"kubernetes.io/projected/01ab3dd5-8196-46d0-ad33-122e2ca51def-kube-api-access-w7l8j\") pod \"01ab3dd5-8196-46d0-ad33-122e2ca51def\" (UID: \"01ab3dd5-8196-46d0-ad33-122e2ca51def\") " Dec 10 22:49:50 crc kubenswrapper[4791]: I1210 22:49:50.991891 4791 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/31d8b7a1-420e-4252-a5b7-eebe8a111292-proxy-tls\") pod \"31d8b7a1-420e-4252-a5b7-eebe8a111292\" (UID: \"31d8b7a1-420e-4252-a5b7-eebe8a111292\") " Dec 10 22:49:50 crc kubenswrapper[4791]: I1210 22:49:50.991913 4791 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-apiservice-cert\") pod \"308be0ea-9f5f-4b29-aeb1-5abd31a0b17b\" (UID: \"308be0ea-9f5f-4b29-aeb1-5abd31a0b17b\") " Dec 10 22:49:50 crc kubenswrapper[4791]: I1210 22:49:50.991937 4791 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/96b93a3a-6083-4aea-8eab-fe1aa8245ad9-metrics-tls\") pod \"96b93a3a-6083-4aea-8eab-fe1aa8245ad9\" (UID: \"96b93a3a-6083-4aea-8eab-fe1aa8245ad9\") " Dec 10 22:49:50 crc kubenswrapper[4791]: I1210 22:49:50.991957 4791 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/496e6271-fb68-4057-954e-a0d97a4afa3f-kube-api-access\") pod \"496e6271-fb68-4057-954e-a0d97a4afa3f\" (UID: \"496e6271-fb68-4057-954e-a0d97a4afa3f\") " Dec 10 22:49:50 crc kubenswrapper[4791]: I1210 22:49:50.991982 4791 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/09efc573-dbb6-4249-bd59-9b87aba8dd28-serving-cert\") pod \"09efc573-dbb6-4249-bd59-9b87aba8dd28\" (UID: \"09efc573-dbb6-4249-bd59-9b87aba8dd28\") " Dec 10 22:49:50 crc kubenswrapper[4791]: I1210 22:49:50.992005 4791 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"default-certificate\" (UniqueName: \"kubernetes.io/secret/c03ee662-fb2f-4fc4-a2c1-af487c19d254-default-certificate\") pod \"c03ee662-fb2f-4fc4-a2c1-af487c19d254\" (UID: \"c03ee662-fb2f-4fc4-a2c1-af487c19d254\") " Dec 10 22:49:50 crc kubenswrapper[4791]: I1210 22:49:50.992028 4791 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/925f1c65-6136-48ba-85aa-3a3b50560753-ovnkube-config\") pod \"925f1c65-6136-48ba-85aa-3a3b50560753\" (UID: \"925f1c65-6136-48ba-85aa-3a3b50560753\") " Dec 10 22:49:50 crc kubenswrapper[4791]: I1210 22:49:50.992050 4791 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"available-featuregates\" (UniqueName: \"kubernetes.io/empty-dir/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d-available-featuregates\") pod \"bc5039c0-ea34-426b-a2b7-fbbc87b49a6d\" (UID: \"bc5039c0-ea34-426b-a2b7-fbbc87b49a6d\") " Dec 10 22:49:50 crc kubenswrapper[4791]: I1210 22:49:50.992104 4791 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-v47cf\" (UniqueName: \"kubernetes.io/projected/c03ee662-fb2f-4fc4-a2c1-af487c19d254-kube-api-access-v47cf\") pod \"c03ee662-fb2f-4fc4-a2c1-af487c19d254\" (UID: \"c03ee662-fb2f-4fc4-a2c1-af487c19d254\") " Dec 10 22:49:50 crc kubenswrapper[4791]: I1210 22:49:50.992138 4791 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovnkube-script-lib\" (UniqueName: \"kubernetes.io/configmap/6ea678ab-3438-413e-bfe3-290ae7725660-ovnkube-script-lib\") pod \"6ea678ab-3438-413e-bfe3-290ae7725660\" (UID: \"6ea678ab-3438-413e-bfe3-290ae7725660\") " Dec 10 22:49:50 crc kubenswrapper[4791]: I1210 22:49:50.992159 4791 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"samples-operator-tls\" (UniqueName: \"kubernetes.io/secret/a0128f3a-b052-44ed-a84e-c4c8aaf17c13-samples-operator-tls\") pod \"a0128f3a-b052-44ed-a84e-c4c8aaf17c13\" (UID: \"a0128f3a-b052-44ed-a84e-c4c8aaf17c13\") " Dec 10 22:49:50 crc kubenswrapper[4791]: I1210 22:49:50.992181 4791 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"console-serving-cert\" (UniqueName: \"kubernetes.io/secret/43509403-f426-496e-be36-56cef71462f5-console-serving-cert\") pod \"43509403-f426-496e-be36-56cef71462f5\" (UID: \"43509403-f426-496e-be36-56cef71462f5\") " Dec 10 22:49:50 crc kubenswrapper[4791]: I1210 22:49:50.992203 4791 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-user-template-error\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-template-error\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") " Dec 10 22:49:50 crc kubenswrapper[4791]: I1210 22:49:50.992224 4791 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-etcd-client\") pod \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\" (UID: \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\") " Dec 10 22:49:50 crc kubenswrapper[4791]: I1210 22:49:50.992245 4791 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-service-ca\" (UniqueName: \"kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-service-ca\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") " Dec 10 22:49:50 crc kubenswrapper[4791]: I1210 22:49:50.992267 4791 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-cfbct\" (UniqueName: \"kubernetes.io/projected/57a731c4-ef35-47a8-b875-bfb08a7f8011-kube-api-access-cfbct\") pod \"57a731c4-ef35-47a8-b875-bfb08a7f8011\" (UID: \"57a731c4-ef35-47a8-b875-bfb08a7f8011\") " Dec 10 22:49:50 crc kubenswrapper[4791]: I1210 22:49:50.992287 4791 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/87cf06ed-a83f-41a7-828d-70653580a8cb-config-volume\") pod \"87cf06ed-a83f-41a7-828d-70653580a8cb\" (UID: \"87cf06ed-a83f-41a7-828d-70653580a8cb\") " Dec 10 22:49:50 crc kubenswrapper[4791]: I1210 22:49:50.992305 4791 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/8f668bae-612b-4b75-9490-919e737c6a3b-bound-sa-token\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 10 22:49:50 crc kubenswrapper[4791]: I1210 22:49:50.992324 4791 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/20b0d48f-5fd6-431c-a545-e3c800c7b866-cert\") pod \"20b0d48f-5fd6-431c-a545-e3c800c7b866\" (UID: \"20b0d48f-5fd6-431c-a545-e3c800c7b866\") " Dec 10 22:49:50 crc kubenswrapper[4791]: I1210 22:49:50.992360 4791 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/57a731c4-ef35-47a8-b875-bfb08a7f8011-catalog-content\") pod \"57a731c4-ef35-47a8-b875-bfb08a7f8011\" (UID: \"57a731c4-ef35-47a8-b875-bfb08a7f8011\") " Dec 10 22:49:50 crc kubenswrapper[4791]: I1210 22:49:50.992381 4791 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-fqsjt\" (UniqueName: \"kubernetes.io/projected/efdd0498-1daa-4136-9a4a-3b948c2293fc-kube-api-access-fqsjt\") pod \"efdd0498-1daa-4136-9a4a-3b948c2293fc\" (UID: \"efdd0498-1daa-4136-9a4a-3b948c2293fc\") " Dec 10 22:49:50 crc kubenswrapper[4791]: I1210 22:49:50.992403 4791 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/496e6271-fb68-4057-954e-a0d97a4afa3f-serving-cert\") pod \"496e6271-fb68-4057-954e-a0d97a4afa3f\" (UID: \"496e6271-fb68-4057-954e-a0d97a4afa3f\") " Dec 10 22:49:50 crc kubenswrapper[4791]: I1210 22:49:50.992423 4791 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-lzf88\" (UniqueName: \"kubernetes.io/projected/0b574797-001e-440a-8f4e-c0be86edad0f-kube-api-access-lzf88\") pod \"0b574797-001e-440a-8f4e-c0be86edad0f\" (UID: \"0b574797-001e-440a-8f4e-c0be86edad0f\") " Dec 10 22:49:50 crc kubenswrapper[4791]: I1210 22:49:50.992441 4791 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-service-ca\") pod \"43509403-f426-496e-be36-56cef71462f5\" (UID: \"43509403-f426-496e-be36-56cef71462f5\") " Dec 10 22:49:50 crc kubenswrapper[4791]: I1210 22:49:50.992463 4791 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/22c825df-677d-4ca6-82db-3454ed06e783-config\") pod \"22c825df-677d-4ca6-82db-3454ed06e783\" (UID: \"22c825df-677d-4ca6-82db-3454ed06e783\") " Dec 10 22:49:50 crc kubenswrapper[4791]: I1210 22:49:50.992484 4791 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b-serving-cert\") pod \"8cea82b4-6893-4ddc-af9f-1bb5ae425c5b\" (UID: \"8cea82b4-6893-4ddc-af9f-1bb5ae425c5b\") " Dec 10 22:49:50 crc kubenswrapper[4791]: I1210 22:49:50.992505 4791 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"installation-pull-secrets\" (UniqueName: \"kubernetes.io/secret/8f668bae-612b-4b75-9490-919e737c6a3b-installation-pull-secrets\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 10 22:49:50 crc kubenswrapper[4791]: I1210 22:49:50.992528 4791 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-2w9zh\" (UniqueName: \"kubernetes.io/projected/4bb40260-dbaa-4fb0-84df-5e680505d512-kube-api-access-2w9zh\") pod \"4bb40260-dbaa-4fb0-84df-5e680505d512\" (UID: \"4bb40260-dbaa-4fb0-84df-5e680505d512\") " Dec 10 22:49:50 crc kubenswrapper[4791]: I1210 22:49:50.992552 4791 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-wxkg8\" (UniqueName: \"kubernetes.io/projected/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59-kube-api-access-wxkg8\") pod \"3cb93b32-e0ae-4377-b9c8-fdb9842c6d59\" (UID: \"3cb93b32-e0ae-4377-b9c8-fdb9842c6d59\") " Dec 10 22:49:50 crc kubenswrapper[4791]: I1210 22:49:50.992608 4791 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-fcqwp\" (UniqueName: \"kubernetes.io/projected/5fe579f8-e8a6-4643-bce5-a661393c4dde-kube-api-access-fcqwp\") pod \"5fe579f8-e8a6-4643-bce5-a661393c4dde\" (UID: \"5fe579f8-e8a6-4643-bce5-a661393c4dde\") " Dec 10 22:49:50 crc kubenswrapper[4791]: I1210 22:49:50.992634 4791 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-249nr\" (UniqueName: \"kubernetes.io/projected/b6312bbd-5731-4ea0-a20f-81d5a57df44a-kube-api-access-249nr\") pod \"b6312bbd-5731-4ea0-a20f-81d5a57df44a\" (UID: \"b6312bbd-5731-4ea0-a20f-81d5a57df44a\") " Dec 10 22:49:50 crc kubenswrapper[4791]: I1210 22:49:50.992660 4791 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/6509e943-70c6-444c-bc41-48a544e36fbd-trusted-ca-bundle\") pod \"6509e943-70c6-444c-bc41-48a544e36fbd\" (UID: \"6509e943-70c6-444c-bc41-48a544e36fbd\") " Dec 10 22:49:50 crc kubenswrapper[4791]: I1210 22:49:50.992685 4791 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-session\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-session\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") " Dec 10 22:49:50 crc kubenswrapper[4791]: I1210 22:49:50.992707 4791 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/9d4552c7-cd75-42dd-8880-30dd377c49a4-config\") pod \"9d4552c7-cd75-42dd-8880-30dd377c49a4\" (UID: \"9d4552c7-cd75-42dd-8880-30dd377c49a4\") " Dec 10 22:49:50 crc kubenswrapper[4791]: I1210 22:49:50.992730 4791 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"mcc-auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/0b574797-001e-440a-8f4e-c0be86edad0f-mcc-auth-proxy-config\") pod \"0b574797-001e-440a-8f4e-c0be86edad0f\" (UID: \"0b574797-001e-440a-8f4e-c0be86edad0f\") " Dec 10 22:49:50 crc kubenswrapper[4791]: I1210 22:49:50.992754 4791 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/e7e6199b-1264-4501-8953-767f51328d08-serving-cert\") pod \"e7e6199b-1264-4501-8953-767f51328d08\" (UID: \"e7e6199b-1264-4501-8953-767f51328d08\") " Dec 10 22:49:50 crc kubenswrapper[4791]: I1210 22:49:50.992778 4791 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-webhook-cert\") pod \"308be0ea-9f5f-4b29-aeb1-5abd31a0b17b\" (UID: \"308be0ea-9f5f-4b29-aeb1-5abd31a0b17b\") " Dec 10 22:49:50 crc kubenswrapper[4791]: I1210 22:49:50.992801 4791 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"node-bootstrap-token\" (UniqueName: \"kubernetes.io/secret/5fe579f8-e8a6-4643-bce5-a661393c4dde-node-bootstrap-token\") pod \"5fe579f8-e8a6-4643-bce5-a661393c4dde\" (UID: \"5fe579f8-e8a6-4643-bce5-a661393c4dde\") " Dec 10 22:49:50 crc kubenswrapper[4791]: I1210 22:49:50.992831 4791 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ca-trust-extracted\" (UniqueName: \"kubernetes.io/empty-dir/8f668bae-612b-4b75-9490-919e737c6a3b-ca-trust-extracted\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 10 22:49:50 crc kubenswrapper[4791]: I1210 22:49:50.992856 4791 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"profile-collector-cert\" (UniqueName: \"kubernetes.io/secret/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9-profile-collector-cert\") pod \"f88749ec-7931-4ee7-b3fc-1ec5e11f92e9\" (UID: \"f88749ec-7931-4ee7-b3fc-1ec5e11f92e9\") " Dec 10 22:49:50 crc kubenswrapper[4791]: I1210 22:49:50.992881 4791 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/8f668bae-612b-4b75-9490-919e737c6a3b-trusted-ca\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 10 22:49:50 crc kubenswrapper[4791]: I1210 22:49:50.992905 4791 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/1386a44e-36a2-460c-96d0-0359d2b6f0f5-serving-cert\") pod \"1386a44e-36a2-460c-96d0-0359d2b6f0f5\" (UID: \"1386a44e-36a2-460c-96d0-0359d2b6f0f5\") " Dec 10 22:49:50 crc kubenswrapper[4791]: I1210 22:49:50.992929 4791 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-nzwt7\" (UniqueName: \"kubernetes.io/projected/96b93a3a-6083-4aea-8eab-fe1aa8245ad9-kube-api-access-nzwt7\") pod \"96b93a3a-6083-4aea-8eab-fe1aa8245ad9\" (UID: \"96b93a3a-6083-4aea-8eab-fe1aa8245ad9\") " Dec 10 22:49:50 crc kubenswrapper[4791]: I1210 22:49:50.992952 4791 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-dbsvg\" (UniqueName: \"kubernetes.io/projected/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9-kube-api-access-dbsvg\") pod \"f88749ec-7931-4ee7-b3fc-1ec5e11f92e9\" (UID: \"f88749ec-7931-4ee7-b3fc-1ec5e11f92e9\") " Dec 10 22:49:50 crc kubenswrapper[4791]: I1210 22:49:50.992975 4791 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"mcd-auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/fda69060-fa79-4696-b1a6-7980f124bf7c-mcd-auth-proxy-config\") pod \"fda69060-fa79-4696-b1a6-7980f124bf7c\" (UID: \"fda69060-fa79-4696-b1a6-7980f124bf7c\") " Dec 10 22:49:50 crc kubenswrapper[4791]: I1210 22:49:50.992999 4791 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-user-template-login\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-template-login\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") " Dec 10 22:49:50 crc kubenswrapper[4791]: I1210 22:49:50.993022 4791 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/1386a44e-36a2-460c-96d0-0359d2b6f0f5-kube-api-access\") pod \"1386a44e-36a2-460c-96d0-0359d2b6f0f5\" (UID: \"1386a44e-36a2-460c-96d0-0359d2b6f0f5\") " Dec 10 22:49:50 crc kubenswrapper[4791]: I1210 22:49:50.993045 4791 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/925f1c65-6136-48ba-85aa-3a3b50560753-env-overrides\") pod \"925f1c65-6136-48ba-85aa-3a3b50560753\" (UID: \"925f1c65-6136-48ba-85aa-3a3b50560753\") " Dec 10 22:49:50 crc kubenswrapper[4791]: I1210 22:49:50.993066 4791 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"audit\" (UniqueName: \"kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-audit\") pod \"1bf7eb37-55a3-4c65-b768-a94c82151e69\" (UID: \"1bf7eb37-55a3-4c65-b768-a94c82151e69\") " Dec 10 22:49:50 crc kubenswrapper[4791]: I1210 22:49:50.993090 4791 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"control-plane-machine-set-operator-tls\" (UniqueName: \"kubernetes.io/secret/6731426b-95fe-49ff-bb5f-40441049fde2-control-plane-machine-set-operator-tls\") pod \"6731426b-95fe-49ff-bb5f-40441049fde2\" (UID: \"6731426b-95fe-49ff-bb5f-40441049fde2\") " Dec 10 22:49:50 crc kubenswrapper[4791]: I1210 22:49:50.993113 4791 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"service-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/6509e943-70c6-444c-bc41-48a544e36fbd-service-ca-bundle\") pod \"6509e943-70c6-444c-bc41-48a544e36fbd\" (UID: \"6509e943-70c6-444c-bc41-48a544e36fbd\") " Dec 10 22:49:50 crc kubenswrapper[4791]: I1210 22:49:50.993135 4791 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-d4lsv\" (UniqueName: \"kubernetes.io/projected/25e176fe-21b4-4974-b1ed-c8b94f112a7f-kube-api-access-d4lsv\") pod \"25e176fe-21b4-4974-b1ed-c8b94f112a7f\" (UID: \"25e176fe-21b4-4974-b1ed-c8b94f112a7f\") " Dec 10 22:49:50 crc kubenswrapper[4791]: I1210 22:49:50.993159 4791 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-2d4wz\" (UniqueName: \"kubernetes.io/projected/5441d097-087c-4d9a-baa8-b210afa90fc9-kube-api-access-2d4wz\") pod \"5441d097-087c-4d9a-baa8-b210afa90fc9\" (UID: \"5441d097-087c-4d9a-baa8-b210afa90fc9\") " Dec 10 22:49:50 crc kubenswrapper[4791]: I1210 22:49:50.993183 4791 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d-utilities\") pod \"b11524ee-3fca-4b1b-9cdf-6da289fdbc7d\" (UID: \"b11524ee-3fca-4b1b-9cdf-6da289fdbc7d\") " Dec 10 22:49:50 crc kubenswrapper[4791]: I1210 22:49:50.993204 4791 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"image-import-ca\" (UniqueName: \"kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-image-import-ca\") pod \"1bf7eb37-55a3-4c65-b768-a94c82151e69\" (UID: \"1bf7eb37-55a3-4c65-b768-a94c82151e69\") " Dec 10 22:49:50 crc kubenswrapper[4791]: I1210 22:49:50.993227 4791 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-zkvpv\" (UniqueName: \"kubernetes.io/projected/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-kube-api-access-zkvpv\") pod \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\" (UID: \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\") " Dec 10 22:49:50 crc kubenswrapper[4791]: I1210 22:49:50.993251 4791 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/57a731c4-ef35-47a8-b875-bfb08a7f8011-utilities\") pod \"57a731c4-ef35-47a8-b875-bfb08a7f8011\" (UID: \"57a731c4-ef35-47a8-b875-bfb08a7f8011\") " Dec 10 22:49:50 crc kubenswrapper[4791]: I1210 22:49:50.993271 4791 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"encryption-config\" (UniqueName: \"kubernetes.io/secret/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-encryption-config\") pod \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\" (UID: \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\") " Dec 10 22:49:50 crc kubenswrapper[4791]: I1210 22:49:50.993295 4791 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/0b78653f-4ff9-4508-8672-245ed9b561e3-serving-cert\") pod \"0b78653f-4ff9-4508-8672-245ed9b561e3\" (UID: \"0b78653f-4ff9-4508-8672-245ed9b561e3\") " Dec 10 22:49:50 crc kubenswrapper[4791]: I1210 22:49:50.993318 4791 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"profile-collector-cert\" (UniqueName: \"kubernetes.io/secret/b6312bbd-5731-4ea0-a20f-81d5a57df44a-profile-collector-cert\") pod \"b6312bbd-5731-4ea0-a20f-81d5a57df44a\" (UID: \"b6312bbd-5731-4ea0-a20f-81d5a57df44a\") " Dec 10 22:49:50 crc kubenswrapper[4791]: I1210 22:49:50.993360 4791 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovn-control-plane-metrics-cert\" (UniqueName: \"kubernetes.io/secret/925f1c65-6136-48ba-85aa-3a3b50560753-ovn-control-plane-metrics-cert\") pod \"925f1c65-6136-48ba-85aa-3a3b50560753\" (UID: \"925f1c65-6136-48ba-85aa-3a3b50560753\") " Dec 10 22:49:50 crc kubenswrapper[4791]: I1210 22:49:50.993384 4791 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-s4n52\" (UniqueName: \"kubernetes.io/projected/925f1c65-6136-48ba-85aa-3a3b50560753-kube-api-access-s4n52\") pod \"925f1c65-6136-48ba-85aa-3a3b50560753\" (UID: \"925f1c65-6136-48ba-85aa-3a3b50560753\") " Dec 10 22:49:50 crc kubenswrapper[4791]: I1210 22:49:50.993405 4791 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"srv-cert\" (UniqueName: \"kubernetes.io/secret/b6312bbd-5731-4ea0-a20f-81d5a57df44a-srv-cert\") pod \"b6312bbd-5731-4ea0-a20f-81d5a57df44a\" (UID: \"b6312bbd-5731-4ea0-a20f-81d5a57df44a\") " Dec 10 22:49:50 crc kubenswrapper[4791]: I1210 22:49:50.993429 4791 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/01ab3dd5-8196-46d0-ad33-122e2ca51def-serving-cert\") pod \"01ab3dd5-8196-46d0-ad33-122e2ca51def\" (UID: \"01ab3dd5-8196-46d0-ad33-122e2ca51def\") " Dec 10 22:49:50 crc kubenswrapper[4791]: I1210 22:49:50.993455 4791 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-pj782\" (UniqueName: \"kubernetes.io/projected/b6cd30de-2eeb-49a2-ab40-9167f4560ff5-kube-api-access-pj782\") pod \"b6cd30de-2eeb-49a2-ab40-9167f4560ff5\" (UID: \"b6cd30de-2eeb-49a2-ab40-9167f4560ff5\") " Dec 10 22:49:50 crc kubenswrapper[4791]: I1210 22:49:50.993484 4791 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/0b574797-001e-440a-8f4e-c0be86edad0f-proxy-tls\") pod \"0b574797-001e-440a-8f4e-c0be86edad0f\" (UID: \"0b574797-001e-440a-8f4e-c0be86edad0f\") " Dec 10 22:49:50 crc kubenswrapper[4791]: I1210 22:49:50.993507 4791 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/a31745f5-9847-4afe-82a5-3161cc66ca93-metrics-tls\") pod \"a31745f5-9847-4afe-82a5-3161cc66ca93\" (UID: \"a31745f5-9847-4afe-82a5-3161cc66ca93\") " Dec 10 22:49:50 crc kubenswrapper[4791]: I1210 22:49:50.993529 4791 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"image-registry-operator-tls\" (UniqueName: \"kubernetes.io/secret/bf126b07-da06-4140-9a57-dfd54fc6b486-image-registry-operator-tls\") pod \"bf126b07-da06-4140-9a57-dfd54fc6b486\" (UID: \"bf126b07-da06-4140-9a57-dfd54fc6b486\") " Dec 10 22:49:50 crc kubenswrapper[4791]: I1210 22:49:50.993630 4791 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 10 22:49:50 crc kubenswrapper[4791]: I1210 22:49:50.993657 4791 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/87cf06ed-a83f-41a7-828d-70653580a8cb-metrics-tls\") pod \"87cf06ed-a83f-41a7-828d-70653580a8cb\" (UID: \"87cf06ed-a83f-41a7-828d-70653580a8cb\") " Dec 10 22:49:50 crc kubenswrapper[4791]: I1210 22:49:50.993721 4791 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/01ab3dd5-8196-46d0-ad33-122e2ca51def-config\") pod \"01ab3dd5-8196-46d0-ad33-122e2ca51def\" (UID: \"01ab3dd5-8196-46d0-ad33-122e2ca51def\") " Dec 10 22:49:50 crc kubenswrapper[4791]: I1210 22:49:50.993746 4791 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"srv-cert\" (UniqueName: \"kubernetes.io/secret/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9-srv-cert\") pod \"f88749ec-7931-4ee7-b3fc-1ec5e11f92e9\" (UID: \"f88749ec-7931-4ee7-b3fc-1ec5e11f92e9\") " Dec 10 22:49:50 crc kubenswrapper[4791]: I1210 22:49:50.993769 4791 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"registry-certificates\" (UniqueName: \"kubernetes.io/configmap/8f668bae-612b-4b75-9490-919e737c6a3b-registry-certificates\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 10 22:49:50 crc kubenswrapper[4791]: I1210 22:49:50.993791 4791 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"multus-daemon-config\" (UniqueName: \"kubernetes.io/configmap/4bb40260-dbaa-4fb0-84df-5e680505d512-multus-daemon-config\") pod \"4bb40260-dbaa-4fb0-84df-5e680505d512\" (UID: \"4bb40260-dbaa-4fb0-84df-5e680505d512\") " Dec 10 22:49:50 crc kubenswrapper[4791]: I1210 22:49:50.993815 4791 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/6ea678ab-3438-413e-bfe3-290ae7725660-ovnkube-config\") pod \"6ea678ab-3438-413e-bfe3-290ae7725660\" (UID: \"6ea678ab-3438-413e-bfe3-290ae7725660\") " Dec 10 22:49:50 crc kubenswrapper[4791]: I1210 22:49:50.993836 4791 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etcd-service-ca\" (UniqueName: \"kubernetes.io/configmap/09efc573-dbb6-4249-bd59-9b87aba8dd28-etcd-service-ca\") pod \"09efc573-dbb6-4249-bd59-9b87aba8dd28\" (UID: \"09efc573-dbb6-4249-bd59-9b87aba8dd28\") " Dec 10 22:49:50 crc kubenswrapper[4791]: I1210 22:49:50.993857 4791 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-audit-policies\") pod \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\" (UID: \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\") " Dec 10 22:49:50 crc kubenswrapper[4791]: I1210 22:49:50.993881 4791 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/1d611f23-29be-4491-8495-bee1670e935f-catalog-content\") pod \"1d611f23-29be-4491-8495-bee1670e935f\" (UID: \"1d611f23-29be-4491-8495-bee1670e935f\") " Dec 10 22:49:50 crc kubenswrapper[4791]: I1210 22:49:50.993903 4791 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"oauth-serving-cert\" (UniqueName: \"kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-oauth-serving-cert\") pod \"43509403-f426-496e-be36-56cef71462f5\" (UID: \"43509403-f426-496e-be36-56cef71462f5\") " Dec 10 22:49:50 crc kubenswrapper[4791]: I1210 22:49:50.993925 4791 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-w4xd4\" (UniqueName: \"kubernetes.io/projected/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b-kube-api-access-w4xd4\") pod \"8cea82b4-6893-4ddc-af9f-1bb5ae425c5b\" (UID: \"8cea82b4-6893-4ddc-af9f-1bb5ae425c5b\") " Dec 10 22:49:50 crc kubenswrapper[4791]: I1210 22:49:50.993949 4791 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/0b78653f-4ff9-4508-8672-245ed9b561e3-service-ca\") pod \"0b78653f-4ff9-4508-8672-245ed9b561e3\" (UID: \"0b78653f-4ff9-4508-8672-245ed9b561e3\") " Dec 10 22:49:50 crc kubenswrapper[4791]: I1210 22:49:50.993971 4791 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d-serving-cert\") pod \"bc5039c0-ea34-426b-a2b7-fbbc87b49a6d\" (UID: \"bc5039c0-ea34-426b-a2b7-fbbc87b49a6d\") " Dec 10 22:49:50 crc kubenswrapper[4791]: I1210 22:49:50.993993 4791 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-x2m85\" (UniqueName: \"kubernetes.io/projected/cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d-kube-api-access-x2m85\") pod \"cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d\" (UID: \"cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d\") " Dec 10 22:49:50 crc kubenswrapper[4791]: I1210 22:49:50.994015 4791 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"console-oauth-config\" (UniqueName: \"kubernetes.io/secret/43509403-f426-496e-be36-56cef71462f5-console-oauth-config\") pod \"43509403-f426-496e-be36-56cef71462f5\" (UID: \"43509403-f426-496e-be36-56cef71462f5\") " Dec 10 22:49:50 crc kubenswrapper[4791]: I1210 22:49:50.994038 4791 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-6g6sz\" (UniqueName: \"kubernetes.io/projected/6509e943-70c6-444c-bc41-48a544e36fbd-kube-api-access-6g6sz\") pod \"6509e943-70c6-444c-bc41-48a544e36fbd\" (UID: \"6509e943-70c6-444c-bc41-48a544e36fbd\") " Dec 10 22:49:50 crc kubenswrapper[4791]: I1210 22:49:50.994060 4791 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/c03ee662-fb2f-4fc4-a2c1-af487c19d254-metrics-certs\") pod \"c03ee662-fb2f-4fc4-a2c1-af487c19d254\" (UID: \"c03ee662-fb2f-4fc4-a2c1-af487c19d254\") " Dec 10 22:49:50 crc kubenswrapper[4791]: I1210 22:49:50.994083 4791 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-xcgwh\" (UniqueName: \"kubernetes.io/projected/fda69060-fa79-4696-b1a6-7980f124bf7c-kube-api-access-xcgwh\") pod \"fda69060-fa79-4696-b1a6-7980f124bf7c\" (UID: \"fda69060-fa79-4696-b1a6-7980f124bf7c\") " Dec 10 22:49:50 crc kubenswrapper[4791]: I1210 22:49:50.994109 4791 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etcd-ca\" (UniqueName: \"kubernetes.io/configmap/09efc573-dbb6-4249-bd59-9b87aba8dd28-etcd-ca\") pod \"09efc573-dbb6-4249-bd59-9b87aba8dd28\" (UID: \"09efc573-dbb6-4249-bd59-9b87aba8dd28\") " Dec 10 22:49:50 crc kubenswrapper[4791]: I1210 22:49:50.994132 4791 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-ocp-branding-template\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-ocp-branding-template\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") " Dec 10 22:49:50 crc kubenswrapper[4791]: I1210 22:49:50.994154 4791 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c-serving-cert\") pod \"210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c\" (UID: \"210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c\") " Dec 10 22:49:50 crc kubenswrapper[4791]: I1210 22:49:50.994177 4791 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/5441d097-087c-4d9a-baa8-b210afa90fc9-client-ca\") pod \"5441d097-087c-4d9a-baa8-b210afa90fc9\" (UID: \"5441d097-087c-4d9a-baa8-b210afa90fc9\") " Dec 10 22:49:50 crc kubenswrapper[4791]: I1210 22:49:50.994200 4791 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serviceca\" (UniqueName: \"kubernetes.io/configmap/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59-serviceca\") pod \"3cb93b32-e0ae-4377-b9c8-fdb9842c6d59\" (UID: \"3cb93b32-e0ae-4377-b9c8-fdb9842c6d59\") " Dec 10 22:49:50 crc kubenswrapper[4791]: I1210 22:49:50.994222 4791 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-mg5zb\" (UniqueName: \"kubernetes.io/projected/6402fda4-df10-493c-b4e5-d0569419652d-kube-api-access-mg5zb\") pod \"6402fda4-df10-493c-b4e5-d0569419652d\" (UID: \"6402fda4-df10-493c-b4e5-d0569419652d\") " Dec 10 22:49:50 crc kubenswrapper[4791]: I1210 22:49:50.994244 4791 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/a31745f5-9847-4afe-82a5-3161cc66ca93-trusted-ca\") pod \"a31745f5-9847-4afe-82a5-3161cc66ca93\" (UID: \"a31745f5-9847-4afe-82a5-3161cc66ca93\") " Dec 10 22:49:50 crc kubenswrapper[4791]: I1210 22:49:50.994267 4791 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/09efc573-dbb6-4249-bd59-9b87aba8dd28-etcd-client\") pod \"09efc573-dbb6-4249-bd59-9b87aba8dd28\" (UID: \"09efc573-dbb6-4249-bd59-9b87aba8dd28\") " Dec 10 22:49:50 crc kubenswrapper[4791]: I1210 22:49:50.994288 4791 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-4d4hj\" (UniqueName: \"kubernetes.io/projected/3ab1a177-2de0-46d9-b765-d0d0649bb42e-kube-api-access-4d4hj\") pod \"3ab1a177-2de0-46d9-b765-d0d0649bb42e\" (UID: \"3ab1a177-2de0-46d9-b765-d0d0649bb42e\") " Dec 10 22:49:50 crc kubenswrapper[4791]: I1210 22:49:50.994320 4791 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-sb6h7\" (UniqueName: \"kubernetes.io/projected/1bf7eb37-55a3-4c65-b768-a94c82151e69-kube-api-access-sb6h7\") pod \"1bf7eb37-55a3-4c65-b768-a94c82151e69\" (UID: \"1bf7eb37-55a3-4c65-b768-a94c82151e69\") " Dec 10 22:49:50 crc kubenswrapper[4791]: I1210 22:49:50.994365 4791 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-vt5rc\" (UniqueName: \"kubernetes.io/projected/44663579-783b-4372-86d6-acf235a62d72-kube-api-access-vt5rc\") pod \"44663579-783b-4372-86d6-acf235a62d72\" (UID: \"44663579-783b-4372-86d6-acf235a62d72\") " Dec 10 22:49:50 crc kubenswrapper[4791]: I1210 22:49:50.994389 4791 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"certs\" (UniqueName: \"kubernetes.io/secret/5fe579f8-e8a6-4643-bce5-a661393c4dde-certs\") pod \"5fe579f8-e8a6-4643-bce5-a661393c4dde\" (UID: \"5fe579f8-e8a6-4643-bce5-a661393c4dde\") " Dec 10 22:49:50 crc kubenswrapper[4791]: I1210 22:49:50.994415 4791 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/fda69060-fa79-4696-b1a6-7980f124bf7c-proxy-tls\") pod \"fda69060-fa79-4696-b1a6-7980f124bf7c\" (UID: \"fda69060-fa79-4696-b1a6-7980f124bf7c\") " Dec 10 22:49:50 crc kubenswrapper[4791]: I1210 22:49:50.994438 4791 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etcd-serving-ca\" (UniqueName: \"kubernetes.io/configmap/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-etcd-serving-ca\") pod \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\" (UID: \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\") " Dec 10 22:49:50 crc kubenswrapper[4791]: I1210 22:49:50.994461 4791 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"console-config\" (UniqueName: \"kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-console-config\") pod \"43509403-f426-496e-be36-56cef71462f5\" (UID: \"43509403-f426-496e-be36-56cef71462f5\") " Dec 10 22:49:50 crc kubenswrapper[4791]: I1210 22:49:50.994482 4791 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/5225d0e4-402f-4861-b410-819f433b1803-catalog-content\") pod \"5225d0e4-402f-4861-b410-819f433b1803\" (UID: \"5225d0e4-402f-4861-b410-819f433b1803\") " Dec 10 22:49:50 crc kubenswrapper[4791]: I1210 22:49:50.994502 4791 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-qg5z5\" (UniqueName: \"kubernetes.io/projected/43509403-f426-496e-be36-56cef71462f5-kube-api-access-qg5z5\") pod \"43509403-f426-496e-be36-56cef71462f5\" (UID: \"43509403-f426-496e-be36-56cef71462f5\") " Dec 10 22:49:50 crc kubenswrapper[4791]: I1210 22:49:50.994526 4791 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-htfz6\" (UniqueName: \"kubernetes.io/projected/6ea678ab-3438-413e-bfe3-290ae7725660-kube-api-access-htfz6\") pod \"6ea678ab-3438-413e-bfe3-290ae7725660\" (UID: \"6ea678ab-3438-413e-bfe3-290ae7725660\") " Dec 10 22:49:50 crc kubenswrapper[4791]: I1210 22:49:50.994546 4791 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/5441d097-087c-4d9a-baa8-b210afa90fc9-config\") pod \"5441d097-087c-4d9a-baa8-b210afa90fc9\" (UID: \"5441d097-087c-4d9a-baa8-b210afa90fc9\") " Dec 10 22:49:50 crc kubenswrapper[4791]: I1210 22:49:50.994567 4791 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/e7e6199b-1264-4501-8953-767f51328d08-kube-api-access\") pod \"e7e6199b-1264-4501-8953-767f51328d08\" (UID: \"e7e6199b-1264-4501-8953-767f51328d08\") " Dec 10 22:49:50 crc kubenswrapper[4791]: I1210 22:49:50.994592 4791 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/22c825df-677d-4ca6-82db-3454ed06e783-auth-proxy-config\") pod \"22c825df-677d-4ca6-82db-3454ed06e783\" (UID: \"22c825df-677d-4ca6-82db-3454ed06e783\") " Dec 10 22:49:50 crc kubenswrapper[4791]: I1210 22:49:50.994615 4791 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"marketplace-trusted-ca\" (UniqueName: \"kubernetes.io/configmap/b6cd30de-2eeb-49a2-ab40-9167f4560ff5-marketplace-trusted-ca\") pod \"b6cd30de-2eeb-49a2-ab40-9167f4560ff5\" (UID: \"b6cd30de-2eeb-49a2-ab40-9167f4560ff5\") " Dec 10 22:49:50 crc kubenswrapper[4791]: I1210 22:49:50.994637 4791 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/31d8b7a1-420e-4252-a5b7-eebe8a111292-auth-proxy-config\") pod \"31d8b7a1-420e-4252-a5b7-eebe8a111292\" (UID: \"31d8b7a1-420e-4252-a5b7-eebe8a111292\") " Dec 10 22:49:50 crc kubenswrapper[4791]: I1210 22:49:50.994659 4791 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"cni-binary-copy\" (UniqueName: \"kubernetes.io/configmap/7bb08738-c794-4ee8-9972-3a62ca171029-cni-binary-copy\") pod \"7bb08738-c794-4ee8-9972-3a62ca171029\" (UID: \"7bb08738-c794-4ee8-9972-3a62ca171029\") " Dec 10 22:49:50 crc kubenswrapper[4791]: I1210 22:49:50.994683 4791 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-xcphl\" (UniqueName: \"kubernetes.io/projected/7583ce53-e0fe-4a16-9e4d-50516596a136-kube-api-access-xcphl\") pod \"7583ce53-e0fe-4a16-9e4d-50516596a136\" (UID: \"7583ce53-e0fe-4a16-9e4d-50516596a136\") " Dec 10 22:49:50 crc kubenswrapper[4791]: I1210 22:49:50.994708 4791 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/5441d097-087c-4d9a-baa8-b210afa90fc9-serving-cert\") pod \"5441d097-087c-4d9a-baa8-b210afa90fc9\" (UID: \"5441d097-087c-4d9a-baa8-b210afa90fc9\") " Dec 10 22:49:50 crc kubenswrapper[4791]: I1210 22:49:50.994729 4791 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-serving-cert\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-serving-cert\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") " Dec 10 22:49:50 crc kubenswrapper[4791]: I1210 22:49:50.994751 4791 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/1bf7eb37-55a3-4c65-b768-a94c82151e69-etcd-client\") pod \"1bf7eb37-55a3-4c65-b768-a94c82151e69\" (UID: \"1bf7eb37-55a3-4c65-b768-a94c82151e69\") " Dec 10 22:49:50 crc kubenswrapper[4791]: I1210 22:49:50.994775 4791 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovn-node-metrics-cert\" (UniqueName: \"kubernetes.io/secret/6ea678ab-3438-413e-bfe3-290ae7725660-ovn-node-metrics-cert\") pod \"6ea678ab-3438-413e-bfe3-290ae7725660\" (UID: \"6ea678ab-3438-413e-bfe3-290ae7725660\") " Dec 10 22:49:50 crc kubenswrapper[4791]: I1210 22:49:50.994798 4791 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-tk88c\" (UniqueName: \"kubernetes.io/projected/7539238d-5fe0-46ed-884e-1c3b566537ec-kube-api-access-tk88c\") pod \"7539238d-5fe0-46ed-884e-1c3b566537ec\" (UID: \"7539238d-5fe0-46ed-884e-1c3b566537ec\") " Dec 10 22:49:50 crc kubenswrapper[4791]: I1210 22:49:50.994821 4791 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/e7e6199b-1264-4501-8953-767f51328d08-config\") pod \"e7e6199b-1264-4501-8953-767f51328d08\" (UID: \"e7e6199b-1264-4501-8953-767f51328d08\") " Dec 10 22:49:50 crc kubenswrapper[4791]: I1210 22:49:50.994843 4791 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"machine-approver-tls\" (UniqueName: \"kubernetes.io/secret/22c825df-677d-4ca6-82db-3454ed06e783-machine-approver-tls\") pod \"22c825df-677d-4ca6-82db-3454ed06e783\" (UID: \"22c825df-677d-4ca6-82db-3454ed06e783\") " Dec 10 22:49:50 crc kubenswrapper[4791]: I1210 22:49:50.994869 4791 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-user-template-provider-selection\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-template-provider-selection\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") " Dec 10 22:49:50 crc kubenswrapper[4791]: I1210 22:49:50.994897 4791 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-audit-policies\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") " Dec 10 22:49:50 crc kubenswrapper[4791]: I1210 22:49:50.994920 4791 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/7539238d-5fe0-46ed-884e-1c3b566537ec-config\") pod \"7539238d-5fe0-46ed-884e-1c3b566537ec\" (UID: \"7539238d-5fe0-46ed-884e-1c3b566537ec\") " Dec 10 22:49:50 crc kubenswrapper[4791]: I1210 22:49:50.994945 4791 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-8tdtz\" (UniqueName: \"kubernetes.io/projected/09efc573-dbb6-4249-bd59-9b87aba8dd28-kube-api-access-8tdtz\") pod \"09efc573-dbb6-4249-bd59-9b87aba8dd28\" (UID: \"09efc573-dbb6-4249-bd59-9b87aba8dd28\") " Dec 10 22:49:50 crc kubenswrapper[4791]: I1210 22:49:50.994969 4791 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-gf66m\" (UniqueName: \"kubernetes.io/projected/a0128f3a-b052-44ed-a84e-c4c8aaf17c13-kube-api-access-gf66m\") pod \"a0128f3a-b052-44ed-a84e-c4c8aaf17c13\" (UID: \"a0128f3a-b052-44ed-a84e-c4c8aaf17c13\") " Dec 10 22:49:50 crc kubenswrapper[4791]: I1210 22:49:50.994992 4791 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-6ccd8\" (UniqueName: \"kubernetes.io/projected/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-kube-api-access-6ccd8\") pod \"308be0ea-9f5f-4b29-aeb1-5abd31a0b17b\" (UID: \"308be0ea-9f5f-4b29-aeb1-5abd31a0b17b\") " Dec 10 22:49:50 crc kubenswrapper[4791]: I1210 22:49:50.995015 4791 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-x4zgh\" (UniqueName: \"kubernetes.io/projected/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d-kube-api-access-x4zgh\") pod \"b11524ee-3fca-4b1b-9cdf-6da289fdbc7d\" (UID: \"b11524ee-3fca-4b1b-9cdf-6da289fdbc7d\") " Dec 10 22:49:50 crc kubenswrapper[4791]: I1210 22:49:50.995038 4791 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/7539238d-5fe0-46ed-884e-1c3b566537ec-serving-cert\") pod \"7539238d-5fe0-46ed-884e-1c3b566537ec\" (UID: \"7539238d-5fe0-46ed-884e-1c3b566537ec\") " Dec 10 22:49:50 crc kubenswrapper[4791]: I1210 22:49:50.995060 4791 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-config\") pod \"1bf7eb37-55a3-4c65-b768-a94c82151e69\" (UID: \"1bf7eb37-55a3-4c65-b768-a94c82151e69\") " Dec 10 22:49:50 crc kubenswrapper[4791]: I1210 22:49:50.995118 4791 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-trusted-ca-bundle\") pod \"1bf7eb37-55a3-4c65-b768-a94c82151e69\" (UID: \"1bf7eb37-55a3-4c65-b768-a94c82151e69\") " Dec 10 22:49:50 crc kubenswrapper[4791]: I1210 22:49:50.995144 4791 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-d6qdx\" (UniqueName: \"kubernetes.io/projected/87cf06ed-a83f-41a7-828d-70653580a8cb-kube-api-access-d6qdx\") pod \"87cf06ed-a83f-41a7-828d-70653580a8cb\" (UID: \"87cf06ed-a83f-41a7-828d-70653580a8cb\") " Dec 10 22:49:50 crc kubenswrapper[4791]: I1210 22:49:50.995166 4791 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"images\" (UniqueName: \"kubernetes.io/configmap/31d8b7a1-420e-4252-a5b7-eebe8a111292-images\") pod \"31d8b7a1-420e-4252-a5b7-eebe8a111292\" (UID: \"31d8b7a1-420e-4252-a5b7-eebe8a111292\") " Dec 10 22:49:50 crc kubenswrapper[4791]: I1210 22:49:50.995189 4791 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/5225d0e4-402f-4861-b410-819f433b1803-utilities\") pod \"5225d0e4-402f-4861-b410-819f433b1803\" (UID: \"5225d0e4-402f-4861-b410-819f433b1803\") " Dec 10 22:49:50 crc kubenswrapper[4791]: I1210 22:49:50.995222 4791 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"stats-auth\" (UniqueName: \"kubernetes.io/secret/c03ee662-fb2f-4fc4-a2c1-af487c19d254-stats-auth\") pod \"c03ee662-fb2f-4fc4-a2c1-af487c19d254\" (UID: \"c03ee662-fb2f-4fc4-a2c1-af487c19d254\") " Dec 10 22:49:50 crc kubenswrapper[4791]: I1210 22:49:50.995249 4791 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-jkwtn\" (UniqueName: \"kubernetes.io/projected/5b88f790-22fa-440e-b583-365168c0b23d-kube-api-access-jkwtn\") pod \"5b88f790-22fa-440e-b583-365168c0b23d\" (UID: \"5b88f790-22fa-440e-b583-365168c0b23d\") " Dec 10 22:49:50 crc kubenswrapper[4791]: I1210 22:49:50.995273 4791 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"registry-tls\" (UniqueName: \"kubernetes.io/projected/8f668bae-612b-4b75-9490-919e737c6a3b-registry-tls\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 10 22:49:50 crc kubenswrapper[4791]: I1210 22:49:50.995300 4791 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/496e6271-fb68-4057-954e-a0d97a4afa3f-config\") pod \"496e6271-fb68-4057-954e-a0d97a4afa3f\" (UID: \"496e6271-fb68-4057-954e-a0d97a4afa3f\") " Dec 10 22:49:50 crc kubenswrapper[4791]: I1210 22:49:50.995324 4791 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/6509e943-70c6-444c-bc41-48a544e36fbd-config\") pod \"6509e943-70c6-444c-bc41-48a544e36fbd\" (UID: \"6509e943-70c6-444c-bc41-48a544e36fbd\") " Dec 10 22:49:50 crc kubenswrapper[4791]: I1210 22:49:50.995367 4791 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/7583ce53-e0fe-4a16-9e4d-50516596a136-serving-cert\") pod \"7583ce53-e0fe-4a16-9e4d-50516596a136\" (UID: \"7583ce53-e0fe-4a16-9e4d-50516596a136\") " Dec 10 22:49:50 crc kubenswrapper[4791]: I1210 22:49:50.995390 4791 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-user-idp-0-file-data\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-idp-0-file-data\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") " Dec 10 22:49:50 crc kubenswrapper[4791]: I1210 22:49:50.995412 4791 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"encryption-config\" (UniqueName: \"kubernetes.io/secret/1bf7eb37-55a3-4c65-b768-a94c82151e69-encryption-config\") pod \"1bf7eb37-55a3-4c65-b768-a94c82151e69\" (UID: \"1bf7eb37-55a3-4c65-b768-a94c82151e69\") " Dec 10 22:49:50 crc kubenswrapper[4791]: I1210 22:49:50.995434 4791 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"service-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/c03ee662-fb2f-4fc4-a2c1-af487c19d254-service-ca-bundle\") pod \"c03ee662-fb2f-4fc4-a2c1-af487c19d254\" (UID: \"c03ee662-fb2f-4fc4-a2c1-af487c19d254\") " Dec 10 22:49:50 crc kubenswrapper[4791]: I1210 22:49:50.995457 4791 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-serving-cert\") pod \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\" (UID: \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\") " Dec 10 22:49:50 crc kubenswrapper[4791]: I1210 22:49:50.995483 4791 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-w9rds\" (UniqueName: \"kubernetes.io/projected/20b0d48f-5fd6-431c-a545-e3c800c7b866-kube-api-access-w9rds\") pod \"20b0d48f-5fd6-431c-a545-e3c800c7b866\" (UID: \"20b0d48f-5fd6-431c-a545-e3c800c7b866\") " Dec 10 22:49:50 crc kubenswrapper[4791]: I1210 22:49:50.995508 4791 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-trusted-ca-bundle\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") " Dec 10 22:49:50 crc kubenswrapper[4791]: I1210 22:49:50.995531 4791 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/7583ce53-e0fe-4a16-9e4d-50516596a136-config\") pod \"7583ce53-e0fe-4a16-9e4d-50516596a136\" (UID: \"7583ce53-e0fe-4a16-9e4d-50516596a136\") " Dec 10 22:49:50 crc kubenswrapper[4791]: I1210 22:49:50.995555 4791 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"webhook-certs\" (UniqueName: \"kubernetes.io/secret/efdd0498-1daa-4136-9a4a-3b948c2293fc-webhook-certs\") pod \"efdd0498-1daa-4136-9a4a-3b948c2293fc\" (UID: \"efdd0498-1daa-4136-9a4a-3b948c2293fc\") " Dec 10 22:49:50 crc kubenswrapper[4791]: I1210 22:49:50.995581 4791 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/6ea678ab-3438-413e-bfe3-290ae7725660-env-overrides\") pod \"6ea678ab-3438-413e-bfe3-290ae7725660\" (UID: \"6ea678ab-3438-413e-bfe3-290ae7725660\") " Dec 10 22:49:50 crc kubenswrapper[4791]: I1210 22:49:50.995604 4791 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/1d611f23-29be-4491-8495-bee1670e935f-utilities\") pod \"1d611f23-29be-4491-8495-bee1670e935f\" (UID: \"1d611f23-29be-4491-8495-bee1670e935f\") " Dec 10 22:49:50 crc kubenswrapper[4791]: I1210 22:49:50.995628 4791 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-pcxfs\" (UniqueName: \"kubernetes.io/projected/9d4552c7-cd75-42dd-8880-30dd377c49a4-kube-api-access-pcxfs\") pod \"9d4552c7-cd75-42dd-8880-30dd377c49a4\" (UID: \"9d4552c7-cd75-42dd-8880-30dd377c49a4\") " Dec 10 22:49:50 crc kubenswrapper[4791]: I1210 22:49:50.995653 4791 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/9d4552c7-cd75-42dd-8880-30dd377c49a4-trusted-ca\") pod \"9d4552c7-cd75-42dd-8880-30dd377c49a4\" (UID: \"9d4552c7-cd75-42dd-8880-30dd377c49a4\") " Dec 10 22:49:50 crc kubenswrapper[4791]: I1210 22:49:50.995678 4791 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b-config\") pod \"8cea82b4-6893-4ddc-af9f-1bb5ae425c5b\" (UID: \"8cea82b4-6893-4ddc-af9f-1bb5ae425c5b\") " Dec 10 22:49:50 crc kubenswrapper[4791]: I1210 22:49:50.995701 4791 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-kfwg7\" (UniqueName: \"kubernetes.io/projected/8f668bae-612b-4b75-9490-919e737c6a3b-kube-api-access-kfwg7\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 10 22:49:50 crc kubenswrapper[4791]: I1210 22:49:50.995728 4791 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d-catalog-content\") pod \"b11524ee-3fca-4b1b-9cdf-6da289fdbc7d\" (UID: \"b11524ee-3fca-4b1b-9cdf-6da289fdbc7d\") " Dec 10 22:49:50 crc kubenswrapper[4791]: I1210 22:49:50.995763 4791 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-279lb\" (UniqueName: \"kubernetes.io/projected/7bb08738-c794-4ee8-9972-3a62ca171029-kube-api-access-279lb\") pod \"7bb08738-c794-4ee8-9972-3a62ca171029\" (UID: \"7bb08738-c794-4ee8-9972-3a62ca171029\") " Dec 10 22:49:50 crc kubenswrapper[4791]: I1210 22:49:50.995789 4791 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-mnrrd\" (UniqueName: \"kubernetes.io/projected/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d-kube-api-access-mnrrd\") pod \"bc5039c0-ea34-426b-a2b7-fbbc87b49a6d\" (UID: \"bc5039c0-ea34-426b-a2b7-fbbc87b49a6d\") " Dec 10 22:49:50 crc kubenswrapper[4791]: I1210 22:49:50.995812 4791 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-lz9wn\" (UniqueName: \"kubernetes.io/projected/a31745f5-9847-4afe-82a5-3161cc66ca93-kube-api-access-lz9wn\") pod \"a31745f5-9847-4afe-82a5-3161cc66ca93\" (UID: \"a31745f5-9847-4afe-82a5-3161cc66ca93\") " Dec 10 22:49:50 crc kubenswrapper[4791]: I1210 22:49:50.995835 4791 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-zgdk5\" (UniqueName: \"kubernetes.io/projected/31d8b7a1-420e-4252-a5b7-eebe8a111292-kube-api-access-zgdk5\") pod \"31d8b7a1-420e-4252-a5b7-eebe8a111292\" (UID: \"31d8b7a1-420e-4252-a5b7-eebe8a111292\") " Dec 10 22:49:50 crc kubenswrapper[4791]: I1210 22:49:50.995859 4791 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/5b88f790-22fa-440e-b583-365168c0b23d-metrics-certs\") pod \"5b88f790-22fa-440e-b583-365168c0b23d\" (UID: \"5b88f790-22fa-440e-b583-365168c0b23d\") " Dec 10 22:49:50 crc kubenswrapper[4791]: I1210 22:49:50.995885 4791 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"marketplace-operator-metrics\" (UniqueName: \"kubernetes.io/secret/b6cd30de-2eeb-49a2-ab40-9167f4560ff5-marketplace-operator-metrics\") pod \"b6cd30de-2eeb-49a2-ab40-9167f4560ff5\" (UID: \"b6cd30de-2eeb-49a2-ab40-9167f4560ff5\") " Dec 10 22:49:50 crc kubenswrapper[4791]: I1210 22:49:50.995907 4791 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"tmpfs\" (UniqueName: \"kubernetes.io/empty-dir/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-tmpfs\") pod \"308be0ea-9f5f-4b29-aeb1-5abd31a0b17b\" (UID: \"308be0ea-9f5f-4b29-aeb1-5abd31a0b17b\") " Dec 10 22:49:50 crc kubenswrapper[4791]: I1210 22:49:50.995929 4791 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-7c4vf\" (UniqueName: \"kubernetes.io/projected/22c825df-677d-4ca6-82db-3454ed06e783-kube-api-access-7c4vf\") pod \"22c825df-677d-4ca6-82db-3454ed06e783\" (UID: \"22c825df-677d-4ca6-82db-3454ed06e783\") " Dec 10 22:49:50 crc kubenswrapper[4791]: I1210 22:49:50.995953 4791 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/1386a44e-36a2-460c-96d0-0359d2b6f0f5-config\") pod \"1386a44e-36a2-460c-96d0-0359d2b6f0f5\" (UID: \"1386a44e-36a2-460c-96d0-0359d2b6f0f5\") " Dec 10 22:49:50 crc kubenswrapper[4791]: I1210 22:49:50.995976 4791 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/7583ce53-e0fe-4a16-9e4d-50516596a136-client-ca\") pod \"7583ce53-e0fe-4a16-9e4d-50516596a136\" (UID: \"7583ce53-e0fe-4a16-9e4d-50516596a136\") " Dec 10 22:49:50 crc kubenswrapper[4791]: I1210 22:49:50.996001 4791 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-jhbk2\" (UniqueName: \"kubernetes.io/projected/bd23aa5c-e532-4e53-bccf-e79f130c5ae8-kube-api-access-jhbk2\") pod \"bd23aa5c-e532-4e53-bccf-e79f130c5ae8\" (UID: \"bd23aa5c-e532-4e53-bccf-e79f130c5ae8\") " Dec 10 22:49:50 crc kubenswrapper[4791]: I1210 22:49:50.996025 4791 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-qs4fp\" (UniqueName: \"kubernetes.io/projected/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c-kube-api-access-qs4fp\") pod \"210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c\" (UID: \"210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c\") " Dec 10 22:49:50 crc kubenswrapper[4791]: I1210 22:49:50.996048 4791 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"cni-binary-copy\" (UniqueName: \"kubernetes.io/configmap/4bb40260-dbaa-4fb0-84df-5e680505d512-cni-binary-copy\") pod \"4bb40260-dbaa-4fb0-84df-5e680505d512\" (UID: \"4bb40260-dbaa-4fb0-84df-5e680505d512\") " Dec 10 22:49:50 crc kubenswrapper[4791]: I1210 22:49:50.996070 4791 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"images\" (UniqueName: \"kubernetes.io/configmap/6402fda4-df10-493c-b4e5-d0569419652d-images\") pod \"6402fda4-df10-493c-b4e5-d0569419652d\" (UID: \"6402fda4-df10-493c-b4e5-d0569419652d\") " Dec 10 22:49:50 crc kubenswrapper[4791]: I1210 22:49:50.996094 4791 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/9d4552c7-cd75-42dd-8880-30dd377c49a4-serving-cert\") pod \"9d4552c7-cd75-42dd-8880-30dd377c49a4\" (UID: \"9d4552c7-cd75-42dd-8880-30dd377c49a4\") " Dec 10 22:49:50 crc kubenswrapper[4791]: I1210 22:49:50.996126 4791 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/1bf7eb37-55a3-4c65-b768-a94c82151e69-serving-cert\") pod \"1bf7eb37-55a3-4c65-b768-a94c82151e69\" (UID: \"1bf7eb37-55a3-4c65-b768-a94c82151e69\") " Dec 10 22:49:50 crc kubenswrapper[4791]: I1210 22:49:50.996153 4791 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-rnphk\" (UniqueName: \"kubernetes.io/projected/bf126b07-da06-4140-9a57-dfd54fc6b486-kube-api-access-rnphk\") pod \"bf126b07-da06-4140-9a57-dfd54fc6b486\" (UID: \"bf126b07-da06-4140-9a57-dfd54fc6b486\") " Dec 10 22:49:50 crc kubenswrapper[4791]: I1210 22:49:50.996178 4791 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/09efc573-dbb6-4249-bd59-9b87aba8dd28-config\") pod \"09efc573-dbb6-4249-bd59-9b87aba8dd28\" (UID: \"09efc573-dbb6-4249-bd59-9b87aba8dd28\") " Dec 10 22:49:50 crc kubenswrapper[4791]: I1210 22:49:50.996205 4791 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/6402fda4-df10-493c-b4e5-d0569419652d-config\") pod \"6402fda4-df10-493c-b4e5-d0569419652d\" (UID: \"6402fda4-df10-493c-b4e5-d0569419652d\") " Dec 10 22:49:50 crc kubenswrapper[4791]: I1210 22:49:50.996234 4791 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c-config\") pod \"210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c\" (UID: \"210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c\") " Dec 10 22:49:50 crc kubenswrapper[4791]: I1210 22:49:50.996258 4791 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"signing-key\" (UniqueName: \"kubernetes.io/secret/25e176fe-21b4-4974-b1ed-c8b94f112a7f-signing-key\") pod \"25e176fe-21b4-4974-b1ed-c8b94f112a7f\" (UID: \"25e176fe-21b4-4974-b1ed-c8b94f112a7f\") " Dec 10 22:49:50 crc kubenswrapper[4791]: I1210 22:49:50.996283 4791 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/6509e943-70c6-444c-bc41-48a544e36fbd-serving-cert\") pod \"6509e943-70c6-444c-bc41-48a544e36fbd\" (UID: \"6509e943-70c6-444c-bc41-48a544e36fbd\") " Dec 10 22:49:50 crc kubenswrapper[4791]: I1210 22:49:50.996306 4791 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/7583ce53-e0fe-4a16-9e4d-50516596a136-proxy-ca-bundles\") pod \"7583ce53-e0fe-4a16-9e4d-50516596a136\" (UID: \"7583ce53-e0fe-4a16-9e4d-50516596a136\") " Dec 10 22:49:50 crc kubenswrapper[4791]: I1210 22:49:50.996333 4791 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"cni-sysctl-allowlist\" (UniqueName: \"kubernetes.io/configmap/7bb08738-c794-4ee8-9972-3a62ca171029-cni-sysctl-allowlist\") pod \"7bb08738-c794-4ee8-9972-3a62ca171029\" (UID: \"7bb08738-c794-4ee8-9972-3a62ca171029\") " Dec 10 22:49:50 crc kubenswrapper[4791]: I1210 22:49:50.996777 4791 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-pjr6v\" (UniqueName: \"kubernetes.io/projected/49ef4625-1d3a-4a9f-b595-c2433d32326d-kube-api-access-pjr6v\") pod \"49ef4625-1d3a-4a9f-b595-c2433d32326d\" (UID: \"49ef4625-1d3a-4a9f-b595-c2433d32326d\") " Dec 10 22:49:50 crc kubenswrapper[4791]: I1210 22:49:50.996807 4791 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"signing-cabundle\" (UniqueName: \"kubernetes.io/configmap/25e176fe-21b4-4974-b1ed-c8b94f112a7f-signing-cabundle\") pod \"25e176fe-21b4-4974-b1ed-c8b94f112a7f\" (UID: \"25e176fe-21b4-4974-b1ed-c8b94f112a7f\") " Dec 10 22:49:50 crc kubenswrapper[4791]: I1210 22:49:50.996859 4791 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cqllr\" (UniqueName: \"kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr\") pod \"network-check-target-xd92c\" (UID: \"3b6479f0-333b-4a96-9adf-2099afdc2447\") " pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 10 22:49:50 crc kubenswrapper[4791]: I1210 22:49:50.996892 4791 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-s2kz5\" (UniqueName: \"kubernetes.io/projected/ef543e1b-8068-4ea3-b32a-61027b32e95d-kube-api-access-s2kz5\") pod \"network-node-identity-vrzqb\" (UID: \"ef543e1b-8068-4ea3-b32a-61027b32e95d\") " pod="openshift-network-node-identity/network-node-identity-vrzqb" Dec 10 22:49:50 crc kubenswrapper[4791]: I1210 22:49:50.996918 4791 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-rdwmf\" (UniqueName: \"kubernetes.io/projected/37a5e44f-9a88-4405-be8a-b645485e7312-kube-api-access-rdwmf\") pod \"network-operator-58b4c7f79c-55gtf\" (UID: \"37a5e44f-9a88-4405-be8a-b645485e7312\") " pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" Dec 10 22:49:50 crc kubenswrapper[4791]: I1210 22:49:50.996946 4791 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-s2dwl\" (UniqueName: \"kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl\") pod \"network-check-source-55646444c4-trplf\" (UID: \"9d751cbb-f2e2-430d-9754-c882a5e924a5\") " pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 10 22:49:50 crc kubenswrapper[4791]: I1210 22:49:50.996970 4791 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovnkube-identity-cm\" (UniqueName: \"kubernetes.io/configmap/ef543e1b-8068-4ea3-b32a-61027b32e95d-ovnkube-identity-cm\") pod \"network-node-identity-vrzqb\" (UID: \"ef543e1b-8068-4ea3-b32a-61027b32e95d\") " pod="openshift-network-node-identity/network-node-identity-vrzqb" Dec 10 22:49:50 crc kubenswrapper[4791]: I1210 22:49:50.997000 4791 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serviceca\" (UniqueName: \"kubernetes.io/configmap/b9f00f25-3fa1-4b63-9710-fd352224b01b-serviceca\") pod \"node-ca-vczjq\" (UID: \"b9f00f25-3fa1-4b63-9710-fd352224b01b\") " pod="openshift-image-registry/node-ca-vczjq" Dec 10 22:49:50 crc kubenswrapper[4791]: I1210 22:49:50.997021 4791 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host\" (UniqueName: \"kubernetes.io/host-path/b9f00f25-3fa1-4b63-9710-fd352224b01b-host\") pod \"node-ca-vczjq\" (UID: \"b9f00f25-3fa1-4b63-9710-fd352224b01b\") " pod="openshift-image-registry/node-ca-vczjq" Dec 10 22:49:50 crc kubenswrapper[4791]: I1210 22:49:50.997045 4791 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-slash\" (UniqueName: \"kubernetes.io/host-path/d75a4c96-2883-4a0b-bab2-0fab2b6c0b49-host-slash\") pod \"iptables-alerter-4ln5h\" (UID: \"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\") " pod="openshift-network-operator/iptables-alerter-4ln5h" Dec 10 22:49:50 crc kubenswrapper[4791]: I1210 22:49:50.997068 4791 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-etc-kube\" (UniqueName: \"kubernetes.io/host-path/37a5e44f-9a88-4405-be8a-b645485e7312-host-etc-kube\") pod \"network-operator-58b4c7f79c-55gtf\" (UID: \"37a5e44f-9a88-4405-be8a-b645485e7312\") " pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" Dec 10 22:49:50 crc kubenswrapper[4791]: I1210 22:49:50.997093 4791 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/37a5e44f-9a88-4405-be8a-b645485e7312-metrics-tls\") pod \"network-operator-58b4c7f79c-55gtf\" (UID: \"37a5e44f-9a88-4405-be8a-b645485e7312\") " pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" Dec 10 22:49:51 crc kubenswrapper[4791]: I1210 22:49:50.997119 4791 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nginx-conf\" (UniqueName: \"kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 10 22:49:51 crc kubenswrapper[4791]: I1210 22:49:50.997144 4791 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/ef543e1b-8068-4ea3-b32a-61027b32e95d-env-overrides\") pod \"network-node-identity-vrzqb\" (UID: \"ef543e1b-8068-4ea3-b32a-61027b32e95d\") " pod="openshift-network-node-identity/network-node-identity-vrzqb" Dec 10 22:49:51 crc kubenswrapper[4791]: I1210 22:49:50.997169 4791 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"iptables-alerter-script\" (UniqueName: \"kubernetes.io/configmap/d75a4c96-2883-4a0b-bab2-0fab2b6c0b49-iptables-alerter-script\") pod \"iptables-alerter-4ln5h\" (UID: \"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\") " pod="openshift-network-operator/iptables-alerter-4ln5h" Dec 10 22:49:51 crc kubenswrapper[4791]: I1210 22:49:50.997199 4791 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/ef543e1b-8068-4ea3-b32a-61027b32e95d-webhook-cert\") pod \"network-node-identity-vrzqb\" (UID: \"ef543e1b-8068-4ea3-b32a-61027b32e95d\") " pod="openshift-network-node-identity/network-node-identity-vrzqb" Dec 10 22:49:51 crc kubenswrapper[4791]: I1210 22:49:50.997227 4791 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"networking-console-plugin-cert\" (UniqueName: \"kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 10 22:49:51 crc kubenswrapper[4791]: I1210 22:49:50.997255 4791 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-rczfb\" (UniqueName: \"kubernetes.io/projected/d75a4c96-2883-4a0b-bab2-0fab2b6c0b49-kube-api-access-rczfb\") pod \"iptables-alerter-4ln5h\" (UID: \"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\") " pod="openshift-network-operator/iptables-alerter-4ln5h" Dec 10 22:49:51 crc kubenswrapper[4791]: I1210 22:49:50.997281 4791 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"hosts-file\" (UniqueName: \"kubernetes.io/host-path/c0ab74e0-42a9-4f42-ac80-16030cf4ffff-hosts-file\") pod \"node-resolver-tchxk\" (UID: \"c0ab74e0-42a9-4f42-ac80-16030cf4ffff\") " pod="openshift-dns/node-resolver-tchxk" Dec 10 22:49:51 crc kubenswrapper[4791]: I1210 22:49:50.997304 4791 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-zn2qt\" (UniqueName: \"kubernetes.io/projected/c0ab74e0-42a9-4f42-ac80-16030cf4ffff-kube-api-access-zn2qt\") pod \"node-resolver-tchxk\" (UID: \"c0ab74e0-42a9-4f42-ac80-16030cf4ffff\") " pod="openshift-dns/node-resolver-tchxk" Dec 10 22:49:51 crc kubenswrapper[4791]: I1210 22:49:50.997328 4791 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-lrrpv\" (UniqueName: \"kubernetes.io/projected/b9f00f25-3fa1-4b63-9710-fd352224b01b-kube-api-access-lrrpv\") pod \"node-ca-vczjq\" (UID: \"b9f00f25-3fa1-4b63-9710-fd352224b01b\") " pod="openshift-image-registry/node-ca-vczjq" Dec 10 22:49:51 crc kubenswrapper[4791]: I1210 22:49:50.991752 4791 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-kube-api-access-ngvvp" (OuterVolumeSpecName: "kube-api-access-ngvvp") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "kube-api-access-ngvvp". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 10 22:49:51 crc kubenswrapper[4791]: I1210 22:49:50.991790 4791 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/6731426b-95fe-49ff-bb5f-40441049fde2-kube-api-access-x7zkh" (OuterVolumeSpecName: "kube-api-access-x7zkh") pod "6731426b-95fe-49ff-bb5f-40441049fde2" (UID: "6731426b-95fe-49ff-bb5f-40441049fde2"). InnerVolumeSpecName "kube-api-access-x7zkh". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 10 22:49:51 crc kubenswrapper[4791]: I1210 22:49:51.000621 4791 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-ocp-branding-template" (OuterVolumeSpecName: "v4-0-config-system-ocp-branding-template") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "v4-0-config-system-ocp-branding-template". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 10 22:49:51 crc kubenswrapper[4791]: I1210 22:49:51.000740 4791 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9-srv-cert" (OuterVolumeSpecName: "srv-cert") pod "f88749ec-7931-4ee7-b3fc-1ec5e11f92e9" (UID: "f88749ec-7931-4ee7-b3fc-1ec5e11f92e9"). InnerVolumeSpecName "srv-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 10 22:49:51 crc kubenswrapper[4791]: I1210 22:49:51.000746 4791 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/9d4552c7-cd75-42dd-8880-30dd377c49a4-kube-api-access-pcxfs" (OuterVolumeSpecName: "kube-api-access-pcxfs") pod "9d4552c7-cd75-42dd-8880-30dd377c49a4" (UID: "9d4552c7-cd75-42dd-8880-30dd377c49a4"). InnerVolumeSpecName "kube-api-access-pcxfs". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 10 22:49:51 crc kubenswrapper[4791]: I1210 22:49:51.000799 4791 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c-config" (OuterVolumeSpecName: "config") pod "210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c" (UID: "210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 10 22:49:51 crc kubenswrapper[4791]: I1210 22:49:51.000883 4791 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/44663579-783b-4372-86d6-acf235a62d72-kube-api-access-vt5rc" (OuterVolumeSpecName: "kube-api-access-vt5rc") pod "44663579-783b-4372-86d6-acf235a62d72" (UID: "44663579-783b-4372-86d6-acf235a62d72"). InnerVolumeSpecName "kube-api-access-vt5rc". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 10 22:49:51 crc kubenswrapper[4791]: I1210 22:49:51.000949 4791 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/09efc573-dbb6-4249-bd59-9b87aba8dd28-etcd-service-ca" (OuterVolumeSpecName: "etcd-service-ca") pod "09efc573-dbb6-4249-bd59-9b87aba8dd28" (UID: "09efc573-dbb6-4249-bd59-9b87aba8dd28"). InnerVolumeSpecName "etcd-service-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 10 22:49:51 crc kubenswrapper[4791]: I1210 22:49:51.000960 4791 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/09efc573-dbb6-4249-bd59-9b87aba8dd28-etcd-ca" (OuterVolumeSpecName: "etcd-ca") pod "09efc573-dbb6-4249-bd59-9b87aba8dd28" (UID: "09efc573-dbb6-4249-bd59-9b87aba8dd28"). InnerVolumeSpecName "etcd-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 10 22:49:51 crc kubenswrapper[4791]: I1210 22:49:51.001228 4791 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/25e176fe-21b4-4974-b1ed-c8b94f112a7f-signing-key" (OuterVolumeSpecName: "signing-key") pod "25e176fe-21b4-4974-b1ed-c8b94f112a7f" (UID: "25e176fe-21b4-4974-b1ed-c8b94f112a7f"). InnerVolumeSpecName "signing-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 10 22:49:51 crc kubenswrapper[4791]: I1210 22:49:51.001240 4791 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/5fe579f8-e8a6-4643-bce5-a661393c4dde-certs" (OuterVolumeSpecName: "certs") pod "5fe579f8-e8a6-4643-bce5-a661393c4dde" (UID: "5fe579f8-e8a6-4643-bce5-a661393c4dde"). InnerVolumeSpecName "certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 10 22:49:51 crc kubenswrapper[4791]: I1210 22:49:51.001276 4791 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c" (UID: "210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 10 22:49:51 crc kubenswrapper[4791]: I1210 22:49:51.001458 4791 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/6402fda4-df10-493c-b4e5-d0569419652d-images" (OuterVolumeSpecName: "images") pod "6402fda4-df10-493c-b4e5-d0569419652d" (UID: "6402fda4-df10-493c-b4e5-d0569419652d"). InnerVolumeSpecName "images". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 10 22:49:51 crc kubenswrapper[4791]: I1210 22:49:50.991930 4791 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/5225d0e4-402f-4861-b410-819f433b1803-kube-api-access-9xfj7" (OuterVolumeSpecName: "kube-api-access-9xfj7") pod "5225d0e4-402f-4861-b410-819f433b1803" (UID: "5225d0e4-402f-4861-b410-819f433b1803"). InnerVolumeSpecName "kube-api-access-9xfj7". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 10 22:49:51 crc kubenswrapper[4791]: I1210 22:49:50.991937 4791 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/a31745f5-9847-4afe-82a5-3161cc66ca93-bound-sa-token" (OuterVolumeSpecName: "bound-sa-token") pod "a31745f5-9847-4afe-82a5-3161cc66ca93" (UID: "a31745f5-9847-4afe-82a5-3161cc66ca93"). InnerVolumeSpecName "bound-sa-token". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 10 22:49:51 crc kubenswrapper[4791]: I1210 22:49:50.991990 4791 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/6402fda4-df10-493c-b4e5-d0569419652d-machine-api-operator-tls" (OuterVolumeSpecName: "machine-api-operator-tls") pod "6402fda4-df10-493c-b4e5-d0569419652d" (UID: "6402fda4-df10-493c-b4e5-d0569419652d"). InnerVolumeSpecName "machine-api-operator-tls". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 10 22:49:51 crc kubenswrapper[4791]: I1210 22:49:50.992017 4791 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/0b78653f-4ff9-4508-8672-245ed9b561e3-kube-api-access" (OuterVolumeSpecName: "kube-api-access") pod "0b78653f-4ff9-4508-8672-245ed9b561e3" (UID: "0b78653f-4ff9-4508-8672-245ed9b561e3"). InnerVolumeSpecName "kube-api-access". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 10 22:49:51 crc kubenswrapper[4791]: I1210 22:49:50.992077 4791 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/3ab1a177-2de0-46d9-b765-d0d0649bb42e-package-server-manager-serving-cert" (OuterVolumeSpecName: "package-server-manager-serving-cert") pod "3ab1a177-2de0-46d9-b765-d0d0649bb42e" (UID: "3ab1a177-2de0-46d9-b765-d0d0649bb42e"). InnerVolumeSpecName "package-server-manager-serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 10 22:49:51 crc kubenswrapper[4791]: I1210 22:49:50.992151 4791 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/1d611f23-29be-4491-8495-bee1670e935f-kube-api-access-bf2bz" (OuterVolumeSpecName: "kube-api-access-bf2bz") pod "1d611f23-29be-4491-8495-bee1670e935f" (UID: "1d611f23-29be-4491-8495-bee1670e935f"). InnerVolumeSpecName "kube-api-access-bf2bz". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 10 22:49:51 crc kubenswrapper[4791]: I1210 22:49:50.992226 4791 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-router-certs" (OuterVolumeSpecName: "v4-0-config-system-router-certs") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "v4-0-config-system-router-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 10 22:49:51 crc kubenswrapper[4791]: I1210 22:49:50.992381 4791 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/c03ee662-fb2f-4fc4-a2c1-af487c19d254-default-certificate" (OuterVolumeSpecName: "default-certificate") pod "c03ee662-fb2f-4fc4-a2c1-af487c19d254" (UID: "c03ee662-fb2f-4fc4-a2c1-af487c19d254"). InnerVolumeSpecName "default-certificate". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 10 22:49:51 crc kubenswrapper[4791]: I1210 22:49:50.992505 4791 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-trusted-ca-bundle" (OuterVolumeSpecName: "trusted-ca-bundle") pod "09ae3b1a-e8e7-4524-b54b-61eab6f9239a" (UID: "09ae3b1a-e8e7-4524-b54b-61eab6f9239a"). InnerVolumeSpecName "trusted-ca-bundle". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 10 22:49:51 crc kubenswrapper[4791]: I1210 22:49:50.992548 4791 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/bf126b07-da06-4140-9a57-dfd54fc6b486-bound-sa-token" (OuterVolumeSpecName: "bound-sa-token") pod "bf126b07-da06-4140-9a57-dfd54fc6b486" (UID: "bf126b07-da06-4140-9a57-dfd54fc6b486"). InnerVolumeSpecName "bound-sa-token". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 10 22:49:51 crc kubenswrapper[4791]: I1210 22:49:50.992623 4791 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/bf126b07-da06-4140-9a57-dfd54fc6b486-trusted-ca" (OuterVolumeSpecName: "trusted-ca") pod "bf126b07-da06-4140-9a57-dfd54fc6b486" (UID: "bf126b07-da06-4140-9a57-dfd54fc6b486"). InnerVolumeSpecName "trusted-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 10 22:49:51 crc kubenswrapper[4791]: I1210 22:49:50.992648 4791 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-trusted-ca-bundle" (OuterVolumeSpecName: "trusted-ca-bundle") pod "43509403-f426-496e-be36-56cef71462f5" (UID: "43509403-f426-496e-be36-56cef71462f5"). InnerVolumeSpecName "trusted-ca-bundle". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 10 22:49:51 crc kubenswrapper[4791]: I1210 22:49:50.992723 4791 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/96b93a3a-6083-4aea-8eab-fe1aa8245ad9-metrics-tls" (OuterVolumeSpecName: "metrics-tls") pod "96b93a3a-6083-4aea-8eab-fe1aa8245ad9" (UID: "96b93a3a-6083-4aea-8eab-fe1aa8245ad9"). InnerVolumeSpecName "metrics-tls". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 10 22:49:51 crc kubenswrapper[4791]: I1210 22:49:50.992739 4791 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-apiservice-cert" (OuterVolumeSpecName: "apiservice-cert") pod "308be0ea-9f5f-4b29-aeb1-5abd31a0b17b" (UID: "308be0ea-9f5f-4b29-aeb1-5abd31a0b17b"). InnerVolumeSpecName "apiservice-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 10 22:49:51 crc kubenswrapper[4791]: I1210 22:49:50.992752 4791 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-cliconfig" (OuterVolumeSpecName: "v4-0-config-system-cliconfig") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "v4-0-config-system-cliconfig". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 10 22:49:51 crc kubenswrapper[4791]: I1210 22:49:50.992829 4791 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/496e6271-fb68-4057-954e-a0d97a4afa3f-kube-api-access" (OuterVolumeSpecName: "kube-api-access") pod "496e6271-fb68-4057-954e-a0d97a4afa3f" (UID: "496e6271-fb68-4057-954e-a0d97a4afa3f"). InnerVolumeSpecName "kube-api-access". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 10 22:49:51 crc kubenswrapper[4791]: I1210 22:49:50.992864 4791 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/01ab3dd5-8196-46d0-ad33-122e2ca51def-kube-api-access-w7l8j" (OuterVolumeSpecName: "kube-api-access-w7l8j") pod "01ab3dd5-8196-46d0-ad33-122e2ca51def" (UID: "01ab3dd5-8196-46d0-ad33-122e2ca51def"). InnerVolumeSpecName "kube-api-access-w7l8j". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 10 22:49:51 crc kubenswrapper[4791]: I1210 22:49:50.992873 4791 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/09efc573-dbb6-4249-bd59-9b87aba8dd28-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "09efc573-dbb6-4249-bd59-9b87aba8dd28" (UID: "09efc573-dbb6-4249-bd59-9b87aba8dd28"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 10 22:49:51 crc kubenswrapper[4791]: I1210 22:49:50.992915 4791 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-etcd-serving-ca" (OuterVolumeSpecName: "etcd-serving-ca") pod "1bf7eb37-55a3-4c65-b768-a94c82151e69" (UID: "1bf7eb37-55a3-4c65-b768-a94c82151e69"). InnerVolumeSpecName "etcd-serving-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 10 22:49:51 crc kubenswrapper[4791]: I1210 22:49:50.992939 4791 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/31d8b7a1-420e-4252-a5b7-eebe8a111292-proxy-tls" (OuterVolumeSpecName: "proxy-tls") pod "31d8b7a1-420e-4252-a5b7-eebe8a111292" (UID: "31d8b7a1-420e-4252-a5b7-eebe8a111292"). InnerVolumeSpecName "proxy-tls". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 10 22:49:51 crc kubenswrapper[4791]: I1210 22:49:50.993051 4791 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-template-error" (OuterVolumeSpecName: "v4-0-config-user-template-error") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "v4-0-config-user-template-error". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 10 22:49:51 crc kubenswrapper[4791]: I1210 22:49:50.993073 4791 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/5fe579f8-e8a6-4643-bce5-a661393c4dde-kube-api-access-fcqwp" (OuterVolumeSpecName: "kube-api-access-fcqwp") pod "5fe579f8-e8a6-4643-bce5-a661393c4dde" (UID: "5fe579f8-e8a6-4643-bce5-a661393c4dde"). InnerVolumeSpecName "kube-api-access-fcqwp". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 10 22:49:51 crc kubenswrapper[4791]: I1210 22:49:51.001956 4791 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b-kube-api-access-w4xd4" (OuterVolumeSpecName: "kube-api-access-w4xd4") pod "8cea82b4-6893-4ddc-af9f-1bb5ae425c5b" (UID: "8cea82b4-6893-4ddc-af9f-1bb5ae425c5b"). InnerVolumeSpecName "kube-api-access-w4xd4". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 10 22:49:51 crc kubenswrapper[4791]: I1210 22:49:51.001958 4791 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-webhook-cert" (OuterVolumeSpecName: "webhook-cert") pod "308be0ea-9f5f-4b29-aeb1-5abd31a0b17b" (UID: "308be0ea-9f5f-4b29-aeb1-5abd31a0b17b"). InnerVolumeSpecName "webhook-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 10 22:49:51 crc kubenswrapper[4791]: I1210 22:49:51.002491 4791 swap_util.go:74] "error creating dir to test if tmpfs noswap is enabled. Assuming not supported" mount path="" error="stat /var/lib/kubelet/plugins/kubernetes.io/empty-dir: no such file or directory" Dec 10 22:49:51 crc kubenswrapper[4791]: I1210 22:49:51.005468 4791 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-10T22:49:50Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-10T22:49:50Z\\\",\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Dec 10 22:49:51 crc kubenswrapper[4791]: I1210 22:49:51.006425 4791 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovnkube-identity-cm\" (UniqueName: \"kubernetes.io/configmap/ef543e1b-8068-4ea3-b32a-61027b32e95d-ovnkube-identity-cm\") pod \"network-node-identity-vrzqb\" (UID: \"ef543e1b-8068-4ea3-b32a-61027b32e95d\") " pod="openshift-network-node-identity/network-node-identity-vrzqb" Dec 10 22:49:51 crc kubenswrapper[4791]: I1210 22:49:51.002517 4791 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/7583ce53-e0fe-4a16-9e4d-50516596a136-proxy-ca-bundles" (OuterVolumeSpecName: "proxy-ca-bundles") pod "7583ce53-e0fe-4a16-9e4d-50516596a136" (UID: "7583ce53-e0fe-4a16-9e4d-50516596a136"). InnerVolumeSpecName "proxy-ca-bundles". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 10 22:49:51 crc kubenswrapper[4791]: I1210 22:49:51.008903 4791 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/a31745f5-9847-4afe-82a5-3161cc66ca93-kube-api-access-lz9wn" (OuterVolumeSpecName: "kube-api-access-lz9wn") pod "a31745f5-9847-4afe-82a5-3161cc66ca93" (UID: "a31745f5-9847-4afe-82a5-3161cc66ca93"). InnerVolumeSpecName "kube-api-access-lz9wn". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 10 22:49:51 crc kubenswrapper[4791]: I1210 22:49:51.003185 4791 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/5441d097-087c-4d9a-baa8-b210afa90fc9-client-ca" (OuterVolumeSpecName: "client-ca") pod "5441d097-087c-4d9a-baa8-b210afa90fc9" (UID: "5441d097-087c-4d9a-baa8-b210afa90fc9"). InnerVolumeSpecName "client-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 10 22:49:51 crc kubenswrapper[4791]: I1210 22:49:50.993136 4791 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/a0128f3a-b052-44ed-a84e-c4c8aaf17c13-samples-operator-tls" (OuterVolumeSpecName: "samples-operator-tls") pod "a0128f3a-b052-44ed-a84e-c4c8aaf17c13" (UID: "a0128f3a-b052-44ed-a84e-c4c8aaf17c13"). InnerVolumeSpecName "samples-operator-tls". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 10 22:49:51 crc kubenswrapper[4791]: I1210 22:49:50.993221 4791 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/43509403-f426-496e-be36-56cef71462f5-console-serving-cert" (OuterVolumeSpecName: "console-serving-cert") pod "43509403-f426-496e-be36-56cef71462f5" (UID: "43509403-f426-496e-be36-56cef71462f5"). InnerVolumeSpecName "console-serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 10 22:49:51 crc kubenswrapper[4791]: I1210 22:49:50.993279 4791 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/5fe579f8-e8a6-4643-bce5-a661393c4dde-node-bootstrap-token" (OuterVolumeSpecName: "node-bootstrap-token") pod "5fe579f8-e8a6-4643-bce5-a661393c4dde" (UID: "5fe579f8-e8a6-4643-bce5-a661393c4dde"). InnerVolumeSpecName "node-bootstrap-token". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 10 22:49:51 crc kubenswrapper[4791]: I1210 22:49:51.009718 4791 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/bf126b07-da06-4140-9a57-dfd54fc6b486-image-registry-operator-tls" (OuterVolumeSpecName: "image-registry-operator-tls") pod "bf126b07-da06-4140-9a57-dfd54fc6b486" (UID: "bf126b07-da06-4140-9a57-dfd54fc6b486"). InnerVolumeSpecName "image-registry-operator-tls". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 10 22:49:51 crc kubenswrapper[4791]: I1210 22:49:50.993288 4791 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "8cea82b4-6893-4ddc-af9f-1bb5ae425c5b" (UID: "8cea82b4-6893-4ddc-af9f-1bb5ae425c5b"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 10 22:49:51 crc kubenswrapper[4791]: I1210 22:49:50.993333 4791 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/22c825df-677d-4ca6-82db-3454ed06e783-config" (OuterVolumeSpecName: "config") pod "22c825df-677d-4ca6-82db-3454ed06e783" (UID: "22c825df-677d-4ca6-82db-3454ed06e783"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 10 22:49:51 crc kubenswrapper[4791]: I1210 22:49:50.993389 4791 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/efdd0498-1daa-4136-9a4a-3b948c2293fc-kube-api-access-fqsjt" (OuterVolumeSpecName: "kube-api-access-fqsjt") pod "efdd0498-1daa-4136-9a4a-3b948c2293fc" (UID: "efdd0498-1daa-4136-9a4a-3b948c2293fc"). InnerVolumeSpecName "kube-api-access-fqsjt". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 10 22:49:51 crc kubenswrapper[4791]: I1210 22:49:50.993481 4791 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/b6312bbd-5731-4ea0-a20f-81d5a57df44a-kube-api-access-249nr" (OuterVolumeSpecName: "kube-api-access-249nr") pod "b6312bbd-5731-4ea0-a20f-81d5a57df44a" (UID: "b6312bbd-5731-4ea0-a20f-81d5a57df44a"). InnerVolumeSpecName "kube-api-access-249nr". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 10 22:49:51 crc kubenswrapper[4791]: I1210 22:49:50.993533 4791 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/8f668bae-612b-4b75-9490-919e737c6a3b-installation-pull-secrets" (OuterVolumeSpecName: "installation-pull-secrets") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b"). InnerVolumeSpecName "installation-pull-secrets". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 10 22:49:51 crc kubenswrapper[4791]: I1210 22:49:50.993538 4791 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/496e6271-fb68-4057-954e-a0d97a4afa3f-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "496e6271-fb68-4057-954e-a0d97a4afa3f" (UID: "496e6271-fb68-4057-954e-a0d97a4afa3f"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 10 22:49:51 crc kubenswrapper[4791]: I1210 22:49:50.993535 4791 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/6ea678ab-3438-413e-bfe3-290ae7725660-ovnkube-script-lib" (OuterVolumeSpecName: "ovnkube-script-lib") pod "6ea678ab-3438-413e-bfe3-290ae7725660" (UID: "6ea678ab-3438-413e-bfe3-290ae7725660"). InnerVolumeSpecName "ovnkube-script-lib". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 10 22:49:51 crc kubenswrapper[4791]: I1210 22:49:50.993609 4791 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/925f1c65-6136-48ba-85aa-3a3b50560753-ovnkube-config" (OuterVolumeSpecName: "ovnkube-config") pod "925f1c65-6136-48ba-85aa-3a3b50560753" (UID: "925f1c65-6136-48ba-85aa-3a3b50560753"). InnerVolumeSpecName "ovnkube-config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 10 22:49:51 crc kubenswrapper[4791]: I1210 22:49:50.993628 4791 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d-available-featuregates" (OuterVolumeSpecName: "available-featuregates") pod "bc5039c0-ea34-426b-a2b7-fbbc87b49a6d" (UID: "bc5039c0-ea34-426b-a2b7-fbbc87b49a6d"). InnerVolumeSpecName "available-featuregates". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 10 22:49:51 crc kubenswrapper[4791]: I1210 22:49:50.993669 4791 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-etcd-client" (OuterVolumeSpecName: "etcd-client") pod "09ae3b1a-e8e7-4524-b54b-61eab6f9239a" (UID: "09ae3b1a-e8e7-4524-b54b-61eab6f9239a"). InnerVolumeSpecName "etcd-client". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 10 22:49:51 crc kubenswrapper[4791]: I1210 22:49:50.993728 4791 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/0b574797-001e-440a-8f4e-c0be86edad0f-kube-api-access-lzf88" (OuterVolumeSpecName: "kube-api-access-lzf88") pod "0b574797-001e-440a-8f4e-c0be86edad0f" (UID: "0b574797-001e-440a-8f4e-c0be86edad0f"). InnerVolumeSpecName "kube-api-access-lzf88". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 10 22:49:51 crc kubenswrapper[4791]: I1210 22:49:50.993733 4791 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/4bb40260-dbaa-4fb0-84df-5e680505d512-kube-api-access-2w9zh" (OuterVolumeSpecName: "kube-api-access-2w9zh") pod "4bb40260-dbaa-4fb0-84df-5e680505d512" (UID: "4bb40260-dbaa-4fb0-84df-5e680505d512"). InnerVolumeSpecName "kube-api-access-2w9zh". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 10 22:49:51 crc kubenswrapper[4791]: I1210 22:49:50.993802 4791 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/925f1c65-6136-48ba-85aa-3a3b50560753-ovn-control-plane-metrics-cert" (OuterVolumeSpecName: "ovn-control-plane-metrics-cert") pod "925f1c65-6136-48ba-85aa-3a3b50560753" (UID: "925f1c65-6136-48ba-85aa-3a3b50560753"). InnerVolumeSpecName "ovn-control-plane-metrics-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 10 22:49:51 crc kubenswrapper[4791]: I1210 22:49:50.993844 4791 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/5441d097-087c-4d9a-baa8-b210afa90fc9-kube-api-access-2d4wz" (OuterVolumeSpecName: "kube-api-access-2d4wz") pod "5441d097-087c-4d9a-baa8-b210afa90fc9" (UID: "5441d097-087c-4d9a-baa8-b210afa90fc9"). InnerVolumeSpecName "kube-api-access-2d4wz". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 10 22:49:51 crc kubenswrapper[4791]: I1210 22:49:50.993904 4791 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/8f668bae-612b-4b75-9490-919e737c6a3b-bound-sa-token" (OuterVolumeSpecName: "bound-sa-token") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b"). InnerVolumeSpecName "bound-sa-token". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 10 22:49:51 crc kubenswrapper[4791]: I1210 22:49:50.993916 4791 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59-kube-api-access-wxkg8" (OuterVolumeSpecName: "kube-api-access-wxkg8") pod "3cb93b32-e0ae-4377-b9c8-fdb9842c6d59" (UID: "3cb93b32-e0ae-4377-b9c8-fdb9842c6d59"). InnerVolumeSpecName "kube-api-access-wxkg8". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 10 22:49:51 crc kubenswrapper[4791]: I1210 22:49:50.993998 4791 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-service-ca" (OuterVolumeSpecName: "service-ca") pod "43509403-f426-496e-be36-56cef71462f5" (UID: "43509403-f426-496e-be36-56cef71462f5"). InnerVolumeSpecName "service-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 10 22:49:51 crc kubenswrapper[4791]: I1210 22:49:50.994064 4791 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d-utilities" (OuterVolumeSpecName: "utilities") pod "b11524ee-3fca-4b1b-9cdf-6da289fdbc7d" (UID: "b11524ee-3fca-4b1b-9cdf-6da289fdbc7d"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 10 22:49:51 crc kubenswrapper[4791]: I1210 22:49:50.994098 4791 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/87cf06ed-a83f-41a7-828d-70653580a8cb-config-volume" (OuterVolumeSpecName: "config-volume") pod "87cf06ed-a83f-41a7-828d-70653580a8cb" (UID: "87cf06ed-a83f-41a7-828d-70653580a8cb"). InnerVolumeSpecName "config-volume". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 10 22:49:51 crc kubenswrapper[4791]: I1210 22:49:50.994142 4791 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/20b0d48f-5fd6-431c-a545-e3c800c7b866-cert" (OuterVolumeSpecName: "cert") pod "20b0d48f-5fd6-431c-a545-e3c800c7b866" (UID: "20b0d48f-5fd6-431c-a545-e3c800c7b866"). InnerVolumeSpecName "cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 10 22:49:51 crc kubenswrapper[4791]: I1210 22:49:50.994189 4791 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/6509e943-70c6-444c-bc41-48a544e36fbd-trusted-ca-bundle" (OuterVolumeSpecName: "trusted-ca-bundle") pod "6509e943-70c6-444c-bc41-48a544e36fbd" (UID: "6509e943-70c6-444c-bc41-48a544e36fbd"). InnerVolumeSpecName "trusted-ca-bundle". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 10 22:49:51 crc kubenswrapper[4791]: I1210 22:49:50.994309 4791 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-service-ca" (OuterVolumeSpecName: "v4-0-config-system-service-ca") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "v4-0-config-system-service-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 10 22:49:51 crc kubenswrapper[4791]: I1210 22:49:50.994327 4791 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/57a731c4-ef35-47a8-b875-bfb08a7f8011-kube-api-access-cfbct" (OuterVolumeSpecName: "kube-api-access-cfbct") pod "57a731c4-ef35-47a8-b875-bfb08a7f8011" (UID: "57a731c4-ef35-47a8-b875-bfb08a7f8011"). InnerVolumeSpecName "kube-api-access-cfbct". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 10 22:49:51 crc kubenswrapper[4791]: I1210 22:49:50.994423 4791 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-session" (OuterVolumeSpecName: "v4-0-config-system-session") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "v4-0-config-system-session". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 10 22:49:51 crc kubenswrapper[4791]: I1210 22:49:50.994506 4791 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-template-login" (OuterVolumeSpecName: "v4-0-config-user-template-login") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "v4-0-config-user-template-login". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 10 22:49:51 crc kubenswrapper[4791]: I1210 22:49:50.994533 4791 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/1386a44e-36a2-460c-96d0-0359d2b6f0f5-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "1386a44e-36a2-460c-96d0-0359d2b6f0f5" (UID: "1386a44e-36a2-460c-96d0-0359d2b6f0f5"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 10 22:49:51 crc kubenswrapper[4791]: I1210 22:49:50.994555 4791 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-image-import-ca" (OuterVolumeSpecName: "image-import-ca") pod "1bf7eb37-55a3-4c65-b768-a94c82151e69" (UID: "1bf7eb37-55a3-4c65-b768-a94c82151e69"). InnerVolumeSpecName "image-import-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 10 22:49:51 crc kubenswrapper[4791]: I1210 22:49:50.994568 4791 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-encryption-config" (OuterVolumeSpecName: "encryption-config") pod "09ae3b1a-e8e7-4524-b54b-61eab6f9239a" (UID: "09ae3b1a-e8e7-4524-b54b-61eab6f9239a"). InnerVolumeSpecName "encryption-config". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 10 22:49:51 crc kubenswrapper[4791]: I1210 22:49:50.994712 4791 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9-profile-collector-cert" (OuterVolumeSpecName: "profile-collector-cert") pod "f88749ec-7931-4ee7-b3fc-1ec5e11f92e9" (UID: "f88749ec-7931-4ee7-b3fc-1ec5e11f92e9"). InnerVolumeSpecName "profile-collector-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 10 22:49:51 crc kubenswrapper[4791]: I1210 22:49:50.994718 4791 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/0b78653f-4ff9-4508-8672-245ed9b561e3-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "0b78653f-4ff9-4508-8672-245ed9b561e3" (UID: "0b78653f-4ff9-4508-8672-245ed9b561e3"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 10 22:49:51 crc kubenswrapper[4791]: I1210 22:49:50.994740 4791 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/96b93a3a-6083-4aea-8eab-fe1aa8245ad9-kube-api-access-nzwt7" (OuterVolumeSpecName: "kube-api-access-nzwt7") pod "96b93a3a-6083-4aea-8eab-fe1aa8245ad9" (UID: "96b93a3a-6083-4aea-8eab-fe1aa8245ad9"). InnerVolumeSpecName "kube-api-access-nzwt7". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 10 22:49:51 crc kubenswrapper[4791]: I1210 22:49:50.994787 4791 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-kube-api-access-zkvpv" (OuterVolumeSpecName: "kube-api-access-zkvpv") pod "09ae3b1a-e8e7-4524-b54b-61eab6f9239a" (UID: "09ae3b1a-e8e7-4524-b54b-61eab6f9239a"). InnerVolumeSpecName "kube-api-access-zkvpv". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 10 22:49:51 crc kubenswrapper[4791]: I1210 22:49:50.994884 4791 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/9d4552c7-cd75-42dd-8880-30dd377c49a4-config" (OuterVolumeSpecName: "config") pod "9d4552c7-cd75-42dd-8880-30dd377c49a4" (UID: "9d4552c7-cd75-42dd-8880-30dd377c49a4"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 10 22:49:51 crc kubenswrapper[4791]: I1210 22:49:50.994891 4791 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/b6312bbd-5731-4ea0-a20f-81d5a57df44a-profile-collector-cert" (OuterVolumeSpecName: "profile-collector-cert") pod "b6312bbd-5731-4ea0-a20f-81d5a57df44a" (UID: "b6312bbd-5731-4ea0-a20f-81d5a57df44a"). InnerVolumeSpecName "profile-collector-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 10 22:49:51 crc kubenswrapper[4791]: I1210 22:49:50.995156 4791 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/57a731c4-ef35-47a8-b875-bfb08a7f8011-utilities" (OuterVolumeSpecName: "utilities") pod "57a731c4-ef35-47a8-b875-bfb08a7f8011" (UID: "57a731c4-ef35-47a8-b875-bfb08a7f8011"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 10 22:49:51 crc kubenswrapper[4791]: I1210 22:49:50.995427 4791 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9-kube-api-access-dbsvg" (OuterVolumeSpecName: "kube-api-access-dbsvg") pod "f88749ec-7931-4ee7-b3fc-1ec5e11f92e9" (UID: "f88749ec-7931-4ee7-b3fc-1ec5e11f92e9"). InnerVolumeSpecName "kube-api-access-dbsvg". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 10 22:49:51 crc kubenswrapper[4791]: I1210 22:49:50.995611 4791 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/0b574797-001e-440a-8f4e-c0be86edad0f-mcc-auth-proxy-config" (OuterVolumeSpecName: "mcc-auth-proxy-config") pod "0b574797-001e-440a-8f4e-c0be86edad0f" (UID: "0b574797-001e-440a-8f4e-c0be86edad0f"). InnerVolumeSpecName "mcc-auth-proxy-config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 10 22:49:51 crc kubenswrapper[4791]: I1210 22:49:50.996182 4791 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/8f668bae-612b-4b75-9490-919e737c6a3b-trusted-ca" (OuterVolumeSpecName: "trusted-ca") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b"). InnerVolumeSpecName "trusted-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 10 22:49:51 crc kubenswrapper[4791]: I1210 22:49:50.996195 4791 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/1386a44e-36a2-460c-96d0-0359d2b6f0f5-kube-api-access" (OuterVolumeSpecName: "kube-api-access") pod "1386a44e-36a2-460c-96d0-0359d2b6f0f5" (UID: "1386a44e-36a2-460c-96d0-0359d2b6f0f5"). InnerVolumeSpecName "kube-api-access". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 10 22:49:51 crc kubenswrapper[4791]: I1210 22:49:50.996420 4791 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/925f1c65-6136-48ba-85aa-3a3b50560753-kube-api-access-s4n52" (OuterVolumeSpecName: "kube-api-access-s4n52") pod "925f1c65-6136-48ba-85aa-3a3b50560753" (UID: "925f1c65-6136-48ba-85aa-3a3b50560753"). InnerVolumeSpecName "kube-api-access-s4n52". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 10 22:49:51 crc kubenswrapper[4791]: I1210 22:49:50.996546 4791 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d-kube-api-access-x2m85" (OuterVolumeSpecName: "kube-api-access-x2m85") pod "cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d" (UID: "cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d"). InnerVolumeSpecName "kube-api-access-x2m85". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 10 22:49:51 crc kubenswrapper[4791]: I1210 22:49:50.996775 4791 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/fda69060-fa79-4696-b1a6-7980f124bf7c-mcd-auth-proxy-config" (OuterVolumeSpecName: "mcd-auth-proxy-config") pod "fda69060-fa79-4696-b1a6-7980f124bf7c" (UID: "fda69060-fa79-4696-b1a6-7980f124bf7c"). InnerVolumeSpecName "mcd-auth-proxy-config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 10 22:49:51 crc kubenswrapper[4791]: I1210 22:49:50.997799 4791 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/925f1c65-6136-48ba-85aa-3a3b50560753-env-overrides" (OuterVolumeSpecName: "env-overrides") pod "925f1c65-6136-48ba-85aa-3a3b50560753" (UID: "925f1c65-6136-48ba-85aa-3a3b50560753"). InnerVolumeSpecName "env-overrides". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 10 22:49:51 crc kubenswrapper[4791]: I1210 22:49:50.998880 4791 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/3ab1a177-2de0-46d9-b765-d0d0649bb42e-kube-api-access-4d4hj" (OuterVolumeSpecName: "kube-api-access-4d4hj") pod "3ab1a177-2de0-46d9-b765-d0d0649bb42e" (UID: "3ab1a177-2de0-46d9-b765-d0d0649bb42e"). InnerVolumeSpecName "kube-api-access-4d4hj". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 10 22:49:51 crc kubenswrapper[4791]: I1210 22:49:50.999462 4791 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "bc5039c0-ea34-426b-a2b7-fbbc87b49a6d" (UID: "bc5039c0-ea34-426b-a2b7-fbbc87b49a6d"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 10 22:49:51 crc kubenswrapper[4791]: I1210 22:49:50.999854 4791 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/6509e943-70c6-444c-bc41-48a544e36fbd-kube-api-access-6g6sz" (OuterVolumeSpecName: "kube-api-access-6g6sz") pod "6509e943-70c6-444c-bc41-48a544e36fbd" (UID: "6509e943-70c6-444c-bc41-48a544e36fbd"). InnerVolumeSpecName "kube-api-access-6g6sz". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 10 22:49:51 crc kubenswrapper[4791]: I1210 22:49:50.999953 4791 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/1d611f23-29be-4491-8495-bee1670e935f-utilities" (OuterVolumeSpecName: "utilities") pod "1d611f23-29be-4491-8495-bee1670e935f" (UID: "1d611f23-29be-4491-8495-bee1670e935f"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 10 22:49:51 crc kubenswrapper[4791]: I1210 22:49:51.000076 4791 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/6731426b-95fe-49ff-bb5f-40441049fde2-control-plane-machine-set-operator-tls" (OuterVolumeSpecName: "control-plane-machine-set-operator-tls") pod "6731426b-95fe-49ff-bb5f-40441049fde2" (UID: "6731426b-95fe-49ff-bb5f-40441049fde2"). InnerVolumeSpecName "control-plane-machine-set-operator-tls". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 10 22:49:51 crc kubenswrapper[4791]: I1210 22:49:51.000085 4791 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/c03ee662-fb2f-4fc4-a2c1-af487c19d254-metrics-certs" (OuterVolumeSpecName: "metrics-certs") pod "c03ee662-fb2f-4fc4-a2c1-af487c19d254" (UID: "c03ee662-fb2f-4fc4-a2c1-af487c19d254"). InnerVolumeSpecName "metrics-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 10 22:49:51 crc kubenswrapper[4791]: I1210 22:49:51.001731 4791 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/8f668bae-612b-4b75-9490-919e737c6a3b-registry-certificates" (OuterVolumeSpecName: "registry-certificates") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b"). InnerVolumeSpecName "registry-certificates". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 10 22:49:51 crc kubenswrapper[4791]: I1210 22:49:51.003195 4791 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/7bb08738-c794-4ee8-9972-3a62ca171029-cni-sysctl-allowlist" (OuterVolumeSpecName: "cni-sysctl-allowlist") pod "7bb08738-c794-4ee8-9972-3a62ca171029" (UID: "7bb08738-c794-4ee8-9972-3a62ca171029"). InnerVolumeSpecName "cni-sysctl-allowlist". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 10 22:49:51 crc kubenswrapper[4791]: I1210 22:49:51.003520 4791 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/49ef4625-1d3a-4a9f-b595-c2433d32326d-kube-api-access-pjr6v" (OuterVolumeSpecName: "kube-api-access-pjr6v") pod "49ef4625-1d3a-4a9f-b595-c2433d32326d" (UID: "49ef4625-1d3a-4a9f-b595-c2433d32326d"). InnerVolumeSpecName "kube-api-access-pjr6v". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 10 22:49:51 crc kubenswrapper[4791]: I1210 22:49:51.003708 4791 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-tmpfs" (OuterVolumeSpecName: "tmpfs") pod "308be0ea-9f5f-4b29-aeb1-5abd31a0b17b" (UID: "308be0ea-9f5f-4b29-aeb1-5abd31a0b17b"). InnerVolumeSpecName "tmpfs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 10 22:49:51 crc kubenswrapper[4791]: I1210 22:49:51.003996 4791 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/01ab3dd5-8196-46d0-ad33-122e2ca51def-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "01ab3dd5-8196-46d0-ad33-122e2ca51def" (UID: "01ab3dd5-8196-46d0-ad33-122e2ca51def"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 10 22:49:51 crc kubenswrapper[4791]: I1210 22:49:51.004178 4791 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/25e176fe-21b4-4974-b1ed-c8b94f112a7f-signing-cabundle" (OuterVolumeSpecName: "signing-cabundle") pod "25e176fe-21b4-4974-b1ed-c8b94f112a7f" (UID: "25e176fe-21b4-4974-b1ed-c8b94f112a7f"). InnerVolumeSpecName "signing-cabundle". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 10 22:49:51 crc kubenswrapper[4791]: I1210 22:49:51.004271 4791 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/b6cd30de-2eeb-49a2-ab40-9167f4560ff5-kube-api-access-pj782" (OuterVolumeSpecName: "kube-api-access-pj782") pod "b6cd30de-2eeb-49a2-ab40-9167f4560ff5" (UID: "b6cd30de-2eeb-49a2-ab40-9167f4560ff5"). InnerVolumeSpecName "kube-api-access-pj782". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 10 22:49:51 crc kubenswrapper[4791]: I1210 22:49:51.004547 4791 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59-serviceca" (OuterVolumeSpecName: "serviceca") pod "3cb93b32-e0ae-4377-b9c8-fdb9842c6d59" (UID: "3cb93b32-e0ae-4377-b9c8-fdb9842c6d59"). InnerVolumeSpecName "serviceca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 10 22:49:51 crc kubenswrapper[4791]: I1210 22:49:51.004555 4791 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/fda69060-fa79-4696-b1a6-7980f124bf7c-kube-api-access-xcgwh" (OuterVolumeSpecName: "kube-api-access-xcgwh") pod "fda69060-fa79-4696-b1a6-7980f124bf7c" (UID: "fda69060-fa79-4696-b1a6-7980f124bf7c"). InnerVolumeSpecName "kube-api-access-xcgwh". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 10 22:49:51 crc kubenswrapper[4791]: I1210 22:49:51.004798 4791 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b-config" (OuterVolumeSpecName: "config") pod "8cea82b4-6893-4ddc-af9f-1bb5ae425c5b" (UID: "8cea82b4-6893-4ddc-af9f-1bb5ae425c5b"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 10 22:49:51 crc kubenswrapper[4791]: I1210 22:49:51.005045 4791 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/0b78653f-4ff9-4508-8672-245ed9b561e3-service-ca" (OuterVolumeSpecName: "service-ca") pod "0b78653f-4ff9-4508-8672-245ed9b561e3" (UID: "0b78653f-4ff9-4508-8672-245ed9b561e3"). InnerVolumeSpecName "service-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 10 22:49:51 crc kubenswrapper[4791]: I1210 22:49:51.005078 4791 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/01ab3dd5-8196-46d0-ad33-122e2ca51def-config" (OuterVolumeSpecName: "config") pod "01ab3dd5-8196-46d0-ad33-122e2ca51def" (UID: "01ab3dd5-8196-46d0-ad33-122e2ca51def"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 10 22:49:51 crc kubenswrapper[4791]: I1210 22:49:51.005099 4791 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/87cf06ed-a83f-41a7-828d-70653580a8cb-metrics-tls" (OuterVolumeSpecName: "metrics-tls") pod "87cf06ed-a83f-41a7-828d-70653580a8cb" (UID: "87cf06ed-a83f-41a7-828d-70653580a8cb"). InnerVolumeSpecName "metrics-tls". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 10 22:49:51 crc kubenswrapper[4791]: I1210 22:49:51.005188 4791 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/8f668bae-612b-4b75-9490-919e737c6a3b-kube-api-access-kfwg7" (OuterVolumeSpecName: "kube-api-access-kfwg7") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b"). InnerVolumeSpecName "kube-api-access-kfwg7". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 10 22:49:51 crc kubenswrapper[4791]: I1210 22:49:51.005366 4791 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/fda69060-fa79-4696-b1a6-7980f124bf7c-proxy-tls" (OuterVolumeSpecName: "proxy-tls") pod "fda69060-fa79-4696-b1a6-7980f124bf7c" (UID: "fda69060-fa79-4696-b1a6-7980f124bf7c"). InnerVolumeSpecName "proxy-tls". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 10 22:49:51 crc kubenswrapper[4791]: I1210 22:49:51.006855 4791 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/7583ce53-e0fe-4a16-9e4d-50516596a136-client-ca" (OuterVolumeSpecName: "client-ca") pod "7583ce53-e0fe-4a16-9e4d-50516596a136" (UID: "7583ce53-e0fe-4a16-9e4d-50516596a136"). InnerVolumeSpecName "client-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 10 22:49:51 crc kubenswrapper[4791]: I1210 22:49:51.007230 4791 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/22c825df-677d-4ca6-82db-3454ed06e783-kube-api-access-7c4vf" (OuterVolumeSpecName: "kube-api-access-7c4vf") pod "22c825df-677d-4ca6-82db-3454ed06e783" (UID: "22c825df-677d-4ca6-82db-3454ed06e783"). InnerVolumeSpecName "kube-api-access-7c4vf". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 10 22:49:51 crc kubenswrapper[4791]: I1210 22:49:51.007453 4791 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/6402fda4-df10-493c-b4e5-d0569419652d-kube-api-access-mg5zb" (OuterVolumeSpecName: "kube-api-access-mg5zb") pod "6402fda4-df10-493c-b4e5-d0569419652d" (UID: "6402fda4-df10-493c-b4e5-d0569419652d"). InnerVolumeSpecName "kube-api-access-mg5zb". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 10 22:49:51 crc kubenswrapper[4791]: I1210 22:49:51.008009 4791 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/1386a44e-36a2-460c-96d0-0359d2b6f0f5-config" (OuterVolumeSpecName: "config") pod "1386a44e-36a2-460c-96d0-0359d2b6f0f5" (UID: "1386a44e-36a2-460c-96d0-0359d2b6f0f5"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 10 22:49:51 crc kubenswrapper[4791]: I1210 22:49:51.008182 4791 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/bd23aa5c-e532-4e53-bccf-e79f130c5ae8-kube-api-access-jhbk2" (OuterVolumeSpecName: "kube-api-access-jhbk2") pod "bd23aa5c-e532-4e53-bccf-e79f130c5ae8" (UID: "bd23aa5c-e532-4e53-bccf-e79f130c5ae8"). InnerVolumeSpecName "kube-api-access-jhbk2". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 10 22:49:51 crc kubenswrapper[4791]: I1210 22:49:51.008487 4791 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/bf126b07-da06-4140-9a57-dfd54fc6b486-kube-api-access-rnphk" (OuterVolumeSpecName: "kube-api-access-rnphk") pod "bf126b07-da06-4140-9a57-dfd54fc6b486" (UID: "bf126b07-da06-4140-9a57-dfd54fc6b486"). InnerVolumeSpecName "kube-api-access-rnphk". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 10 22:49:51 crc kubenswrapper[4791]: I1210 22:49:51.008581 4791 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c-kube-api-access-qs4fp" (OuterVolumeSpecName: "kube-api-access-qs4fp") pod "210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c" (UID: "210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c"). InnerVolumeSpecName "kube-api-access-qs4fp". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 10 22:49:51 crc kubenswrapper[4791]: E1210 22:49:51.008836 4791 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-10 22:49:51.508785906 +0000 UTC m=+25.938403589 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 10 22:49:51 crc kubenswrapper[4791]: E1210 22:49:51.009831 4791 secret.go:188] Couldn't get secret openshift-network-console/networking-console-plugin-cert: object "openshift-network-console"/"networking-console-plugin-cert" not registered Dec 10 22:49:51 crc kubenswrapper[4791]: I1210 22:49:51.010629 4791 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/7bb08738-c794-4ee8-9972-3a62ca171029-kube-api-access-279lb" (OuterVolumeSpecName: "kube-api-access-279lb") pod "7bb08738-c794-4ee8-9972-3a62ca171029" (UID: "7bb08738-c794-4ee8-9972-3a62ca171029"). InnerVolumeSpecName "kube-api-access-279lb". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 10 22:49:51 crc kubenswrapper[4791]: I1210 22:49:51.010674 4791 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/e7e6199b-1264-4501-8953-767f51328d08-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "e7e6199b-1264-4501-8953-767f51328d08" (UID: "e7e6199b-1264-4501-8953-767f51328d08"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 10 22:49:51 crc kubenswrapper[4791]: I1210 22:49:51.010923 4791 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d-kube-api-access-mnrrd" (OuterVolumeSpecName: "kube-api-access-mnrrd") pod "bc5039c0-ea34-426b-a2b7-fbbc87b49a6d" (UID: "bc5039c0-ea34-426b-a2b7-fbbc87b49a6d"). InnerVolumeSpecName "kube-api-access-mnrrd". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 10 22:49:51 crc kubenswrapper[4791]: I1210 22:49:51.015862 4791 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/a31745f5-9847-4afe-82a5-3161cc66ca93-trusted-ca" (OuterVolumeSpecName: "trusted-ca") pod "a31745f5-9847-4afe-82a5-3161cc66ca93" (UID: "a31745f5-9847-4afe-82a5-3161cc66ca93"). InnerVolumeSpecName "trusted-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 10 22:49:51 crc kubenswrapper[4791]: I1210 22:49:51.016261 4791 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/25e176fe-21b4-4974-b1ed-c8b94f112a7f-kube-api-access-d4lsv" (OuterVolumeSpecName: "kube-api-access-d4lsv") pod "25e176fe-21b4-4974-b1ed-c8b94f112a7f" (UID: "25e176fe-21b4-4974-b1ed-c8b94f112a7f"). InnerVolumeSpecName "kube-api-access-d4lsv". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 10 22:49:51 crc kubenswrapper[4791]: I1210 22:49:50.993086 4791 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/c03ee662-fb2f-4fc4-a2c1-af487c19d254-kube-api-access-v47cf" (OuterVolumeSpecName: "kube-api-access-v47cf") pod "c03ee662-fb2f-4fc4-a2c1-af487c19d254" (UID: "c03ee662-fb2f-4fc4-a2c1-af487c19d254"). InnerVolumeSpecName "kube-api-access-v47cf". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 10 22:49:51 crc kubenswrapper[4791]: I1210 22:49:51.019025 4791 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/ef543e1b-8068-4ea3-b32a-61027b32e95d-env-overrides\") pod \"network-node-identity-vrzqb\" (UID: \"ef543e1b-8068-4ea3-b32a-61027b32e95d\") " pod="openshift-network-node-identity/network-node-identity-vrzqb" Dec 10 22:49:51 crc kubenswrapper[4791]: E1210 22:49:51.019126 4791 configmap.go:193] Couldn't get configMap openshift-network-console/networking-console-plugin: object "openshift-network-console"/"networking-console-plugin" not registered Dec 10 22:49:51 crc kubenswrapper[4791]: I1210 22:49:51.019385 4791 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/31d8b7a1-420e-4252-a5b7-eebe8a111292-kube-api-access-zgdk5" (OuterVolumeSpecName: "kube-api-access-zgdk5") pod "31d8b7a1-420e-4252-a5b7-eebe8a111292" (UID: "31d8b7a1-420e-4252-a5b7-eebe8a111292"). InnerVolumeSpecName "kube-api-access-zgdk5". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 10 22:49:51 crc kubenswrapper[4791]: E1210 22:49:51.019579 4791 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2025-12-10 22:49:51.519553159 +0000 UTC m=+25.949170772 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "networking-console-plugin-cert" (UniqueName: "kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin-cert" not registered Dec 10 22:49:51 crc kubenswrapper[4791]: E1210 22:49:51.019700 4791 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2025-12-10 22:49:51.519688593 +0000 UTC m=+25.949306206 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "nginx-conf" (UniqueName: "kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin" not registered Dec 10 22:49:51 crc kubenswrapper[4791]: I1210 22:49:51.020025 4791 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/6402fda4-df10-493c-b4e5-d0569419652d-config" (OuterVolumeSpecName: "config") pod "6402fda4-df10-493c-b4e5-d0569419652d" (UID: "6402fda4-df10-493c-b4e5-d0569419652d"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 10 22:49:51 crc kubenswrapper[4791]: I1210 22:49:51.020378 4791 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/b6cd30de-2eeb-49a2-ab40-9167f4560ff5-marketplace-operator-metrics" (OuterVolumeSpecName: "marketplace-operator-metrics") pod "b6cd30de-2eeb-49a2-ab40-9167f4560ff5" (UID: "b6cd30de-2eeb-49a2-ab40-9167f4560ff5"). InnerVolumeSpecName "marketplace-operator-metrics". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 10 22:49:51 crc kubenswrapper[4791]: I1210 22:49:51.020726 4791 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/8f668bae-612b-4b75-9490-919e737c6a3b-registry-tls" (OuterVolumeSpecName: "registry-tls") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b"). InnerVolumeSpecName "registry-tls". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 10 22:49:51 crc kubenswrapper[4791]: I1210 22:49:51.021572 4791 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-kube-api-access-6ccd8" (OuterVolumeSpecName: "kube-api-access-6ccd8") pod "308be0ea-9f5f-4b29-aeb1-5abd31a0b17b" (UID: "308be0ea-9f5f-4b29-aeb1-5abd31a0b17b"). InnerVolumeSpecName "kube-api-access-6ccd8". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 10 22:49:51 crc kubenswrapper[4791]: I1210 22:49:51.021801 4791 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d-kube-api-access-x4zgh" (OuterVolumeSpecName: "kube-api-access-x4zgh") pod "b11524ee-3fca-4b1b-9cdf-6da289fdbc7d" (UID: "b11524ee-3fca-4b1b-9cdf-6da289fdbc7d"). InnerVolumeSpecName "kube-api-access-x4zgh". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 10 22:49:51 crc kubenswrapper[4791]: I1210 22:49:51.022158 4791 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"iptables-alerter-script\" (UniqueName: \"kubernetes.io/configmap/d75a4c96-2883-4a0b-bab2-0fab2b6c0b49-iptables-alerter-script\") pod \"iptables-alerter-4ln5h\" (UID: \"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\") " pod="openshift-network-operator/iptables-alerter-4ln5h" Dec 10 22:49:51 crc kubenswrapper[4791]: I1210 22:49:51.022739 4791 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/5441d097-087c-4d9a-baa8-b210afa90fc9-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "5441d097-087c-4d9a-baa8-b210afa90fc9" (UID: "5441d097-087c-4d9a-baa8-b210afa90fc9"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 10 22:49:51 crc kubenswrapper[4791]: I1210 22:49:51.023257 4791 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/5b88f790-22fa-440e-b583-365168c0b23d-metrics-certs" (OuterVolumeSpecName: "metrics-certs") pod "5b88f790-22fa-440e-b583-365168c0b23d" (UID: "5b88f790-22fa-440e-b583-365168c0b23d"). InnerVolumeSpecName "metrics-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 10 22:49:51 crc kubenswrapper[4791]: I1210 22:49:51.025016 4791 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/e7e6199b-1264-4501-8953-767f51328d08-config" (OuterVolumeSpecName: "config") pod "e7e6199b-1264-4501-8953-767f51328d08" (UID: "e7e6199b-1264-4501-8953-767f51328d08"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 10 22:49:51 crc kubenswrapper[4791]: I1210 22:49:51.025662 4791 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-serving-cert" (OuterVolumeSpecName: "v4-0-config-system-serving-cert") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "v4-0-config-system-serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 10 22:49:51 crc kubenswrapper[4791]: I1210 22:49:51.025756 4791 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "09ae3b1a-e8e7-4524-b54b-61eab6f9239a" (UID: "09ae3b1a-e8e7-4524-b54b-61eab6f9239a"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 10 22:49:51 crc kubenswrapper[4791]: I1210 22:49:51.025997 4791 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/20b0d48f-5fd6-431c-a545-e3c800c7b866-kube-api-access-w9rds" (OuterVolumeSpecName: "kube-api-access-w9rds") pod "20b0d48f-5fd6-431c-a545-e3c800c7b866" (UID: "20b0d48f-5fd6-431c-a545-e3c800c7b866"). InnerVolumeSpecName "kube-api-access-w9rds". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 10 22:49:51 crc kubenswrapper[4791]: I1210 22:49:51.026455 4791 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/7539238d-5fe0-46ed-884e-1c3b566537ec-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "7539238d-5fe0-46ed-884e-1c3b566537ec" (UID: "7539238d-5fe0-46ed-884e-1c3b566537ec"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 10 22:49:51 crc kubenswrapper[4791]: I1210 22:49:51.026616 4791 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/09efc573-dbb6-4249-bd59-9b87aba8dd28-etcd-client" (OuterVolumeSpecName: "etcd-client") pod "09efc573-dbb6-4249-bd59-9b87aba8dd28" (UID: "09efc573-dbb6-4249-bd59-9b87aba8dd28"). InnerVolumeSpecName "etcd-client". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 10 22:49:51 crc kubenswrapper[4791]: I1210 22:49:51.026833 4791 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/a0128f3a-b052-44ed-a84e-c4c8aaf17c13-kube-api-access-gf66m" (OuterVolumeSpecName: "kube-api-access-gf66m") pod "a0128f3a-b052-44ed-a84e-c4c8aaf17c13" (UID: "a0128f3a-b052-44ed-a84e-c4c8aaf17c13"). InnerVolumeSpecName "kube-api-access-gf66m". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 10 22:49:51 crc kubenswrapper[4791]: I1210 22:49:51.027509 4791 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/87cf06ed-a83f-41a7-828d-70653580a8cb-kube-api-access-d6qdx" (OuterVolumeSpecName: "kube-api-access-d6qdx") pod "87cf06ed-a83f-41a7-828d-70653580a8cb" (UID: "87cf06ed-a83f-41a7-828d-70653580a8cb"). InnerVolumeSpecName "kube-api-access-d6qdx". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 10 22:49:51 crc kubenswrapper[4791]: I1210 22:49:51.027454 4791 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-oauth-serving-cert" (OuterVolumeSpecName: "oauth-serving-cert") pod "43509403-f426-496e-be36-56cef71462f5" (UID: "43509403-f426-496e-be36-56cef71462f5"). InnerVolumeSpecName "oauth-serving-cert". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 10 22:49:51 crc kubenswrapper[4791]: I1210 22:49:51.027787 4791 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/6ea678ab-3438-413e-bfe3-290ae7725660-kube-api-access-htfz6" (OuterVolumeSpecName: "kube-api-access-htfz6") pod "6ea678ab-3438-413e-bfe3-290ae7725660" (UID: "6ea678ab-3438-413e-bfe3-290ae7725660"). InnerVolumeSpecName "kube-api-access-htfz6". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 10 22:49:51 crc kubenswrapper[4791]: E1210 22:49:51.028059 4791 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Dec 10 22:49:51 crc kubenswrapper[4791]: E1210 22:49:51.028160 4791 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Dec 10 22:49:51 crc kubenswrapper[4791]: E1210 22:49:51.028232 4791 projected.go:194] Error preparing data for projected volume kube-api-access-s2dwl for pod openshift-network-diagnostics/network-check-source-55646444c4-trplf: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Dec 10 22:49:51 crc kubenswrapper[4791]: E1210 22:49:51.028382 4791 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl podName:9d751cbb-f2e2-430d-9754-c882a5e924a5 nodeName:}" failed. No retries permitted until 2025-12-10 22:49:51.528331232 +0000 UTC m=+25.957948845 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "kube-api-access-s2dwl" (UniqueName: "kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl") pod "network-check-source-55646444c4-trplf" (UID: "9d751cbb-f2e2-430d-9754-c882a5e924a5") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Dec 10 22:49:51 crc kubenswrapper[4791]: I1210 22:49:51.028457 4791 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/5225d0e4-402f-4861-b410-819f433b1803-utilities" (OuterVolumeSpecName: "utilities") pod "5225d0e4-402f-4861-b410-819f433b1803" (UID: "5225d0e4-402f-4861-b410-819f433b1803"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 10 22:49:51 crc kubenswrapper[4791]: I1210 22:49:51.028120 4791 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/efdd0498-1daa-4136-9a4a-3b948c2293fc-webhook-certs" (OuterVolumeSpecName: "webhook-certs") pod "efdd0498-1daa-4136-9a4a-3b948c2293fc" (UID: "efdd0498-1daa-4136-9a4a-3b948c2293fc"). InnerVolumeSpecName "webhook-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 10 22:49:51 crc kubenswrapper[4791]: I1210 22:49:51.028136 4791 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-audit-policies" (OuterVolumeSpecName: "audit-policies") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "audit-policies". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 10 22:49:51 crc kubenswrapper[4791]: I1210 22:49:51.028580 4791 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/6ea678ab-3438-413e-bfe3-290ae7725660-ovn-node-metrics-cert" (OuterVolumeSpecName: "ovn-node-metrics-cert") pod "6ea678ab-3438-413e-bfe3-290ae7725660" (UID: "6ea678ab-3438-413e-bfe3-290ae7725660"). InnerVolumeSpecName "ovn-node-metrics-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 10 22:49:51 crc kubenswrapper[4791]: I1210 22:49:51.028915 4791 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/c03ee662-fb2f-4fc4-a2c1-af487c19d254-stats-auth" (OuterVolumeSpecName: "stats-auth") pod "c03ee662-fb2f-4fc4-a2c1-af487c19d254" (UID: "c03ee662-fb2f-4fc4-a2c1-af487c19d254"). InnerVolumeSpecName "stats-auth". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 10 22:49:51 crc kubenswrapper[4791]: I1210 22:49:51.028862 4791 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-template-provider-selection" (OuterVolumeSpecName: "v4-0-config-user-template-provider-selection") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "v4-0-config-user-template-provider-selection". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 10 22:49:51 crc kubenswrapper[4791]: I1210 22:49:51.029275 4791 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/6ea678ab-3438-413e-bfe3-290ae7725660-ovnkube-config" (OuterVolumeSpecName: "ovnkube-config") pod "6ea678ab-3438-413e-bfe3-290ae7725660" (UID: "6ea678ab-3438-413e-bfe3-290ae7725660"). InnerVolumeSpecName "ovnkube-config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 10 22:49:51 crc kubenswrapper[4791]: I1210 22:49:51.029290 4791 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/5b88f790-22fa-440e-b583-365168c0b23d-kube-api-access-jkwtn" (OuterVolumeSpecName: "kube-api-access-jkwtn") pod "5b88f790-22fa-440e-b583-365168c0b23d" (UID: "5b88f790-22fa-440e-b583-365168c0b23d"). InnerVolumeSpecName "kube-api-access-jkwtn". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 10 22:49:51 crc kubenswrapper[4791]: I1210 22:49:51.029578 4791 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/0b574797-001e-440a-8f4e-c0be86edad0f-proxy-tls" (OuterVolumeSpecName: "proxy-tls") pod "0b574797-001e-440a-8f4e-c0be86edad0f" (UID: "0b574797-001e-440a-8f4e-c0be86edad0f"). InnerVolumeSpecName "proxy-tls". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 10 22:49:51 crc kubenswrapper[4791]: I1210 22:49:51.030637 4791 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-trusted-ca-bundle" (OuterVolumeSpecName: "v4-0-config-system-trusted-ca-bundle") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "v4-0-config-system-trusted-ca-bundle". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 10 22:49:51 crc kubenswrapper[4791]: I1210 22:49:51.030674 4791 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/a31745f5-9847-4afe-82a5-3161cc66ca93-metrics-tls" (OuterVolumeSpecName: "metrics-tls") pod "a31745f5-9847-4afe-82a5-3161cc66ca93" (UID: "a31745f5-9847-4afe-82a5-3161cc66ca93"). InnerVolumeSpecName "metrics-tls". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 10 22:49:51 crc kubenswrapper[4791]: I1210 22:49:51.030700 4791 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/4bb40260-dbaa-4fb0-84df-5e680505d512-cni-binary-copy" (OuterVolumeSpecName: "cni-binary-copy") pod "4bb40260-dbaa-4fb0-84df-5e680505d512" (UID: "4bb40260-dbaa-4fb0-84df-5e680505d512"). InnerVolumeSpecName "cni-binary-copy". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 10 22:49:51 crc kubenswrapper[4791]: I1210 22:49:51.030906 4791 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-etcd-serving-ca" (OuterVolumeSpecName: "etcd-serving-ca") pod "09ae3b1a-e8e7-4524-b54b-61eab6f9239a" (UID: "09ae3b1a-e8e7-4524-b54b-61eab6f9239a"). InnerVolumeSpecName "etcd-serving-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 10 22:49:51 crc kubenswrapper[4791]: I1210 22:49:51.031000 4791 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-audit-policies" (OuterVolumeSpecName: "audit-policies") pod "09ae3b1a-e8e7-4524-b54b-61eab6f9239a" (UID: "09ae3b1a-e8e7-4524-b54b-61eab6f9239a"). InnerVolumeSpecName "audit-policies". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 10 22:49:51 crc kubenswrapper[4791]: I1210 22:49:51.031091 4791 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/6509e943-70c6-444c-bc41-48a544e36fbd-service-ca-bundle" (OuterVolumeSpecName: "service-ca-bundle") pod "6509e943-70c6-444c-bc41-48a544e36fbd" (UID: "6509e943-70c6-444c-bc41-48a544e36fbd"). InnerVolumeSpecName "service-ca-bundle". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 10 22:49:51 crc kubenswrapper[4791]: I1210 22:49:51.031150 4791 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/6509e943-70c6-444c-bc41-48a544e36fbd-config" (OuterVolumeSpecName: "config") pod "6509e943-70c6-444c-bc41-48a544e36fbd" (UID: "6509e943-70c6-444c-bc41-48a544e36fbd"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 10 22:49:51 crc kubenswrapper[4791]: I1210 22:49:51.031694 4791 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/7bb08738-c794-4ee8-9972-3a62ca171029-cni-binary-copy" (OuterVolumeSpecName: "cni-binary-copy") pod "7bb08738-c794-4ee8-9972-3a62ca171029" (UID: "7bb08738-c794-4ee8-9972-3a62ca171029"). InnerVolumeSpecName "cni-binary-copy". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 10 22:49:51 crc kubenswrapper[4791]: I1210 22:49:51.031761 4791 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/b6cd30de-2eeb-49a2-ab40-9167f4560ff5-marketplace-trusted-ca" (OuterVolumeSpecName: "marketplace-trusted-ca") pod "b6cd30de-2eeb-49a2-ab40-9167f4560ff5" (UID: "b6cd30de-2eeb-49a2-ab40-9167f4560ff5"). InnerVolumeSpecName "marketplace-trusted-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 10 22:49:51 crc kubenswrapper[4791]: I1210 22:49:51.031769 4791 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-rdwmf\" (UniqueName: \"kubernetes.io/projected/37a5e44f-9a88-4405-be8a-b645485e7312-kube-api-access-rdwmf\") pod \"network-operator-58b4c7f79c-55gtf\" (UID: \"37a5e44f-9a88-4405-be8a-b645485e7312\") " pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" Dec 10 22:49:51 crc kubenswrapper[4791]: I1210 22:49:51.031821 4791 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/ef543e1b-8068-4ea3-b32a-61027b32e95d-webhook-cert\") pod \"network-node-identity-vrzqb\" (UID: \"ef543e1b-8068-4ea3-b32a-61027b32e95d\") " pod="openshift-network-node-identity/network-node-identity-vrzqb" Dec 10 22:49:51 crc kubenswrapper[4791]: I1210 22:49:51.031880 4791 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/6509e943-70c6-444c-bc41-48a544e36fbd-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "6509e943-70c6-444c-bc41-48a544e36fbd" (UID: "6509e943-70c6-444c-bc41-48a544e36fbd"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 10 22:49:51 crc kubenswrapper[4791]: I1210 22:49:51.031895 4791 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/e7e6199b-1264-4501-8953-767f51328d08-kube-api-access" (OuterVolumeSpecName: "kube-api-access") pod "e7e6199b-1264-4501-8953-767f51328d08" (UID: "e7e6199b-1264-4501-8953-767f51328d08"). InnerVolumeSpecName "kube-api-access". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 10 22:49:51 crc kubenswrapper[4791]: I1210 22:49:51.033758 4791 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/09efc573-dbb6-4249-bd59-9b87aba8dd28-config" (OuterVolumeSpecName: "config") pod "09efc573-dbb6-4249-bd59-9b87aba8dd28" (UID: "09efc573-dbb6-4249-bd59-9b87aba8dd28"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 10 22:49:51 crc kubenswrapper[4791]: I1210 22:49:51.035225 4791 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/6ea678ab-3438-413e-bfe3-290ae7725660-env-overrides" (OuterVolumeSpecName: "env-overrides") pod "6ea678ab-3438-413e-bfe3-290ae7725660" (UID: "6ea678ab-3438-413e-bfe3-290ae7725660"). InnerVolumeSpecName "env-overrides". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 10 22:49:51 crc kubenswrapper[4791]: I1210 22:49:51.036127 4791 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/37a5e44f-9a88-4405-be8a-b645485e7312-metrics-tls\") pod \"network-operator-58b4c7f79c-55gtf\" (UID: \"37a5e44f-9a88-4405-be8a-b645485e7312\") " pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" Dec 10 22:49:51 crc kubenswrapper[4791]: I1210 22:49:51.036139 4791 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/7539238d-5fe0-46ed-884e-1c3b566537ec-kube-api-access-tk88c" (OuterVolumeSpecName: "kube-api-access-tk88c") pod "7539238d-5fe0-46ed-884e-1c3b566537ec" (UID: "7539238d-5fe0-46ed-884e-1c3b566537ec"). InnerVolumeSpecName "kube-api-access-tk88c". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 10 22:49:51 crc kubenswrapper[4791]: I1210 22:49:51.036205 4791 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/7539238d-5fe0-46ed-884e-1c3b566537ec-config" (OuterVolumeSpecName: "config") pod "7539238d-5fe0-46ed-884e-1c3b566537ec" (UID: "7539238d-5fe0-46ed-884e-1c3b566537ec"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 10 22:49:51 crc kubenswrapper[4791]: I1210 22:49:51.038546 4791 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-idp-0-file-data" (OuterVolumeSpecName: "v4-0-config-user-idp-0-file-data") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "v4-0-config-user-idp-0-file-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 10 22:49:51 crc kubenswrapper[4791]: I1210 22:49:51.038737 4791 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/b6312bbd-5731-4ea0-a20f-81d5a57df44a-srv-cert" (OuterVolumeSpecName: "srv-cert") pod "b6312bbd-5731-4ea0-a20f-81d5a57df44a" (UID: "b6312bbd-5731-4ea0-a20f-81d5a57df44a"). InnerVolumeSpecName "srv-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 10 22:49:51 crc kubenswrapper[4791]: I1210 22:49:51.042138 4791 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/9d4552c7-cd75-42dd-8880-30dd377c49a4-trusted-ca" (OuterVolumeSpecName: "trusted-ca") pod "9d4552c7-cd75-42dd-8880-30dd377c49a4" (UID: "9d4552c7-cd75-42dd-8880-30dd377c49a4"). InnerVolumeSpecName "trusted-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 10 22:49:51 crc kubenswrapper[4791]: I1210 22:49:51.042486 4791 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/1bf7eb37-55a3-4c65-b768-a94c82151e69-kube-api-access-sb6h7" (OuterVolumeSpecName: "kube-api-access-sb6h7") pod "1bf7eb37-55a3-4c65-b768-a94c82151e69" (UID: "1bf7eb37-55a3-4c65-b768-a94c82151e69"). InnerVolumeSpecName "kube-api-access-sb6h7". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 10 22:49:51 crc kubenswrapper[4791]: I1210 22:49:51.042913 4791 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/9d4552c7-cd75-42dd-8880-30dd377c49a4-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "9d4552c7-cd75-42dd-8880-30dd377c49a4" (UID: "9d4552c7-cd75-42dd-8880-30dd377c49a4"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 10 22:49:51 crc kubenswrapper[4791]: E1210 22:49:51.043586 4791 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Dec 10 22:49:51 crc kubenswrapper[4791]: I1210 22:49:51.043597 4791 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/4bb40260-dbaa-4fb0-84df-5e680505d512-multus-daemon-config" (OuterVolumeSpecName: "multus-daemon-config") pod "4bb40260-dbaa-4fb0-84df-5e680505d512" (UID: "4bb40260-dbaa-4fb0-84df-5e680505d512"). InnerVolumeSpecName "multus-daemon-config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 10 22:49:51 crc kubenswrapper[4791]: E1210 22:49:51.043618 4791 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Dec 10 22:49:51 crc kubenswrapper[4791]: E1210 22:49:51.043662 4791 projected.go:194] Error preparing data for projected volume kube-api-access-cqllr for pod openshift-network-diagnostics/network-check-target-xd92c: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Dec 10 22:49:51 crc kubenswrapper[4791]: E1210 22:49:51.043730 4791 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr podName:3b6479f0-333b-4a96-9adf-2099afdc2447 nodeName:}" failed. No retries permitted until 2025-12-10 22:49:51.543705322 +0000 UTC m=+25.973323135 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "kube-api-access-cqllr" (UniqueName: "kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr") pod "network-check-target-xd92c" (UID: "3b6479f0-333b-4a96-9adf-2099afdc2447") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Dec 10 22:49:51 crc kubenswrapper[4791]: I1210 22:49:51.046607 4791 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/57a731c4-ef35-47a8-b875-bfb08a7f8011-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "57a731c4-ef35-47a8-b875-bfb08a7f8011" (UID: "57a731c4-ef35-47a8-b875-bfb08a7f8011"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 10 22:49:51 crc kubenswrapper[4791]: I1210 22:49:51.050628 4791 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/22c825df-677d-4ca6-82db-3454ed06e783-machine-approver-tls" (OuterVolumeSpecName: "machine-approver-tls") pod "22c825df-677d-4ca6-82db-3454ed06e783" (UID: "22c825df-677d-4ca6-82db-3454ed06e783"). InnerVolumeSpecName "machine-approver-tls". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 10 22:49:51 crc kubenswrapper[4791]: I1210 22:49:51.052389 4791 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/22c825df-677d-4ca6-82db-3454ed06e783-auth-proxy-config" (OuterVolumeSpecName: "auth-proxy-config") pod "22c825df-677d-4ca6-82db-3454ed06e783" (UID: "22c825df-677d-4ca6-82db-3454ed06e783"). InnerVolumeSpecName "auth-proxy-config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 10 22:49:51 crc kubenswrapper[4791]: I1210 22:49:51.052732 4791 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-config" (OuterVolumeSpecName: "config") pod "1bf7eb37-55a3-4c65-b768-a94c82151e69" (UID: "1bf7eb37-55a3-4c65-b768-a94c82151e69"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 10 22:49:51 crc kubenswrapper[4791]: I1210 22:49:51.052960 4791 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/1bf7eb37-55a3-4c65-b768-a94c82151e69-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "1bf7eb37-55a3-4c65-b768-a94c82151e69" (UID: "1bf7eb37-55a3-4c65-b768-a94c82151e69"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 10 22:49:51 crc kubenswrapper[4791]: I1210 22:49:51.054933 4791 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-rczfb\" (UniqueName: \"kubernetes.io/projected/d75a4c96-2883-4a0b-bab2-0fab2b6c0b49-kube-api-access-rczfb\") pod \"iptables-alerter-4ln5h\" (UID: \"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\") " pod="openshift-network-operator/iptables-alerter-4ln5h" Dec 10 22:49:51 crc kubenswrapper[4791]: I1210 22:49:51.061480 4791 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-s2kz5\" (UniqueName: \"kubernetes.io/projected/ef543e1b-8068-4ea3-b32a-61027b32e95d-kube-api-access-s2kz5\") pod \"network-node-identity-vrzqb\" (UID: \"ef543e1b-8068-4ea3-b32a-61027b32e95d\") " pod="openshift-network-node-identity/network-node-identity-vrzqb" Dec 10 22:49:51 crc kubenswrapper[4791]: I1210 22:49:51.062282 4791 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/7583ce53-e0fe-4a16-9e4d-50516596a136-config" (OuterVolumeSpecName: "config") pod "7583ce53-e0fe-4a16-9e4d-50516596a136" (UID: "7583ce53-e0fe-4a16-9e4d-50516596a136"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 10 22:49:51 crc kubenswrapper[4791]: I1210 22:49:51.066682 4791 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/1bf7eb37-55a3-4c65-b768-a94c82151e69-etcd-client" (OuterVolumeSpecName: "etcd-client") pod "1bf7eb37-55a3-4c65-b768-a94c82151e69" (UID: "1bf7eb37-55a3-4c65-b768-a94c82151e69"). InnerVolumeSpecName "etcd-client". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 10 22:49:51 crc kubenswrapper[4791]: I1210 22:49:51.066692 4791 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/7583ce53-e0fe-4a16-9e4d-50516596a136-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "7583ce53-e0fe-4a16-9e4d-50516596a136" (UID: "7583ce53-e0fe-4a16-9e4d-50516596a136"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 10 22:49:51 crc kubenswrapper[4791]: I1210 22:49:51.067043 4791 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-audit" (OuterVolumeSpecName: "audit") pod "1bf7eb37-55a3-4c65-b768-a94c82151e69" (UID: "1bf7eb37-55a3-4c65-b768-a94c82151e69"). InnerVolumeSpecName "audit". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 10 22:49:51 crc kubenswrapper[4791]: I1210 22:49:51.073525 4791 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/c03ee662-fb2f-4fc4-a2c1-af487c19d254-service-ca-bundle" (OuterVolumeSpecName: "service-ca-bundle") pod "c03ee662-fb2f-4fc4-a2c1-af487c19d254" (UID: "c03ee662-fb2f-4fc4-a2c1-af487c19d254"). InnerVolumeSpecName "service-ca-bundle". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 10 22:49:51 crc kubenswrapper[4791]: I1210 22:49:51.073822 4791 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/496e6271-fb68-4057-954e-a0d97a4afa3f-config" (OuterVolumeSpecName: "config") pod "496e6271-fb68-4057-954e-a0d97a4afa3f" (UID: "496e6271-fb68-4057-954e-a0d97a4afa3f"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 10 22:49:51 crc kubenswrapper[4791]: I1210 22:49:51.074816 4791 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-trusted-ca-bundle" (OuterVolumeSpecName: "trusted-ca-bundle") pod "1bf7eb37-55a3-4c65-b768-a94c82151e69" (UID: "1bf7eb37-55a3-4c65-b768-a94c82151e69"). InnerVolumeSpecName "trusted-ca-bundle". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 10 22:49:51 crc kubenswrapper[4791]: I1210 22:49:51.075063 4791 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/31d8b7a1-420e-4252-a5b7-eebe8a111292-images" (OuterVolumeSpecName: "images") pod "31d8b7a1-420e-4252-a5b7-eebe8a111292" (UID: "31d8b7a1-420e-4252-a5b7-eebe8a111292"). InnerVolumeSpecName "images". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 10 22:49:51 crc kubenswrapper[4791]: I1210 22:49:51.076590 4791 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/31d8b7a1-420e-4252-a5b7-eebe8a111292-auth-proxy-config" (OuterVolumeSpecName: "auth-proxy-config") pod "31d8b7a1-420e-4252-a5b7-eebe8a111292" (UID: "31d8b7a1-420e-4252-a5b7-eebe8a111292"). InnerVolumeSpecName "auth-proxy-config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 10 22:49:51 crc kubenswrapper[4791]: I1210 22:49:51.076697 4791 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-console-config" (OuterVolumeSpecName: "console-config") pod "43509403-f426-496e-be36-56cef71462f5" (UID: "43509403-f426-496e-be36-56cef71462f5"). InnerVolumeSpecName "console-config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 10 22:49:51 crc kubenswrapper[4791]: I1210 22:49:51.076858 4791 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/43509403-f426-496e-be36-56cef71462f5-console-oauth-config" (OuterVolumeSpecName: "console-oauth-config") pod "43509403-f426-496e-be36-56cef71462f5" (UID: "43509403-f426-496e-be36-56cef71462f5"). InnerVolumeSpecName "console-oauth-config". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 10 22:49:51 crc kubenswrapper[4791]: I1210 22:49:51.077125 4791 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/1bf7eb37-55a3-4c65-b768-a94c82151e69-encryption-config" (OuterVolumeSpecName: "encryption-config") pod "1bf7eb37-55a3-4c65-b768-a94c82151e69" (UID: "1bf7eb37-55a3-4c65-b768-a94c82151e69"). InnerVolumeSpecName "encryption-config". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 10 22:49:51 crc kubenswrapper[4791]: I1210 22:49:51.077916 4791 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/5441d097-087c-4d9a-baa8-b210afa90fc9-config" (OuterVolumeSpecName: "config") pod "5441d097-087c-4d9a-baa8-b210afa90fc9" (UID: "5441d097-087c-4d9a-baa8-b210afa90fc9"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 10 22:49:51 crc kubenswrapper[4791]: I1210 22:49:51.078105 4791 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/09efc573-dbb6-4249-bd59-9b87aba8dd28-kube-api-access-8tdtz" (OuterVolumeSpecName: "kube-api-access-8tdtz") pod "09efc573-dbb6-4249-bd59-9b87aba8dd28" (UID: "09efc573-dbb6-4249-bd59-9b87aba8dd28"). InnerVolumeSpecName "kube-api-access-8tdtz". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 10 22:49:51 crc kubenswrapper[4791]: I1210 22:49:51.078598 4791 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/43509403-f426-496e-be36-56cef71462f5-kube-api-access-qg5z5" (OuterVolumeSpecName: "kube-api-access-qg5z5") pod "43509403-f426-496e-be36-56cef71462f5" (UID: "43509403-f426-496e-be36-56cef71462f5"). InnerVolumeSpecName "kube-api-access-qg5z5". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 10 22:49:51 crc kubenswrapper[4791]: I1210 22:49:51.080035 4791 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f8349fea-c49b-49a1-b5ee-032c5dffe021\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T22:49:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T22:49:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T22:49:25Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T22:49:25Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T22:49:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://acafe12186df99d1059a31260de0eafce35ff7771bcb8f5c6753b72665d6f60b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T22:49:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ea4c346507862a8c12756a3061d6e25e788a8d79f4621b62f81f4322185c82a8\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T22:49:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://2e6e324ec6f4f7bb69fdfab4ea3ac1aa55e480dd8e2a8984b0fe5b4c9f66636e\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T22:49:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://1b2e11ff2bf8789fa1d0d23ae648b3e6ef653106e5bc4bbb5dd6447f7cd6f05d\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d6aa7f8c5d908d90008f2df49860ac8a3480136da567e3d85f86135d779204f9\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-12-10T22:49:44Z\\\",\\\"message\\\":\\\"W1210 22:49:33.845733 1 cmd.go:257] Using insecure, self-signed certificates\\\\nI1210 22:49:33.846108 1 crypto.go:601] Generating new CA for check-endpoints-signer@1765406973 cert, and key in /tmp/serving-cert-2585701456/serving-signer.crt, /tmp/serving-cert-2585701456/serving-signer.key\\\\nI1210 22:49:34.529748 1 observer_polling.go:159] Starting file observer\\\\nW1210 22:49:34.532225 1 builder.go:272] unable to get owner reference (falling back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": dial tcp [::1]:6443: connect: connection refused\\\\nI1210 22:49:34.532599 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1210 22:49:34.533977 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-2585701456/tls.crt::/tmp/serving-cert-2585701456/tls.key\\\\\\\"\\\\nF1210 22:49:44.774710 1 cmd.go:182] error initializing delegating authentication: unable to load configmap based request-header-client-ca-file: Get \\\\\\\"https://localhost:6443/api/v1/namespaces/kube-system/configmaps/extension-apiserver-authentication\\\\\\\": net/http: TLS handshake timeout\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-10T22:49:33Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T22:49:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://0c4d036ab269f62561fe7ef1586012cb3a16e3583a08b9488a0616ec00c8e0ae\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T22:49:32Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c5856d6b26c19fda13b5c9ac2892a7e2f9c6f79feb8173ecfd9b33b673327d67\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c5856d6b26c19fda13b5c9ac2892a7e2f9c6f79feb8173ecfd9b33b673327d67\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-10T22:49:27Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-10T22:49:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-10T22:49:25Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Dec 10 22:49:51 crc kubenswrapper[4791]: I1210 22:49:51.077914 4791 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "b11524ee-3fca-4b1b-9cdf-6da289fdbc7d" (UID: "b11524ee-3fca-4b1b-9cdf-6da289fdbc7d"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 10 22:49:51 crc kubenswrapper[4791]: I1210 22:49:51.082706 4791 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/7583ce53-e0fe-4a16-9e4d-50516596a136-kube-api-access-xcphl" (OuterVolumeSpecName: "kube-api-access-xcphl") pod "7583ce53-e0fe-4a16-9e4d-50516596a136" (UID: "7583ce53-e0fe-4a16-9e4d-50516596a136"). InnerVolumeSpecName "kube-api-access-xcphl". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 10 22:49:51 crc kubenswrapper[4791]: I1210 22:49:51.090674 4791 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/1d611f23-29be-4491-8495-bee1670e935f-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "1d611f23-29be-4491-8495-bee1670e935f" (UID: "1d611f23-29be-4491-8495-bee1670e935f"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 10 22:49:51 crc kubenswrapper[4791]: I1210 22:49:51.091539 4791 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-vczjq" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b9f00f25-3fa1-4b63-9710-fd352224b01b\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T22:49:50Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T22:49:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T22:49:50Z\\\",\\\"message\\\":\\\"containers with unready status: [node-ca]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T22:49:50Z\\\",\\\"message\\\":\\\"containers with unready status: [node-ca]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lrrpv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-10T22:49:50Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-vczjq\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Dec 10 22:49:51 crc kubenswrapper[4791]: I1210 22:49:51.100718 4791 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host\" (UniqueName: \"kubernetes.io/host-path/b9f00f25-3fa1-4b63-9710-fd352224b01b-host\") pod \"node-ca-vczjq\" (UID: \"b9f00f25-3fa1-4b63-9710-fd352224b01b\") " pod="openshift-image-registry/node-ca-vczjq" Dec 10 22:49:51 crc kubenswrapper[4791]: I1210 22:49:51.100759 4791 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-slash\" (UniqueName: \"kubernetes.io/host-path/d75a4c96-2883-4a0b-bab2-0fab2b6c0b49-host-slash\") pod \"iptables-alerter-4ln5h\" (UID: \"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\") " pod="openshift-network-operator/iptables-alerter-4ln5h" Dec 10 22:49:51 crc kubenswrapper[4791]: I1210 22:49:51.100775 4791 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-etc-kube\" (UniqueName: \"kubernetes.io/host-path/37a5e44f-9a88-4405-be8a-b645485e7312-host-etc-kube\") pod \"network-operator-58b4c7f79c-55gtf\" (UID: \"37a5e44f-9a88-4405-be8a-b645485e7312\") " pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" Dec 10 22:49:51 crc kubenswrapper[4791]: I1210 22:49:51.100798 4791 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"hosts-file\" (UniqueName: \"kubernetes.io/host-path/c0ab74e0-42a9-4f42-ac80-16030cf4ffff-hosts-file\") pod \"node-resolver-tchxk\" (UID: \"c0ab74e0-42a9-4f42-ac80-16030cf4ffff\") " pod="openshift-dns/node-resolver-tchxk" Dec 10 22:49:51 crc kubenswrapper[4791]: I1210 22:49:51.100812 4791 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-zn2qt\" (UniqueName: \"kubernetes.io/projected/c0ab74e0-42a9-4f42-ac80-16030cf4ffff-kube-api-access-zn2qt\") pod \"node-resolver-tchxk\" (UID: \"c0ab74e0-42a9-4f42-ac80-16030cf4ffff\") " pod="openshift-dns/node-resolver-tchxk" Dec 10 22:49:51 crc kubenswrapper[4791]: I1210 22:49:51.100839 4791 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-lrrpv\" (UniqueName: \"kubernetes.io/projected/b9f00f25-3fa1-4b63-9710-fd352224b01b-kube-api-access-lrrpv\") pod \"node-ca-vczjq\" (UID: \"b9f00f25-3fa1-4b63-9710-fd352224b01b\") " pod="openshift-image-registry/node-ca-vczjq" Dec 10 22:49:51 crc kubenswrapper[4791]: I1210 22:49:51.100867 4791 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serviceca\" (UniqueName: \"kubernetes.io/configmap/b9f00f25-3fa1-4b63-9710-fd352224b01b-serviceca\") pod \"node-ca-vczjq\" (UID: \"b9f00f25-3fa1-4b63-9710-fd352224b01b\") " pod="openshift-image-registry/node-ca-vczjq" Dec 10 22:49:51 crc kubenswrapper[4791]: I1210 22:49:51.100958 4791 reconciler_common.go:293] "Volume detached for volume \"registry-certificates\" (UniqueName: \"kubernetes.io/configmap/8f668bae-612b-4b75-9490-919e737c6a3b-registry-certificates\") on node \"crc\" DevicePath \"\"" Dec 10 22:49:51 crc kubenswrapper[4791]: I1210 22:49:51.100969 4791 reconciler_common.go:293] "Volume detached for volume \"multus-daemon-config\" (UniqueName: \"kubernetes.io/configmap/4bb40260-dbaa-4fb0-84df-5e680505d512-multus-daemon-config\") on node \"crc\" DevicePath \"\"" Dec 10 22:49:51 crc kubenswrapper[4791]: I1210 22:49:51.100978 4791 reconciler_common.go:293] "Volume detached for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/6ea678ab-3438-413e-bfe3-290ae7725660-ovnkube-config\") on node \"crc\" DevicePath \"\"" Dec 10 22:49:51 crc kubenswrapper[4791]: I1210 22:49:51.100986 4791 reconciler_common.go:293] "Volume detached for volume \"etcd-service-ca\" (UniqueName: \"kubernetes.io/configmap/09efc573-dbb6-4249-bd59-9b87aba8dd28-etcd-service-ca\") on node \"crc\" DevicePath \"\"" Dec 10 22:49:51 crc kubenswrapper[4791]: I1210 22:49:51.100994 4791 reconciler_common.go:293] "Volume detached for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-audit-policies\") on node \"crc\" DevicePath \"\"" Dec 10 22:49:51 crc kubenswrapper[4791]: I1210 22:49:51.101003 4791 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/1d611f23-29be-4491-8495-bee1670e935f-catalog-content\") on node \"crc\" DevicePath \"\"" Dec 10 22:49:51 crc kubenswrapper[4791]: I1210 22:49:51.101011 4791 reconciler_common.go:293] "Volume detached for volume \"oauth-serving-cert\" (UniqueName: \"kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-oauth-serving-cert\") on node \"crc\" DevicePath \"\"" Dec 10 22:49:51 crc kubenswrapper[4791]: I1210 22:49:51.101019 4791 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-w4xd4\" (UniqueName: \"kubernetes.io/projected/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b-kube-api-access-w4xd4\") on node \"crc\" DevicePath \"\"" Dec 10 22:49:51 crc kubenswrapper[4791]: I1210 22:49:51.101033 4791 reconciler_common.go:293] "Volume detached for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/0b78653f-4ff9-4508-8672-245ed9b561e3-service-ca\") on node \"crc\" DevicePath \"\"" Dec 10 22:49:51 crc kubenswrapper[4791]: I1210 22:49:51.101041 4791 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d-serving-cert\") on node \"crc\" DevicePath \"\"" Dec 10 22:49:51 crc kubenswrapper[4791]: I1210 22:49:51.101050 4791 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-x2m85\" (UniqueName: \"kubernetes.io/projected/cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d-kube-api-access-x2m85\") on node \"crc\" DevicePath \"\"" Dec 10 22:49:51 crc kubenswrapper[4791]: I1210 22:49:51.101058 4791 reconciler_common.go:293] "Volume detached for volume \"console-oauth-config\" (UniqueName: \"kubernetes.io/secret/43509403-f426-496e-be36-56cef71462f5-console-oauth-config\") on node \"crc\" DevicePath \"\"" Dec 10 22:49:51 crc kubenswrapper[4791]: I1210 22:49:51.101066 4791 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-6g6sz\" (UniqueName: \"kubernetes.io/projected/6509e943-70c6-444c-bc41-48a544e36fbd-kube-api-access-6g6sz\") on node \"crc\" DevicePath \"\"" Dec 10 22:49:51 crc kubenswrapper[4791]: I1210 22:49:51.101074 4791 reconciler_common.go:293] "Volume detached for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/c03ee662-fb2f-4fc4-a2c1-af487c19d254-metrics-certs\") on node \"crc\" DevicePath \"\"" Dec 10 22:49:51 crc kubenswrapper[4791]: I1210 22:49:51.101082 4791 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-xcgwh\" (UniqueName: \"kubernetes.io/projected/fda69060-fa79-4696-b1a6-7980f124bf7c-kube-api-access-xcgwh\") on node \"crc\" DevicePath \"\"" Dec 10 22:49:51 crc kubenswrapper[4791]: I1210 22:49:51.101095 4791 reconciler_common.go:293] "Volume detached for volume \"etcd-ca\" (UniqueName: \"kubernetes.io/configmap/09efc573-dbb6-4249-bd59-9b87aba8dd28-etcd-ca\") on node \"crc\" DevicePath \"\"" Dec 10 22:49:51 crc kubenswrapper[4791]: I1210 22:49:51.101103 4791 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-ocp-branding-template\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-ocp-branding-template\") on node \"crc\" DevicePath \"\"" Dec 10 22:49:51 crc kubenswrapper[4791]: I1210 22:49:51.101113 4791 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c-serving-cert\") on node \"crc\" DevicePath \"\"" Dec 10 22:49:51 crc kubenswrapper[4791]: I1210 22:49:51.101121 4791 reconciler_common.go:293] "Volume detached for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/09efc573-dbb6-4249-bd59-9b87aba8dd28-etcd-client\") on node \"crc\" DevicePath \"\"" Dec 10 22:49:51 crc kubenswrapper[4791]: I1210 22:49:51.101128 4791 reconciler_common.go:293] "Volume detached for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/5441d097-087c-4d9a-baa8-b210afa90fc9-client-ca\") on node \"crc\" DevicePath \"\"" Dec 10 22:49:51 crc kubenswrapper[4791]: I1210 22:49:51.101136 4791 reconciler_common.go:293] "Volume detached for volume \"serviceca\" (UniqueName: \"kubernetes.io/configmap/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59-serviceca\") on node \"crc\" DevicePath \"\"" Dec 10 22:49:51 crc kubenswrapper[4791]: I1210 22:49:51.101144 4791 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-mg5zb\" (UniqueName: \"kubernetes.io/projected/6402fda4-df10-493c-b4e5-d0569419652d-kube-api-access-mg5zb\") on node \"crc\" DevicePath \"\"" Dec 10 22:49:51 crc kubenswrapper[4791]: I1210 22:49:51.101152 4791 reconciler_common.go:293] "Volume detached for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/a31745f5-9847-4afe-82a5-3161cc66ca93-trusted-ca\") on node \"crc\" DevicePath \"\"" Dec 10 22:49:51 crc kubenswrapper[4791]: I1210 22:49:51.101160 4791 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-4d4hj\" (UniqueName: \"kubernetes.io/projected/3ab1a177-2de0-46d9-b765-d0d0649bb42e-kube-api-access-4d4hj\") on node \"crc\" DevicePath \"\"" Dec 10 22:49:51 crc kubenswrapper[4791]: I1210 22:49:51.101168 4791 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-sb6h7\" (UniqueName: \"kubernetes.io/projected/1bf7eb37-55a3-4c65-b768-a94c82151e69-kube-api-access-sb6h7\") on node \"crc\" DevicePath \"\"" Dec 10 22:49:51 crc kubenswrapper[4791]: I1210 22:49:51.101175 4791 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-vt5rc\" (UniqueName: \"kubernetes.io/projected/44663579-783b-4372-86d6-acf235a62d72-kube-api-access-vt5rc\") on node \"crc\" DevicePath \"\"" Dec 10 22:49:51 crc kubenswrapper[4791]: I1210 22:49:51.101182 4791 reconciler_common.go:293] "Volume detached for volume \"certs\" (UniqueName: \"kubernetes.io/secret/5fe579f8-e8a6-4643-bce5-a661393c4dde-certs\") on node \"crc\" DevicePath \"\"" Dec 10 22:49:51 crc kubenswrapper[4791]: I1210 22:49:51.101193 4791 reconciler_common.go:293] "Volume detached for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/fda69060-fa79-4696-b1a6-7980f124bf7c-proxy-tls\") on node \"crc\" DevicePath \"\"" Dec 10 22:49:51 crc kubenswrapper[4791]: I1210 22:49:51.101200 4791 reconciler_common.go:293] "Volume detached for volume \"etcd-serving-ca\" (UniqueName: \"kubernetes.io/configmap/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-etcd-serving-ca\") on node \"crc\" DevicePath \"\"" Dec 10 22:49:51 crc kubenswrapper[4791]: I1210 22:49:51.101208 4791 reconciler_common.go:293] "Volume detached for volume \"console-config\" (UniqueName: \"kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-console-config\") on node \"crc\" DevicePath \"\"" Dec 10 22:49:51 crc kubenswrapper[4791]: I1210 22:49:51.101216 4791 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-qg5z5\" (UniqueName: \"kubernetes.io/projected/43509403-f426-496e-be36-56cef71462f5-kube-api-access-qg5z5\") on node \"crc\" DevicePath \"\"" Dec 10 22:49:51 crc kubenswrapper[4791]: I1210 22:49:51.101224 4791 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-htfz6\" (UniqueName: \"kubernetes.io/projected/6ea678ab-3438-413e-bfe3-290ae7725660-kube-api-access-htfz6\") on node \"crc\" DevicePath \"\"" Dec 10 22:49:51 crc kubenswrapper[4791]: I1210 22:49:51.101232 4791 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/5441d097-087c-4d9a-baa8-b210afa90fc9-config\") on node \"crc\" DevicePath \"\"" Dec 10 22:49:51 crc kubenswrapper[4791]: I1210 22:49:51.101239 4791 reconciler_common.go:293] "Volume detached for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/e7e6199b-1264-4501-8953-767f51328d08-kube-api-access\") on node \"crc\" DevicePath \"\"" Dec 10 22:49:51 crc kubenswrapper[4791]: I1210 22:49:51.101247 4791 reconciler_common.go:293] "Volume detached for volume \"auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/22c825df-677d-4ca6-82db-3454ed06e783-auth-proxy-config\") on node \"crc\" DevicePath \"\"" Dec 10 22:49:51 crc kubenswrapper[4791]: I1210 22:49:51.101254 4791 reconciler_common.go:293] "Volume detached for volume \"marketplace-trusted-ca\" (UniqueName: \"kubernetes.io/configmap/b6cd30de-2eeb-49a2-ab40-9167f4560ff5-marketplace-trusted-ca\") on node \"crc\" DevicePath \"\"" Dec 10 22:49:51 crc kubenswrapper[4791]: I1210 22:49:51.101262 4791 reconciler_common.go:293] "Volume detached for volume \"auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/31d8b7a1-420e-4252-a5b7-eebe8a111292-auth-proxy-config\") on node \"crc\" DevicePath \"\"" Dec 10 22:49:51 crc kubenswrapper[4791]: I1210 22:49:51.101270 4791 reconciler_common.go:293] "Volume detached for volume \"cni-binary-copy\" (UniqueName: \"kubernetes.io/configmap/7bb08738-c794-4ee8-9972-3a62ca171029-cni-binary-copy\") on node \"crc\" DevicePath \"\"" Dec 10 22:49:51 crc kubenswrapper[4791]: I1210 22:49:51.101278 4791 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-xcphl\" (UniqueName: \"kubernetes.io/projected/7583ce53-e0fe-4a16-9e4d-50516596a136-kube-api-access-xcphl\") on node \"crc\" DevicePath \"\"" Dec 10 22:49:51 crc kubenswrapper[4791]: I1210 22:49:51.101285 4791 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/5441d097-087c-4d9a-baa8-b210afa90fc9-serving-cert\") on node \"crc\" DevicePath \"\"" Dec 10 22:49:51 crc kubenswrapper[4791]: I1210 22:49:51.101293 4791 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-serving-cert\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-serving-cert\") on node \"crc\" DevicePath \"\"" Dec 10 22:49:51 crc kubenswrapper[4791]: I1210 22:49:51.101311 4791 reconciler_common.go:293] "Volume detached for volume \"machine-approver-tls\" (UniqueName: \"kubernetes.io/secret/22c825df-677d-4ca6-82db-3454ed06e783-machine-approver-tls\") on node \"crc\" DevicePath \"\"" Dec 10 22:49:51 crc kubenswrapper[4791]: I1210 22:49:51.101320 4791 reconciler_common.go:293] "Volume detached for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/1bf7eb37-55a3-4c65-b768-a94c82151e69-etcd-client\") on node \"crc\" DevicePath \"\"" Dec 10 22:49:51 crc kubenswrapper[4791]: I1210 22:49:51.101330 4791 reconciler_common.go:293] "Volume detached for volume \"ovn-node-metrics-cert\" (UniqueName: \"kubernetes.io/secret/6ea678ab-3438-413e-bfe3-290ae7725660-ovn-node-metrics-cert\") on node \"crc\" DevicePath \"\"" Dec 10 22:49:51 crc kubenswrapper[4791]: I1210 22:49:51.101368 4791 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-tk88c\" (UniqueName: \"kubernetes.io/projected/7539238d-5fe0-46ed-884e-1c3b566537ec-kube-api-access-tk88c\") on node \"crc\" DevicePath \"\"" Dec 10 22:49:51 crc kubenswrapper[4791]: I1210 22:49:51.101376 4791 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/e7e6199b-1264-4501-8953-767f51328d08-config\") on node \"crc\" DevicePath \"\"" Dec 10 22:49:51 crc kubenswrapper[4791]: I1210 22:49:51.101386 4791 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-user-template-provider-selection\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-template-provider-selection\") on node \"crc\" DevicePath \"\"" Dec 10 22:49:51 crc kubenswrapper[4791]: I1210 22:49:51.101394 4791 reconciler_common.go:293] "Volume detached for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-audit-policies\") on node \"crc\" DevicePath \"\"" Dec 10 22:49:51 crc kubenswrapper[4791]: I1210 22:49:51.101401 4791 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/7539238d-5fe0-46ed-884e-1c3b566537ec-config\") on node \"crc\" DevicePath \"\"" Dec 10 22:49:51 crc kubenswrapper[4791]: I1210 22:49:51.101409 4791 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-8tdtz\" (UniqueName: \"kubernetes.io/projected/09efc573-dbb6-4249-bd59-9b87aba8dd28-kube-api-access-8tdtz\") on node \"crc\" DevicePath \"\"" Dec 10 22:49:51 crc kubenswrapper[4791]: I1210 22:49:51.101417 4791 reconciler_common.go:293] "Volume detached for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-trusted-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 10 22:49:51 crc kubenswrapper[4791]: I1210 22:49:51.101425 4791 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-gf66m\" (UniqueName: \"kubernetes.io/projected/a0128f3a-b052-44ed-a84e-c4c8aaf17c13-kube-api-access-gf66m\") on node \"crc\" DevicePath \"\"" Dec 10 22:49:51 crc kubenswrapper[4791]: I1210 22:49:51.101433 4791 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-6ccd8\" (UniqueName: \"kubernetes.io/projected/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-kube-api-access-6ccd8\") on node \"crc\" DevicePath \"\"" Dec 10 22:49:51 crc kubenswrapper[4791]: I1210 22:49:51.101440 4791 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-x4zgh\" (UniqueName: \"kubernetes.io/projected/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d-kube-api-access-x4zgh\") on node \"crc\" DevicePath \"\"" Dec 10 22:49:51 crc kubenswrapper[4791]: I1210 22:49:51.101454 4791 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/7539238d-5fe0-46ed-884e-1c3b566537ec-serving-cert\") on node \"crc\" DevicePath \"\"" Dec 10 22:49:51 crc kubenswrapper[4791]: I1210 22:49:51.101462 4791 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-config\") on node \"crc\" DevicePath \"\"" Dec 10 22:49:51 crc kubenswrapper[4791]: I1210 22:49:51.101470 4791 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-d6qdx\" (UniqueName: \"kubernetes.io/projected/87cf06ed-a83f-41a7-828d-70653580a8cb-kube-api-access-d6qdx\") on node \"crc\" DevicePath \"\"" Dec 10 22:49:51 crc kubenswrapper[4791]: I1210 22:49:51.101481 4791 reconciler_common.go:293] "Volume detached for volume \"images\" (UniqueName: \"kubernetes.io/configmap/31d8b7a1-420e-4252-a5b7-eebe8a111292-images\") on node \"crc\" DevicePath \"\"" Dec 10 22:49:51 crc kubenswrapper[4791]: I1210 22:49:51.101488 4791 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/5225d0e4-402f-4861-b410-819f433b1803-utilities\") on node \"crc\" DevicePath \"\"" Dec 10 22:49:51 crc kubenswrapper[4791]: I1210 22:49:51.101496 4791 reconciler_common.go:293] "Volume detached for volume \"stats-auth\" (UniqueName: \"kubernetes.io/secret/c03ee662-fb2f-4fc4-a2c1-af487c19d254-stats-auth\") on node \"crc\" DevicePath \"\"" Dec 10 22:49:51 crc kubenswrapper[4791]: I1210 22:49:51.101504 4791 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/7583ce53-e0fe-4a16-9e4d-50516596a136-serving-cert\") on node \"crc\" DevicePath \"\"" Dec 10 22:49:51 crc kubenswrapper[4791]: I1210 22:49:51.101512 4791 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-jkwtn\" (UniqueName: \"kubernetes.io/projected/5b88f790-22fa-440e-b583-365168c0b23d-kube-api-access-jkwtn\") on node \"crc\" DevicePath \"\"" Dec 10 22:49:51 crc kubenswrapper[4791]: I1210 22:49:51.101519 4791 reconciler_common.go:293] "Volume detached for volume \"registry-tls\" (UniqueName: \"kubernetes.io/projected/8f668bae-612b-4b75-9490-919e737c6a3b-registry-tls\") on node \"crc\" DevicePath \"\"" Dec 10 22:49:51 crc kubenswrapper[4791]: I1210 22:49:51.101526 4791 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/496e6271-fb68-4057-954e-a0d97a4afa3f-config\") on node \"crc\" DevicePath \"\"" Dec 10 22:49:51 crc kubenswrapper[4791]: I1210 22:49:51.101534 4791 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/6509e943-70c6-444c-bc41-48a544e36fbd-config\") on node \"crc\" DevicePath \"\"" Dec 10 22:49:51 crc kubenswrapper[4791]: I1210 22:49:51.101542 4791 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-user-idp-0-file-data\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-idp-0-file-data\") on node \"crc\" DevicePath \"\"" Dec 10 22:49:51 crc kubenswrapper[4791]: I1210 22:49:51.101550 4791 reconciler_common.go:293] "Volume detached for volume \"encryption-config\" (UniqueName: \"kubernetes.io/secret/1bf7eb37-55a3-4c65-b768-a94c82151e69-encryption-config\") on node \"crc\" DevicePath \"\"" Dec 10 22:49:51 crc kubenswrapper[4791]: I1210 22:49:51.101562 4791 reconciler_common.go:293] "Volume detached for volume \"service-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/c03ee662-fb2f-4fc4-a2c1-af487c19d254-service-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 10 22:49:51 crc kubenswrapper[4791]: I1210 22:49:51.101570 4791 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-serving-cert\") on node \"crc\" DevicePath \"\"" Dec 10 22:49:51 crc kubenswrapper[4791]: I1210 22:49:51.101577 4791 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-w9rds\" (UniqueName: \"kubernetes.io/projected/20b0d48f-5fd6-431c-a545-e3c800c7b866-kube-api-access-w9rds\") on node \"crc\" DevicePath \"\"" Dec 10 22:49:51 crc kubenswrapper[4791]: I1210 22:49:51.101586 4791 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-trusted-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 10 22:49:51 crc kubenswrapper[4791]: I1210 22:49:51.101593 4791 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/7583ce53-e0fe-4a16-9e4d-50516596a136-config\") on node \"crc\" DevicePath \"\"" Dec 10 22:49:51 crc kubenswrapper[4791]: I1210 22:49:51.101604 4791 reconciler_common.go:293] "Volume detached for volume \"webhook-certs\" (UniqueName: \"kubernetes.io/secret/efdd0498-1daa-4136-9a4a-3b948c2293fc-webhook-certs\") on node \"crc\" DevicePath \"\"" Dec 10 22:49:51 crc kubenswrapper[4791]: I1210 22:49:51.101633 4791 reconciler_common.go:293] "Volume detached for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/6ea678ab-3438-413e-bfe3-290ae7725660-env-overrides\") on node \"crc\" DevicePath \"\"" Dec 10 22:49:51 crc kubenswrapper[4791]: I1210 22:49:51.101641 4791 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/1d611f23-29be-4491-8495-bee1670e935f-utilities\") on node \"crc\" DevicePath \"\"" Dec 10 22:49:51 crc kubenswrapper[4791]: I1210 22:49:51.101649 4791 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-pcxfs\" (UniqueName: \"kubernetes.io/projected/9d4552c7-cd75-42dd-8880-30dd377c49a4-kube-api-access-pcxfs\") on node \"crc\" DevicePath \"\"" Dec 10 22:49:51 crc kubenswrapper[4791]: I1210 22:49:51.101656 4791 reconciler_common.go:293] "Volume detached for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/9d4552c7-cd75-42dd-8880-30dd377c49a4-trusted-ca\") on node \"crc\" DevicePath \"\"" Dec 10 22:49:51 crc kubenswrapper[4791]: I1210 22:49:51.101664 4791 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b-config\") on node \"crc\" DevicePath \"\"" Dec 10 22:49:51 crc kubenswrapper[4791]: I1210 22:49:51.101681 4791 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-kfwg7\" (UniqueName: \"kubernetes.io/projected/8f668bae-612b-4b75-9490-919e737c6a3b-kube-api-access-kfwg7\") on node \"crc\" DevicePath \"\"" Dec 10 22:49:51 crc kubenswrapper[4791]: I1210 22:49:51.101689 4791 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d-catalog-content\") on node \"crc\" DevicePath \"\"" Dec 10 22:49:51 crc kubenswrapper[4791]: I1210 22:49:51.101715 4791 reconciler_common.go:293] "Volume detached for volume \"tmpfs\" (UniqueName: \"kubernetes.io/empty-dir/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-tmpfs\") on node \"crc\" DevicePath \"\"" Dec 10 22:49:51 crc kubenswrapper[4791]: I1210 22:49:51.101723 4791 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-279lb\" (UniqueName: \"kubernetes.io/projected/7bb08738-c794-4ee8-9972-3a62ca171029-kube-api-access-279lb\") on node \"crc\" DevicePath \"\"" Dec 10 22:49:51 crc kubenswrapper[4791]: I1210 22:49:51.101731 4791 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-mnrrd\" (UniqueName: \"kubernetes.io/projected/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d-kube-api-access-mnrrd\") on node \"crc\" DevicePath \"\"" Dec 10 22:49:51 crc kubenswrapper[4791]: I1210 22:49:51.101740 4791 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-lz9wn\" (UniqueName: \"kubernetes.io/projected/a31745f5-9847-4afe-82a5-3161cc66ca93-kube-api-access-lz9wn\") on node \"crc\" DevicePath \"\"" Dec 10 22:49:51 crc kubenswrapper[4791]: I1210 22:49:51.101747 4791 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-zgdk5\" (UniqueName: \"kubernetes.io/projected/31d8b7a1-420e-4252-a5b7-eebe8a111292-kube-api-access-zgdk5\") on node \"crc\" DevicePath \"\"" Dec 10 22:49:51 crc kubenswrapper[4791]: I1210 22:49:51.101755 4791 reconciler_common.go:293] "Volume detached for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/5b88f790-22fa-440e-b583-365168c0b23d-metrics-certs\") on node \"crc\" DevicePath \"\"" Dec 10 22:49:51 crc kubenswrapper[4791]: I1210 22:49:51.101766 4791 reconciler_common.go:293] "Volume detached for volume \"marketplace-operator-metrics\" (UniqueName: \"kubernetes.io/secret/b6cd30de-2eeb-49a2-ab40-9167f4560ff5-marketplace-operator-metrics\") on node \"crc\" DevicePath \"\"" Dec 10 22:49:51 crc kubenswrapper[4791]: I1210 22:49:51.101774 4791 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-7c4vf\" (UniqueName: \"kubernetes.io/projected/22c825df-677d-4ca6-82db-3454ed06e783-kube-api-access-7c4vf\") on node \"crc\" DevicePath \"\"" Dec 10 22:49:51 crc kubenswrapper[4791]: I1210 22:49:51.101782 4791 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/1386a44e-36a2-460c-96d0-0359d2b6f0f5-config\") on node \"crc\" DevicePath \"\"" Dec 10 22:49:51 crc kubenswrapper[4791]: I1210 22:49:51.101790 4791 reconciler_common.go:293] "Volume detached for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/7583ce53-e0fe-4a16-9e4d-50516596a136-client-ca\") on node \"crc\" DevicePath \"\"" Dec 10 22:49:51 crc kubenswrapper[4791]: I1210 22:49:51.101797 4791 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-jhbk2\" (UniqueName: \"kubernetes.io/projected/bd23aa5c-e532-4e53-bccf-e79f130c5ae8-kube-api-access-jhbk2\") on node \"crc\" DevicePath \"\"" Dec 10 22:49:51 crc kubenswrapper[4791]: I1210 22:49:51.101805 4791 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-qs4fp\" (UniqueName: \"kubernetes.io/projected/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c-kube-api-access-qs4fp\") on node \"crc\" DevicePath \"\"" Dec 10 22:49:51 crc kubenswrapper[4791]: I1210 22:49:51.101821 4791 reconciler_common.go:293] "Volume detached for volume \"cni-binary-copy\" (UniqueName: \"kubernetes.io/configmap/4bb40260-dbaa-4fb0-84df-5e680505d512-cni-binary-copy\") on node \"crc\" DevicePath \"\"" Dec 10 22:49:51 crc kubenswrapper[4791]: I1210 22:49:51.101829 4791 reconciler_common.go:293] "Volume detached for volume \"images\" (UniqueName: \"kubernetes.io/configmap/6402fda4-df10-493c-b4e5-d0569419652d-images\") on node \"crc\" DevicePath \"\"" Dec 10 22:49:51 crc kubenswrapper[4791]: I1210 22:49:51.101844 4791 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/9d4552c7-cd75-42dd-8880-30dd377c49a4-serving-cert\") on node \"crc\" DevicePath \"\"" Dec 10 22:49:51 crc kubenswrapper[4791]: I1210 22:49:51.101852 4791 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/1bf7eb37-55a3-4c65-b768-a94c82151e69-serving-cert\") on node \"crc\" DevicePath \"\"" Dec 10 22:49:51 crc kubenswrapper[4791]: I1210 22:49:51.101860 4791 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-rnphk\" (UniqueName: \"kubernetes.io/projected/bf126b07-da06-4140-9a57-dfd54fc6b486-kube-api-access-rnphk\") on node \"crc\" DevicePath \"\"" Dec 10 22:49:51 crc kubenswrapper[4791]: I1210 22:49:51.101867 4791 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/09efc573-dbb6-4249-bd59-9b87aba8dd28-config\") on node \"crc\" DevicePath \"\"" Dec 10 22:49:51 crc kubenswrapper[4791]: I1210 22:49:51.101875 4791 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/6402fda4-df10-493c-b4e5-d0569419652d-config\") on node \"crc\" DevicePath \"\"" Dec 10 22:49:51 crc kubenswrapper[4791]: I1210 22:49:51.101883 4791 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c-config\") on node \"crc\" DevicePath \"\"" Dec 10 22:49:51 crc kubenswrapper[4791]: I1210 22:49:51.101891 4791 reconciler_common.go:293] "Volume detached for volume \"signing-cabundle\" (UniqueName: \"kubernetes.io/configmap/25e176fe-21b4-4974-b1ed-c8b94f112a7f-signing-cabundle\") on node \"crc\" DevicePath \"\"" Dec 10 22:49:51 crc kubenswrapper[4791]: I1210 22:49:51.101899 4791 reconciler_common.go:293] "Volume detached for volume \"signing-key\" (UniqueName: \"kubernetes.io/secret/25e176fe-21b4-4974-b1ed-c8b94f112a7f-signing-key\") on node \"crc\" DevicePath \"\"" Dec 10 22:49:51 crc kubenswrapper[4791]: I1210 22:49:51.101915 4791 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/6509e943-70c6-444c-bc41-48a544e36fbd-serving-cert\") on node \"crc\" DevicePath \"\"" Dec 10 22:49:51 crc kubenswrapper[4791]: I1210 22:49:51.101922 4791 reconciler_common.go:293] "Volume detached for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/7583ce53-e0fe-4a16-9e4d-50516596a136-proxy-ca-bundles\") on node \"crc\" DevicePath \"\"" Dec 10 22:49:51 crc kubenswrapper[4791]: I1210 22:49:51.101930 4791 reconciler_common.go:293] "Volume detached for volume \"cni-sysctl-allowlist\" (UniqueName: \"kubernetes.io/configmap/7bb08738-c794-4ee8-9972-3a62ca171029-cni-sysctl-allowlist\") on node \"crc\" DevicePath \"\"" Dec 10 22:49:51 crc kubenswrapper[4791]: I1210 22:49:51.101952 4791 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-pjr6v\" (UniqueName: \"kubernetes.io/projected/49ef4625-1d3a-4a9f-b595-c2433d32326d-kube-api-access-pjr6v\") on node \"crc\" DevicePath \"\"" Dec 10 22:49:51 crc kubenswrapper[4791]: I1210 22:49:51.101960 4791 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-ngvvp\" (UniqueName: \"kubernetes.io/projected/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-kube-api-access-ngvvp\") on node \"crc\" DevicePath \"\"" Dec 10 22:49:51 crc kubenswrapper[4791]: I1210 22:49:51.101968 4791 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-x7zkh\" (UniqueName: \"kubernetes.io/projected/6731426b-95fe-49ff-bb5f-40441049fde2-kube-api-access-x7zkh\") on node \"crc\" DevicePath \"\"" Dec 10 22:49:51 crc kubenswrapper[4791]: I1210 22:49:51.101975 4791 reconciler_common.go:293] "Volume detached for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/0b78653f-4ff9-4508-8672-245ed9b561e3-kube-api-access\") on node \"crc\" DevicePath \"\"" Dec 10 22:49:51 crc kubenswrapper[4791]: I1210 22:49:51.101983 4791 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-cliconfig\" (UniqueName: \"kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-cliconfig\") on node \"crc\" DevicePath \"\"" Dec 10 22:49:51 crc kubenswrapper[4791]: I1210 22:49:51.101991 4791 reconciler_common.go:293] "Volume detached for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/a31745f5-9847-4afe-82a5-3161cc66ca93-bound-sa-token\") on node \"crc\" DevicePath \"\"" Dec 10 22:49:51 crc kubenswrapper[4791]: I1210 22:49:51.101998 4791 reconciler_common.go:293] "Volume detached for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-trusted-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 10 22:49:51 crc kubenswrapper[4791]: I1210 22:49:51.102006 4791 reconciler_common.go:293] "Volume detached for volume \"machine-api-operator-tls\" (UniqueName: \"kubernetes.io/secret/6402fda4-df10-493c-b4e5-d0569419652d-machine-api-operator-tls\") on node \"crc\" DevicePath \"\"" Dec 10 22:49:51 crc kubenswrapper[4791]: I1210 22:49:51.102015 4791 reconciler_common.go:293] "Volume detached for volume \"package-server-manager-serving-cert\" (UniqueName: \"kubernetes.io/secret/3ab1a177-2de0-46d9-b765-d0d0649bb42e-package-server-manager-serving-cert\") on node \"crc\" DevicePath \"\"" Dec 10 22:49:51 crc kubenswrapper[4791]: I1210 22:49:51.102024 4791 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-9xfj7\" (UniqueName: \"kubernetes.io/projected/5225d0e4-402f-4861-b410-819f433b1803-kube-api-access-9xfj7\") on node \"crc\" DevicePath \"\"" Dec 10 22:49:51 crc kubenswrapper[4791]: I1210 22:49:51.102032 4791 reconciler_common.go:293] "Volume detached for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/bf126b07-da06-4140-9a57-dfd54fc6b486-trusted-ca\") on node \"crc\" DevicePath \"\"" Dec 10 22:49:51 crc kubenswrapper[4791]: I1210 22:49:51.102055 4791 reconciler_common.go:293] "Volume detached for volume \"etcd-serving-ca\" (UniqueName: \"kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-etcd-serving-ca\") on node \"crc\" DevicePath \"\"" Dec 10 22:49:51 crc kubenswrapper[4791]: I1210 22:49:51.102063 4791 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-router-certs\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-router-certs\") on node \"crc\" DevicePath \"\"" Dec 10 22:49:51 crc kubenswrapper[4791]: I1210 22:49:51.102073 4791 reconciler_common.go:293] "Volume detached for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-trusted-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 10 22:49:51 crc kubenswrapper[4791]: I1210 22:49:51.102081 4791 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-bf2bz\" (UniqueName: \"kubernetes.io/projected/1d611f23-29be-4491-8495-bee1670e935f-kube-api-access-bf2bz\") on node \"crc\" DevicePath \"\"" Dec 10 22:49:51 crc kubenswrapper[4791]: I1210 22:49:51.102088 4791 reconciler_common.go:293] "Volume detached for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/bf126b07-da06-4140-9a57-dfd54fc6b486-bound-sa-token\") on node \"crc\" DevicePath \"\"" Dec 10 22:49:51 crc kubenswrapper[4791]: I1210 22:49:51.102096 4791 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-w7l8j\" (UniqueName: \"kubernetes.io/projected/01ab3dd5-8196-46d0-ad33-122e2ca51def-kube-api-access-w7l8j\") on node \"crc\" DevicePath \"\"" Dec 10 22:49:51 crc kubenswrapper[4791]: I1210 22:49:51.102103 4791 reconciler_common.go:293] "Volume detached for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/31d8b7a1-420e-4252-a5b7-eebe8a111292-proxy-tls\") on node \"crc\" DevicePath \"\"" Dec 10 22:49:51 crc kubenswrapper[4791]: I1210 22:49:51.102119 4791 reconciler_common.go:293] "Volume detached for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-apiservice-cert\") on node \"crc\" DevicePath \"\"" Dec 10 22:49:51 crc kubenswrapper[4791]: I1210 22:49:51.102126 4791 reconciler_common.go:293] "Volume detached for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/96b93a3a-6083-4aea-8eab-fe1aa8245ad9-metrics-tls\") on node \"crc\" DevicePath \"\"" Dec 10 22:49:51 crc kubenswrapper[4791]: I1210 22:49:51.102150 4791 reconciler_common.go:293] "Volume detached for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/496e6271-fb68-4057-954e-a0d97a4afa3f-kube-api-access\") on node \"crc\" DevicePath \"\"" Dec 10 22:49:51 crc kubenswrapper[4791]: I1210 22:49:51.102157 4791 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/09efc573-dbb6-4249-bd59-9b87aba8dd28-serving-cert\") on node \"crc\" DevicePath \"\"" Dec 10 22:49:51 crc kubenswrapper[4791]: I1210 22:49:51.102165 4791 reconciler_common.go:293] "Volume detached for volume \"default-certificate\" (UniqueName: \"kubernetes.io/secret/c03ee662-fb2f-4fc4-a2c1-af487c19d254-default-certificate\") on node \"crc\" DevicePath \"\"" Dec 10 22:49:51 crc kubenswrapper[4791]: I1210 22:49:51.102175 4791 reconciler_common.go:293] "Volume detached for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/925f1c65-6136-48ba-85aa-3a3b50560753-ovnkube-config\") on node \"crc\" DevicePath \"\"" Dec 10 22:49:51 crc kubenswrapper[4791]: I1210 22:49:51.102182 4791 reconciler_common.go:293] "Volume detached for volume \"available-featuregates\" (UniqueName: \"kubernetes.io/empty-dir/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d-available-featuregates\") on node \"crc\" DevicePath \"\"" Dec 10 22:49:51 crc kubenswrapper[4791]: I1210 22:49:51.102192 4791 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-v47cf\" (UniqueName: \"kubernetes.io/projected/c03ee662-fb2f-4fc4-a2c1-af487c19d254-kube-api-access-v47cf\") on node \"crc\" DevicePath \"\"" Dec 10 22:49:51 crc kubenswrapper[4791]: I1210 22:49:51.102199 4791 reconciler_common.go:293] "Volume detached for volume \"ovnkube-script-lib\" (UniqueName: \"kubernetes.io/configmap/6ea678ab-3438-413e-bfe3-290ae7725660-ovnkube-script-lib\") on node \"crc\" DevicePath \"\"" Dec 10 22:49:51 crc kubenswrapper[4791]: I1210 22:49:51.102207 4791 reconciler_common.go:293] "Volume detached for volume \"samples-operator-tls\" (UniqueName: \"kubernetes.io/secret/a0128f3a-b052-44ed-a84e-c4c8aaf17c13-samples-operator-tls\") on node \"crc\" DevicePath \"\"" Dec 10 22:49:51 crc kubenswrapper[4791]: I1210 22:49:51.102214 4791 reconciler_common.go:293] "Volume detached for volume \"console-serving-cert\" (UniqueName: \"kubernetes.io/secret/43509403-f426-496e-be36-56cef71462f5-console-serving-cert\") on node \"crc\" DevicePath \"\"" Dec 10 22:49:51 crc kubenswrapper[4791]: I1210 22:49:51.102221 4791 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-user-template-error\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-template-error\") on node \"crc\" DevicePath \"\"" Dec 10 22:49:51 crc kubenswrapper[4791]: I1210 22:49:51.102231 4791 reconciler_common.go:293] "Volume detached for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-etcd-client\") on node \"crc\" DevicePath \"\"" Dec 10 22:49:51 crc kubenswrapper[4791]: I1210 22:49:51.102241 4791 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-service-ca\" (UniqueName: \"kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-service-ca\") on node \"crc\" DevicePath \"\"" Dec 10 22:49:51 crc kubenswrapper[4791]: I1210 22:49:51.102249 4791 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-cfbct\" (UniqueName: \"kubernetes.io/projected/57a731c4-ef35-47a8-b875-bfb08a7f8011-kube-api-access-cfbct\") on node \"crc\" DevicePath \"\"" Dec 10 22:49:51 crc kubenswrapper[4791]: I1210 22:49:51.102257 4791 reconciler_common.go:293] "Volume detached for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/87cf06ed-a83f-41a7-828d-70653580a8cb-config-volume\") on node \"crc\" DevicePath \"\"" Dec 10 22:49:51 crc kubenswrapper[4791]: I1210 22:49:51.102264 4791 reconciler_common.go:293] "Volume detached for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/8f668bae-612b-4b75-9490-919e737c6a3b-bound-sa-token\") on node \"crc\" DevicePath \"\"" Dec 10 22:49:51 crc kubenswrapper[4791]: I1210 22:49:51.102272 4791 reconciler_common.go:293] "Volume detached for volume \"cert\" (UniqueName: \"kubernetes.io/secret/20b0d48f-5fd6-431c-a545-e3c800c7b866-cert\") on node \"crc\" DevicePath \"\"" Dec 10 22:49:51 crc kubenswrapper[4791]: I1210 22:49:51.102279 4791 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/57a731c4-ef35-47a8-b875-bfb08a7f8011-catalog-content\") on node \"crc\" DevicePath \"\"" Dec 10 22:49:51 crc kubenswrapper[4791]: I1210 22:49:51.102287 4791 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-fqsjt\" (UniqueName: \"kubernetes.io/projected/efdd0498-1daa-4136-9a4a-3b948c2293fc-kube-api-access-fqsjt\") on node \"crc\" DevicePath \"\"" Dec 10 22:49:51 crc kubenswrapper[4791]: I1210 22:49:51.102303 4791 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/496e6271-fb68-4057-954e-a0d97a4afa3f-serving-cert\") on node \"crc\" DevicePath \"\"" Dec 10 22:49:51 crc kubenswrapper[4791]: I1210 22:49:51.102310 4791 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-lzf88\" (UniqueName: \"kubernetes.io/projected/0b574797-001e-440a-8f4e-c0be86edad0f-kube-api-access-lzf88\") on node \"crc\" DevicePath \"\"" Dec 10 22:49:51 crc kubenswrapper[4791]: I1210 22:49:51.102320 4791 reconciler_common.go:293] "Volume detached for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-service-ca\") on node \"crc\" DevicePath \"\"" Dec 10 22:49:51 crc kubenswrapper[4791]: I1210 22:49:51.102328 4791 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/22c825df-677d-4ca6-82db-3454ed06e783-config\") on node \"crc\" DevicePath \"\"" Dec 10 22:49:51 crc kubenswrapper[4791]: I1210 22:49:51.102335 4791 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b-serving-cert\") on node \"crc\" DevicePath \"\"" Dec 10 22:49:51 crc kubenswrapper[4791]: I1210 22:49:51.102359 4791 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-249nr\" (UniqueName: \"kubernetes.io/projected/b6312bbd-5731-4ea0-a20f-81d5a57df44a-kube-api-access-249nr\") on node \"crc\" DevicePath \"\"" Dec 10 22:49:51 crc kubenswrapper[4791]: I1210 22:49:51.102367 4791 reconciler_common.go:293] "Volume detached for volume \"installation-pull-secrets\" (UniqueName: \"kubernetes.io/secret/8f668bae-612b-4b75-9490-919e737c6a3b-installation-pull-secrets\") on node \"crc\" DevicePath \"\"" Dec 10 22:49:51 crc kubenswrapper[4791]: I1210 22:49:51.102378 4791 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-2w9zh\" (UniqueName: \"kubernetes.io/projected/4bb40260-dbaa-4fb0-84df-5e680505d512-kube-api-access-2w9zh\") on node \"crc\" DevicePath \"\"" Dec 10 22:49:51 crc kubenswrapper[4791]: I1210 22:49:51.102385 4791 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-wxkg8\" (UniqueName: \"kubernetes.io/projected/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59-kube-api-access-wxkg8\") on node \"crc\" DevicePath \"\"" Dec 10 22:49:51 crc kubenswrapper[4791]: I1210 22:49:51.102398 4791 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-fcqwp\" (UniqueName: \"kubernetes.io/projected/5fe579f8-e8a6-4643-bce5-a661393c4dde-kube-api-access-fcqwp\") on node \"crc\" DevicePath \"\"" Dec 10 22:49:51 crc kubenswrapper[4791]: I1210 22:49:51.102406 4791 reconciler_common.go:293] "Volume detached for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/6509e943-70c6-444c-bc41-48a544e36fbd-trusted-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 10 22:49:51 crc kubenswrapper[4791]: I1210 22:49:51.102416 4791 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-session\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-session\") on node \"crc\" DevicePath \"\"" Dec 10 22:49:51 crc kubenswrapper[4791]: I1210 22:49:51.102424 4791 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/9d4552c7-cd75-42dd-8880-30dd377c49a4-config\") on node \"crc\" DevicePath \"\"" Dec 10 22:49:51 crc kubenswrapper[4791]: I1210 22:49:51.102434 4791 reconciler_common.go:293] "Volume detached for volume \"mcc-auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/0b574797-001e-440a-8f4e-c0be86edad0f-mcc-auth-proxy-config\") on node \"crc\" DevicePath \"\"" Dec 10 22:49:51 crc kubenswrapper[4791]: I1210 22:49:51.102442 4791 reconciler_common.go:293] "Volume detached for volume \"profile-collector-cert\" (UniqueName: \"kubernetes.io/secret/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9-profile-collector-cert\") on node \"crc\" DevicePath \"\"" Dec 10 22:49:51 crc kubenswrapper[4791]: I1210 22:49:51.102449 4791 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/e7e6199b-1264-4501-8953-767f51328d08-serving-cert\") on node \"crc\" DevicePath \"\"" Dec 10 22:49:51 crc kubenswrapper[4791]: I1210 22:49:51.102457 4791 reconciler_common.go:293] "Volume detached for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-webhook-cert\") on node \"crc\" DevicePath \"\"" Dec 10 22:49:51 crc kubenswrapper[4791]: I1210 22:49:51.102467 4791 reconciler_common.go:293] "Volume detached for volume \"node-bootstrap-token\" (UniqueName: \"kubernetes.io/secret/5fe579f8-e8a6-4643-bce5-a661393c4dde-node-bootstrap-token\") on node \"crc\" DevicePath \"\"" Dec 10 22:49:51 crc kubenswrapper[4791]: I1210 22:49:51.102474 4791 reconciler_common.go:293] "Volume detached for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/1386a44e-36a2-460c-96d0-0359d2b6f0f5-kube-api-access\") on node \"crc\" DevicePath \"\"" Dec 10 22:49:51 crc kubenswrapper[4791]: I1210 22:49:51.102482 4791 reconciler_common.go:293] "Volume detached for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/8f668bae-612b-4b75-9490-919e737c6a3b-trusted-ca\") on node \"crc\" DevicePath \"\"" Dec 10 22:49:51 crc kubenswrapper[4791]: I1210 22:49:51.102489 4791 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/1386a44e-36a2-460c-96d0-0359d2b6f0f5-serving-cert\") on node \"crc\" DevicePath \"\"" Dec 10 22:49:51 crc kubenswrapper[4791]: I1210 22:49:51.102496 4791 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-nzwt7\" (UniqueName: \"kubernetes.io/projected/96b93a3a-6083-4aea-8eab-fe1aa8245ad9-kube-api-access-nzwt7\") on node \"crc\" DevicePath \"\"" Dec 10 22:49:51 crc kubenswrapper[4791]: I1210 22:49:51.102504 4791 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-dbsvg\" (UniqueName: \"kubernetes.io/projected/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9-kube-api-access-dbsvg\") on node \"crc\" DevicePath \"\"" Dec 10 22:49:51 crc kubenswrapper[4791]: I1210 22:49:51.102512 4791 reconciler_common.go:293] "Volume detached for volume \"mcd-auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/fda69060-fa79-4696-b1a6-7980f124bf7c-mcd-auth-proxy-config\") on node \"crc\" DevicePath \"\"" Dec 10 22:49:51 crc kubenswrapper[4791]: I1210 22:49:51.102520 4791 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-user-template-login\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-template-login\") on node \"crc\" DevicePath \"\"" Dec 10 22:49:51 crc kubenswrapper[4791]: I1210 22:49:51.102528 4791 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-2d4wz\" (UniqueName: \"kubernetes.io/projected/5441d097-087c-4d9a-baa8-b210afa90fc9-kube-api-access-2d4wz\") on node \"crc\" DevicePath \"\"" Dec 10 22:49:51 crc kubenswrapper[4791]: I1210 22:49:51.102547 4791 reconciler_common.go:293] "Volume detached for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/925f1c65-6136-48ba-85aa-3a3b50560753-env-overrides\") on node \"crc\" DevicePath \"\"" Dec 10 22:49:51 crc kubenswrapper[4791]: I1210 22:49:51.102554 4791 reconciler_common.go:293] "Volume detached for volume \"audit\" (UniqueName: \"kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-audit\") on node \"crc\" DevicePath \"\"" Dec 10 22:49:51 crc kubenswrapper[4791]: I1210 22:49:51.102562 4791 reconciler_common.go:293] "Volume detached for volume \"control-plane-machine-set-operator-tls\" (UniqueName: \"kubernetes.io/secret/6731426b-95fe-49ff-bb5f-40441049fde2-control-plane-machine-set-operator-tls\") on node \"crc\" DevicePath \"\"" Dec 10 22:49:51 crc kubenswrapper[4791]: I1210 22:49:51.102570 4791 reconciler_common.go:293] "Volume detached for volume \"service-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/6509e943-70c6-444c-bc41-48a544e36fbd-service-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 10 22:49:51 crc kubenswrapper[4791]: I1210 22:49:51.102581 4791 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-d4lsv\" (UniqueName: \"kubernetes.io/projected/25e176fe-21b4-4974-b1ed-c8b94f112a7f-kube-api-access-d4lsv\") on node \"crc\" DevicePath \"\"" Dec 10 22:49:51 crc kubenswrapper[4791]: I1210 22:49:51.102589 4791 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/0b78653f-4ff9-4508-8672-245ed9b561e3-serving-cert\") on node \"crc\" DevicePath \"\"" Dec 10 22:49:51 crc kubenswrapper[4791]: I1210 22:49:51.102598 4791 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d-utilities\") on node \"crc\" DevicePath \"\"" Dec 10 22:49:51 crc kubenswrapper[4791]: I1210 22:49:51.102606 4791 reconciler_common.go:293] "Volume detached for volume \"image-import-ca\" (UniqueName: \"kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-image-import-ca\") on node \"crc\" DevicePath \"\"" Dec 10 22:49:51 crc kubenswrapper[4791]: I1210 22:49:51.102614 4791 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-zkvpv\" (UniqueName: \"kubernetes.io/projected/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-kube-api-access-zkvpv\") on node \"crc\" DevicePath \"\"" Dec 10 22:49:51 crc kubenswrapper[4791]: I1210 22:49:51.102625 4791 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/57a731c4-ef35-47a8-b875-bfb08a7f8011-utilities\") on node \"crc\" DevicePath \"\"" Dec 10 22:49:51 crc kubenswrapper[4791]: I1210 22:49:51.102632 4791 reconciler_common.go:293] "Volume detached for volume \"encryption-config\" (UniqueName: \"kubernetes.io/secret/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-encryption-config\") on node \"crc\" DevicePath \"\"" Dec 10 22:49:51 crc kubenswrapper[4791]: I1210 22:49:51.102640 4791 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/01ab3dd5-8196-46d0-ad33-122e2ca51def-serving-cert\") on node \"crc\" DevicePath \"\"" Dec 10 22:49:51 crc kubenswrapper[4791]: I1210 22:49:51.102647 4791 reconciler_common.go:293] "Volume detached for volume \"profile-collector-cert\" (UniqueName: \"kubernetes.io/secret/b6312bbd-5731-4ea0-a20f-81d5a57df44a-profile-collector-cert\") on node \"crc\" DevicePath \"\"" Dec 10 22:49:51 crc kubenswrapper[4791]: I1210 22:49:51.102656 4791 reconciler_common.go:293] "Volume detached for volume \"ovn-control-plane-metrics-cert\" (UniqueName: \"kubernetes.io/secret/925f1c65-6136-48ba-85aa-3a3b50560753-ovn-control-plane-metrics-cert\") on node \"crc\" DevicePath \"\"" Dec 10 22:49:51 crc kubenswrapper[4791]: I1210 22:49:51.102666 4791 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-s4n52\" (UniqueName: \"kubernetes.io/projected/925f1c65-6136-48ba-85aa-3a3b50560753-kube-api-access-s4n52\") on node \"crc\" DevicePath \"\"" Dec 10 22:49:51 crc kubenswrapper[4791]: I1210 22:49:51.102674 4791 reconciler_common.go:293] "Volume detached for volume \"srv-cert\" (UniqueName: \"kubernetes.io/secret/b6312bbd-5731-4ea0-a20f-81d5a57df44a-srv-cert\") on node \"crc\" DevicePath \"\"" Dec 10 22:49:51 crc kubenswrapper[4791]: I1210 22:49:51.102681 4791 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-pj782\" (UniqueName: \"kubernetes.io/projected/b6cd30de-2eeb-49a2-ab40-9167f4560ff5-kube-api-access-pj782\") on node \"crc\" DevicePath \"\"" Dec 10 22:49:51 crc kubenswrapper[4791]: I1210 22:49:51.102705 4791 reconciler_common.go:293] "Volume detached for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/0b574797-001e-440a-8f4e-c0be86edad0f-proxy-tls\") on node \"crc\" DevicePath \"\"" Dec 10 22:49:51 crc kubenswrapper[4791]: I1210 22:49:51.102722 4791 reconciler_common.go:293] "Volume detached for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/a31745f5-9847-4afe-82a5-3161cc66ca93-metrics-tls\") on node \"crc\" DevicePath \"\"" Dec 10 22:49:51 crc kubenswrapper[4791]: I1210 22:49:51.102745 4791 reconciler_common.go:293] "Volume detached for volume \"image-registry-operator-tls\" (UniqueName: \"kubernetes.io/secret/bf126b07-da06-4140-9a57-dfd54fc6b486-image-registry-operator-tls\") on node \"crc\" DevicePath \"\"" Dec 10 22:49:51 crc kubenswrapper[4791]: I1210 22:49:51.103639 4791 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serviceca\" (UniqueName: \"kubernetes.io/configmap/b9f00f25-3fa1-4b63-9710-fd352224b01b-serviceca\") pod \"node-ca-vczjq\" (UID: \"b9f00f25-3fa1-4b63-9710-fd352224b01b\") " pod="openshift-image-registry/node-ca-vczjq" Dec 10 22:49:51 crc kubenswrapper[4791]: I1210 22:49:51.103682 4791 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host\" (UniqueName: \"kubernetes.io/host-path/b9f00f25-3fa1-4b63-9710-fd352224b01b-host\") pod \"node-ca-vczjq\" (UID: \"b9f00f25-3fa1-4b63-9710-fd352224b01b\") " pod="openshift-image-registry/node-ca-vczjq" Dec 10 22:49:51 crc kubenswrapper[4791]: I1210 22:49:51.103706 4791 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-slash\" (UniqueName: \"kubernetes.io/host-path/d75a4c96-2883-4a0b-bab2-0fab2b6c0b49-host-slash\") pod \"iptables-alerter-4ln5h\" (UID: \"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\") " pod="openshift-network-operator/iptables-alerter-4ln5h" Dec 10 22:49:51 crc kubenswrapper[4791]: I1210 22:49:51.103742 4791 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-etc-kube\" (UniqueName: \"kubernetes.io/host-path/37a5e44f-9a88-4405-be8a-b645485e7312-host-etc-kube\") pod \"network-operator-58b4c7f79c-55gtf\" (UID: \"37a5e44f-9a88-4405-be8a-b645485e7312\") " pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" Dec 10 22:49:51 crc kubenswrapper[4791]: I1210 22:49:51.103855 4791 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"hosts-file\" (UniqueName: \"kubernetes.io/host-path/c0ab74e0-42a9-4f42-ac80-16030cf4ffff-hosts-file\") pod \"node-resolver-tchxk\" (UID: \"c0ab74e0-42a9-4f42-ac80-16030cf4ffff\") " pod="openshift-dns/node-resolver-tchxk" Dec 10 22:49:51 crc kubenswrapper[4791]: I1210 22:49:51.104092 4791 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/8f668bae-612b-4b75-9490-919e737c6a3b-ca-trust-extracted" (OuterVolumeSpecName: "ca-trust-extracted") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b"). InnerVolumeSpecName "ca-trust-extracted". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 10 22:49:51 crc kubenswrapper[4791]: I1210 22:49:51.104190 4791 reconciler_common.go:293] "Volume detached for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/87cf06ed-a83f-41a7-828d-70653580a8cb-metrics-tls\") on node \"crc\" DevicePath \"\"" Dec 10 22:49:51 crc kubenswrapper[4791]: I1210 22:49:51.104242 4791 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/01ab3dd5-8196-46d0-ad33-122e2ca51def-config\") on node \"crc\" DevicePath \"\"" Dec 10 22:49:51 crc kubenswrapper[4791]: I1210 22:49:51.104255 4791 reconciler_common.go:293] "Volume detached for volume \"srv-cert\" (UniqueName: \"kubernetes.io/secret/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9-srv-cert\") on node \"crc\" DevicePath \"\"" Dec 10 22:49:51 crc kubenswrapper[4791]: I1210 22:49:51.107597 4791 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-10T22:49:50Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-10T22:49:50Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Dec 10 22:49:51 crc kubenswrapper[4791]: I1210 22:49:51.118242 4791 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/5225d0e4-402f-4861-b410-819f433b1803-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "5225d0e4-402f-4861-b410-819f433b1803" (UID: "5225d0e4-402f-4861-b410-819f433b1803"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 10 22:49:51 crc kubenswrapper[4791]: I1210 22:49:51.134099 4791 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-10T22:49:50Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-10T22:49:50Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Dec 10 22:49:51 crc kubenswrapper[4791]: I1210 22:49:51.137850 4791 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-lrrpv\" (UniqueName: \"kubernetes.io/projected/b9f00f25-3fa1-4b63-9710-fd352224b01b-kube-api-access-lrrpv\") pod \"node-ca-vczjq\" (UID: \"b9f00f25-3fa1-4b63-9710-fd352224b01b\") " pod="openshift-image-registry/node-ca-vczjq" Dec 10 22:49:51 crc kubenswrapper[4791]: I1210 22:49:51.138823 4791 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-zn2qt\" (UniqueName: \"kubernetes.io/projected/c0ab74e0-42a9-4f42-ac80-16030cf4ffff-kube-api-access-zn2qt\") pod \"node-resolver-tchxk\" (UID: \"c0ab74e0-42a9-4f42-ac80-16030cf4ffff\") " pod="openshift-dns/node-resolver-tchxk" Dec 10 22:49:51 crc kubenswrapper[4791]: I1210 22:49:51.145787 4791 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-tchxk" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"c0ab74e0-42a9-4f42-ac80-16030cf4ffff\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T22:49:50Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T22:49:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T22:49:50Z\\\",\\\"message\\\":\\\"containers with unready status: [dns-node-resolver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T22:49:50Z\\\",\\\"message\\\":\\\"containers with unready status: [dns-node-resolver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zn2qt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-10T22:49:50Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-tchxk\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Dec 10 22:49:51 crc kubenswrapper[4791]: I1210 22:49:51.156558 4791 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-10T22:49:50Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-operator]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-10T22:49:50Z\\\",\\\"message\\\":\\\"containers with unready status: [network-operator]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":false,\\\"restartCount\\\":5,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Dec 10 22:49:51 crc kubenswrapper[4791]: I1210 22:49:51.163538 4791 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" Dec 10 22:49:51 crc kubenswrapper[4791]: I1210 22:49:51.175335 4791 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-node-identity/network-node-identity-vrzqb" Dec 10 22:49:51 crc kubenswrapper[4791]: E1210 22:49:51.179250 4791 kubelet.go:1929] "Failed creating a mirror pod for" err="pods \"kube-apiserver-crc\" already exists" pod="openshift-kube-apiserver/kube-apiserver-crc" Dec 10 22:49:51 crc kubenswrapper[4791]: E1210 22:49:51.180043 4791 kubelet.go:1929] "Failed creating a mirror pod for" err="pods \"kube-controller-manager-crc\" already exists" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Dec 10 22:49:51 crc kubenswrapper[4791]: I1210 22:49:51.183597 4791 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-operator/iptables-alerter-4ln5h" Dec 10 22:49:51 crc kubenswrapper[4791]: I1210 22:49:51.190763 4791 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-dns/node-resolver-tchxk" Dec 10 22:49:51 crc kubenswrapper[4791]: I1210 22:49:51.205758 4791 reconciler_common.go:293] "Volume detached for volume \"ca-trust-extracted\" (UniqueName: \"kubernetes.io/empty-dir/8f668bae-612b-4b75-9490-919e737c6a3b-ca-trust-extracted\") on node \"crc\" DevicePath \"\"" Dec 10 22:49:51 crc kubenswrapper[4791]: I1210 22:49:51.205795 4791 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/5225d0e4-402f-4861-b410-819f433b1803-catalog-content\") on node \"crc\" DevicePath \"\"" Dec 10 22:49:51 crc kubenswrapper[4791]: I1210 22:49:51.215585 4791 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-image-registry/node-ca-vczjq" Dec 10 22:49:51 crc kubenswrapper[4791]: W1210 22:49:51.229615 4791 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podc0ab74e0_42a9_4f42_ac80_16030cf4ffff.slice/crio-274a2d7c02eee2d83ecf84953663519fe920414b624e92f4e28108fb844af339 WatchSource:0}: Error finding container 274a2d7c02eee2d83ecf84953663519fe920414b624e92f4e28108fb844af339: Status 404 returned error can't find the container with id 274a2d7c02eee2d83ecf84953663519fe920414b624e92f4e28108fb844af339 Dec 10 22:49:51 crc kubenswrapper[4791]: I1210 22:49:51.510099 4791 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 10 22:49:51 crc kubenswrapper[4791]: E1210 22:49:51.510272 4791 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-10 22:49:52.510257253 +0000 UTC m=+26.939874866 (durationBeforeRetry 1s). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 10 22:49:51 crc kubenswrapper[4791]: I1210 22:49:51.611322 4791 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nginx-conf\" (UniqueName: \"kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 10 22:49:51 crc kubenswrapper[4791]: I1210 22:49:51.611405 4791 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"networking-console-plugin-cert\" (UniqueName: \"kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 10 22:49:51 crc kubenswrapper[4791]: I1210 22:49:51.611437 4791 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cqllr\" (UniqueName: \"kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr\") pod \"network-check-target-xd92c\" (UID: \"3b6479f0-333b-4a96-9adf-2099afdc2447\") " pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 10 22:49:51 crc kubenswrapper[4791]: E1210 22:49:51.611452 4791 configmap.go:193] Couldn't get configMap openshift-network-console/networking-console-plugin: object "openshift-network-console"/"networking-console-plugin" not registered Dec 10 22:49:51 crc kubenswrapper[4791]: E1210 22:49:51.611536 4791 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2025-12-10 22:49:52.611517061 +0000 UTC m=+27.041134674 (durationBeforeRetry 1s). Error: MountVolume.SetUp failed for volume "nginx-conf" (UniqueName: "kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin" not registered Dec 10 22:49:51 crc kubenswrapper[4791]: E1210 22:49:51.611590 4791 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Dec 10 22:49:51 crc kubenswrapper[4791]: E1210 22:49:51.611612 4791 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Dec 10 22:49:51 crc kubenswrapper[4791]: E1210 22:49:51.611625 4791 projected.go:194] Error preparing data for projected volume kube-api-access-s2dwl for pod openshift-network-diagnostics/network-check-source-55646444c4-trplf: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Dec 10 22:49:51 crc kubenswrapper[4791]: E1210 22:49:51.611688 4791 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl podName:9d751cbb-f2e2-430d-9754-c882a5e924a5 nodeName:}" failed. No retries permitted until 2025-12-10 22:49:52.611671215 +0000 UTC m=+27.041288928 (durationBeforeRetry 1s). Error: MountVolume.SetUp failed for volume "kube-api-access-s2dwl" (UniqueName: "kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl") pod "network-check-source-55646444c4-trplf" (UID: "9d751cbb-f2e2-430d-9754-c882a5e924a5") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Dec 10 22:49:51 crc kubenswrapper[4791]: E1210 22:49:51.611758 4791 secret.go:188] Couldn't get secret openshift-network-console/networking-console-plugin-cert: object "openshift-network-console"/"networking-console-plugin-cert" not registered Dec 10 22:49:51 crc kubenswrapper[4791]: E1210 22:49:51.611786 4791 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2025-12-10 22:49:52.611777958 +0000 UTC m=+27.041395711 (durationBeforeRetry 1s). Error: MountVolume.SetUp failed for volume "networking-console-plugin-cert" (UniqueName: "kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin-cert" not registered Dec 10 22:49:51 crc kubenswrapper[4791]: I1210 22:49:51.611471 4791 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-s2dwl\" (UniqueName: \"kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl\") pod \"network-check-source-55646444c4-trplf\" (UID: \"9d751cbb-f2e2-430d-9754-c882a5e924a5\") " pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 10 22:49:51 crc kubenswrapper[4791]: E1210 22:49:51.611837 4791 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Dec 10 22:49:51 crc kubenswrapper[4791]: E1210 22:49:51.611886 4791 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Dec 10 22:49:51 crc kubenswrapper[4791]: E1210 22:49:51.611898 4791 projected.go:194] Error preparing data for projected volume kube-api-access-cqllr for pod openshift-network-diagnostics/network-check-target-xd92c: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Dec 10 22:49:51 crc kubenswrapper[4791]: E1210 22:49:51.611939 4791 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr podName:3b6479f0-333b-4a96-9adf-2099afdc2447 nodeName:}" failed. No retries permitted until 2025-12-10 22:49:52.611931321 +0000 UTC m=+27.041548934 (durationBeforeRetry 1s). Error: MountVolume.SetUp failed for volume "kube-api-access-cqllr" (UniqueName: "kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr") pod "network-check-target-xd92c" (UID: "3b6479f0-333b-4a96-9adf-2099afdc2447") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Dec 10 22:49:51 crc kubenswrapper[4791]: I1210 22:49:51.629884 4791 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-machine-config-operator/machine-config-daemon-5rb5l"] Dec 10 22:49:51 crc kubenswrapper[4791]: I1210 22:49:51.630465 4791 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-config-operator/machine-config-daemon-5rb5l" Dec 10 22:49:51 crc kubenswrapper[4791]: I1210 22:49:51.630671 4791 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-multus/multus-additional-cni-plugins-rw584"] Dec 10 22:49:51 crc kubenswrapper[4791]: I1210 22:49:51.631492 4791 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/multus-additional-cni-plugins-rw584" Dec 10 22:49:51 crc kubenswrapper[4791]: I1210 22:49:51.634275 4791 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-multus"/"cni-copy-resources" Dec 10 22:49:51 crc kubenswrapper[4791]: I1210 22:49:51.634297 4791 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-config-operator"/"openshift-service-ca.crt" Dec 10 22:49:51 crc kubenswrapper[4791]: I1210 22:49:51.634359 4791 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"proxy-tls" Dec 10 22:49:51 crc kubenswrapper[4791]: I1210 22:49:51.634456 4791 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-config-operator"/"kube-rbac-proxy" Dec 10 22:49:51 crc kubenswrapper[4791]: I1210 22:49:51.634646 4791 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-multus"/"default-cni-sysctl-allowlist" Dec 10 22:49:51 crc kubenswrapper[4791]: I1210 22:49:51.635756 4791 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-config-operator"/"kube-root-ca.crt" Dec 10 22:49:51 crc kubenswrapper[4791]: I1210 22:49:51.635975 4791 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-multus"/"openshift-service-ca.crt" Dec 10 22:49:51 crc kubenswrapper[4791]: I1210 22:49:51.635978 4791 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-multus"/"multus-ancillary-tools-dockercfg-vnmsz" Dec 10 22:49:51 crc kubenswrapper[4791]: I1210 22:49:51.637828 4791 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"machine-config-daemon-dockercfg-r5tcq" Dec 10 22:49:51 crc kubenswrapper[4791]: I1210 22:49:51.637880 4791 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-multus"/"kube-root-ca.crt" Dec 10 22:49:51 crc kubenswrapper[4791]: I1210 22:49:51.640738 4791 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-multus/multus-4nwkq"] Dec 10 22:49:51 crc kubenswrapper[4791]: I1210 22:49:51.641232 4791 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/multus-4nwkq" Dec 10 22:49:51 crc kubenswrapper[4791]: I1210 22:49:51.643795 4791 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-multus"/"multus-daemon-config" Dec 10 22:49:51 crc kubenswrapper[4791]: I1210 22:49:51.643950 4791 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-multus"/"default-dockercfg-2q5b6" Dec 10 22:49:51 crc kubenswrapper[4791]: I1210 22:49:51.648247 4791 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-10T22:49:50Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-10T22:49:50Z\\\",\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Dec 10 22:49:51 crc kubenswrapper[4791]: I1210 22:49:51.663225 4791 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f8349fea-c49b-49a1-b5ee-032c5dffe021\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T22:49:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T22:49:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T22:49:25Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T22:49:25Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T22:49:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://acafe12186df99d1059a31260de0eafce35ff7771bcb8f5c6753b72665d6f60b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T22:49:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ea4c346507862a8c12756a3061d6e25e788a8d79f4621b62f81f4322185c82a8\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T22:49:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://2e6e324ec6f4f7bb69fdfab4ea3ac1aa55e480dd8e2a8984b0fe5b4c9f66636e\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T22:49:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://1b2e11ff2bf8789fa1d0d23ae648b3e6ef653106e5bc4bbb5dd6447f7cd6f05d\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d6aa7f8c5d908d90008f2df49860ac8a3480136da567e3d85f86135d779204f9\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-12-10T22:49:44Z\\\",\\\"message\\\":\\\"W1210 22:49:33.845733 1 cmd.go:257] Using insecure, self-signed certificates\\\\nI1210 22:49:33.846108 1 crypto.go:601] Generating new CA for check-endpoints-signer@1765406973 cert, and key in /tmp/serving-cert-2585701456/serving-signer.crt, /tmp/serving-cert-2585701456/serving-signer.key\\\\nI1210 22:49:34.529748 1 observer_polling.go:159] Starting file observer\\\\nW1210 22:49:34.532225 1 builder.go:272] unable to get owner reference (falling back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": dial tcp [::1]:6443: connect: connection refused\\\\nI1210 22:49:34.532599 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1210 22:49:34.533977 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-2585701456/tls.crt::/tmp/serving-cert-2585701456/tls.key\\\\\\\"\\\\nF1210 22:49:44.774710 1 cmd.go:182] error initializing delegating authentication: unable to load configmap based request-header-client-ca-file: Get \\\\\\\"https://localhost:6443/api/v1/namespaces/kube-system/configmaps/extension-apiserver-authentication\\\\\\\": net/http: TLS handshake timeout\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-10T22:49:33Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T22:49:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://0c4d036ab269f62561fe7ef1586012cb3a16e3583a08b9488a0616ec00c8e0ae\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T22:49:32Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c5856d6b26c19fda13b5c9ac2892a7e2f9c6f79feb8173ecfd9b33b673327d67\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c5856d6b26c19fda13b5c9ac2892a7e2f9c6f79feb8173ecfd9b33b673327d67\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-10T22:49:27Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-10T22:49:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-10T22:49:25Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Dec 10 22:49:51 crc kubenswrapper[4791]: I1210 22:49:51.672511 4791 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-10T22:49:50Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-10T22:49:50Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Dec 10 22:49:51 crc kubenswrapper[4791]: I1210 22:49:51.680504 4791 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-5rb5l" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ba35653c-6e06-4cee-a4d6-137764090d18\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T22:49:51Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T22:49:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T22:49:51Z\\\",\\\"message\\\":\\\"containers with unready status: [machine-config-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T22:49:51Z\\\",\\\"message\\\":\\\"containers with unready status: [machine-config-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fvvrs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fvvrs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-10T22:49:51Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-5rb5l\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Dec 10 22:49:51 crc kubenswrapper[4791]: I1210 22:49:51.689192 4791 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"698bbe9a-fbe5-436d-a835-5e80df77ba24\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T22:49:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T22:49:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T22:49:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T22:49:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T22:49:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://224d73713312376a58ec6cc2bc8fdb6ffa44c4749ae0e7691d5fe27b9afca09f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T22:49:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://372eeaffd40992c6f06efc4df6dfcb44b31de64ee8256c21508cdd7c74ecb96e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T22:49:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://76b6013b062cac5b0b4d12a94f38ce7b43f9c7ea216063455d88a956ac04d26e\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T22:49:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://78e950e122e761d950da1a3c60c06476e53920e8795f6c61db22c01e22d5136e\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T22:49:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-10T22:49:25Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Dec 10 22:49:51 crc kubenswrapper[4791]: I1210 22:49:51.698145 4791 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-10T22:49:50Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [webhook approver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-10T22:49:50Z\\\",\\\"message\\\":\\\"containers with unready status: [webhook approver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":false,\\\"restartCount\\\":6,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Dec 10 22:49:51 crc kubenswrapper[4791]: I1210 22:49:51.704500 4791 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-vczjq" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b9f00f25-3fa1-4b63-9710-fd352224b01b\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T22:49:50Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T22:49:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T22:49:50Z\\\",\\\"message\\\":\\\"containers with unready status: [node-ca]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T22:49:50Z\\\",\\\"message\\\":\\\"containers with unready status: [node-ca]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lrrpv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-10T22:49:50Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-vczjq\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Dec 10 22:49:51 crc kubenswrapper[4791]: I1210 22:49:51.712321 4791 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/ba35653c-6e06-4cee-a4d6-137764090d18-proxy-tls\") pod \"machine-config-daemon-5rb5l\" (UID: \"ba35653c-6e06-4cee-a4d6-137764090d18\") " pod="openshift-machine-config-operator/machine-config-daemon-5rb5l" Dec 10 22:49:51 crc kubenswrapper[4791]: I1210 22:49:51.712384 4791 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"system-cni-dir\" (UniqueName: \"kubernetes.io/host-path/ff9beb90-69b0-4732-bf37-0b81f58ecc98-system-cni-dir\") pod \"multus-additional-cni-plugins-rw584\" (UID: \"ff9beb90-69b0-4732-bf37-0b81f58ecc98\") " pod="openshift-multus/multus-additional-cni-plugins-rw584" Dec 10 22:49:51 crc kubenswrapper[4791]: I1210 22:49:51.712406 4791 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cni-sysctl-allowlist\" (UniqueName: \"kubernetes.io/configmap/ff9beb90-69b0-4732-bf37-0b81f58ecc98-cni-sysctl-allowlist\") pod \"multus-additional-cni-plugins-rw584\" (UID: \"ff9beb90-69b0-4732-bf37-0b81f58ecc98\") " pod="openshift-multus/multus-additional-cni-plugins-rw584" Dec 10 22:49:51 crc kubenswrapper[4791]: I1210 22:49:51.712438 4791 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rootfs\" (UniqueName: \"kubernetes.io/host-path/ba35653c-6e06-4cee-a4d6-137764090d18-rootfs\") pod \"machine-config-daemon-5rb5l\" (UID: \"ba35653c-6e06-4cee-a4d6-137764090d18\") " pod="openshift-machine-config-operator/machine-config-daemon-5rb5l" Dec 10 22:49:51 crc kubenswrapper[4791]: I1210 22:49:51.712470 4791 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-fvvrs\" (UniqueName: \"kubernetes.io/projected/ba35653c-6e06-4cee-a4d6-137764090d18-kube-api-access-fvvrs\") pod \"machine-config-daemon-5rb5l\" (UID: \"ba35653c-6e06-4cee-a4d6-137764090d18\") " pod="openshift-machine-config-operator/machine-config-daemon-5rb5l" Dec 10 22:49:51 crc kubenswrapper[4791]: I1210 22:49:51.712486 4791 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-fksqn\" (UniqueName: \"kubernetes.io/projected/ff9beb90-69b0-4732-bf37-0b81f58ecc98-kube-api-access-fksqn\") pod \"multus-additional-cni-plugins-rw584\" (UID: \"ff9beb90-69b0-4732-bf37-0b81f58ecc98\") " pod="openshift-multus/multus-additional-cni-plugins-rw584" Dec 10 22:49:51 crc kubenswrapper[4791]: I1210 22:49:51.712503 4791 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cnibin\" (UniqueName: \"kubernetes.io/host-path/ff9beb90-69b0-4732-bf37-0b81f58ecc98-cnibin\") pod \"multus-additional-cni-plugins-rw584\" (UID: \"ff9beb90-69b0-4732-bf37-0b81f58ecc98\") " pod="openshift-multus/multus-additional-cni-plugins-rw584" Dec 10 22:49:51 crc kubenswrapper[4791]: I1210 22:49:51.712538 4791 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"os-release\" (UniqueName: \"kubernetes.io/host-path/ff9beb90-69b0-4732-bf37-0b81f58ecc98-os-release\") pod \"multus-additional-cni-plugins-rw584\" (UID: \"ff9beb90-69b0-4732-bf37-0b81f58ecc98\") " pod="openshift-multus/multus-additional-cni-plugins-rw584" Dec 10 22:49:51 crc kubenswrapper[4791]: I1210 22:49:51.712593 4791 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"tuning-conf-dir\" (UniqueName: \"kubernetes.io/host-path/ff9beb90-69b0-4732-bf37-0b81f58ecc98-tuning-conf-dir\") pod \"multus-additional-cni-plugins-rw584\" (UID: \"ff9beb90-69b0-4732-bf37-0b81f58ecc98\") " pod="openshift-multus/multus-additional-cni-plugins-rw584" Dec 10 22:49:51 crc kubenswrapper[4791]: I1210 22:49:51.712654 4791 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"mcd-auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/ba35653c-6e06-4cee-a4d6-137764090d18-mcd-auth-proxy-config\") pod \"machine-config-daemon-5rb5l\" (UID: \"ba35653c-6e06-4cee-a4d6-137764090d18\") " pod="openshift-machine-config-operator/machine-config-daemon-5rb5l" Dec 10 22:49:51 crc kubenswrapper[4791]: I1210 22:49:51.712673 4791 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cni-binary-copy\" (UniqueName: \"kubernetes.io/configmap/ff9beb90-69b0-4732-bf37-0b81f58ecc98-cni-binary-copy\") pod \"multus-additional-cni-plugins-rw584\" (UID: \"ff9beb90-69b0-4732-bf37-0b81f58ecc98\") " pod="openshift-multus/multus-additional-cni-plugins-rw584" Dec 10 22:49:51 crc kubenswrapper[4791]: I1210 22:49:51.717945 4791 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-10T22:49:50Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-operator]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-10T22:49:50Z\\\",\\\"message\\\":\\\"containers with unready status: [network-operator]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":false,\\\"restartCount\\\":5,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T22:49:51Z is after 2025-08-24T17:21:41Z" Dec 10 22:49:51 crc kubenswrapper[4791]: I1210 22:49:51.729241 4791 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-10T22:49:50Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-10T22:49:50Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T22:49:51Z is after 2025-08-24T17:21:41Z" Dec 10 22:49:51 crc kubenswrapper[4791]: I1210 22:49:51.740708 4791 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-10T22:49:50Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-10T22:49:50Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T22:49:51Z is after 2025-08-24T17:21:41Z" Dec 10 22:49:51 crc kubenswrapper[4791]: I1210 22:49:51.752566 4791 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-tchxk" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"c0ab74e0-42a9-4f42-ac80-16030cf4ffff\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T22:49:50Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T22:49:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T22:49:50Z\\\",\\\"message\\\":\\\"containers with unready status: [dns-node-resolver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T22:49:50Z\\\",\\\"message\\\":\\\"containers with unready status: [dns-node-resolver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zn2qt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-10T22:49:50Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-tchxk\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T22:49:51Z is after 2025-08-24T17:21:41Z" Dec 10 22:49:51 crc kubenswrapper[4791]: I1210 22:49:51.762759 4791 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-10T22:49:50Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-10T22:49:50Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T22:49:51Z is after 2025-08-24T17:21:41Z" Dec 10 22:49:51 crc kubenswrapper[4791]: I1210 22:49:51.774165 4791 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-10T22:49:50Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-10T22:49:50Z\\\",\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T22:49:51Z is after 2025-08-24T17:21:41Z" Dec 10 22:49:51 crc kubenswrapper[4791]: I1210 22:49:51.788201 4791 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f8349fea-c49b-49a1-b5ee-032c5dffe021\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T22:49:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T22:49:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T22:49:25Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T22:49:25Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T22:49:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://acafe12186df99d1059a31260de0eafce35ff7771bcb8f5c6753b72665d6f60b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T22:49:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ea4c346507862a8c12756a3061d6e25e788a8d79f4621b62f81f4322185c82a8\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T22:49:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://2e6e324ec6f4f7bb69fdfab4ea3ac1aa55e480dd8e2a8984b0fe5b4c9f66636e\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T22:49:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://1b2e11ff2bf8789fa1d0d23ae648b3e6ef653106e5bc4bbb5dd6447f7cd6f05d\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d6aa7f8c5d908d90008f2df49860ac8a3480136da567e3d85f86135d779204f9\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-12-10T22:49:44Z\\\",\\\"message\\\":\\\"W1210 22:49:33.845733 1 cmd.go:257] Using insecure, self-signed certificates\\\\nI1210 22:49:33.846108 1 crypto.go:601] Generating new CA for check-endpoints-signer@1765406973 cert, and key in /tmp/serving-cert-2585701456/serving-signer.crt, /tmp/serving-cert-2585701456/serving-signer.key\\\\nI1210 22:49:34.529748 1 observer_polling.go:159] Starting file observer\\\\nW1210 22:49:34.532225 1 builder.go:272] unable to get owner reference (falling back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": dial tcp [::1]:6443: connect: connection refused\\\\nI1210 22:49:34.532599 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1210 22:49:34.533977 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-2585701456/tls.crt::/tmp/serving-cert-2585701456/tls.key\\\\\\\"\\\\nF1210 22:49:44.774710 1 cmd.go:182] error initializing delegating authentication: unable to load configmap based request-header-client-ca-file: Get \\\\\\\"https://localhost:6443/api/v1/namespaces/kube-system/configmaps/extension-apiserver-authentication\\\\\\\": net/http: TLS handshake timeout\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-10T22:49:33Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T22:49:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://0c4d036ab269f62561fe7ef1586012cb3a16e3583a08b9488a0616ec00c8e0ae\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T22:49:32Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c5856d6b26c19fda13b5c9ac2892a7e2f9c6f79feb8173ecfd9b33b673327d67\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c5856d6b26c19fda13b5c9ac2892a7e2f9c6f79feb8173ecfd9b33b673327d67\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-10T22:49:27Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-10T22:49:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-10T22:49:25Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T22:49:51Z is after 2025-08-24T17:21:41Z" Dec 10 22:49:51 crc kubenswrapper[4791]: I1210 22:49:51.798722 4791 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-5rb5l" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ba35653c-6e06-4cee-a4d6-137764090d18\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T22:49:51Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T22:49:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T22:49:51Z\\\",\\\"message\\\":\\\"containers with unready status: [machine-config-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T22:49:51Z\\\",\\\"message\\\":\\\"containers with unready status: [machine-config-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fvvrs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fvvrs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-10T22:49:51Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-5rb5l\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T22:49:51Z is after 2025-08-24T17:21:41Z" Dec 10 22:49:51 crc kubenswrapper[4791]: I1210 22:49:51.811424 4791 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-10T22:49:50Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [webhook approver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-10T22:49:50Z\\\",\\\"message\\\":\\\"containers with unready status: [webhook approver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":false,\\\"restartCount\\\":6,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T22:49:51Z is after 2025-08-24T17:21:41Z" Dec 10 22:49:51 crc kubenswrapper[4791]: I1210 22:49:51.813758 4791 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-kubernetes\" (UniqueName: \"kubernetes.io/host-path/672aa28c-8169-49ed-87b8-21187d13a80c-etc-kubernetes\") pod \"multus-4nwkq\" (UID: \"672aa28c-8169-49ed-87b8-21187d13a80c\") " pod="openshift-multus/multus-4nwkq" Dec 10 22:49:51 crc kubenswrapper[4791]: I1210 22:49:51.813817 4791 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-fksqn\" (UniqueName: \"kubernetes.io/projected/ff9beb90-69b0-4732-bf37-0b81f58ecc98-kube-api-access-fksqn\") pod \"multus-additional-cni-plugins-rw584\" (UID: \"ff9beb90-69b0-4732-bf37-0b81f58ecc98\") " pod="openshift-multus/multus-additional-cni-plugins-rw584" Dec 10 22:49:51 crc kubenswrapper[4791]: I1210 22:49:51.813840 4791 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"os-release\" (UniqueName: \"kubernetes.io/host-path/ff9beb90-69b0-4732-bf37-0b81f58ecc98-os-release\") pod \"multus-additional-cni-plugins-rw584\" (UID: \"ff9beb90-69b0-4732-bf37-0b81f58ecc98\") " pod="openshift-multus/multus-additional-cni-plugins-rw584" Dec 10 22:49:51 crc kubenswrapper[4791]: I1210 22:49:51.813860 4791 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"tuning-conf-dir\" (UniqueName: \"kubernetes.io/host-path/ff9beb90-69b0-4732-bf37-0b81f58ecc98-tuning-conf-dir\") pod \"multus-additional-cni-plugins-rw584\" (UID: \"ff9beb90-69b0-4732-bf37-0b81f58ecc98\") " pod="openshift-multus/multus-additional-cni-plugins-rw584" Dec 10 22:49:51 crc kubenswrapper[4791]: I1210 22:49:51.813878 4791 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cnibin\" (UniqueName: \"kubernetes.io/host-path/672aa28c-8169-49ed-87b8-21187d13a80c-cnibin\") pod \"multus-4nwkq\" (UID: \"672aa28c-8169-49ed-87b8-21187d13a80c\") " pod="openshift-multus/multus-4nwkq" Dec 10 22:49:51 crc kubenswrapper[4791]: I1210 22:49:51.813896 4791 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"os-release\" (UniqueName: \"kubernetes.io/host-path/672aa28c-8169-49ed-87b8-21187d13a80c-os-release\") pod \"multus-4nwkq\" (UID: \"672aa28c-8169-49ed-87b8-21187d13a80c\") " pod="openshift-multus/multus-4nwkq" Dec 10 22:49:51 crc kubenswrapper[4791]: I1210 22:49:51.813912 4791 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-run-k8s-cni-cncf-io\" (UniqueName: \"kubernetes.io/host-path/672aa28c-8169-49ed-87b8-21187d13a80c-host-run-k8s-cni-cncf-io\") pod \"multus-4nwkq\" (UID: \"672aa28c-8169-49ed-87b8-21187d13a80c\") " pod="openshift-multus/multus-4nwkq" Dec 10 22:49:51 crc kubenswrapper[4791]: I1210 22:49:51.813932 4791 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"mcd-auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/ba35653c-6e06-4cee-a4d6-137764090d18-mcd-auth-proxy-config\") pod \"machine-config-daemon-5rb5l\" (UID: \"ba35653c-6e06-4cee-a4d6-137764090d18\") " pod="openshift-machine-config-operator/machine-config-daemon-5rb5l" Dec 10 22:49:51 crc kubenswrapper[4791]: I1210 22:49:51.813948 4791 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"hostroot\" (UniqueName: \"kubernetes.io/host-path/672aa28c-8169-49ed-87b8-21187d13a80c-hostroot\") pod \"multus-4nwkq\" (UID: \"672aa28c-8169-49ed-87b8-21187d13a80c\") " pod="openshift-multus/multus-4nwkq" Dec 10 22:49:51 crc kubenswrapper[4791]: I1210 22:49:51.813964 4791 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-var-lib-kubelet\" (UniqueName: \"kubernetes.io/host-path/672aa28c-8169-49ed-87b8-21187d13a80c-host-var-lib-kubelet\") pod \"multus-4nwkq\" (UID: \"672aa28c-8169-49ed-87b8-21187d13a80c\") " pod="openshift-multus/multus-4nwkq" Dec 10 22:49:51 crc kubenswrapper[4791]: I1210 22:49:51.813982 4791 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"multus-conf-dir\" (UniqueName: \"kubernetes.io/host-path/672aa28c-8169-49ed-87b8-21187d13a80c-multus-conf-dir\") pod \"multus-4nwkq\" (UID: \"672aa28c-8169-49ed-87b8-21187d13a80c\") " pod="openshift-multus/multus-4nwkq" Dec 10 22:49:51 crc kubenswrapper[4791]: I1210 22:49:51.814001 4791 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-xl62j\" (UniqueName: \"kubernetes.io/projected/672aa28c-8169-49ed-87b8-21187d13a80c-kube-api-access-xl62j\") pod \"multus-4nwkq\" (UID: \"672aa28c-8169-49ed-87b8-21187d13a80c\") " pod="openshift-multus/multus-4nwkq" Dec 10 22:49:51 crc kubenswrapper[4791]: I1210 22:49:51.814018 4791 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cni-sysctl-allowlist\" (UniqueName: \"kubernetes.io/configmap/ff9beb90-69b0-4732-bf37-0b81f58ecc98-cni-sysctl-allowlist\") pod \"multus-additional-cni-plugins-rw584\" (UID: \"ff9beb90-69b0-4732-bf37-0b81f58ecc98\") " pod="openshift-multus/multus-additional-cni-plugins-rw584" Dec 10 22:49:51 crc kubenswrapper[4791]: I1210 22:49:51.814048 4791 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rootfs\" (UniqueName: \"kubernetes.io/host-path/ba35653c-6e06-4cee-a4d6-137764090d18-rootfs\") pod \"machine-config-daemon-5rb5l\" (UID: \"ba35653c-6e06-4cee-a4d6-137764090d18\") " pod="openshift-machine-config-operator/machine-config-daemon-5rb5l" Dec 10 22:49:51 crc kubenswrapper[4791]: I1210 22:49:51.814068 4791 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"system-cni-dir\" (UniqueName: \"kubernetes.io/host-path/672aa28c-8169-49ed-87b8-21187d13a80c-system-cni-dir\") pod \"multus-4nwkq\" (UID: \"672aa28c-8169-49ed-87b8-21187d13a80c\") " pod="openshift-multus/multus-4nwkq" Dec 10 22:49:51 crc kubenswrapper[4791]: I1210 22:49:51.814084 4791 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"multus-cni-dir\" (UniqueName: \"kubernetes.io/host-path/672aa28c-8169-49ed-87b8-21187d13a80c-multus-cni-dir\") pod \"multus-4nwkq\" (UID: \"672aa28c-8169-49ed-87b8-21187d13a80c\") " pod="openshift-multus/multus-4nwkq" Dec 10 22:49:51 crc kubenswrapper[4791]: I1210 22:49:51.814106 4791 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-fvvrs\" (UniqueName: \"kubernetes.io/projected/ba35653c-6e06-4cee-a4d6-137764090d18-kube-api-access-fvvrs\") pod \"machine-config-daemon-5rb5l\" (UID: \"ba35653c-6e06-4cee-a4d6-137764090d18\") " pod="openshift-machine-config-operator/machine-config-daemon-5rb5l" Dec 10 22:49:51 crc kubenswrapper[4791]: I1210 22:49:51.814099 4791 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"os-release\" (UniqueName: \"kubernetes.io/host-path/ff9beb90-69b0-4732-bf37-0b81f58ecc98-os-release\") pod \"multus-additional-cni-plugins-rw584\" (UID: \"ff9beb90-69b0-4732-bf37-0b81f58ecc98\") " pod="openshift-multus/multus-additional-cni-plugins-rw584" Dec 10 22:49:51 crc kubenswrapper[4791]: I1210 22:49:51.814128 4791 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-run-netns\" (UniqueName: \"kubernetes.io/host-path/672aa28c-8169-49ed-87b8-21187d13a80c-host-run-netns\") pod \"multus-4nwkq\" (UID: \"672aa28c-8169-49ed-87b8-21187d13a80c\") " pod="openshift-multus/multus-4nwkq" Dec 10 22:49:51 crc kubenswrapper[4791]: I1210 22:49:51.814202 4791 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cnibin\" (UniqueName: \"kubernetes.io/host-path/ff9beb90-69b0-4732-bf37-0b81f58ecc98-cnibin\") pod \"multus-additional-cni-plugins-rw584\" (UID: \"ff9beb90-69b0-4732-bf37-0b81f58ecc98\") " pod="openshift-multus/multus-additional-cni-plugins-rw584" Dec 10 22:49:51 crc kubenswrapper[4791]: I1210 22:49:51.814299 4791 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rootfs\" (UniqueName: \"kubernetes.io/host-path/ba35653c-6e06-4cee-a4d6-137764090d18-rootfs\") pod \"machine-config-daemon-5rb5l\" (UID: \"ba35653c-6e06-4cee-a4d6-137764090d18\") " pod="openshift-machine-config-operator/machine-config-daemon-5rb5l" Dec 10 22:49:51 crc kubenswrapper[4791]: I1210 22:49:51.814368 4791 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-run-multus-certs\" (UniqueName: \"kubernetes.io/host-path/672aa28c-8169-49ed-87b8-21187d13a80c-host-run-multus-certs\") pod \"multus-4nwkq\" (UID: \"672aa28c-8169-49ed-87b8-21187d13a80c\") " pod="openshift-multus/multus-4nwkq" Dec 10 22:49:51 crc kubenswrapper[4791]: I1210 22:49:51.814421 4791 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cni-binary-copy\" (UniqueName: \"kubernetes.io/configmap/ff9beb90-69b0-4732-bf37-0b81f58ecc98-cni-binary-copy\") pod \"multus-additional-cni-plugins-rw584\" (UID: \"ff9beb90-69b0-4732-bf37-0b81f58ecc98\") " pod="openshift-multus/multus-additional-cni-plugins-rw584" Dec 10 22:49:51 crc kubenswrapper[4791]: I1210 22:49:51.814448 4791 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-var-lib-cni-bin\" (UniqueName: \"kubernetes.io/host-path/672aa28c-8169-49ed-87b8-21187d13a80c-host-var-lib-cni-bin\") pod \"multus-4nwkq\" (UID: \"672aa28c-8169-49ed-87b8-21187d13a80c\") " pod="openshift-multus/multus-4nwkq" Dec 10 22:49:51 crc kubenswrapper[4791]: I1210 22:49:51.814486 4791 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cnibin\" (UniqueName: \"kubernetes.io/host-path/ff9beb90-69b0-4732-bf37-0b81f58ecc98-cnibin\") pod \"multus-additional-cni-plugins-rw584\" (UID: \"ff9beb90-69b0-4732-bf37-0b81f58ecc98\") " pod="openshift-multus/multus-additional-cni-plugins-rw584" Dec 10 22:49:51 crc kubenswrapper[4791]: I1210 22:49:51.814856 4791 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cni-sysctl-allowlist\" (UniqueName: \"kubernetes.io/configmap/ff9beb90-69b0-4732-bf37-0b81f58ecc98-cni-sysctl-allowlist\") pod \"multus-additional-cni-plugins-rw584\" (UID: \"ff9beb90-69b0-4732-bf37-0b81f58ecc98\") " pod="openshift-multus/multus-additional-cni-plugins-rw584" Dec 10 22:49:51 crc kubenswrapper[4791]: I1210 22:49:51.814903 4791 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/ba35653c-6e06-4cee-a4d6-137764090d18-proxy-tls\") pod \"machine-config-daemon-5rb5l\" (UID: \"ba35653c-6e06-4cee-a4d6-137764090d18\") " pod="openshift-machine-config-operator/machine-config-daemon-5rb5l" Dec 10 22:49:51 crc kubenswrapper[4791]: I1210 22:49:51.814919 4791 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"mcd-auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/ba35653c-6e06-4cee-a4d6-137764090d18-mcd-auth-proxy-config\") pod \"machine-config-daemon-5rb5l\" (UID: \"ba35653c-6e06-4cee-a4d6-137764090d18\") " pod="openshift-machine-config-operator/machine-config-daemon-5rb5l" Dec 10 22:49:51 crc kubenswrapper[4791]: I1210 22:49:51.814962 4791 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"multus-socket-dir-parent\" (UniqueName: \"kubernetes.io/host-path/672aa28c-8169-49ed-87b8-21187d13a80c-multus-socket-dir-parent\") pod \"multus-4nwkq\" (UID: \"672aa28c-8169-49ed-87b8-21187d13a80c\") " pod="openshift-multus/multus-4nwkq" Dec 10 22:49:51 crc kubenswrapper[4791]: I1210 22:49:51.815133 4791 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"system-cni-dir\" (UniqueName: \"kubernetes.io/host-path/ff9beb90-69b0-4732-bf37-0b81f58ecc98-system-cni-dir\") pod \"multus-additional-cni-plugins-rw584\" (UID: \"ff9beb90-69b0-4732-bf37-0b81f58ecc98\") " pod="openshift-multus/multus-additional-cni-plugins-rw584" Dec 10 22:49:51 crc kubenswrapper[4791]: I1210 22:49:51.815213 4791 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-var-lib-cni-multus\" (UniqueName: \"kubernetes.io/host-path/672aa28c-8169-49ed-87b8-21187d13a80c-host-var-lib-cni-multus\") pod \"multus-4nwkq\" (UID: \"672aa28c-8169-49ed-87b8-21187d13a80c\") " pod="openshift-multus/multus-4nwkq" Dec 10 22:49:51 crc kubenswrapper[4791]: I1210 22:49:51.815246 4791 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"multus-daemon-config\" (UniqueName: \"kubernetes.io/configmap/672aa28c-8169-49ed-87b8-21187d13a80c-multus-daemon-config\") pod \"multus-4nwkq\" (UID: \"672aa28c-8169-49ed-87b8-21187d13a80c\") " pod="openshift-multus/multus-4nwkq" Dec 10 22:49:51 crc kubenswrapper[4791]: I1210 22:49:51.815311 4791 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"system-cni-dir\" (UniqueName: \"kubernetes.io/host-path/ff9beb90-69b0-4732-bf37-0b81f58ecc98-system-cni-dir\") pod \"multus-additional-cni-plugins-rw584\" (UID: \"ff9beb90-69b0-4732-bf37-0b81f58ecc98\") " pod="openshift-multus/multus-additional-cni-plugins-rw584" Dec 10 22:49:51 crc kubenswrapper[4791]: I1210 22:49:51.815322 4791 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cni-binary-copy\" (UniqueName: \"kubernetes.io/configmap/ff9beb90-69b0-4732-bf37-0b81f58ecc98-cni-binary-copy\") pod \"multus-additional-cni-plugins-rw584\" (UID: \"ff9beb90-69b0-4732-bf37-0b81f58ecc98\") " pod="openshift-multus/multus-additional-cni-plugins-rw584" Dec 10 22:49:51 crc kubenswrapper[4791]: I1210 22:49:51.815379 4791 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cni-binary-copy\" (UniqueName: \"kubernetes.io/configmap/672aa28c-8169-49ed-87b8-21187d13a80c-cni-binary-copy\") pod \"multus-4nwkq\" (UID: \"672aa28c-8169-49ed-87b8-21187d13a80c\") " pod="openshift-multus/multus-4nwkq" Dec 10 22:49:51 crc kubenswrapper[4791]: I1210 22:49:51.815436 4791 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"tuning-conf-dir\" (UniqueName: \"kubernetes.io/host-path/ff9beb90-69b0-4732-bf37-0b81f58ecc98-tuning-conf-dir\") pod \"multus-additional-cni-plugins-rw584\" (UID: \"ff9beb90-69b0-4732-bf37-0b81f58ecc98\") " pod="openshift-multus/multus-additional-cni-plugins-rw584" Dec 10 22:49:51 crc kubenswrapper[4791]: I1210 22:49:51.819450 4791 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/ba35653c-6e06-4cee-a4d6-137764090d18-proxy-tls\") pod \"machine-config-daemon-5rb5l\" (UID: \"ba35653c-6e06-4cee-a4d6-137764090d18\") " pod="openshift-machine-config-operator/machine-config-daemon-5rb5l" Dec 10 22:49:51 crc kubenswrapper[4791]: I1210 22:49:51.826007 4791 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-vczjq" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b9f00f25-3fa1-4b63-9710-fd352224b01b\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T22:49:50Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T22:49:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T22:49:50Z\\\",\\\"message\\\":\\\"containers with unready status: [node-ca]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T22:49:50Z\\\",\\\"message\\\":\\\"containers with unready status: [node-ca]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lrrpv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-10T22:49:50Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-vczjq\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T22:49:51Z is after 2025-08-24T17:21:41Z" Dec 10 22:49:51 crc kubenswrapper[4791]: I1210 22:49:51.835653 4791 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-fvvrs\" (UniqueName: \"kubernetes.io/projected/ba35653c-6e06-4cee-a4d6-137764090d18-kube-api-access-fvvrs\") pod \"machine-config-daemon-5rb5l\" (UID: \"ba35653c-6e06-4cee-a4d6-137764090d18\") " pod="openshift-machine-config-operator/machine-config-daemon-5rb5l" Dec 10 22:49:51 crc kubenswrapper[4791]: I1210 22:49:51.835942 4791 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-fksqn\" (UniqueName: \"kubernetes.io/projected/ff9beb90-69b0-4732-bf37-0b81f58ecc98-kube-api-access-fksqn\") pod \"multus-additional-cni-plugins-rw584\" (UID: \"ff9beb90-69b0-4732-bf37-0b81f58ecc98\") " pod="openshift-multus/multus-additional-cni-plugins-rw584" Dec 10 22:49:51 crc kubenswrapper[4791]: I1210 22:49:51.846937 4791 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-4nwkq" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"672aa28c-8169-49ed-87b8-21187d13a80c\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T22:49:51Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T22:49:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T22:49:51Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T22:49:51Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xl62j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-10T22:49:51Z\\\"}}\" for pod \"openshift-multus\"/\"multus-4nwkq\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T22:49:51Z is after 2025-08-24T17:21:41Z" Dec 10 22:49:51 crc kubenswrapper[4791]: I1210 22:49:51.867725 4791 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"698bbe9a-fbe5-436d-a835-5e80df77ba24\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T22:49:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T22:49:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T22:49:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T22:49:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T22:49:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://224d73713312376a58ec6cc2bc8fdb6ffa44c4749ae0e7691d5fe27b9afca09f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T22:49:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://372eeaffd40992c6f06efc4df6dfcb44b31de64ee8256c21508cdd7c74ecb96e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T22:49:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://76b6013b062cac5b0b4d12a94f38ce7b43f9c7ea216063455d88a956ac04d26e\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T22:49:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://78e950e122e761d950da1a3c60c06476e53920e8795f6c61db22c01e22d5136e\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T22:49:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-10T22:49:25Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T22:49:51Z is after 2025-08-24T17:21:41Z" Dec 10 22:49:51 crc kubenswrapper[4791]: I1210 22:49:51.881940 4791 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-10T22:49:50Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-10T22:49:50Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T22:49:51Z is after 2025-08-24T17:21:41Z" Dec 10 22:49:51 crc kubenswrapper[4791]: I1210 22:49:51.888584 4791 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="01ab3dd5-8196-46d0-ad33-122e2ca51def" path="/var/lib/kubelet/pods/01ab3dd5-8196-46d0-ad33-122e2ca51def/volumes" Dec 10 22:49:51 crc kubenswrapper[4791]: I1210 22:49:51.889079 4791 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="09ae3b1a-e8e7-4524-b54b-61eab6f9239a" path="/var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes" Dec 10 22:49:51 crc kubenswrapper[4791]: I1210 22:49:51.889901 4791 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="09efc573-dbb6-4249-bd59-9b87aba8dd28" path="/var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes" Dec 10 22:49:51 crc kubenswrapper[4791]: I1210 22:49:51.890495 4791 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="0b574797-001e-440a-8f4e-c0be86edad0f" path="/var/lib/kubelet/pods/0b574797-001e-440a-8f4e-c0be86edad0f/volumes" Dec 10 22:49:51 crc kubenswrapper[4791]: I1210 22:49:51.891029 4791 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="0b78653f-4ff9-4508-8672-245ed9b561e3" path="/var/lib/kubelet/pods/0b78653f-4ff9-4508-8672-245ed9b561e3/volumes" Dec 10 22:49:51 crc kubenswrapper[4791]: I1210 22:49:51.891502 4791 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="1386a44e-36a2-460c-96d0-0359d2b6f0f5" path="/var/lib/kubelet/pods/1386a44e-36a2-460c-96d0-0359d2b6f0f5/volumes" Dec 10 22:49:51 crc kubenswrapper[4791]: I1210 22:49:51.892075 4791 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="1bf7eb37-55a3-4c65-b768-a94c82151e69" path="/var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes" Dec 10 22:49:51 crc kubenswrapper[4791]: I1210 22:49:51.892637 4791 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="1d611f23-29be-4491-8495-bee1670e935f" path="/var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes" Dec 10 22:49:51 crc kubenswrapper[4791]: I1210 22:49:51.893216 4791 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="20b0d48f-5fd6-431c-a545-e3c800c7b866" path="/var/lib/kubelet/pods/20b0d48f-5fd6-431c-a545-e3c800c7b866/volumes" Dec 10 22:49:51 crc kubenswrapper[4791]: I1210 22:49:51.893752 4791 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c" path="/var/lib/kubelet/pods/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c/volumes" Dec 10 22:49:51 crc kubenswrapper[4791]: I1210 22:49:51.894258 4791 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="22c825df-677d-4ca6-82db-3454ed06e783" path="/var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/volumes" Dec 10 22:49:51 crc kubenswrapper[4791]: I1210 22:49:51.894889 4791 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="25e176fe-21b4-4974-b1ed-c8b94f112a7f" path="/var/lib/kubelet/pods/25e176fe-21b4-4974-b1ed-c8b94f112a7f/volumes" Dec 10 22:49:51 crc kubenswrapper[4791]: I1210 22:49:51.897504 4791 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="308be0ea-9f5f-4b29-aeb1-5abd31a0b17b" path="/var/lib/kubelet/pods/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b/volumes" Dec 10 22:49:51 crc kubenswrapper[4791]: I1210 22:49:51.898090 4791 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="31d8b7a1-420e-4252-a5b7-eebe8a111292" path="/var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/volumes" Dec 10 22:49:51 crc kubenswrapper[4791]: I1210 22:49:51.898424 4791 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-10T22:49:50Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-10T22:49:50Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T22:49:51Z is after 2025-08-24T17:21:41Z" Dec 10 22:49:51 crc kubenswrapper[4791]: I1210 22:49:51.898707 4791 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="3ab1a177-2de0-46d9-b765-d0d0649bb42e" path="/var/lib/kubelet/pods/3ab1a177-2de0-46d9-b765-d0d0649bb42e/volumes" Dec 10 22:49:51 crc kubenswrapper[4791]: I1210 22:49:51.899784 4791 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="3cb93b32-e0ae-4377-b9c8-fdb9842c6d59" path="/var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/volumes" Dec 10 22:49:51 crc kubenswrapper[4791]: I1210 22:49:51.900357 4791 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="43509403-f426-496e-be36-56cef71462f5" path="/var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes" Dec 10 22:49:51 crc kubenswrapper[4791]: I1210 22:49:51.901328 4791 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="44663579-783b-4372-86d6-acf235a62d72" path="/var/lib/kubelet/pods/44663579-783b-4372-86d6-acf235a62d72/volumes" Dec 10 22:49:51 crc kubenswrapper[4791]: I1210 22:49:51.901931 4791 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="496e6271-fb68-4057-954e-a0d97a4afa3f" path="/var/lib/kubelet/pods/496e6271-fb68-4057-954e-a0d97a4afa3f/volumes" Dec 10 22:49:51 crc kubenswrapper[4791]: I1210 22:49:51.902482 4791 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" path="/var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes" Dec 10 22:49:51 crc kubenswrapper[4791]: I1210 22:49:51.903559 4791 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="49ef4625-1d3a-4a9f-b595-c2433d32326d" path="/var/lib/kubelet/pods/49ef4625-1d3a-4a9f-b595-c2433d32326d/volumes" Dec 10 22:49:51 crc kubenswrapper[4791]: I1210 22:49:51.904179 4791 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="4bb40260-dbaa-4fb0-84df-5e680505d512" path="/var/lib/kubelet/pods/4bb40260-dbaa-4fb0-84df-5e680505d512/volumes" Dec 10 22:49:51 crc kubenswrapper[4791]: I1210 22:49:51.904752 4791 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="5225d0e4-402f-4861-b410-819f433b1803" path="/var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes" Dec 10 22:49:51 crc kubenswrapper[4791]: I1210 22:49:51.905826 4791 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="5441d097-087c-4d9a-baa8-b210afa90fc9" path="/var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes" Dec 10 22:49:51 crc kubenswrapper[4791]: I1210 22:49:51.906228 4791 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="57a731c4-ef35-47a8-b875-bfb08a7f8011" path="/var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes" Dec 10 22:49:51 crc kubenswrapper[4791]: I1210 22:49:51.907229 4791 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="5b88f790-22fa-440e-b583-365168c0b23d" path="/var/lib/kubelet/pods/5b88f790-22fa-440e-b583-365168c0b23d/volumes" Dec 10 22:49:51 crc kubenswrapper[4791]: I1210 22:49:51.907869 4791 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="5fe579f8-e8a6-4643-bce5-a661393c4dde" path="/var/lib/kubelet/pods/5fe579f8-e8a6-4643-bce5-a661393c4dde/volumes" Dec 10 22:49:51 crc kubenswrapper[4791]: I1210 22:49:51.908735 4791 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="6402fda4-df10-493c-b4e5-d0569419652d" path="/var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/volumes" Dec 10 22:49:51 crc kubenswrapper[4791]: I1210 22:49:51.909264 4791 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="6509e943-70c6-444c-bc41-48a544e36fbd" path="/var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes" Dec 10 22:49:51 crc kubenswrapper[4791]: I1210 22:49:51.910090 4791 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="6731426b-95fe-49ff-bb5f-40441049fde2" path="/var/lib/kubelet/pods/6731426b-95fe-49ff-bb5f-40441049fde2/volumes" Dec 10 22:49:51 crc kubenswrapper[4791]: I1210 22:49:51.910548 4791 kubelet_volumes.go:152] "Cleaned up orphaned volume subpath from pod" podUID="6ea678ab-3438-413e-bfe3-290ae7725660" path="/var/lib/kubelet/pods/6ea678ab-3438-413e-bfe3-290ae7725660/volume-subpaths/run-systemd/ovnkube-controller/6" Dec 10 22:49:51 crc kubenswrapper[4791]: I1210 22:49:51.910647 4791 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="6ea678ab-3438-413e-bfe3-290ae7725660" path="/var/lib/kubelet/pods/6ea678ab-3438-413e-bfe3-290ae7725660/volumes" Dec 10 22:49:51 crc kubenswrapper[4791]: I1210 22:49:51.912653 4791 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="7539238d-5fe0-46ed-884e-1c3b566537ec" path="/var/lib/kubelet/pods/7539238d-5fe0-46ed-884e-1c3b566537ec/volumes" Dec 10 22:49:51 crc kubenswrapper[4791]: I1210 22:49:51.913195 4791 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="7583ce53-e0fe-4a16-9e4d-50516596a136" path="/var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes" Dec 10 22:49:51 crc kubenswrapper[4791]: I1210 22:49:51.913752 4791 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="7bb08738-c794-4ee8-9972-3a62ca171029" path="/var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/volumes" Dec 10 22:49:51 crc kubenswrapper[4791]: I1210 22:49:51.915051 4791 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-tchxk" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"c0ab74e0-42a9-4f42-ac80-16030cf4ffff\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T22:49:50Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T22:49:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T22:49:50Z\\\",\\\"message\\\":\\\"containers with unready status: [dns-node-resolver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T22:49:50Z\\\",\\\"message\\\":\\\"containers with unready status: [dns-node-resolver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zn2qt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-10T22:49:50Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-tchxk\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T22:49:51Z is after 2025-08-24T17:21:41Z" Dec 10 22:49:51 crc kubenswrapper[4791]: I1210 22:49:51.915224 4791 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="87cf06ed-a83f-41a7-828d-70653580a8cb" path="/var/lib/kubelet/pods/87cf06ed-a83f-41a7-828d-70653580a8cb/volumes" Dec 10 22:49:51 crc kubenswrapper[4791]: I1210 22:49:51.916208 4791 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"system-cni-dir\" (UniqueName: \"kubernetes.io/host-path/672aa28c-8169-49ed-87b8-21187d13a80c-system-cni-dir\") pod \"multus-4nwkq\" (UID: \"672aa28c-8169-49ed-87b8-21187d13a80c\") " pod="openshift-multus/multus-4nwkq" Dec 10 22:49:51 crc kubenswrapper[4791]: I1210 22:49:51.916256 4791 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"multus-cni-dir\" (UniqueName: \"kubernetes.io/host-path/672aa28c-8169-49ed-87b8-21187d13a80c-multus-cni-dir\") pod \"multus-4nwkq\" (UID: \"672aa28c-8169-49ed-87b8-21187d13a80c\") " pod="openshift-multus/multus-4nwkq" Dec 10 22:49:51 crc kubenswrapper[4791]: I1210 22:49:51.916276 4791 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-run-netns\" (UniqueName: \"kubernetes.io/host-path/672aa28c-8169-49ed-87b8-21187d13a80c-host-run-netns\") pod \"multus-4nwkq\" (UID: \"672aa28c-8169-49ed-87b8-21187d13a80c\") " pod="openshift-multus/multus-4nwkq" Dec 10 22:49:51 crc kubenswrapper[4791]: I1210 22:49:51.916303 4791 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-run-multus-certs\" (UniqueName: \"kubernetes.io/host-path/672aa28c-8169-49ed-87b8-21187d13a80c-host-run-multus-certs\") pod \"multus-4nwkq\" (UID: \"672aa28c-8169-49ed-87b8-21187d13a80c\") " pod="openshift-multus/multus-4nwkq" Dec 10 22:49:51 crc kubenswrapper[4791]: I1210 22:49:51.916331 4791 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-var-lib-cni-bin\" (UniqueName: \"kubernetes.io/host-path/672aa28c-8169-49ed-87b8-21187d13a80c-host-var-lib-cni-bin\") pod \"multus-4nwkq\" (UID: \"672aa28c-8169-49ed-87b8-21187d13a80c\") " pod="openshift-multus/multus-4nwkq" Dec 10 22:49:51 crc kubenswrapper[4791]: I1210 22:49:51.916350 4791 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"system-cni-dir\" (UniqueName: \"kubernetes.io/host-path/672aa28c-8169-49ed-87b8-21187d13a80c-system-cni-dir\") pod \"multus-4nwkq\" (UID: \"672aa28c-8169-49ed-87b8-21187d13a80c\") " pod="openshift-multus/multus-4nwkq" Dec 10 22:49:51 crc kubenswrapper[4791]: I1210 22:49:51.916254 4791 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="8cea82b4-6893-4ddc-af9f-1bb5ae425c5b" path="/var/lib/kubelet/pods/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b/volumes" Dec 10 22:49:51 crc kubenswrapper[4791]: I1210 22:49:51.916381 4791 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"multus-socket-dir-parent\" (UniqueName: \"kubernetes.io/host-path/672aa28c-8169-49ed-87b8-21187d13a80c-multus-socket-dir-parent\") pod \"multus-4nwkq\" (UID: \"672aa28c-8169-49ed-87b8-21187d13a80c\") " pod="openshift-multus/multus-4nwkq" Dec 10 22:49:51 crc kubenswrapper[4791]: I1210 22:49:51.916413 4791 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-var-lib-cni-multus\" (UniqueName: \"kubernetes.io/host-path/672aa28c-8169-49ed-87b8-21187d13a80c-host-var-lib-cni-multus\") pod \"multus-4nwkq\" (UID: \"672aa28c-8169-49ed-87b8-21187d13a80c\") " pod="openshift-multus/multus-4nwkq" Dec 10 22:49:51 crc kubenswrapper[4791]: I1210 22:49:51.916434 4791 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"multus-daemon-config\" (UniqueName: \"kubernetes.io/configmap/672aa28c-8169-49ed-87b8-21187d13a80c-multus-daemon-config\") pod \"multus-4nwkq\" (UID: \"672aa28c-8169-49ed-87b8-21187d13a80c\") " pod="openshift-multus/multus-4nwkq" Dec 10 22:49:51 crc kubenswrapper[4791]: I1210 22:49:51.916438 4791 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-run-multus-certs\" (UniqueName: \"kubernetes.io/host-path/672aa28c-8169-49ed-87b8-21187d13a80c-host-run-multus-certs\") pod \"multus-4nwkq\" (UID: \"672aa28c-8169-49ed-87b8-21187d13a80c\") " pod="openshift-multus/multus-4nwkq" Dec 10 22:49:51 crc kubenswrapper[4791]: I1210 22:49:51.916462 4791 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-var-lib-cni-bin\" (UniqueName: \"kubernetes.io/host-path/672aa28c-8169-49ed-87b8-21187d13a80c-host-var-lib-cni-bin\") pod \"multus-4nwkq\" (UID: \"672aa28c-8169-49ed-87b8-21187d13a80c\") " pod="openshift-multus/multus-4nwkq" Dec 10 22:49:51 crc kubenswrapper[4791]: I1210 22:49:51.916457 4791 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cni-binary-copy\" (UniqueName: \"kubernetes.io/configmap/672aa28c-8169-49ed-87b8-21187d13a80c-cni-binary-copy\") pod \"multus-4nwkq\" (UID: \"672aa28c-8169-49ed-87b8-21187d13a80c\") " pod="openshift-multus/multus-4nwkq" Dec 10 22:49:51 crc kubenswrapper[4791]: I1210 22:49:51.916469 4791 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-run-netns\" (UniqueName: \"kubernetes.io/host-path/672aa28c-8169-49ed-87b8-21187d13a80c-host-run-netns\") pod \"multus-4nwkq\" (UID: \"672aa28c-8169-49ed-87b8-21187d13a80c\") " pod="openshift-multus/multus-4nwkq" Dec 10 22:49:51 crc kubenswrapper[4791]: I1210 22:49:51.916539 4791 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-kubernetes\" (UniqueName: \"kubernetes.io/host-path/672aa28c-8169-49ed-87b8-21187d13a80c-etc-kubernetes\") pod \"multus-4nwkq\" (UID: \"672aa28c-8169-49ed-87b8-21187d13a80c\") " pod="openshift-multus/multus-4nwkq" Dec 10 22:49:51 crc kubenswrapper[4791]: I1210 22:49:51.916543 4791 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-var-lib-cni-multus\" (UniqueName: \"kubernetes.io/host-path/672aa28c-8169-49ed-87b8-21187d13a80c-host-var-lib-cni-multus\") pod \"multus-4nwkq\" (UID: \"672aa28c-8169-49ed-87b8-21187d13a80c\") " pod="openshift-multus/multus-4nwkq" Dec 10 22:49:51 crc kubenswrapper[4791]: I1210 22:49:51.916572 4791 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cnibin\" (UniqueName: \"kubernetes.io/host-path/672aa28c-8169-49ed-87b8-21187d13a80c-cnibin\") pod \"multus-4nwkq\" (UID: \"672aa28c-8169-49ed-87b8-21187d13a80c\") " pod="openshift-multus/multus-4nwkq" Dec 10 22:49:51 crc kubenswrapper[4791]: I1210 22:49:51.916590 4791 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-run-k8s-cni-cncf-io\" (UniqueName: \"kubernetes.io/host-path/672aa28c-8169-49ed-87b8-21187d13a80c-host-run-k8s-cni-cncf-io\") pod \"multus-4nwkq\" (UID: \"672aa28c-8169-49ed-87b8-21187d13a80c\") " pod="openshift-multus/multus-4nwkq" Dec 10 22:49:51 crc kubenswrapper[4791]: I1210 22:49:51.916603 4791 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-kubernetes\" (UniqueName: \"kubernetes.io/host-path/672aa28c-8169-49ed-87b8-21187d13a80c-etc-kubernetes\") pod \"multus-4nwkq\" (UID: \"672aa28c-8169-49ed-87b8-21187d13a80c\") " pod="openshift-multus/multus-4nwkq" Dec 10 22:49:51 crc kubenswrapper[4791]: I1210 22:49:51.916607 4791 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"os-release\" (UniqueName: \"kubernetes.io/host-path/672aa28c-8169-49ed-87b8-21187d13a80c-os-release\") pod \"multus-4nwkq\" (UID: \"672aa28c-8169-49ed-87b8-21187d13a80c\") " pod="openshift-multus/multus-4nwkq" Dec 10 22:49:51 crc kubenswrapper[4791]: I1210 22:49:51.916627 4791 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cnibin\" (UniqueName: \"kubernetes.io/host-path/672aa28c-8169-49ed-87b8-21187d13a80c-cnibin\") pod \"multus-4nwkq\" (UID: \"672aa28c-8169-49ed-87b8-21187d13a80c\") " pod="openshift-multus/multus-4nwkq" Dec 10 22:49:51 crc kubenswrapper[4791]: I1210 22:49:51.916645 4791 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"hostroot\" (UniqueName: \"kubernetes.io/host-path/672aa28c-8169-49ed-87b8-21187d13a80c-hostroot\") pod \"multus-4nwkq\" (UID: \"672aa28c-8169-49ed-87b8-21187d13a80c\") " pod="openshift-multus/multus-4nwkq" Dec 10 22:49:51 crc kubenswrapper[4791]: I1210 22:49:51.916646 4791 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"multus-cni-dir\" (UniqueName: \"kubernetes.io/host-path/672aa28c-8169-49ed-87b8-21187d13a80c-multus-cni-dir\") pod \"multus-4nwkq\" (UID: \"672aa28c-8169-49ed-87b8-21187d13a80c\") " pod="openshift-multus/multus-4nwkq" Dec 10 22:49:51 crc kubenswrapper[4791]: I1210 22:49:51.916687 4791 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"multus-conf-dir\" (UniqueName: \"kubernetes.io/host-path/672aa28c-8169-49ed-87b8-21187d13a80c-multus-conf-dir\") pod \"multus-4nwkq\" (UID: \"672aa28c-8169-49ed-87b8-21187d13a80c\") " pod="openshift-multus/multus-4nwkq" Dec 10 22:49:51 crc kubenswrapper[4791]: I1210 22:49:51.916688 4791 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"hostroot\" (UniqueName: \"kubernetes.io/host-path/672aa28c-8169-49ed-87b8-21187d13a80c-hostroot\") pod \"multus-4nwkq\" (UID: \"672aa28c-8169-49ed-87b8-21187d13a80c\") " pod="openshift-multus/multus-4nwkq" Dec 10 22:49:51 crc kubenswrapper[4791]: I1210 22:49:51.916666 4791 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"multus-conf-dir\" (UniqueName: \"kubernetes.io/host-path/672aa28c-8169-49ed-87b8-21187d13a80c-multus-conf-dir\") pod \"multus-4nwkq\" (UID: \"672aa28c-8169-49ed-87b8-21187d13a80c\") " pod="openshift-multus/multus-4nwkq" Dec 10 22:49:51 crc kubenswrapper[4791]: I1210 22:49:51.916659 4791 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"multus-socket-dir-parent\" (UniqueName: \"kubernetes.io/host-path/672aa28c-8169-49ed-87b8-21187d13a80c-multus-socket-dir-parent\") pod \"multus-4nwkq\" (UID: \"672aa28c-8169-49ed-87b8-21187d13a80c\") " pod="openshift-multus/multus-4nwkq" Dec 10 22:49:51 crc kubenswrapper[4791]: I1210 22:49:51.916663 4791 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-run-k8s-cni-cncf-io\" (UniqueName: \"kubernetes.io/host-path/672aa28c-8169-49ed-87b8-21187d13a80c-host-run-k8s-cni-cncf-io\") pod \"multus-4nwkq\" (UID: \"672aa28c-8169-49ed-87b8-21187d13a80c\") " pod="openshift-multus/multus-4nwkq" Dec 10 22:49:51 crc kubenswrapper[4791]: I1210 22:49:51.916736 4791 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"os-release\" (UniqueName: \"kubernetes.io/host-path/672aa28c-8169-49ed-87b8-21187d13a80c-os-release\") pod \"multus-4nwkq\" (UID: \"672aa28c-8169-49ed-87b8-21187d13a80c\") " pod="openshift-multus/multus-4nwkq" Dec 10 22:49:51 crc kubenswrapper[4791]: I1210 22:49:51.916735 4791 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-xl62j\" (UniqueName: \"kubernetes.io/projected/672aa28c-8169-49ed-87b8-21187d13a80c-kube-api-access-xl62j\") pod \"multus-4nwkq\" (UID: \"672aa28c-8169-49ed-87b8-21187d13a80c\") " pod="openshift-multus/multus-4nwkq" Dec 10 22:49:51 crc kubenswrapper[4791]: I1210 22:49:51.916776 4791 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-var-lib-kubelet\" (UniqueName: \"kubernetes.io/host-path/672aa28c-8169-49ed-87b8-21187d13a80c-host-var-lib-kubelet\") pod \"multus-4nwkq\" (UID: \"672aa28c-8169-49ed-87b8-21187d13a80c\") " pod="openshift-multus/multus-4nwkq" Dec 10 22:49:51 crc kubenswrapper[4791]: I1210 22:49:51.916825 4791 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-var-lib-kubelet\" (UniqueName: \"kubernetes.io/host-path/672aa28c-8169-49ed-87b8-21187d13a80c-host-var-lib-kubelet\") pod \"multus-4nwkq\" (UID: \"672aa28c-8169-49ed-87b8-21187d13a80c\") " pod="openshift-multus/multus-4nwkq" Dec 10 22:49:51 crc kubenswrapper[4791]: I1210 22:49:51.917185 4791 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"multus-daemon-config\" (UniqueName: \"kubernetes.io/configmap/672aa28c-8169-49ed-87b8-21187d13a80c-multus-daemon-config\") pod \"multus-4nwkq\" (UID: \"672aa28c-8169-49ed-87b8-21187d13a80c\") " pod="openshift-multus/multus-4nwkq" Dec 10 22:49:51 crc kubenswrapper[4791]: I1210 22:49:51.917222 4791 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cni-binary-copy\" (UniqueName: \"kubernetes.io/configmap/672aa28c-8169-49ed-87b8-21187d13a80c-cni-binary-copy\") pod \"multus-4nwkq\" (UID: \"672aa28c-8169-49ed-87b8-21187d13a80c\") " pod="openshift-multus/multus-4nwkq" Dec 10 22:49:51 crc kubenswrapper[4791]: I1210 22:49:51.917481 4791 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="925f1c65-6136-48ba-85aa-3a3b50560753" path="/var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/volumes" Dec 10 22:49:51 crc kubenswrapper[4791]: I1210 22:49:51.918238 4791 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="96b93a3a-6083-4aea-8eab-fe1aa8245ad9" path="/var/lib/kubelet/pods/96b93a3a-6083-4aea-8eab-fe1aa8245ad9/volumes" Dec 10 22:49:51 crc kubenswrapper[4791]: I1210 22:49:51.919364 4791 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="9d4552c7-cd75-42dd-8880-30dd377c49a4" path="/var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/volumes" Dec 10 22:49:51 crc kubenswrapper[4791]: I1210 22:49:51.919832 4791 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="a0128f3a-b052-44ed-a84e-c4c8aaf17c13" path="/var/lib/kubelet/pods/a0128f3a-b052-44ed-a84e-c4c8aaf17c13/volumes" Dec 10 22:49:51 crc kubenswrapper[4791]: I1210 22:49:51.920839 4791 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="a31745f5-9847-4afe-82a5-3161cc66ca93" path="/var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/volumes" Dec 10 22:49:51 crc kubenswrapper[4791]: I1210 22:49:51.921839 4791 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="b11524ee-3fca-4b1b-9cdf-6da289fdbc7d" path="/var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes" Dec 10 22:49:51 crc kubenswrapper[4791]: I1210 22:49:51.922471 4791 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="b6312bbd-5731-4ea0-a20f-81d5a57df44a" path="/var/lib/kubelet/pods/b6312bbd-5731-4ea0-a20f-81d5a57df44a/volumes" Dec 10 22:49:51 crc kubenswrapper[4791]: I1210 22:49:51.923259 4791 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="b6cd30de-2eeb-49a2-ab40-9167f4560ff5" path="/var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/volumes" Dec 10 22:49:51 crc kubenswrapper[4791]: I1210 22:49:51.923801 4791 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="bc5039c0-ea34-426b-a2b7-fbbc87b49a6d" path="/var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/volumes" Dec 10 22:49:51 crc kubenswrapper[4791]: I1210 22:49:51.924699 4791 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="bd23aa5c-e532-4e53-bccf-e79f130c5ae8" path="/var/lib/kubelet/pods/bd23aa5c-e532-4e53-bccf-e79f130c5ae8/volumes" Dec 10 22:49:51 crc kubenswrapper[4791]: I1210 22:49:51.925443 4791 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="bf126b07-da06-4140-9a57-dfd54fc6b486" path="/var/lib/kubelet/pods/bf126b07-da06-4140-9a57-dfd54fc6b486/volumes" Dec 10 22:49:51 crc kubenswrapper[4791]: I1210 22:49:51.927792 4791 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="c03ee662-fb2f-4fc4-a2c1-af487c19d254" path="/var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/volumes" Dec 10 22:49:51 crc kubenswrapper[4791]: I1210 22:49:51.928248 4791 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d" path="/var/lib/kubelet/pods/cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d/volumes" Dec 10 22:49:51 crc kubenswrapper[4791]: I1210 22:49:51.928726 4791 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="e7e6199b-1264-4501-8953-767f51328d08" path="/var/lib/kubelet/pods/e7e6199b-1264-4501-8953-767f51328d08/volumes" Dec 10 22:49:51 crc kubenswrapper[4791]: I1210 22:49:51.929654 4791 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="efdd0498-1daa-4136-9a4a-3b948c2293fc" path="/var/lib/kubelet/pods/efdd0498-1daa-4136-9a4a-3b948c2293fc/volumes" Dec 10 22:49:51 crc kubenswrapper[4791]: I1210 22:49:51.930200 4791 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="f88749ec-7931-4ee7-b3fc-1ec5e11f92e9" path="/var/lib/kubelet/pods/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9/volumes" Dec 10 22:49:51 crc kubenswrapper[4791]: I1210 22:49:51.931209 4791 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="fda69060-fa79-4696-b1a6-7980f124bf7c" path="/var/lib/kubelet/pods/fda69060-fa79-4696-b1a6-7980f124bf7c/volumes" Dec 10 22:49:51 crc kubenswrapper[4791]: I1210 22:49:51.943817 4791 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-xl62j\" (UniqueName: \"kubernetes.io/projected/672aa28c-8169-49ed-87b8-21187d13a80c-kube-api-access-xl62j\") pod \"multus-4nwkq\" (UID: \"672aa28c-8169-49ed-87b8-21187d13a80c\") " pod="openshift-multus/multus-4nwkq" Dec 10 22:49:51 crc kubenswrapper[4791]: I1210 22:49:51.946701 4791 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-config-operator/machine-config-daemon-5rb5l" Dec 10 22:49:51 crc kubenswrapper[4791]: I1210 22:49:51.957978 4791 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/multus-additional-cni-plugins-rw584" Dec 10 22:49:51 crc kubenswrapper[4791]: W1210 22:49:51.958789 4791 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podba35653c_6e06_4cee_a4d6_137764090d18.slice/crio-ceacaa6180b67c4bb59535370651072d2e73338b82ac9157c1a101f503b1eebe WatchSource:0}: Error finding container ceacaa6180b67c4bb59535370651072d2e73338b82ac9157c1a101f503b1eebe: Status 404 returned error can't find the container with id ceacaa6180b67c4bb59535370651072d2e73338b82ac9157c1a101f503b1eebe Dec 10 22:49:51 crc kubenswrapper[4791]: I1210 22:49:51.959154 4791 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-rw584" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ff9beb90-69b0-4732-bf37-0b81f58ecc98\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T22:49:51Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T22:49:51Z\\\",\\\"message\\\":\\\"containers with incomplete status: [egress-router-binary-copy cni-plugins bond-cni-plugin routeoverride-cni whereabouts-cni-bincopy whereabouts-cni]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T22:49:51Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T22:49:51Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fksqn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fksqn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fksqn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fksqn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fksqn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fksqn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fksqn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-10T22:49:51Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-rw584\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T22:49:51Z is after 2025-08-24T17:21:41Z" Dec 10 22:49:51 crc kubenswrapper[4791]: I1210 22:49:51.965902 4791 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/multus-4nwkq" Dec 10 22:49:51 crc kubenswrapper[4791]: I1210 22:49:51.984227 4791 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-10T22:49:50Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-operator]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-10T22:49:50Z\\\",\\\"message\\\":\\\"containers with unready status: [network-operator]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":false,\\\"restartCount\\\":5,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T22:49:51Z is after 2025-08-24T17:21:41Z" Dec 10 22:49:51 crc kubenswrapper[4791]: W1210 22:49:51.987397 4791 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podff9beb90_69b0_4732_bf37_0b81f58ecc98.slice/crio-a58cc1b3667964658240e5637c4bc7ba428ec03704492f7593476e83b9ae568c WatchSource:0}: Error finding container a58cc1b3667964658240e5637c4bc7ba428ec03704492f7593476e83b9ae568c: Status 404 returned error can't find the container with id a58cc1b3667964658240e5637c4bc7ba428ec03704492f7593476e83b9ae568c Dec 10 22:49:51 crc kubenswrapper[4791]: W1210 22:49:51.997071 4791 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod672aa28c_8169_49ed_87b8_21187d13a80c.slice/crio-9749a30a24e31b417631c8d8fe3a73bf8f3b26d533c736656e22af21292f03e0 WatchSource:0}: Error finding container 9749a30a24e31b417631c8d8fe3a73bf8f3b26d533c736656e22af21292f03e0: Status 404 returned error can't find the container with id 9749a30a24e31b417631c8d8fe3a73bf8f3b26d533c736656e22af21292f03e0 Dec 10 22:49:52 crc kubenswrapper[4791]: I1210 22:49:52.009236 4791 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-ovn-kubernetes/ovnkube-node-zhq64"] Dec 10 22:49:52 crc kubenswrapper[4791]: I1210 22:49:52.009995 4791 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-ovn-kubernetes/ovnkube-node-zhq64" Dec 10 22:49:52 crc kubenswrapper[4791]: I1210 22:49:52.012036 4791 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ovn-kubernetes"/"openshift-service-ca.crt" Dec 10 22:49:52 crc kubenswrapper[4791]: I1210 22:49:52.012445 4791 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ovn-kubernetes"/"ovnkube-config" Dec 10 22:49:52 crc kubenswrapper[4791]: I1210 22:49:52.012744 4791 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ovn-kubernetes"/"kube-root-ca.crt" Dec 10 22:49:52 crc kubenswrapper[4791]: I1210 22:49:52.013203 4791 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ovn-kubernetes"/"ovnkube-script-lib" Dec 10 22:49:52 crc kubenswrapper[4791]: I1210 22:49:52.013469 4791 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ovn-kubernetes"/"ovn-node-metrics-cert" Dec 10 22:49:52 crc kubenswrapper[4791]: I1210 22:49:52.013942 4791 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ovn-kubernetes"/"ovn-kubernetes-node-dockercfg-pwtwl" Dec 10 22:49:52 crc kubenswrapper[4791]: I1210 22:49:52.020740 4791 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ovn-kubernetes"/"env-overrides" Dec 10 22:49:52 crc kubenswrapper[4791]: I1210 22:49:52.054091 4791 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-zhq64" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3cd47739-0fa9-4321-aff1-220f8721a0b3\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T22:49:52Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T22:49:52Z\\\",\\\"message\\\":\\\"containers with incomplete status: [kubecfg-setup]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T22:49:52Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T22:49:52Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wvpw8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wvpw8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wvpw8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wvpw8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wvpw8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wvpw8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wvpw8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wvpw8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wvpw8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-10T22:49:52Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-zhq64\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T22:49:52Z is after 2025-08-24T17:21:41Z" Dec 10 22:49:52 crc kubenswrapper[4791]: I1210 22:49:52.066052 4791 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"698bbe9a-fbe5-436d-a835-5e80df77ba24\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T22:49:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T22:49:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T22:49:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T22:49:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T22:49:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://224d73713312376a58ec6cc2bc8fdb6ffa44c4749ae0e7691d5fe27b9afca09f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T22:49:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://372eeaffd40992c6f06efc4df6dfcb44b31de64ee8256c21508cdd7c74ecb96e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T22:49:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://76b6013b062cac5b0b4d12a94f38ce7b43f9c7ea216063455d88a956ac04d26e\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T22:49:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://78e950e122e761d950da1a3c60c06476e53920e8795f6c61db22c01e22d5136e\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T22:49:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-10T22:49:25Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T22:49:52Z is after 2025-08-24T17:21:41Z" Dec 10 22:49:52 crc kubenswrapper[4791]: I1210 22:49:52.079903 4791 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-10T22:49:50Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [webhook approver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-10T22:49:50Z\\\",\\\"message\\\":\\\"containers with unready status: [webhook approver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":false,\\\"restartCount\\\":6,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T22:49:52Z is after 2025-08-24T17:21:41Z" Dec 10 22:49:52 crc kubenswrapper[4791]: I1210 22:49:52.089331 4791 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-vczjq" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b9f00f25-3fa1-4b63-9710-fd352224b01b\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T22:49:50Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T22:49:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T22:49:50Z\\\",\\\"message\\\":\\\"containers with unready status: [node-ca]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T22:49:50Z\\\",\\\"message\\\":\\\"containers with unready status: [node-ca]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lrrpv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-10T22:49:50Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-vczjq\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T22:49:52Z is after 2025-08-24T17:21:41Z" Dec 10 22:49:52 crc kubenswrapper[4791]: I1210 22:49:52.101376 4791 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-4nwkq" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"672aa28c-8169-49ed-87b8-21187d13a80c\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T22:49:51Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T22:49:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T22:49:51Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T22:49:51Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xl62j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-10T22:49:51Z\\\"}}\" for pod \"openshift-multus\"/\"multus-4nwkq\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T22:49:52Z is after 2025-08-24T17:21:41Z" Dec 10 22:49:52 crc kubenswrapper[4791]: I1210 22:49:52.118188 4791 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-slash\" (UniqueName: \"kubernetes.io/host-path/3cd47739-0fa9-4321-aff1-220f8721a0b3-host-slash\") pod \"ovnkube-node-zhq64\" (UID: \"3cd47739-0fa9-4321-aff1-220f8721a0b3\") " pod="openshift-ovn-kubernetes/ovnkube-node-zhq64" Dec 10 22:49:52 crc kubenswrapper[4791]: I1210 22:49:52.118224 4791 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-cni-bin\" (UniqueName: \"kubernetes.io/host-path/3cd47739-0fa9-4321-aff1-220f8721a0b3-host-cni-bin\") pod \"ovnkube-node-zhq64\" (UID: \"3cd47739-0fa9-4321-aff1-220f8721a0b3\") " pod="openshift-ovn-kubernetes/ovnkube-node-zhq64" Dec 10 22:49:52 crc kubenswrapper[4791]: I1210 22:49:52.118254 4791 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-openvswitch\" (UniqueName: \"kubernetes.io/host-path/3cd47739-0fa9-4321-aff1-220f8721a0b3-run-openvswitch\") pod \"ovnkube-node-zhq64\" (UID: \"3cd47739-0fa9-4321-aff1-220f8721a0b3\") " pod="openshift-ovn-kubernetes/ovnkube-node-zhq64" Dec 10 22:49:52 crc kubenswrapper[4791]: I1210 22:49:52.118268 4791 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-ovn\" (UniqueName: \"kubernetes.io/host-path/3cd47739-0fa9-4321-aff1-220f8721a0b3-run-ovn\") pod \"ovnkube-node-zhq64\" (UID: \"3cd47739-0fa9-4321-aff1-220f8721a0b3\") " pod="openshift-ovn-kubernetes/ovnkube-node-zhq64" Dec 10 22:49:52 crc kubenswrapper[4791]: I1210 22:49:52.118285 4791 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovn-node-metrics-cert\" (UniqueName: \"kubernetes.io/secret/3cd47739-0fa9-4321-aff1-220f8721a0b3-ovn-node-metrics-cert\") pod \"ovnkube-node-zhq64\" (UID: \"3cd47739-0fa9-4321-aff1-220f8721a0b3\") " pod="openshift-ovn-kubernetes/ovnkube-node-zhq64" Dec 10 22:49:52 crc kubenswrapper[4791]: I1210 22:49:52.118302 4791 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-wvpw8\" (UniqueName: \"kubernetes.io/projected/3cd47739-0fa9-4321-aff1-220f8721a0b3-kube-api-access-wvpw8\") pod \"ovnkube-node-zhq64\" (UID: \"3cd47739-0fa9-4321-aff1-220f8721a0b3\") " pod="openshift-ovn-kubernetes/ovnkube-node-zhq64" Dec 10 22:49:52 crc kubenswrapper[4791]: I1210 22:49:52.118318 4791 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"systemd-units\" (UniqueName: \"kubernetes.io/host-path/3cd47739-0fa9-4321-aff1-220f8721a0b3-systemd-units\") pod \"ovnkube-node-zhq64\" (UID: \"3cd47739-0fa9-4321-aff1-220f8721a0b3\") " pod="openshift-ovn-kubernetes/ovnkube-node-zhq64" Dec 10 22:49:52 crc kubenswrapper[4791]: I1210 22:49:52.118336 4791 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-run-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/3cd47739-0fa9-4321-aff1-220f8721a0b3-host-run-ovn-kubernetes\") pod \"ovnkube-node-zhq64\" (UID: \"3cd47739-0fa9-4321-aff1-220f8721a0b3\") " pod="openshift-ovn-kubernetes/ovnkube-node-zhq64" Dec 10 22:49:52 crc kubenswrapper[4791]: I1210 22:49:52.118436 4791 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-openvswitch\" (UniqueName: \"kubernetes.io/host-path/3cd47739-0fa9-4321-aff1-220f8721a0b3-etc-openvswitch\") pod \"ovnkube-node-zhq64\" (UID: \"3cd47739-0fa9-4321-aff1-220f8721a0b3\") " pod="openshift-ovn-kubernetes/ovnkube-node-zhq64" Dec 10 22:49:52 crc kubenswrapper[4791]: I1210 22:49:52.118475 4791 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"node-log\" (UniqueName: \"kubernetes.io/host-path/3cd47739-0fa9-4321-aff1-220f8721a0b3-node-log\") pod \"ovnkube-node-zhq64\" (UID: \"3cd47739-0fa9-4321-aff1-220f8721a0b3\") " pod="openshift-ovn-kubernetes/ovnkube-node-zhq64" Dec 10 22:49:52 crc kubenswrapper[4791]: I1210 22:49:52.118524 4791 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/3cd47739-0fa9-4321-aff1-220f8721a0b3-ovnkube-config\") pod \"ovnkube-node-zhq64\" (UID: \"3cd47739-0fa9-4321-aff1-220f8721a0b3\") " pod="openshift-ovn-kubernetes/ovnkube-node-zhq64" Dec 10 22:49:52 crc kubenswrapper[4791]: I1210 22:49:52.118543 4791 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-run-netns\" (UniqueName: \"kubernetes.io/host-path/3cd47739-0fa9-4321-aff1-220f8721a0b3-host-run-netns\") pod \"ovnkube-node-zhq64\" (UID: \"3cd47739-0fa9-4321-aff1-220f8721a0b3\") " pod="openshift-ovn-kubernetes/ovnkube-node-zhq64" Dec 10 22:49:52 crc kubenswrapper[4791]: I1210 22:49:52.118570 4791 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-kubelet\" (UniqueName: \"kubernetes.io/host-path/3cd47739-0fa9-4321-aff1-220f8721a0b3-host-kubelet\") pod \"ovnkube-node-zhq64\" (UID: \"3cd47739-0fa9-4321-aff1-220f8721a0b3\") " pod="openshift-ovn-kubernetes/ovnkube-node-zhq64" Dec 10 22:49:52 crc kubenswrapper[4791]: I1210 22:49:52.118585 4791 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/3cd47739-0fa9-4321-aff1-220f8721a0b3-env-overrides\") pod \"ovnkube-node-zhq64\" (UID: \"3cd47739-0fa9-4321-aff1-220f8721a0b3\") " pod="openshift-ovn-kubernetes/ovnkube-node-zhq64" Dec 10 22:49:52 crc kubenswrapper[4791]: I1210 22:49:52.118606 4791 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-systemd\" (UniqueName: \"kubernetes.io/host-path/3cd47739-0fa9-4321-aff1-220f8721a0b3-run-systemd\") pod \"ovnkube-node-zhq64\" (UID: \"3cd47739-0fa9-4321-aff1-220f8721a0b3\") " pod="openshift-ovn-kubernetes/ovnkube-node-zhq64" Dec 10 22:49:52 crc kubenswrapper[4791]: I1210 22:49:52.118622 4791 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-var-lib-cni-networks-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/3cd47739-0fa9-4321-aff1-220f8721a0b3-host-var-lib-cni-networks-ovn-kubernetes\") pod \"ovnkube-node-zhq64\" (UID: \"3cd47739-0fa9-4321-aff1-220f8721a0b3\") " pod="openshift-ovn-kubernetes/ovnkube-node-zhq64" Dec 10 22:49:52 crc kubenswrapper[4791]: I1210 22:49:52.118639 4791 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovnkube-script-lib\" (UniqueName: \"kubernetes.io/configmap/3cd47739-0fa9-4321-aff1-220f8721a0b3-ovnkube-script-lib\") pod \"ovnkube-node-zhq64\" (UID: \"3cd47739-0fa9-4321-aff1-220f8721a0b3\") " pod="openshift-ovn-kubernetes/ovnkube-node-zhq64" Dec 10 22:49:52 crc kubenswrapper[4791]: I1210 22:49:52.118683 4791 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-lib-openvswitch\" (UniqueName: \"kubernetes.io/host-path/3cd47739-0fa9-4321-aff1-220f8721a0b3-var-lib-openvswitch\") pod \"ovnkube-node-zhq64\" (UID: \"3cd47739-0fa9-4321-aff1-220f8721a0b3\") " pod="openshift-ovn-kubernetes/ovnkube-node-zhq64" Dec 10 22:49:52 crc kubenswrapper[4791]: I1210 22:49:52.118700 4791 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"log-socket\" (UniqueName: \"kubernetes.io/host-path/3cd47739-0fa9-4321-aff1-220f8721a0b3-log-socket\") pod \"ovnkube-node-zhq64\" (UID: \"3cd47739-0fa9-4321-aff1-220f8721a0b3\") " pod="openshift-ovn-kubernetes/ovnkube-node-zhq64" Dec 10 22:49:52 crc kubenswrapper[4791]: I1210 22:49:52.118714 4791 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-cni-netd\" (UniqueName: \"kubernetes.io/host-path/3cd47739-0fa9-4321-aff1-220f8721a0b3-host-cni-netd\") pod \"ovnkube-node-zhq64\" (UID: \"3cd47739-0fa9-4321-aff1-220f8721a0b3\") " pod="openshift-ovn-kubernetes/ovnkube-node-zhq64" Dec 10 22:49:52 crc kubenswrapper[4791]: I1210 22:49:52.122066 4791 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-10T22:49:50Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-operator]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-10T22:49:50Z\\\",\\\"message\\\":\\\"containers with unready status: [network-operator]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":false,\\\"restartCount\\\":5,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T22:49:52Z is after 2025-08-24T17:21:41Z" Dec 10 22:49:52 crc kubenswrapper[4791]: I1210 22:49:52.133290 4791 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-10T22:49:50Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-10T22:49:50Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T22:49:52Z is after 2025-08-24T17:21:41Z" Dec 10 22:49:52 crc kubenswrapper[4791]: I1210 22:49:52.145079 4791 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-10T22:49:50Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-10T22:49:50Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T22:49:52Z is after 2025-08-24T17:21:41Z" Dec 10 22:49:52 crc kubenswrapper[4791]: I1210 22:49:52.156297 4791 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-tchxk" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"c0ab74e0-42a9-4f42-ac80-16030cf4ffff\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T22:49:50Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T22:49:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T22:49:50Z\\\",\\\"message\\\":\\\"containers with unready status: [dns-node-resolver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T22:49:50Z\\\",\\\"message\\\":\\\"containers with unready status: [dns-node-resolver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zn2qt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-10T22:49:50Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-tchxk\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T22:49:52Z is after 2025-08-24T17:21:41Z" Dec 10 22:49:52 crc kubenswrapper[4791]: I1210 22:49:52.173956 4791 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-rw584" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ff9beb90-69b0-4732-bf37-0b81f58ecc98\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T22:49:51Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T22:49:51Z\\\",\\\"message\\\":\\\"containers with incomplete status: [egress-router-binary-copy cni-plugins bond-cni-plugin routeoverride-cni whereabouts-cni-bincopy whereabouts-cni]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T22:49:51Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T22:49:51Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fksqn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fksqn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fksqn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fksqn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fksqn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fksqn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fksqn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-10T22:49:51Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-rw584\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T22:49:52Z is after 2025-08-24T17:21:41Z" Dec 10 22:49:52 crc kubenswrapper[4791]: I1210 22:49:52.178173 4791 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-image-registry/node-ca-vczjq" event={"ID":"b9f00f25-3fa1-4b63-9710-fd352224b01b","Type":"ContainerStarted","Data":"4698cf4c6b55b41e628448d37c2eb847c2777d0efbd93afcbf367d5e78623b7e"} Dec 10 22:49:52 crc kubenswrapper[4791]: I1210 22:49:52.178217 4791 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-image-registry/node-ca-vczjq" event={"ID":"b9f00f25-3fa1-4b63-9710-fd352224b01b","Type":"ContainerStarted","Data":"c903660e3845f6c7a74b981c7e29ea9d0a2c6d1e86a180f77122f6690a1b9946"} Dec 10 22:49:52 crc kubenswrapper[4791]: I1210 22:49:52.181103 4791 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-additional-cni-plugins-rw584" event={"ID":"ff9beb90-69b0-4732-bf37-0b81f58ecc98","Type":"ContainerStarted","Data":"a58cc1b3667964658240e5637c4bc7ba428ec03704492f7593476e83b9ae568c"} Dec 10 22:49:52 crc kubenswrapper[4791]: I1210 22:49:52.182456 4791 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-5rb5l" event={"ID":"ba35653c-6e06-4cee-a4d6-137764090d18","Type":"ContainerStarted","Data":"185eb2ff44221a514878283ec873773a6c9a9073dc64e296f982e60dee1338e4"} Dec 10 22:49:52 crc kubenswrapper[4791]: I1210 22:49:52.182479 4791 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-5rb5l" event={"ID":"ba35653c-6e06-4cee-a4d6-137764090d18","Type":"ContainerStarted","Data":"ceacaa6180b67c4bb59535370651072d2e73338b82ac9157c1a101f503b1eebe"} Dec 10 22:49:52 crc kubenswrapper[4791]: I1210 22:49:52.183797 4791 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-dns/node-resolver-tchxk" event={"ID":"c0ab74e0-42a9-4f42-ac80-16030cf4ffff","Type":"ContainerStarted","Data":"04e0fd1b53256b7a938e8ccab93e2a4eb73dd1a59e0b9799e09c55e511baf7fe"} Dec 10 22:49:52 crc kubenswrapper[4791]: I1210 22:49:52.183841 4791 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-dns/node-resolver-tchxk" event={"ID":"c0ab74e0-42a9-4f42-ac80-16030cf4ffff","Type":"ContainerStarted","Data":"274a2d7c02eee2d83ecf84953663519fe920414b624e92f4e28108fb844af339"} Dec 10 22:49:52 crc kubenswrapper[4791]: I1210 22:49:52.186571 4791 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" event={"ID":"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49","Type":"ContainerStarted","Data":"3f54f2a8d7fb7cd500f2a5657304ec7a6de79acfd2976df8082c4de0cf261949"} Dec 10 22:49:52 crc kubenswrapper[4791]: I1210 22:49:52.189055 4791 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f8349fea-c49b-49a1-b5ee-032c5dffe021\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T22:49:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T22:49:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T22:49:25Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T22:49:25Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T22:49:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://acafe12186df99d1059a31260de0eafce35ff7771bcb8f5c6753b72665d6f60b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T22:49:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ea4c346507862a8c12756a3061d6e25e788a8d79f4621b62f81f4322185c82a8\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T22:49:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://2e6e324ec6f4f7bb69fdfab4ea3ac1aa55e480dd8e2a8984b0fe5b4c9f66636e\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T22:49:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://1b2e11ff2bf8789fa1d0d23ae648b3e6ef653106e5bc4bbb5dd6447f7cd6f05d\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d6aa7f8c5d908d90008f2df49860ac8a3480136da567e3d85f86135d779204f9\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-12-10T22:49:44Z\\\",\\\"message\\\":\\\"W1210 22:49:33.845733 1 cmd.go:257] Using insecure, self-signed certificates\\\\nI1210 22:49:33.846108 1 crypto.go:601] Generating new CA for check-endpoints-signer@1765406973 cert, and key in /tmp/serving-cert-2585701456/serving-signer.crt, /tmp/serving-cert-2585701456/serving-signer.key\\\\nI1210 22:49:34.529748 1 observer_polling.go:159] Starting file observer\\\\nW1210 22:49:34.532225 1 builder.go:272] unable to get owner reference (falling back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": dial tcp [::1]:6443: connect: connection refused\\\\nI1210 22:49:34.532599 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1210 22:49:34.533977 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-2585701456/tls.crt::/tmp/serving-cert-2585701456/tls.key\\\\\\\"\\\\nF1210 22:49:44.774710 1 cmd.go:182] error initializing delegating authentication: unable to load configmap based request-header-client-ca-file: Get \\\\\\\"https://localhost:6443/api/v1/namespaces/kube-system/configmaps/extension-apiserver-authentication\\\\\\\": net/http: TLS handshake timeout\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-10T22:49:33Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T22:49:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://0c4d036ab269f62561fe7ef1586012cb3a16e3583a08b9488a0616ec00c8e0ae\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T22:49:32Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c5856d6b26c19fda13b5c9ac2892a7e2f9c6f79feb8173ecfd9b33b673327d67\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c5856d6b26c19fda13b5c9ac2892a7e2f9c6f79feb8173ecfd9b33b673327d67\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-10T22:49:27Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-10T22:49:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-10T22:49:25Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T22:49:52Z is after 2025-08-24T17:21:41Z" Dec 10 22:49:52 crc kubenswrapper[4791]: I1210 22:49:52.189424 4791 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" event={"ID":"ef543e1b-8068-4ea3-b32a-61027b32e95d","Type":"ContainerStarted","Data":"3734b110802828632d44263ed0a43cabc4b710833f0be120d72db0940fe783e2"} Dec 10 22:49:52 crc kubenswrapper[4791]: I1210 22:49:52.189450 4791 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" event={"ID":"ef543e1b-8068-4ea3-b32a-61027b32e95d","Type":"ContainerStarted","Data":"794803ac9be3c52509e6dcb10854d8864e1a9e406762e57c7971b07c4da1220e"} Dec 10 22:49:52 crc kubenswrapper[4791]: I1210 22:49:52.189467 4791 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" event={"ID":"ef543e1b-8068-4ea3-b32a-61027b32e95d","Type":"ContainerStarted","Data":"d5caec4216e5c8fd369d36be4c6714eac588fae2baa3d06a21c0df62e6c9979f"} Dec 10 22:49:52 crc kubenswrapper[4791]: I1210 22:49:52.191325 4791 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" event={"ID":"37a5e44f-9a88-4405-be8a-b645485e7312","Type":"ContainerStarted","Data":"20651eb0c16e2c71222f56da1ff755cb0e7b084886acbbb24f0ccaf57041ed85"} Dec 10 22:49:52 crc kubenswrapper[4791]: I1210 22:49:52.191410 4791 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" event={"ID":"37a5e44f-9a88-4405-be8a-b645485e7312","Type":"ContainerStarted","Data":"8ae024a09cb2c50797598ccd8a76bfb4470281f89a249a924963a628f09ad4a7"} Dec 10 22:49:52 crc kubenswrapper[4791]: I1210 22:49:52.193632 4791 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-4nwkq" event={"ID":"672aa28c-8169-49ed-87b8-21187d13a80c","Type":"ContainerStarted","Data":"a2f2a62315529e148a10a33ccb241e9417fbde0fe42afb61cb7ee2438c74636f"} Dec 10 22:49:52 crc kubenswrapper[4791]: I1210 22:49:52.193667 4791 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-4nwkq" event={"ID":"672aa28c-8169-49ed-87b8-21187d13a80c","Type":"ContainerStarted","Data":"9749a30a24e31b417631c8d8fe3a73bf8f3b26d533c736656e22af21292f03e0"} Dec 10 22:49:52 crc kubenswrapper[4791]: I1210 22:49:52.203724 4791 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-10T22:49:50Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-10T22:49:50Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T22:49:52Z is after 2025-08-24T17:21:41Z" Dec 10 22:49:52 crc kubenswrapper[4791]: I1210 22:49:52.215974 4791 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-10T22:49:50Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-10T22:49:50Z\\\",\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T22:49:52Z is after 2025-08-24T17:21:41Z" Dec 10 22:49:52 crc kubenswrapper[4791]: I1210 22:49:52.219093 4791 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-kubelet\" (UniqueName: \"kubernetes.io/host-path/3cd47739-0fa9-4321-aff1-220f8721a0b3-host-kubelet\") pod \"ovnkube-node-zhq64\" (UID: \"3cd47739-0fa9-4321-aff1-220f8721a0b3\") " pod="openshift-ovn-kubernetes/ovnkube-node-zhq64" Dec 10 22:49:52 crc kubenswrapper[4791]: I1210 22:49:52.219130 4791 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/3cd47739-0fa9-4321-aff1-220f8721a0b3-env-overrides\") pod \"ovnkube-node-zhq64\" (UID: \"3cd47739-0fa9-4321-aff1-220f8721a0b3\") " pod="openshift-ovn-kubernetes/ovnkube-node-zhq64" Dec 10 22:49:52 crc kubenswrapper[4791]: I1210 22:49:52.219154 4791 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-systemd\" (UniqueName: \"kubernetes.io/host-path/3cd47739-0fa9-4321-aff1-220f8721a0b3-run-systemd\") pod \"ovnkube-node-zhq64\" (UID: \"3cd47739-0fa9-4321-aff1-220f8721a0b3\") " pod="openshift-ovn-kubernetes/ovnkube-node-zhq64" Dec 10 22:49:52 crc kubenswrapper[4791]: I1210 22:49:52.219178 4791 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-var-lib-cni-networks-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/3cd47739-0fa9-4321-aff1-220f8721a0b3-host-var-lib-cni-networks-ovn-kubernetes\") pod \"ovnkube-node-zhq64\" (UID: \"3cd47739-0fa9-4321-aff1-220f8721a0b3\") " pod="openshift-ovn-kubernetes/ovnkube-node-zhq64" Dec 10 22:49:52 crc kubenswrapper[4791]: I1210 22:49:52.219204 4791 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovnkube-script-lib\" (UniqueName: \"kubernetes.io/configmap/3cd47739-0fa9-4321-aff1-220f8721a0b3-ovnkube-script-lib\") pod \"ovnkube-node-zhq64\" (UID: \"3cd47739-0fa9-4321-aff1-220f8721a0b3\") " pod="openshift-ovn-kubernetes/ovnkube-node-zhq64" Dec 10 22:49:52 crc kubenswrapper[4791]: I1210 22:49:52.219243 4791 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-lib-openvswitch\" (UniqueName: \"kubernetes.io/host-path/3cd47739-0fa9-4321-aff1-220f8721a0b3-var-lib-openvswitch\") pod \"ovnkube-node-zhq64\" (UID: \"3cd47739-0fa9-4321-aff1-220f8721a0b3\") " pod="openshift-ovn-kubernetes/ovnkube-node-zhq64" Dec 10 22:49:52 crc kubenswrapper[4791]: I1210 22:49:52.219274 4791 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"log-socket\" (UniqueName: \"kubernetes.io/host-path/3cd47739-0fa9-4321-aff1-220f8721a0b3-log-socket\") pod \"ovnkube-node-zhq64\" (UID: \"3cd47739-0fa9-4321-aff1-220f8721a0b3\") " pod="openshift-ovn-kubernetes/ovnkube-node-zhq64" Dec 10 22:49:52 crc kubenswrapper[4791]: I1210 22:49:52.219279 4791 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-var-lib-cni-networks-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/3cd47739-0fa9-4321-aff1-220f8721a0b3-host-var-lib-cni-networks-ovn-kubernetes\") pod \"ovnkube-node-zhq64\" (UID: \"3cd47739-0fa9-4321-aff1-220f8721a0b3\") " pod="openshift-ovn-kubernetes/ovnkube-node-zhq64" Dec 10 22:49:52 crc kubenswrapper[4791]: I1210 22:49:52.219296 4791 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-cni-netd\" (UniqueName: \"kubernetes.io/host-path/3cd47739-0fa9-4321-aff1-220f8721a0b3-host-cni-netd\") pod \"ovnkube-node-zhq64\" (UID: \"3cd47739-0fa9-4321-aff1-220f8721a0b3\") " pod="openshift-ovn-kubernetes/ovnkube-node-zhq64" Dec 10 22:49:52 crc kubenswrapper[4791]: I1210 22:49:52.219332 4791 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-cni-netd\" (UniqueName: \"kubernetes.io/host-path/3cd47739-0fa9-4321-aff1-220f8721a0b3-host-cni-netd\") pod \"ovnkube-node-zhq64\" (UID: \"3cd47739-0fa9-4321-aff1-220f8721a0b3\") " pod="openshift-ovn-kubernetes/ovnkube-node-zhq64" Dec 10 22:49:52 crc kubenswrapper[4791]: I1210 22:49:52.219377 4791 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-slash\" (UniqueName: \"kubernetes.io/host-path/3cd47739-0fa9-4321-aff1-220f8721a0b3-host-slash\") pod \"ovnkube-node-zhq64\" (UID: \"3cd47739-0fa9-4321-aff1-220f8721a0b3\") " pod="openshift-ovn-kubernetes/ovnkube-node-zhq64" Dec 10 22:49:52 crc kubenswrapper[4791]: I1210 22:49:52.219395 4791 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-cni-bin\" (UniqueName: \"kubernetes.io/host-path/3cd47739-0fa9-4321-aff1-220f8721a0b3-host-cni-bin\") pod \"ovnkube-node-zhq64\" (UID: \"3cd47739-0fa9-4321-aff1-220f8721a0b3\") " pod="openshift-ovn-kubernetes/ovnkube-node-zhq64" Dec 10 22:49:52 crc kubenswrapper[4791]: I1210 22:49:52.219388 4791 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"log-socket\" (UniqueName: \"kubernetes.io/host-path/3cd47739-0fa9-4321-aff1-220f8721a0b3-log-socket\") pod \"ovnkube-node-zhq64\" (UID: \"3cd47739-0fa9-4321-aff1-220f8721a0b3\") " pod="openshift-ovn-kubernetes/ovnkube-node-zhq64" Dec 10 22:49:52 crc kubenswrapper[4791]: I1210 22:49:52.219432 4791 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-openvswitch\" (UniqueName: \"kubernetes.io/host-path/3cd47739-0fa9-4321-aff1-220f8721a0b3-run-openvswitch\") pod \"ovnkube-node-zhq64\" (UID: \"3cd47739-0fa9-4321-aff1-220f8721a0b3\") " pod="openshift-ovn-kubernetes/ovnkube-node-zhq64" Dec 10 22:49:52 crc kubenswrapper[4791]: I1210 22:49:52.219435 4791 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-slash\" (UniqueName: \"kubernetes.io/host-path/3cd47739-0fa9-4321-aff1-220f8721a0b3-host-slash\") pod \"ovnkube-node-zhq64\" (UID: \"3cd47739-0fa9-4321-aff1-220f8721a0b3\") " pod="openshift-ovn-kubernetes/ovnkube-node-zhq64" Dec 10 22:49:52 crc kubenswrapper[4791]: I1210 22:49:52.219466 4791 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-ovn\" (UniqueName: \"kubernetes.io/host-path/3cd47739-0fa9-4321-aff1-220f8721a0b3-run-ovn\") pod \"ovnkube-node-zhq64\" (UID: \"3cd47739-0fa9-4321-aff1-220f8721a0b3\") " pod="openshift-ovn-kubernetes/ovnkube-node-zhq64" Dec 10 22:49:52 crc kubenswrapper[4791]: I1210 22:49:52.219398 4791 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-lib-openvswitch\" (UniqueName: \"kubernetes.io/host-path/3cd47739-0fa9-4321-aff1-220f8721a0b3-var-lib-openvswitch\") pod \"ovnkube-node-zhq64\" (UID: \"3cd47739-0fa9-4321-aff1-220f8721a0b3\") " pod="openshift-ovn-kubernetes/ovnkube-node-zhq64" Dec 10 22:49:52 crc kubenswrapper[4791]: I1210 22:49:52.219448 4791 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-ovn\" (UniqueName: \"kubernetes.io/host-path/3cd47739-0fa9-4321-aff1-220f8721a0b3-run-ovn\") pod \"ovnkube-node-zhq64\" (UID: \"3cd47739-0fa9-4321-aff1-220f8721a0b3\") " pod="openshift-ovn-kubernetes/ovnkube-node-zhq64" Dec 10 22:49:52 crc kubenswrapper[4791]: I1210 22:49:52.219482 4791 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-openvswitch\" (UniqueName: \"kubernetes.io/host-path/3cd47739-0fa9-4321-aff1-220f8721a0b3-run-openvswitch\") pod \"ovnkube-node-zhq64\" (UID: \"3cd47739-0fa9-4321-aff1-220f8721a0b3\") " pod="openshift-ovn-kubernetes/ovnkube-node-zhq64" Dec 10 22:49:52 crc kubenswrapper[4791]: I1210 22:49:52.219501 4791 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovn-node-metrics-cert\" (UniqueName: \"kubernetes.io/secret/3cd47739-0fa9-4321-aff1-220f8721a0b3-ovn-node-metrics-cert\") pod \"ovnkube-node-zhq64\" (UID: \"3cd47739-0fa9-4321-aff1-220f8721a0b3\") " pod="openshift-ovn-kubernetes/ovnkube-node-zhq64" Dec 10 22:49:52 crc kubenswrapper[4791]: I1210 22:49:52.219520 4791 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-wvpw8\" (UniqueName: \"kubernetes.io/projected/3cd47739-0fa9-4321-aff1-220f8721a0b3-kube-api-access-wvpw8\") pod \"ovnkube-node-zhq64\" (UID: \"3cd47739-0fa9-4321-aff1-220f8721a0b3\") " pod="openshift-ovn-kubernetes/ovnkube-node-zhq64" Dec 10 22:49:52 crc kubenswrapper[4791]: I1210 22:49:52.219537 4791 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"systemd-units\" (UniqueName: \"kubernetes.io/host-path/3cd47739-0fa9-4321-aff1-220f8721a0b3-systemd-units\") pod \"ovnkube-node-zhq64\" (UID: \"3cd47739-0fa9-4321-aff1-220f8721a0b3\") " pod="openshift-ovn-kubernetes/ovnkube-node-zhq64" Dec 10 22:49:52 crc kubenswrapper[4791]: I1210 22:49:52.219552 4791 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-run-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/3cd47739-0fa9-4321-aff1-220f8721a0b3-host-run-ovn-kubernetes\") pod \"ovnkube-node-zhq64\" (UID: \"3cd47739-0fa9-4321-aff1-220f8721a0b3\") " pod="openshift-ovn-kubernetes/ovnkube-node-zhq64" Dec 10 22:49:52 crc kubenswrapper[4791]: I1210 22:49:52.219581 4791 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-openvswitch\" (UniqueName: \"kubernetes.io/host-path/3cd47739-0fa9-4321-aff1-220f8721a0b3-etc-openvswitch\") pod \"ovnkube-node-zhq64\" (UID: \"3cd47739-0fa9-4321-aff1-220f8721a0b3\") " pod="openshift-ovn-kubernetes/ovnkube-node-zhq64" Dec 10 22:49:52 crc kubenswrapper[4791]: I1210 22:49:52.219583 4791 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"systemd-units\" (UniqueName: \"kubernetes.io/host-path/3cd47739-0fa9-4321-aff1-220f8721a0b3-systemd-units\") pod \"ovnkube-node-zhq64\" (UID: \"3cd47739-0fa9-4321-aff1-220f8721a0b3\") " pod="openshift-ovn-kubernetes/ovnkube-node-zhq64" Dec 10 22:49:52 crc kubenswrapper[4791]: I1210 22:49:52.219595 4791 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"node-log\" (UniqueName: \"kubernetes.io/host-path/3cd47739-0fa9-4321-aff1-220f8721a0b3-node-log\") pod \"ovnkube-node-zhq64\" (UID: \"3cd47739-0fa9-4321-aff1-220f8721a0b3\") " pod="openshift-ovn-kubernetes/ovnkube-node-zhq64" Dec 10 22:49:52 crc kubenswrapper[4791]: I1210 22:49:52.219616 4791 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"node-log\" (UniqueName: \"kubernetes.io/host-path/3cd47739-0fa9-4321-aff1-220f8721a0b3-node-log\") pod \"ovnkube-node-zhq64\" (UID: \"3cd47739-0fa9-4321-aff1-220f8721a0b3\") " pod="openshift-ovn-kubernetes/ovnkube-node-zhq64" Dec 10 22:49:52 crc kubenswrapper[4791]: I1210 22:49:52.219627 4791 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/3cd47739-0fa9-4321-aff1-220f8721a0b3-ovnkube-config\") pod \"ovnkube-node-zhq64\" (UID: \"3cd47739-0fa9-4321-aff1-220f8721a0b3\") " pod="openshift-ovn-kubernetes/ovnkube-node-zhq64" Dec 10 22:49:52 crc kubenswrapper[4791]: I1210 22:49:52.219639 4791 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-run-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/3cd47739-0fa9-4321-aff1-220f8721a0b3-host-run-ovn-kubernetes\") pod \"ovnkube-node-zhq64\" (UID: \"3cd47739-0fa9-4321-aff1-220f8721a0b3\") " pod="openshift-ovn-kubernetes/ovnkube-node-zhq64" Dec 10 22:49:52 crc kubenswrapper[4791]: I1210 22:49:52.219660 4791 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-openvswitch\" (UniqueName: \"kubernetes.io/host-path/3cd47739-0fa9-4321-aff1-220f8721a0b3-etc-openvswitch\") pod \"ovnkube-node-zhq64\" (UID: \"3cd47739-0fa9-4321-aff1-220f8721a0b3\") " pod="openshift-ovn-kubernetes/ovnkube-node-zhq64" Dec 10 22:49:52 crc kubenswrapper[4791]: I1210 22:49:52.219655 4791 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-run-netns\" (UniqueName: \"kubernetes.io/host-path/3cd47739-0fa9-4321-aff1-220f8721a0b3-host-run-netns\") pod \"ovnkube-node-zhq64\" (UID: \"3cd47739-0fa9-4321-aff1-220f8721a0b3\") " pod="openshift-ovn-kubernetes/ovnkube-node-zhq64" Dec 10 22:49:52 crc kubenswrapper[4791]: I1210 22:49:52.219477 4791 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-cni-bin\" (UniqueName: \"kubernetes.io/host-path/3cd47739-0fa9-4321-aff1-220f8721a0b3-host-cni-bin\") pod \"ovnkube-node-zhq64\" (UID: \"3cd47739-0fa9-4321-aff1-220f8721a0b3\") " pod="openshift-ovn-kubernetes/ovnkube-node-zhq64" Dec 10 22:49:52 crc kubenswrapper[4791]: I1210 22:49:52.219729 4791 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-run-netns\" (UniqueName: \"kubernetes.io/host-path/3cd47739-0fa9-4321-aff1-220f8721a0b3-host-run-netns\") pod \"ovnkube-node-zhq64\" (UID: \"3cd47739-0fa9-4321-aff1-220f8721a0b3\") " pod="openshift-ovn-kubernetes/ovnkube-node-zhq64" Dec 10 22:49:52 crc kubenswrapper[4791]: I1210 22:49:52.219905 4791 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/3cd47739-0fa9-4321-aff1-220f8721a0b3-env-overrides\") pod \"ovnkube-node-zhq64\" (UID: \"3cd47739-0fa9-4321-aff1-220f8721a0b3\") " pod="openshift-ovn-kubernetes/ovnkube-node-zhq64" Dec 10 22:49:52 crc kubenswrapper[4791]: I1210 22:49:52.220036 4791 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovnkube-script-lib\" (UniqueName: \"kubernetes.io/configmap/3cd47739-0fa9-4321-aff1-220f8721a0b3-ovnkube-script-lib\") pod \"ovnkube-node-zhq64\" (UID: \"3cd47739-0fa9-4321-aff1-220f8721a0b3\") " pod="openshift-ovn-kubernetes/ovnkube-node-zhq64" Dec 10 22:49:52 crc kubenswrapper[4791]: I1210 22:49:52.220084 4791 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-systemd\" (UniqueName: \"kubernetes.io/host-path/3cd47739-0fa9-4321-aff1-220f8721a0b3-run-systemd\") pod \"ovnkube-node-zhq64\" (UID: \"3cd47739-0fa9-4321-aff1-220f8721a0b3\") " pod="openshift-ovn-kubernetes/ovnkube-node-zhq64" Dec 10 22:49:52 crc kubenswrapper[4791]: I1210 22:49:52.220186 4791 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/3cd47739-0fa9-4321-aff1-220f8721a0b3-ovnkube-config\") pod \"ovnkube-node-zhq64\" (UID: \"3cd47739-0fa9-4321-aff1-220f8721a0b3\") " pod="openshift-ovn-kubernetes/ovnkube-node-zhq64" Dec 10 22:49:52 crc kubenswrapper[4791]: I1210 22:49:52.220224 4791 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-kubelet\" (UniqueName: \"kubernetes.io/host-path/3cd47739-0fa9-4321-aff1-220f8721a0b3-host-kubelet\") pod \"ovnkube-node-zhq64\" (UID: \"3cd47739-0fa9-4321-aff1-220f8721a0b3\") " pod="openshift-ovn-kubernetes/ovnkube-node-zhq64" Dec 10 22:49:52 crc kubenswrapper[4791]: I1210 22:49:52.229846 4791 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-5rb5l" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ba35653c-6e06-4cee-a4d6-137764090d18\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T22:49:51Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T22:49:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T22:49:51Z\\\",\\\"message\\\":\\\"containers with unready status: [machine-config-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T22:49:51Z\\\",\\\"message\\\":\\\"containers with unready status: [machine-config-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fvvrs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fvvrs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-10T22:49:51Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-5rb5l\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T22:49:52Z is after 2025-08-24T17:21:41Z" Dec 10 22:49:52 crc kubenswrapper[4791]: I1210 22:49:52.242821 4791 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-5rb5l" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ba35653c-6e06-4cee-a4d6-137764090d18\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T22:49:51Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T22:49:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T22:49:51Z\\\",\\\"message\\\":\\\"containers with unready status: [machine-config-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T22:49:51Z\\\",\\\"message\\\":\\\"containers with unready status: [machine-config-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fvvrs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fvvrs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-10T22:49:51Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-5rb5l\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T22:49:52Z is after 2025-08-24T17:21:41Z" Dec 10 22:49:52 crc kubenswrapper[4791]: I1210 22:49:52.263408 4791 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"698bbe9a-fbe5-436d-a835-5e80df77ba24\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T22:49:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T22:49:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T22:49:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T22:49:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T22:49:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://224d73713312376a58ec6cc2bc8fdb6ffa44c4749ae0e7691d5fe27b9afca09f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T22:49:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://372eeaffd40992c6f06efc4df6dfcb44b31de64ee8256c21508cdd7c74ecb96e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T22:49:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://76b6013b062cac5b0b4d12a94f38ce7b43f9c7ea216063455d88a956ac04d26e\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T22:49:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://78e950e122e761d950da1a3c60c06476e53920e8795f6c61db22c01e22d5136e\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T22:49:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-10T22:49:25Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T22:49:52Z is after 2025-08-24T17:21:41Z" Dec 10 22:49:52 crc kubenswrapper[4791]: I1210 22:49:52.298909 4791 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-10T22:49:52Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-10T22:49:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-10T22:49:52Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3734b110802828632d44263ed0a43cabc4b710833f0be120d72db0940fe783e2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T22:49:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://794803ac9be3c52509e6dcb10854d8864e1a9e406762e57c7971b07c4da1220e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T22:49:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T22:49:52Z is after 2025-08-24T17:21:41Z" Dec 10 22:49:52 crc kubenswrapper[4791]: I1210 22:49:52.336361 4791 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-vczjq" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b9f00f25-3fa1-4b63-9710-fd352224b01b\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T22:49:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T22:49:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T22:49:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T22:49:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4698cf4c6b55b41e628448d37c2eb847c2777d0efbd93afcbf367d5e78623b7e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T22:49:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lrrpv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-10T22:49:50Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-vczjq\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T22:49:52Z is after 2025-08-24T17:21:41Z" Dec 10 22:49:52 crc kubenswrapper[4791]: I1210 22:49:52.379177 4791 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-wvpw8\" (UniqueName: \"kubernetes.io/projected/3cd47739-0fa9-4321-aff1-220f8721a0b3-kube-api-access-wvpw8\") pod \"ovnkube-node-zhq64\" (UID: \"3cd47739-0fa9-4321-aff1-220f8721a0b3\") " pod="openshift-ovn-kubernetes/ovnkube-node-zhq64" Dec 10 22:49:52 crc kubenswrapper[4791]: I1210 22:49:52.379300 4791 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovn-node-metrics-cert\" (UniqueName: \"kubernetes.io/secret/3cd47739-0fa9-4321-aff1-220f8721a0b3-ovn-node-metrics-cert\") pod \"ovnkube-node-zhq64\" (UID: \"3cd47739-0fa9-4321-aff1-220f8721a0b3\") " pod="openshift-ovn-kubernetes/ovnkube-node-zhq64" Dec 10 22:49:52 crc kubenswrapper[4791]: I1210 22:49:52.386715 4791 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-4nwkq" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"672aa28c-8169-49ed-87b8-21187d13a80c\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T22:49:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T22:49:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T22:49:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T22:49:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a2f2a62315529e148a10a33ccb241e9417fbde0fe42afb61cb7ee2438c74636f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T22:49:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xl62j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-10T22:49:51Z\\\"}}\" for pod \"openshift-multus\"/\"multus-4nwkq\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T22:49:52Z is after 2025-08-24T17:21:41Z" Dec 10 22:49:52 crc kubenswrapper[4791]: I1210 22:49:52.429525 4791 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-zhq64" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3cd47739-0fa9-4321-aff1-220f8721a0b3\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T22:49:52Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T22:49:52Z\\\",\\\"message\\\":\\\"containers with incomplete status: [kubecfg-setup]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T22:49:52Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T22:49:52Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wvpw8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wvpw8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wvpw8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wvpw8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wvpw8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wvpw8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wvpw8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wvpw8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wvpw8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-10T22:49:52Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-zhq64\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T22:49:52Z is after 2025-08-24T17:21:41Z" Dec 10 22:49:52 crc kubenswrapper[4791]: I1210 22:49:52.461120 4791 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-10T22:49:52Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-10T22:49:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-10T22:49:52Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://20651eb0c16e2c71222f56da1ff755cb0e7b084886acbbb24f0ccaf57041ed85\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T22:49:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T22:49:52Z is after 2025-08-24T17:21:41Z" Dec 10 22:49:52 crc kubenswrapper[4791]: I1210 22:49:52.499706 4791 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-10T22:49:50Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-10T22:49:50Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T22:49:52Z is after 2025-08-24T17:21:41Z" Dec 10 22:49:52 crc kubenswrapper[4791]: I1210 22:49:52.522985 4791 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 10 22:49:52 crc kubenswrapper[4791]: E1210 22:49:52.523157 4791 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-10 22:49:54.523132619 +0000 UTC m=+28.952750232 (durationBeforeRetry 2s). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 10 22:49:52 crc kubenswrapper[4791]: I1210 22:49:52.539504 4791 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-10T22:49:50Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-10T22:49:50Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T22:49:52Z is after 2025-08-24T17:21:41Z" Dec 10 22:49:52 crc kubenswrapper[4791]: I1210 22:49:52.574670 4791 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-tchxk" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"c0ab74e0-42a9-4f42-ac80-16030cf4ffff\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T22:49:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T22:49:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T22:49:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T22:49:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://04e0fd1b53256b7a938e8ccab93e2a4eb73dd1a59e0b9799e09c55e511baf7fe\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T22:49:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zn2qt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-10T22:49:50Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-tchxk\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T22:49:52Z is after 2025-08-24T17:21:41Z" Dec 10 22:49:52 crc kubenswrapper[4791]: I1210 22:49:52.624368 4791 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"networking-console-plugin-cert\" (UniqueName: \"kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 10 22:49:52 crc kubenswrapper[4791]: I1210 22:49:52.624412 4791 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nginx-conf\" (UniqueName: \"kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 10 22:49:52 crc kubenswrapper[4791]: I1210 22:49:52.624436 4791 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cqllr\" (UniqueName: \"kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr\") pod \"network-check-target-xd92c\" (UID: \"3b6479f0-333b-4a96-9adf-2099afdc2447\") " pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 10 22:49:52 crc kubenswrapper[4791]: I1210 22:49:52.624454 4791 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-s2dwl\" (UniqueName: \"kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl\") pod \"network-check-source-55646444c4-trplf\" (UID: \"9d751cbb-f2e2-430d-9754-c882a5e924a5\") " pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 10 22:49:52 crc kubenswrapper[4791]: E1210 22:49:52.624549 4791 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Dec 10 22:49:52 crc kubenswrapper[4791]: E1210 22:49:52.624551 4791 configmap.go:193] Couldn't get configMap openshift-network-console/networking-console-plugin: object "openshift-network-console"/"networking-console-plugin" not registered Dec 10 22:49:52 crc kubenswrapper[4791]: E1210 22:49:52.624559 4791 secret.go:188] Couldn't get secret openshift-network-console/networking-console-plugin-cert: object "openshift-network-console"/"networking-console-plugin-cert" not registered Dec 10 22:49:52 crc kubenswrapper[4791]: E1210 22:49:52.624622 4791 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2025-12-10 22:49:54.624602852 +0000 UTC m=+29.054220475 (durationBeforeRetry 2s). Error: MountVolume.SetUp failed for volume "nginx-conf" (UniqueName: "kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin" not registered Dec 10 22:49:52 crc kubenswrapper[4791]: E1210 22:49:52.624562 4791 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Dec 10 22:49:52 crc kubenswrapper[4791]: E1210 22:49:52.624641 4791 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2025-12-10 22:49:54.624632033 +0000 UTC m=+29.054249656 (durationBeforeRetry 2s). Error: MountVolume.SetUp failed for volume "networking-console-plugin-cert" (UniqueName: "kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin-cert" not registered Dec 10 22:49:52 crc kubenswrapper[4791]: E1210 22:49:52.624647 4791 projected.go:194] Error preparing data for projected volume kube-api-access-s2dwl for pod openshift-network-diagnostics/network-check-source-55646444c4-trplf: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Dec 10 22:49:52 crc kubenswrapper[4791]: E1210 22:49:52.624693 4791 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl podName:9d751cbb-f2e2-430d-9754-c882a5e924a5 nodeName:}" failed. No retries permitted until 2025-12-10 22:49:54.624672524 +0000 UTC m=+29.054290237 (durationBeforeRetry 2s). Error: MountVolume.SetUp failed for volume "kube-api-access-s2dwl" (UniqueName: "kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl") pod "network-check-source-55646444c4-trplf" (UID: "9d751cbb-f2e2-430d-9754-c882a5e924a5") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Dec 10 22:49:52 crc kubenswrapper[4791]: E1210 22:49:52.624708 4791 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Dec 10 22:49:52 crc kubenswrapper[4791]: E1210 22:49:52.624738 4791 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Dec 10 22:49:52 crc kubenswrapper[4791]: E1210 22:49:52.624751 4791 projected.go:194] Error preparing data for projected volume kube-api-access-cqllr for pod openshift-network-diagnostics/network-check-target-xd92c: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Dec 10 22:49:52 crc kubenswrapper[4791]: E1210 22:49:52.624818 4791 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr podName:3b6479f0-333b-4a96-9adf-2099afdc2447 nodeName:}" failed. No retries permitted until 2025-12-10 22:49:54.624796237 +0000 UTC m=+29.054413840 (durationBeforeRetry 2s). Error: MountVolume.SetUp failed for volume "kube-api-access-cqllr" (UniqueName: "kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr") pod "network-check-target-xd92c" (UID: "3b6479f0-333b-4a96-9adf-2099afdc2447") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Dec 10 22:49:52 crc kubenswrapper[4791]: I1210 22:49:52.631995 4791 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-rw584" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ff9beb90-69b0-4732-bf37-0b81f58ecc98\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T22:49:51Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T22:49:51Z\\\",\\\"message\\\":\\\"containers with incomplete status: [egress-router-binary-copy cni-plugins bond-cni-plugin routeoverride-cni whereabouts-cni-bincopy whereabouts-cni]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T22:49:51Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T22:49:51Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fksqn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fksqn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fksqn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fksqn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fksqn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fksqn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fksqn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-10T22:49:51Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-rw584\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T22:49:52Z is after 2025-08-24T17:21:41Z" Dec 10 22:49:52 crc kubenswrapper[4791]: I1210 22:49:52.663387 4791 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f8349fea-c49b-49a1-b5ee-032c5dffe021\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T22:49:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T22:49:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T22:49:25Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T22:49:25Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T22:49:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://acafe12186df99d1059a31260de0eafce35ff7771bcb8f5c6753b72665d6f60b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T22:49:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ea4c346507862a8c12756a3061d6e25e788a8d79f4621b62f81f4322185c82a8\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T22:49:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://2e6e324ec6f4f7bb69fdfab4ea3ac1aa55e480dd8e2a8984b0fe5b4c9f66636e\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T22:49:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://1b2e11ff2bf8789fa1d0d23ae648b3e6ef653106e5bc4bbb5dd6447f7cd6f05d\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d6aa7f8c5d908d90008f2df49860ac8a3480136da567e3d85f86135d779204f9\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-12-10T22:49:44Z\\\",\\\"message\\\":\\\"W1210 22:49:33.845733 1 cmd.go:257] Using insecure, self-signed certificates\\\\nI1210 22:49:33.846108 1 crypto.go:601] Generating new CA for check-endpoints-signer@1765406973 cert, and key in /tmp/serving-cert-2585701456/serving-signer.crt, /tmp/serving-cert-2585701456/serving-signer.key\\\\nI1210 22:49:34.529748 1 observer_polling.go:159] Starting file observer\\\\nW1210 22:49:34.532225 1 builder.go:272] unable to get owner reference (falling back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": dial tcp [::1]:6443: connect: connection refused\\\\nI1210 22:49:34.532599 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1210 22:49:34.533977 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-2585701456/tls.crt::/tmp/serving-cert-2585701456/tls.key\\\\\\\"\\\\nF1210 22:49:44.774710 1 cmd.go:182] error initializing delegating authentication: unable to load configmap based request-header-client-ca-file: Get \\\\\\\"https://localhost:6443/api/v1/namespaces/kube-system/configmaps/extension-apiserver-authentication\\\\\\\": net/http: TLS handshake timeout\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-10T22:49:33Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T22:49:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://0c4d036ab269f62561fe7ef1586012cb3a16e3583a08b9488a0616ec00c8e0ae\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T22:49:32Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c5856d6b26c19fda13b5c9ac2892a7e2f9c6f79feb8173ecfd9b33b673327d67\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c5856d6b26c19fda13b5c9ac2892a7e2f9c6f79feb8173ecfd9b33b673327d67\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-10T22:49:27Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-10T22:49:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-10T22:49:25Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T22:49:52Z is after 2025-08-24T17:21:41Z" Dec 10 22:49:52 crc kubenswrapper[4791]: I1210 22:49:52.675697 4791 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-ovn-kubernetes/ovnkube-node-zhq64" Dec 10 22:49:52 crc kubenswrapper[4791]: W1210 22:49:52.689460 4791 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod3cd47739_0fa9_4321_aff1_220f8721a0b3.slice/crio-4768e24be0da8b9e4e556b30b09468d0b4d8bd68f0e468160e7300cc2828587d WatchSource:0}: Error finding container 4768e24be0da8b9e4e556b30b09468d0b4d8bd68f0e468160e7300cc2828587d: Status 404 returned error can't find the container with id 4768e24be0da8b9e4e556b30b09468d0b4d8bd68f0e468160e7300cc2828587d Dec 10 22:49:52 crc kubenswrapper[4791]: I1210 22:49:52.709177 4791 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-10T22:49:50Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-10T22:49:50Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T22:49:52Z is after 2025-08-24T17:21:41Z" Dec 10 22:49:52 crc kubenswrapper[4791]: I1210 22:49:52.737368 4791 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-10T22:49:50Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-10T22:49:50Z\\\",\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T22:49:52Z is after 2025-08-24T17:21:41Z" Dec 10 22:49:52 crc kubenswrapper[4791]: I1210 22:49:52.739478 4791 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-etcd/etcd-crc" Dec 10 22:49:52 crc kubenswrapper[4791]: I1210 22:49:52.752478 4791 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-etcd/etcd-crc" Dec 10 22:49:52 crc kubenswrapper[4791]: I1210 22:49:52.782430 4791 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-etcd/etcd-crc"] Dec 10 22:49:52 crc kubenswrapper[4791]: I1210 22:49:52.798524 4791 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f8349fea-c49b-49a1-b5ee-032c5dffe021\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T22:49:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T22:49:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T22:49:25Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T22:49:25Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T22:49:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://acafe12186df99d1059a31260de0eafce35ff7771bcb8f5c6753b72665d6f60b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T22:49:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ea4c346507862a8c12756a3061d6e25e788a8d79f4621b62f81f4322185c82a8\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T22:49:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://2e6e324ec6f4f7bb69fdfab4ea3ac1aa55e480dd8e2a8984b0fe5b4c9f66636e\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T22:49:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://1b2e11ff2bf8789fa1d0d23ae648b3e6ef653106e5bc4bbb5dd6447f7cd6f05d\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d6aa7f8c5d908d90008f2df49860ac8a3480136da567e3d85f86135d779204f9\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-12-10T22:49:44Z\\\",\\\"message\\\":\\\"W1210 22:49:33.845733 1 cmd.go:257] Using insecure, self-signed certificates\\\\nI1210 22:49:33.846108 1 crypto.go:601] Generating new CA for check-endpoints-signer@1765406973 cert, and key in /tmp/serving-cert-2585701456/serving-signer.crt, /tmp/serving-cert-2585701456/serving-signer.key\\\\nI1210 22:49:34.529748 1 observer_polling.go:159] Starting file observer\\\\nW1210 22:49:34.532225 1 builder.go:272] unable to get owner reference (falling back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": dial tcp [::1]:6443: connect: connection refused\\\\nI1210 22:49:34.532599 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1210 22:49:34.533977 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-2585701456/tls.crt::/tmp/serving-cert-2585701456/tls.key\\\\\\\"\\\\nF1210 22:49:44.774710 1 cmd.go:182] error initializing delegating authentication: unable to load configmap based request-header-client-ca-file: Get \\\\\\\"https://localhost:6443/api/v1/namespaces/kube-system/configmaps/extension-apiserver-authentication\\\\\\\": net/http: TLS handshake timeout\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-10T22:49:33Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T22:49:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://0c4d036ab269f62561fe7ef1586012cb3a16e3583a08b9488a0616ec00c8e0ae\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T22:49:32Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c5856d6b26c19fda13b5c9ac2892a7e2f9c6f79feb8173ecfd9b33b673327d67\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c5856d6b26c19fda13b5c9ac2892a7e2f9c6f79feb8173ecfd9b33b673327d67\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-10T22:49:27Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-10T22:49:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-10T22:49:25Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T22:49:52Z is after 2025-08-24T17:21:41Z" Dec 10 22:49:52 crc kubenswrapper[4791]: I1210 22:49:52.837766 4791 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-10T22:49:50Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-10T22:49:50Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T22:49:52Z is after 2025-08-24T17:21:41Z" Dec 10 22:49:52 crc kubenswrapper[4791]: I1210 22:49:52.875789 4791 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-10T22:49:50Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-10T22:49:50Z\\\",\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T22:49:52Z is after 2025-08-24T17:21:41Z" Dec 10 22:49:52 crc kubenswrapper[4791]: I1210 22:49:52.884013 4791 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 10 22:49:52 crc kubenswrapper[4791]: I1210 22:49:52.884094 4791 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 10 22:49:52 crc kubenswrapper[4791]: E1210 22:49:52.884146 4791 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 10 22:49:52 crc kubenswrapper[4791]: E1210 22:49:52.884248 4791 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 10 22:49:52 crc kubenswrapper[4791]: I1210 22:49:52.884013 4791 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 10 22:49:52 crc kubenswrapper[4791]: E1210 22:49:52.884387 4791 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 10 22:49:52 crc kubenswrapper[4791]: I1210 22:49:52.919719 4791 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-5rb5l" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ba35653c-6e06-4cee-a4d6-137764090d18\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T22:49:51Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T22:49:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T22:49:51Z\\\",\\\"message\\\":\\\"containers with unready status: [machine-config-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T22:49:51Z\\\",\\\"message\\\":\\\"containers with unready status: [machine-config-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fvvrs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fvvrs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-10T22:49:51Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-5rb5l\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T22:49:52Z is after 2025-08-24T17:21:41Z" Dec 10 22:49:52 crc kubenswrapper[4791]: I1210 22:49:52.957048 4791 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"698bbe9a-fbe5-436d-a835-5e80df77ba24\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T22:49:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T22:49:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T22:49:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T22:49:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T22:49:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://224d73713312376a58ec6cc2bc8fdb6ffa44c4749ae0e7691d5fe27b9afca09f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T22:49:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://372eeaffd40992c6f06efc4df6dfcb44b31de64ee8256c21508cdd7c74ecb96e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T22:49:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://76b6013b062cac5b0b4d12a94f38ce7b43f9c7ea216063455d88a956ac04d26e\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T22:49:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://78e950e122e761d950da1a3c60c06476e53920e8795f6c61db22c01e22d5136e\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T22:49:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-10T22:49:25Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T22:49:52Z is after 2025-08-24T17:21:41Z" Dec 10 22:49:52 crc kubenswrapper[4791]: I1210 22:49:52.997859 4791 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-10T22:49:52Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-10T22:49:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-10T22:49:52Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3734b110802828632d44263ed0a43cabc4b710833f0be120d72db0940fe783e2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T22:49:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://794803ac9be3c52509e6dcb10854d8864e1a9e406762e57c7971b07c4da1220e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T22:49:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T22:49:52Z is after 2025-08-24T17:21:41Z" Dec 10 22:49:53 crc kubenswrapper[4791]: I1210 22:49:53.045305 4791 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-vczjq" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b9f00f25-3fa1-4b63-9710-fd352224b01b\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T22:49:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T22:49:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T22:49:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T22:49:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4698cf4c6b55b41e628448d37c2eb847c2777d0efbd93afcbf367d5e78623b7e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T22:49:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lrrpv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-10T22:49:50Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-vczjq\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T22:49:53Z is after 2025-08-24T17:21:41Z" Dec 10 22:49:53 crc kubenswrapper[4791]: I1210 22:49:53.084456 4791 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-4nwkq" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"672aa28c-8169-49ed-87b8-21187d13a80c\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T22:49:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T22:49:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T22:49:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T22:49:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a2f2a62315529e148a10a33ccb241e9417fbde0fe42afb61cb7ee2438c74636f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T22:49:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xl62j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-10T22:49:51Z\\\"}}\" for pod \"openshift-multus\"/\"multus-4nwkq\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T22:49:53Z is after 2025-08-24T17:21:41Z" Dec 10 22:49:53 crc kubenswrapper[4791]: I1210 22:49:53.126464 4791 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-zhq64" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3cd47739-0fa9-4321-aff1-220f8721a0b3\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T22:49:52Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T22:49:52Z\\\",\\\"message\\\":\\\"containers with incomplete status: [kubecfg-setup]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T22:49:52Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T22:49:52Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wvpw8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wvpw8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wvpw8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wvpw8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wvpw8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wvpw8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wvpw8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wvpw8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wvpw8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-10T22:49:52Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-zhq64\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T22:49:53Z is after 2025-08-24T17:21:41Z" Dec 10 22:49:53 crc kubenswrapper[4791]: I1210 22:49:53.187851 4791 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-10T22:49:52Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-10T22:49:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-10T22:49:52Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://20651eb0c16e2c71222f56da1ff755cb0e7b084886acbbb24f0ccaf57041ed85\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T22:49:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T22:49:53Z is after 2025-08-24T17:21:41Z" Dec 10 22:49:53 crc kubenswrapper[4791]: I1210 22:49:53.214810 4791 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-10T22:49:50Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-10T22:49:50Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T22:49:53Z is after 2025-08-24T17:21:41Z" Dec 10 22:49:53 crc kubenswrapper[4791]: I1210 22:49:53.214925 4791 generic.go:334] "Generic (PLEG): container finished" podID="ff9beb90-69b0-4732-bf37-0b81f58ecc98" containerID="ce67e54b696bb53ffbbe48c6ce71476b635e0868993e9c427b22f904891433ee" exitCode=0 Dec 10 22:49:53 crc kubenswrapper[4791]: I1210 22:49:53.215024 4791 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-additional-cni-plugins-rw584" event={"ID":"ff9beb90-69b0-4732-bf37-0b81f58ecc98","Type":"ContainerDied","Data":"ce67e54b696bb53ffbbe48c6ce71476b635e0868993e9c427b22f904891433ee"} Dec 10 22:49:53 crc kubenswrapper[4791]: I1210 22:49:53.237115 4791 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-5rb5l" event={"ID":"ba35653c-6e06-4cee-a4d6-137764090d18","Type":"ContainerStarted","Data":"dbb5c875730a98ad2daf316cfafc08fcdce21ac16148154780ac9193ed405dab"} Dec 10 22:49:53 crc kubenswrapper[4791]: I1210 22:49:53.244236 4791 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-10T22:49:50Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-10T22:49:50Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T22:49:53Z is after 2025-08-24T17:21:41Z" Dec 10 22:49:53 crc kubenswrapper[4791]: I1210 22:49:53.255857 4791 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-zhq64" event={"ID":"3cd47739-0fa9-4321-aff1-220f8721a0b3","Type":"ContainerStarted","Data":"3afbd1ffcd2c4b40324aca6e52d6fc822db131c3f35a9af3a6f005c2cffdb7bf"} Dec 10 22:49:53 crc kubenswrapper[4791]: I1210 22:49:53.256213 4791 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-zhq64" event={"ID":"3cd47739-0fa9-4321-aff1-220f8721a0b3","Type":"ContainerStarted","Data":"4768e24be0da8b9e4e556b30b09468d0b4d8bd68f0e468160e7300cc2828587d"} Dec 10 22:49:53 crc kubenswrapper[4791]: I1210 22:49:53.279150 4791 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-tchxk" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"c0ab74e0-42a9-4f42-ac80-16030cf4ffff\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T22:49:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T22:49:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T22:49:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T22:49:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://04e0fd1b53256b7a938e8ccab93e2a4eb73dd1a59e0b9799e09c55e511baf7fe\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T22:49:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zn2qt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-10T22:49:50Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-tchxk\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T22:49:53Z is after 2025-08-24T17:21:41Z" Dec 10 22:49:53 crc kubenswrapper[4791]: I1210 22:49:53.320880 4791 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-rw584" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ff9beb90-69b0-4732-bf37-0b81f58ecc98\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T22:49:51Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T22:49:51Z\\\",\\\"message\\\":\\\"containers with incomplete status: [egress-router-binary-copy cni-plugins bond-cni-plugin routeoverride-cni whereabouts-cni-bincopy whereabouts-cni]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T22:49:51Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T22:49:51Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fksqn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fksqn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fksqn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fksqn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fksqn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fksqn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fksqn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-10T22:49:51Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-rw584\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T22:49:53Z is after 2025-08-24T17:21:41Z" Dec 10 22:49:53 crc kubenswrapper[4791]: I1210 22:49:53.359178 4791 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-10T22:49:52Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-10T22:49:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-10T22:49:52Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3734b110802828632d44263ed0a43cabc4b710833f0be120d72db0940fe783e2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T22:49:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://794803ac9be3c52509e6dcb10854d8864e1a9e406762e57c7971b07c4da1220e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T22:49:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T22:49:53Z is after 2025-08-24T17:21:41Z" Dec 10 22:49:53 crc kubenswrapper[4791]: I1210 22:49:53.397427 4791 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-vczjq" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b9f00f25-3fa1-4b63-9710-fd352224b01b\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T22:49:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T22:49:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T22:49:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T22:49:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4698cf4c6b55b41e628448d37c2eb847c2777d0efbd93afcbf367d5e78623b7e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T22:49:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lrrpv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-10T22:49:50Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-vczjq\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T22:49:53Z is after 2025-08-24T17:21:41Z" Dec 10 22:49:53 crc kubenswrapper[4791]: I1210 22:49:53.439818 4791 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-4nwkq" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"672aa28c-8169-49ed-87b8-21187d13a80c\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T22:49:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T22:49:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T22:49:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T22:49:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a2f2a62315529e148a10a33ccb241e9417fbde0fe42afb61cb7ee2438c74636f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T22:49:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xl62j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-10T22:49:51Z\\\"}}\" for pod \"openshift-multus\"/\"multus-4nwkq\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T22:49:53Z is after 2025-08-24T17:21:41Z" Dec 10 22:49:53 crc kubenswrapper[4791]: I1210 22:49:53.482599 4791 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-zhq64" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3cd47739-0fa9-4321-aff1-220f8721a0b3\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T22:49:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T22:49:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T22:49:52Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T22:49:52Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wvpw8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wvpw8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wvpw8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wvpw8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wvpw8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wvpw8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wvpw8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wvpw8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3afbd1ffcd2c4b40324aca6e52d6fc822db131c3f35a9af3a6f005c2cffdb7bf\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3afbd1ffcd2c4b40324aca6e52d6fc822db131c3f35a9af3a6f005c2cffdb7bf\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-10T22:49:53Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-10T22:49:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wvpw8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-10T22:49:52Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-zhq64\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T22:49:53Z is after 2025-08-24T17:21:41Z" Dec 10 22:49:53 crc kubenswrapper[4791]: I1210 22:49:53.518457 4791 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"698bbe9a-fbe5-436d-a835-5e80df77ba24\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T22:49:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T22:49:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T22:49:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T22:49:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T22:49:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://224d73713312376a58ec6cc2bc8fdb6ffa44c4749ae0e7691d5fe27b9afca09f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T22:49:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://372eeaffd40992c6f06efc4df6dfcb44b31de64ee8256c21508cdd7c74ecb96e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T22:49:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://76b6013b062cac5b0b4d12a94f38ce7b43f9c7ea216063455d88a956ac04d26e\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T22:49:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://78e950e122e761d950da1a3c60c06476e53920e8795f6c61db22c01e22d5136e\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T22:49:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-10T22:49:25Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T22:49:53Z is after 2025-08-24T17:21:41Z" Dec 10 22:49:53 crc kubenswrapper[4791]: I1210 22:49:53.558417 4791 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-10T22:49:50Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-10T22:49:50Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T22:49:53Z is after 2025-08-24T17:21:41Z" Dec 10 22:49:53 crc kubenswrapper[4791]: I1210 22:49:53.598524 4791 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-10T22:49:50Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-10T22:49:50Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T22:49:53Z is after 2025-08-24T17:21:41Z" Dec 10 22:49:53 crc kubenswrapper[4791]: I1210 22:49:53.635754 4791 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-tchxk" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"c0ab74e0-42a9-4f42-ac80-16030cf4ffff\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T22:49:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T22:49:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T22:49:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T22:49:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://04e0fd1b53256b7a938e8ccab93e2a4eb73dd1a59e0b9799e09c55e511baf7fe\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T22:49:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zn2qt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-10T22:49:50Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-tchxk\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T22:49:53Z is after 2025-08-24T17:21:41Z" Dec 10 22:49:53 crc kubenswrapper[4791]: I1210 22:49:53.679372 4791 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-rw584" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ff9beb90-69b0-4732-bf37-0b81f58ecc98\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T22:49:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T22:49:51Z\\\",\\\"message\\\":\\\"containers with incomplete status: [cni-plugins bond-cni-plugin routeoverride-cni whereabouts-cni-bincopy whereabouts-cni]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T22:49:51Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T22:49:51Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fksqn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ce67e54b696bb53ffbbe48c6ce71476b635e0868993e9c427b22f904891433ee\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ce67e54b696bb53ffbbe48c6ce71476b635e0868993e9c427b22f904891433ee\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-10T22:49:52Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-10T22:49:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fksqn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fksqn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fksqn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fksqn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fksqn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fksqn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-10T22:49:51Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-rw584\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T22:49:53Z is after 2025-08-24T17:21:41Z" Dec 10 22:49:53 crc kubenswrapper[4791]: I1210 22:49:53.717727 4791 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-10T22:49:52Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-10T22:49:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-10T22:49:52Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://20651eb0c16e2c71222f56da1ff755cb0e7b084886acbbb24f0ccaf57041ed85\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T22:49:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T22:49:53Z is after 2025-08-24T17:21:41Z" Dec 10 22:49:53 crc kubenswrapper[4791]: I1210 22:49:53.764243 4791 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"90250e13-6c63-443a-a5e5-6e4e341a2289\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T22:49:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T22:49:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T22:49:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T22:49:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T22:49:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ba54f95b1e00843b854db9a10fe6897592159ccacf30ed0f76002093a13eedd3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T22:49:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5aa8ffa518f3a68b60b275f8df4851b2564aa4984b61efdc9470b73072133e43\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T22:49:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://16f989503edc78d9451638e2cf6cf4b19a2cf2b8809b10b237541cb829303091\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T22:49:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://82556824d63ce6e46df14c45689ff1f81ebf729abe902061815c040b323898e9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T22:49:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://61db531940fa6f2effdae5a005b4c91bd2698bd5d1bff9588c171a65ad550eca\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T22:49:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0f65c864d885319ce2a08d61c30be24b3ffb7d23ea30c82ed7faf039cb2c006d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://0f65c864d885319ce2a08d61c30be24b3ffb7d23ea30c82ed7faf039cb2c006d\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-10T22:49:27Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-10T22:49:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e31a76d1d5b2812ba83e24e38268e25ec8bed98fc3da9adcd06bda5822b6ace1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e31a76d1d5b2812ba83e24e38268e25ec8bed98fc3da9adcd06bda5822b6ace1\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-10T22:49:30Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-10T22:49:29Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://b5b5eceae562f3986b543d41de1e977efb85a91d84c98e09159ca6d092eda691\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b5b5eceae562f3986b543d41de1e977efb85a91d84c98e09159ca6d092eda691\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-10T22:49:31Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-10T22:49:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-10T22:49:25Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T22:49:53Z is after 2025-08-24T17:21:41Z" Dec 10 22:49:53 crc kubenswrapper[4791]: I1210 22:49:53.802944 4791 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-10T22:49:50Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-10T22:49:50Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T22:49:53Z is after 2025-08-24T17:21:41Z" Dec 10 22:49:53 crc kubenswrapper[4791]: I1210 22:49:53.839028 4791 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-10T22:49:50Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-10T22:49:50Z\\\",\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T22:49:53Z is after 2025-08-24T17:21:41Z" Dec 10 22:49:53 crc kubenswrapper[4791]: I1210 22:49:53.880103 4791 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f8349fea-c49b-49a1-b5ee-032c5dffe021\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T22:49:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T22:49:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T22:49:25Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T22:49:25Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T22:49:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://acafe12186df99d1059a31260de0eafce35ff7771bcb8f5c6753b72665d6f60b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T22:49:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ea4c346507862a8c12756a3061d6e25e788a8d79f4621b62f81f4322185c82a8\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T22:49:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://2e6e324ec6f4f7bb69fdfab4ea3ac1aa55e480dd8e2a8984b0fe5b4c9f66636e\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T22:49:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://1b2e11ff2bf8789fa1d0d23ae648b3e6ef653106e5bc4bbb5dd6447f7cd6f05d\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d6aa7f8c5d908d90008f2df49860ac8a3480136da567e3d85f86135d779204f9\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-12-10T22:49:44Z\\\",\\\"message\\\":\\\"W1210 22:49:33.845733 1 cmd.go:257] Using insecure, self-signed certificates\\\\nI1210 22:49:33.846108 1 crypto.go:601] Generating new CA for check-endpoints-signer@1765406973 cert, and key in /tmp/serving-cert-2585701456/serving-signer.crt, /tmp/serving-cert-2585701456/serving-signer.key\\\\nI1210 22:49:34.529748 1 observer_polling.go:159] Starting file observer\\\\nW1210 22:49:34.532225 1 builder.go:272] unable to get owner reference (falling back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": dial tcp [::1]:6443: connect: connection refused\\\\nI1210 22:49:34.532599 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1210 22:49:34.533977 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-2585701456/tls.crt::/tmp/serving-cert-2585701456/tls.key\\\\\\\"\\\\nF1210 22:49:44.774710 1 cmd.go:182] error initializing delegating authentication: unable to load configmap based request-header-client-ca-file: Get \\\\\\\"https://localhost:6443/api/v1/namespaces/kube-system/configmaps/extension-apiserver-authentication\\\\\\\": net/http: TLS handshake timeout\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-10T22:49:33Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T22:49:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://0c4d036ab269f62561fe7ef1586012cb3a16e3583a08b9488a0616ec00c8e0ae\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T22:49:32Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c5856d6b26c19fda13b5c9ac2892a7e2f9c6f79feb8173ecfd9b33b673327d67\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c5856d6b26c19fda13b5c9ac2892a7e2f9c6f79feb8173ecfd9b33b673327d67\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-10T22:49:27Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-10T22:49:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-10T22:49:25Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T22:49:53Z is after 2025-08-24T17:21:41Z" Dec 10 22:49:53 crc kubenswrapper[4791]: I1210 22:49:53.915574 4791 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-5rb5l" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ba35653c-6e06-4cee-a4d6-137764090d18\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T22:49:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T22:49:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T22:49:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T22:49:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://dbb5c875730a98ad2daf316cfafc08fcdce21ac16148154780ac9193ed405dab\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T22:49:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fvvrs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://185eb2ff44221a514878283ec873773a6c9a9073dc64e296f982e60dee1338e4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T22:49:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fvvrs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-10T22:49:51Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-5rb5l\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T22:49:53Z is after 2025-08-24T17:21:41Z" Dec 10 22:49:54 crc kubenswrapper[4791]: I1210 22:49:54.261228 4791 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" event={"ID":"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49","Type":"ContainerStarted","Data":"62bfaf5d7a6d1b452cdaf88086447d0c8cb7306e94434e46752be5475124d56a"} Dec 10 22:49:54 crc kubenswrapper[4791]: I1210 22:49:54.266516 4791 generic.go:334] "Generic (PLEG): container finished" podID="3cd47739-0fa9-4321-aff1-220f8721a0b3" containerID="3afbd1ffcd2c4b40324aca6e52d6fc822db131c3f35a9af3a6f005c2cffdb7bf" exitCode=0 Dec 10 22:49:54 crc kubenswrapper[4791]: I1210 22:49:54.266607 4791 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-zhq64" event={"ID":"3cd47739-0fa9-4321-aff1-220f8721a0b3","Type":"ContainerDied","Data":"3afbd1ffcd2c4b40324aca6e52d6fc822db131c3f35a9af3a6f005c2cffdb7bf"} Dec 10 22:49:54 crc kubenswrapper[4791]: I1210 22:49:54.266697 4791 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-zhq64" event={"ID":"3cd47739-0fa9-4321-aff1-220f8721a0b3","Type":"ContainerStarted","Data":"152a7618be8f38fadf4e564b3948980503cedcdb466c37344ccc20f4a3173a5f"} Dec 10 22:49:54 crc kubenswrapper[4791]: I1210 22:49:54.266734 4791 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-zhq64" event={"ID":"3cd47739-0fa9-4321-aff1-220f8721a0b3","Type":"ContainerStarted","Data":"3d18f9cb7207efb1e32de75a30d33e024f656a08f0503db95e5488bc5280971f"} Dec 10 22:49:54 crc kubenswrapper[4791]: I1210 22:49:54.266758 4791 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-zhq64" event={"ID":"3cd47739-0fa9-4321-aff1-220f8721a0b3","Type":"ContainerStarted","Data":"bc721b530c08b433756e978ec0423cc18aabf6771b61eb96f903b4d57ea335ff"} Dec 10 22:49:54 crc kubenswrapper[4791]: I1210 22:49:54.266778 4791 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-zhq64" event={"ID":"3cd47739-0fa9-4321-aff1-220f8721a0b3","Type":"ContainerStarted","Data":"68795a29944e56870ffe94fa732e9dcf0bfb344c0d85d1e590f58a876129c61f"} Dec 10 22:49:54 crc kubenswrapper[4791]: I1210 22:49:54.266795 4791 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-zhq64" event={"ID":"3cd47739-0fa9-4321-aff1-220f8721a0b3","Type":"ContainerStarted","Data":"937137ae9d7480cdc7c59bc4a1df0a4c10e72b3b63693af285278789953ed218"} Dec 10 22:49:54 crc kubenswrapper[4791]: I1210 22:49:54.266811 4791 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-zhq64" event={"ID":"3cd47739-0fa9-4321-aff1-220f8721a0b3","Type":"ContainerStarted","Data":"1a52d19b7671e8b143017c8a981a6f8d3ee3b84650e4b223b814a531ac76f364"} Dec 10 22:49:54 crc kubenswrapper[4791]: I1210 22:49:54.269878 4791 generic.go:334] "Generic (PLEG): container finished" podID="ff9beb90-69b0-4732-bf37-0b81f58ecc98" containerID="818d89bfffb8a110670152c8480684665ea8db578e9768f524fbbe6a92f27c22" exitCode=0 Dec 10 22:49:54 crc kubenswrapper[4791]: I1210 22:49:54.270028 4791 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-additional-cni-plugins-rw584" event={"ID":"ff9beb90-69b0-4732-bf37-0b81f58ecc98","Type":"ContainerDied","Data":"818d89bfffb8a110670152c8480684665ea8db578e9768f524fbbe6a92f27c22"} Dec 10 22:49:54 crc kubenswrapper[4791]: I1210 22:49:54.281598 4791 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"698bbe9a-fbe5-436d-a835-5e80df77ba24\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T22:49:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T22:49:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T22:49:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T22:49:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T22:49:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://224d73713312376a58ec6cc2bc8fdb6ffa44c4749ae0e7691d5fe27b9afca09f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T22:49:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://372eeaffd40992c6f06efc4df6dfcb44b31de64ee8256c21508cdd7c74ecb96e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T22:49:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://76b6013b062cac5b0b4d12a94f38ce7b43f9c7ea216063455d88a956ac04d26e\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T22:49:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://78e950e122e761d950da1a3c60c06476e53920e8795f6c61db22c01e22d5136e\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T22:49:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-10T22:49:25Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T22:49:54Z is after 2025-08-24T17:21:41Z" Dec 10 22:49:54 crc kubenswrapper[4791]: I1210 22:49:54.303076 4791 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-10T22:49:52Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-10T22:49:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-10T22:49:52Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3734b110802828632d44263ed0a43cabc4b710833f0be120d72db0940fe783e2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T22:49:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://794803ac9be3c52509e6dcb10854d8864e1a9e406762e57c7971b07c4da1220e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T22:49:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T22:49:54Z is after 2025-08-24T17:21:41Z" Dec 10 22:49:54 crc kubenswrapper[4791]: I1210 22:49:54.314317 4791 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-vczjq" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b9f00f25-3fa1-4b63-9710-fd352224b01b\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T22:49:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T22:49:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T22:49:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T22:49:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4698cf4c6b55b41e628448d37c2eb847c2777d0efbd93afcbf367d5e78623b7e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T22:49:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lrrpv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-10T22:49:50Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-vczjq\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T22:49:54Z is after 2025-08-24T17:21:41Z" Dec 10 22:49:54 crc kubenswrapper[4791]: I1210 22:49:54.329629 4791 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-4nwkq" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"672aa28c-8169-49ed-87b8-21187d13a80c\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T22:49:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T22:49:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T22:49:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T22:49:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a2f2a62315529e148a10a33ccb241e9417fbde0fe42afb61cb7ee2438c74636f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T22:49:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xl62j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-10T22:49:51Z\\\"}}\" for pod \"openshift-multus\"/\"multus-4nwkq\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T22:49:54Z is after 2025-08-24T17:21:41Z" Dec 10 22:49:54 crc kubenswrapper[4791]: I1210 22:49:54.350541 4791 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-zhq64" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3cd47739-0fa9-4321-aff1-220f8721a0b3\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T22:49:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T22:49:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T22:49:52Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T22:49:52Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wvpw8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wvpw8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wvpw8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wvpw8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wvpw8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wvpw8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wvpw8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wvpw8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3afbd1ffcd2c4b40324aca6e52d6fc822db131c3f35a9af3a6f005c2cffdb7bf\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3afbd1ffcd2c4b40324aca6e52d6fc822db131c3f35a9af3a6f005c2cffdb7bf\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-10T22:49:53Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-10T22:49:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wvpw8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-10T22:49:52Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-zhq64\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T22:49:54Z is after 2025-08-24T17:21:41Z" Dec 10 22:49:54 crc kubenswrapper[4791]: I1210 22:49:54.361521 4791 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-10T22:49:52Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-10T22:49:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-10T22:49:52Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://20651eb0c16e2c71222f56da1ff755cb0e7b084886acbbb24f0ccaf57041ed85\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T22:49:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T22:49:54Z is after 2025-08-24T17:21:41Z" Dec 10 22:49:54 crc kubenswrapper[4791]: I1210 22:49:54.372532 4791 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-10T22:49:50Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-10T22:49:50Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T22:49:54Z is after 2025-08-24T17:21:41Z" Dec 10 22:49:54 crc kubenswrapper[4791]: I1210 22:49:54.387241 4791 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-10T22:49:50Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-10T22:49:50Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T22:49:54Z is after 2025-08-24T17:21:41Z" Dec 10 22:49:54 crc kubenswrapper[4791]: I1210 22:49:54.396753 4791 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-tchxk" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"c0ab74e0-42a9-4f42-ac80-16030cf4ffff\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T22:49:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T22:49:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T22:49:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T22:49:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://04e0fd1b53256b7a938e8ccab93e2a4eb73dd1a59e0b9799e09c55e511baf7fe\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T22:49:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zn2qt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-10T22:49:50Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-tchxk\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T22:49:54Z is after 2025-08-24T17:21:41Z" Dec 10 22:49:54 crc kubenswrapper[4791]: I1210 22:49:54.410083 4791 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-rw584" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ff9beb90-69b0-4732-bf37-0b81f58ecc98\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T22:49:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T22:49:51Z\\\",\\\"message\\\":\\\"containers with incomplete status: [cni-plugins bond-cni-plugin routeoverride-cni whereabouts-cni-bincopy whereabouts-cni]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T22:49:51Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T22:49:51Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fksqn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ce67e54b696bb53ffbbe48c6ce71476b635e0868993e9c427b22f904891433ee\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ce67e54b696bb53ffbbe48c6ce71476b635e0868993e9c427b22f904891433ee\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-10T22:49:52Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-10T22:49:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fksqn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fksqn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fksqn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fksqn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fksqn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fksqn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-10T22:49:51Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-rw584\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T22:49:54Z is after 2025-08-24T17:21:41Z" Dec 10 22:49:54 crc kubenswrapper[4791]: I1210 22:49:54.425483 4791 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f8349fea-c49b-49a1-b5ee-032c5dffe021\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T22:49:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T22:49:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T22:49:25Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T22:49:25Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T22:49:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://acafe12186df99d1059a31260de0eafce35ff7771bcb8f5c6753b72665d6f60b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T22:49:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ea4c346507862a8c12756a3061d6e25e788a8d79f4621b62f81f4322185c82a8\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T22:49:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://2e6e324ec6f4f7bb69fdfab4ea3ac1aa55e480dd8e2a8984b0fe5b4c9f66636e\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T22:49:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://1b2e11ff2bf8789fa1d0d23ae648b3e6ef653106e5bc4bbb5dd6447f7cd6f05d\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d6aa7f8c5d908d90008f2df49860ac8a3480136da567e3d85f86135d779204f9\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-12-10T22:49:44Z\\\",\\\"message\\\":\\\"W1210 22:49:33.845733 1 cmd.go:257] Using insecure, self-signed certificates\\\\nI1210 22:49:33.846108 1 crypto.go:601] Generating new CA for check-endpoints-signer@1765406973 cert, and key in /tmp/serving-cert-2585701456/serving-signer.crt, /tmp/serving-cert-2585701456/serving-signer.key\\\\nI1210 22:49:34.529748 1 observer_polling.go:159] Starting file observer\\\\nW1210 22:49:34.532225 1 builder.go:272] unable to get owner reference (falling back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": dial tcp [::1]:6443: connect: connection refused\\\\nI1210 22:49:34.532599 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1210 22:49:34.533977 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-2585701456/tls.crt::/tmp/serving-cert-2585701456/tls.key\\\\\\\"\\\\nF1210 22:49:44.774710 1 cmd.go:182] error initializing delegating authentication: unable to load configmap based request-header-client-ca-file: Get \\\\\\\"https://localhost:6443/api/v1/namespaces/kube-system/configmaps/extension-apiserver-authentication\\\\\\\": net/http: TLS handshake timeout\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-10T22:49:33Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T22:49:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://0c4d036ab269f62561fe7ef1586012cb3a16e3583a08b9488a0616ec00c8e0ae\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T22:49:32Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c5856d6b26c19fda13b5c9ac2892a7e2f9c6f79feb8173ecfd9b33b673327d67\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c5856d6b26c19fda13b5c9ac2892a7e2f9c6f79feb8173ecfd9b33b673327d67\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-10T22:49:27Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-10T22:49:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-10T22:49:25Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T22:49:54Z is after 2025-08-24T17:21:41Z" Dec 10 22:49:54 crc kubenswrapper[4791]: I1210 22:49:54.443696 4791 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"90250e13-6c63-443a-a5e5-6e4e341a2289\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T22:49:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T22:49:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T22:49:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T22:49:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T22:49:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ba54f95b1e00843b854db9a10fe6897592159ccacf30ed0f76002093a13eedd3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T22:49:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5aa8ffa518f3a68b60b275f8df4851b2564aa4984b61efdc9470b73072133e43\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T22:49:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://16f989503edc78d9451638e2cf6cf4b19a2cf2b8809b10b237541cb829303091\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T22:49:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://82556824d63ce6e46df14c45689ff1f81ebf729abe902061815c040b323898e9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T22:49:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://61db531940fa6f2effdae5a005b4c91bd2698bd5d1bff9588c171a65ad550eca\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T22:49:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0f65c864d885319ce2a08d61c30be24b3ffb7d23ea30c82ed7faf039cb2c006d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://0f65c864d885319ce2a08d61c30be24b3ffb7d23ea30c82ed7faf039cb2c006d\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-10T22:49:27Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-10T22:49:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e31a76d1d5b2812ba83e24e38268e25ec8bed98fc3da9adcd06bda5822b6ace1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e31a76d1d5b2812ba83e24e38268e25ec8bed98fc3da9adcd06bda5822b6ace1\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-10T22:49:30Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-10T22:49:29Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://b5b5eceae562f3986b543d41de1e977efb85a91d84c98e09159ca6d092eda691\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b5b5eceae562f3986b543d41de1e977efb85a91d84c98e09159ca6d092eda691\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-10T22:49:31Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-10T22:49:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-10T22:49:25Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T22:49:54Z is after 2025-08-24T17:21:41Z" Dec 10 22:49:54 crc kubenswrapper[4791]: I1210 22:49:54.456271 4791 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-10T22:49:50Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-10T22:49:50Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T22:49:54Z is after 2025-08-24T17:21:41Z" Dec 10 22:49:54 crc kubenswrapper[4791]: I1210 22:49:54.477071 4791 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-10T22:49:54Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-10T22:49:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-10T22:49:54Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://62bfaf5d7a6d1b452cdaf88086447d0c8cb7306e94434e46752be5475124d56a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T22:49:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T22:49:54Z is after 2025-08-24T17:21:41Z" Dec 10 22:49:54 crc kubenswrapper[4791]: I1210 22:49:54.516971 4791 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-5rb5l" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ba35653c-6e06-4cee-a4d6-137764090d18\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T22:49:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T22:49:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T22:49:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T22:49:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://dbb5c875730a98ad2daf316cfafc08fcdce21ac16148154780ac9193ed405dab\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T22:49:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fvvrs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://185eb2ff44221a514878283ec873773a6c9a9073dc64e296f982e60dee1338e4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T22:49:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fvvrs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-10T22:49:51Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-5rb5l\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T22:49:54Z is after 2025-08-24T17:21:41Z" Dec 10 22:49:54 crc kubenswrapper[4791]: I1210 22:49:54.544370 4791 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 10 22:49:54 crc kubenswrapper[4791]: E1210 22:49:54.544585 4791 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-10 22:49:58.544567409 +0000 UTC m=+32.974185032 (durationBeforeRetry 4s). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 10 22:49:54 crc kubenswrapper[4791]: I1210 22:49:54.556404 4791 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-vczjq" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b9f00f25-3fa1-4b63-9710-fd352224b01b\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T22:49:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T22:49:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T22:49:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T22:49:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4698cf4c6b55b41e628448d37c2eb847c2777d0efbd93afcbf367d5e78623b7e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T22:49:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lrrpv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-10T22:49:50Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-vczjq\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T22:49:54Z is after 2025-08-24T17:21:41Z" Dec 10 22:49:54 crc kubenswrapper[4791]: I1210 22:49:54.603704 4791 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-4nwkq" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"672aa28c-8169-49ed-87b8-21187d13a80c\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T22:49:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T22:49:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T22:49:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T22:49:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a2f2a62315529e148a10a33ccb241e9417fbde0fe42afb61cb7ee2438c74636f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T22:49:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xl62j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-10T22:49:51Z\\\"}}\" for pod \"openshift-multus\"/\"multus-4nwkq\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T22:49:54Z is after 2025-08-24T17:21:41Z" Dec 10 22:49:54 crc kubenswrapper[4791]: I1210 22:49:54.645477 4791 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-s2dwl\" (UniqueName: \"kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl\") pod \"network-check-source-55646444c4-trplf\" (UID: \"9d751cbb-f2e2-430d-9754-c882a5e924a5\") " pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 10 22:49:54 crc kubenswrapper[4791]: I1210 22:49:54.645552 4791 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"networking-console-plugin-cert\" (UniqueName: \"kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 10 22:49:54 crc kubenswrapper[4791]: I1210 22:49:54.645599 4791 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nginx-conf\" (UniqueName: \"kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 10 22:49:54 crc kubenswrapper[4791]: I1210 22:49:54.645638 4791 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cqllr\" (UniqueName: \"kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr\") pod \"network-check-target-xd92c\" (UID: \"3b6479f0-333b-4a96-9adf-2099afdc2447\") " pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 10 22:49:54 crc kubenswrapper[4791]: E1210 22:49:54.645682 4791 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Dec 10 22:49:54 crc kubenswrapper[4791]: E1210 22:49:54.645717 4791 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Dec 10 22:49:54 crc kubenswrapper[4791]: E1210 22:49:54.645728 4791 secret.go:188] Couldn't get secret openshift-network-console/networking-console-plugin-cert: object "openshift-network-console"/"networking-console-plugin-cert" not registered Dec 10 22:49:54 crc kubenswrapper[4791]: E1210 22:49:54.645781 4791 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Dec 10 22:49:54 crc kubenswrapper[4791]: E1210 22:49:54.645805 4791 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Dec 10 22:49:54 crc kubenswrapper[4791]: E1210 22:49:54.645820 4791 projected.go:194] Error preparing data for projected volume kube-api-access-cqllr for pod openshift-network-diagnostics/network-check-target-xd92c: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Dec 10 22:49:54 crc kubenswrapper[4791]: E1210 22:49:54.645820 4791 configmap.go:193] Couldn't get configMap openshift-network-console/networking-console-plugin: object "openshift-network-console"/"networking-console-plugin" not registered Dec 10 22:49:54 crc kubenswrapper[4791]: E1210 22:49:54.645736 4791 projected.go:194] Error preparing data for projected volume kube-api-access-s2dwl for pod openshift-network-diagnostics/network-check-source-55646444c4-trplf: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Dec 10 22:49:54 crc kubenswrapper[4791]: E1210 22:49:54.645846 4791 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2025-12-10 22:49:58.645826257 +0000 UTC m=+33.075443880 (durationBeforeRetry 4s). Error: MountVolume.SetUp failed for volume "networking-console-plugin-cert" (UniqueName: "kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin-cert" not registered Dec 10 22:49:54 crc kubenswrapper[4791]: E1210 22:49:54.645937 4791 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl podName:9d751cbb-f2e2-430d-9754-c882a5e924a5 nodeName:}" failed. No retries permitted until 2025-12-10 22:49:58.64591378 +0000 UTC m=+33.075531443 (durationBeforeRetry 4s). Error: MountVolume.SetUp failed for volume "kube-api-access-s2dwl" (UniqueName: "kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl") pod "network-check-source-55646444c4-trplf" (UID: "9d751cbb-f2e2-430d-9754-c882a5e924a5") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Dec 10 22:49:54 crc kubenswrapper[4791]: E1210 22:49:54.645963 4791 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr podName:3b6479f0-333b-4a96-9adf-2099afdc2447 nodeName:}" failed. No retries permitted until 2025-12-10 22:49:58.645952031 +0000 UTC m=+33.075569744 (durationBeforeRetry 4s). Error: MountVolume.SetUp failed for volume "kube-api-access-cqllr" (UniqueName: "kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr") pod "network-check-target-xd92c" (UID: "3b6479f0-333b-4a96-9adf-2099afdc2447") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Dec 10 22:49:54 crc kubenswrapper[4791]: E1210 22:49:54.645986 4791 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2025-12-10 22:49:58.645975391 +0000 UTC m=+33.075593104 (durationBeforeRetry 4s). Error: MountVolume.SetUp failed for volume "nginx-conf" (UniqueName: "kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin" not registered Dec 10 22:49:54 crc kubenswrapper[4791]: I1210 22:49:54.650831 4791 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-zhq64" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3cd47739-0fa9-4321-aff1-220f8721a0b3\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T22:49:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T22:49:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T22:49:52Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T22:49:52Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wvpw8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wvpw8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wvpw8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wvpw8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wvpw8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wvpw8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wvpw8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wvpw8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3afbd1ffcd2c4b40324aca6e52d6fc822db131c3f35a9af3a6f005c2cffdb7bf\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3afbd1ffcd2c4b40324aca6e52d6fc822db131c3f35a9af3a6f005c2cffdb7bf\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-10T22:49:53Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-10T22:49:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wvpw8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-10T22:49:52Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-zhq64\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T22:49:54Z is after 2025-08-24T17:21:41Z" Dec 10 22:49:54 crc kubenswrapper[4791]: I1210 22:49:54.681847 4791 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"698bbe9a-fbe5-436d-a835-5e80df77ba24\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T22:49:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T22:49:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T22:49:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T22:49:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T22:49:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://224d73713312376a58ec6cc2bc8fdb6ffa44c4749ae0e7691d5fe27b9afca09f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T22:49:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://372eeaffd40992c6f06efc4df6dfcb44b31de64ee8256c21508cdd7c74ecb96e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T22:49:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://76b6013b062cac5b0b4d12a94f38ce7b43f9c7ea216063455d88a956ac04d26e\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T22:49:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://78e950e122e761d950da1a3c60c06476e53920e8795f6c61db22c01e22d5136e\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T22:49:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-10T22:49:25Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T22:49:54Z is after 2025-08-24T17:21:41Z" Dec 10 22:49:54 crc kubenswrapper[4791]: I1210 22:49:54.719636 4791 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-10T22:49:52Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-10T22:49:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-10T22:49:52Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3734b110802828632d44263ed0a43cabc4b710833f0be120d72db0940fe783e2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T22:49:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://794803ac9be3c52509e6dcb10854d8864e1a9e406762e57c7971b07c4da1220e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T22:49:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T22:49:54Z is after 2025-08-24T17:21:41Z" Dec 10 22:49:54 crc kubenswrapper[4791]: I1210 22:49:54.756973 4791 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-tchxk" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"c0ab74e0-42a9-4f42-ac80-16030cf4ffff\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T22:49:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T22:49:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T22:49:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T22:49:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://04e0fd1b53256b7a938e8ccab93e2a4eb73dd1a59e0b9799e09c55e511baf7fe\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T22:49:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zn2qt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-10T22:49:50Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-tchxk\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T22:49:54Z is after 2025-08-24T17:21:41Z" Dec 10 22:49:54 crc kubenswrapper[4791]: I1210 22:49:54.801705 4791 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-rw584" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ff9beb90-69b0-4732-bf37-0b81f58ecc98\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T22:49:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T22:49:51Z\\\",\\\"message\\\":\\\"containers with incomplete status: [bond-cni-plugin routeoverride-cni whereabouts-cni-bincopy whereabouts-cni]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T22:49:51Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T22:49:51Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fksqn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ce67e54b696bb53ffbbe48c6ce71476b635e0868993e9c427b22f904891433ee\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ce67e54b696bb53ffbbe48c6ce71476b635e0868993e9c427b22f904891433ee\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-10T22:49:52Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-10T22:49:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fksqn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://818d89bfffb8a110670152c8480684665ea8db578e9768f524fbbe6a92f27c22\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://818d89bfffb8a110670152c8480684665ea8db578e9768f524fbbe6a92f27c22\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-10T22:49:53Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-10T22:49:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fksqn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fksqn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fksqn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fksqn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fksqn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-10T22:49:51Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-rw584\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T22:49:54Z is after 2025-08-24T17:21:41Z" Dec 10 22:49:54 crc kubenswrapper[4791]: I1210 22:49:54.844383 4791 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-10T22:49:52Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-10T22:49:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-10T22:49:52Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://20651eb0c16e2c71222f56da1ff755cb0e7b084886acbbb24f0ccaf57041ed85\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T22:49:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T22:49:54Z is after 2025-08-24T17:21:41Z" Dec 10 22:49:54 crc kubenswrapper[4791]: I1210 22:49:54.877969 4791 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-10T22:49:50Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-10T22:49:50Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T22:49:54Z is after 2025-08-24T17:21:41Z" Dec 10 22:49:54 crc kubenswrapper[4791]: I1210 22:49:54.884232 4791 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 10 22:49:54 crc kubenswrapper[4791]: I1210 22:49:54.884311 4791 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 10 22:49:54 crc kubenswrapper[4791]: I1210 22:49:54.884232 4791 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 10 22:49:54 crc kubenswrapper[4791]: E1210 22:49:54.884393 4791 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 10 22:49:54 crc kubenswrapper[4791]: E1210 22:49:54.884499 4791 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 10 22:49:54 crc kubenswrapper[4791]: E1210 22:49:54.884596 4791 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 10 22:49:54 crc kubenswrapper[4791]: I1210 22:49:54.916695 4791 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-10T22:49:50Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-10T22:49:50Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T22:49:54Z is after 2025-08-24T17:21:41Z" Dec 10 22:49:54 crc kubenswrapper[4791]: I1210 22:49:54.959170 4791 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-10T22:49:50Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-10T22:49:50Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T22:49:54Z is after 2025-08-24T17:21:41Z" Dec 10 22:49:54 crc kubenswrapper[4791]: I1210 22:49:54.996745 4791 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-10T22:49:54Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-10T22:49:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-10T22:49:54Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://62bfaf5d7a6d1b452cdaf88086447d0c8cb7306e94434e46752be5475124d56a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T22:49:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T22:49:54Z is after 2025-08-24T17:21:41Z" Dec 10 22:49:55 crc kubenswrapper[4791]: I1210 22:49:55.038077 4791 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f8349fea-c49b-49a1-b5ee-032c5dffe021\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T22:49:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T22:49:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T22:49:25Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T22:49:25Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T22:49:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://acafe12186df99d1059a31260de0eafce35ff7771bcb8f5c6753b72665d6f60b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T22:49:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ea4c346507862a8c12756a3061d6e25e788a8d79f4621b62f81f4322185c82a8\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T22:49:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://2e6e324ec6f4f7bb69fdfab4ea3ac1aa55e480dd8e2a8984b0fe5b4c9f66636e\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T22:49:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://1b2e11ff2bf8789fa1d0d23ae648b3e6ef653106e5bc4bbb5dd6447f7cd6f05d\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d6aa7f8c5d908d90008f2df49860ac8a3480136da567e3d85f86135d779204f9\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-12-10T22:49:44Z\\\",\\\"message\\\":\\\"W1210 22:49:33.845733 1 cmd.go:257] Using insecure, self-signed certificates\\\\nI1210 22:49:33.846108 1 crypto.go:601] Generating new CA for check-endpoints-signer@1765406973 cert, and key in /tmp/serving-cert-2585701456/serving-signer.crt, /tmp/serving-cert-2585701456/serving-signer.key\\\\nI1210 22:49:34.529748 1 observer_polling.go:159] Starting file observer\\\\nW1210 22:49:34.532225 1 builder.go:272] unable to get owner reference (falling back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": dial tcp [::1]:6443: connect: connection refused\\\\nI1210 22:49:34.532599 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1210 22:49:34.533977 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-2585701456/tls.crt::/tmp/serving-cert-2585701456/tls.key\\\\\\\"\\\\nF1210 22:49:44.774710 1 cmd.go:182] error initializing delegating authentication: unable to load configmap based request-header-client-ca-file: Get \\\\\\\"https://localhost:6443/api/v1/namespaces/kube-system/configmaps/extension-apiserver-authentication\\\\\\\": net/http: TLS handshake timeout\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-10T22:49:33Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T22:49:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://0c4d036ab269f62561fe7ef1586012cb3a16e3583a08b9488a0616ec00c8e0ae\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T22:49:32Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c5856d6b26c19fda13b5c9ac2892a7e2f9c6f79feb8173ecfd9b33b673327d67\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c5856d6b26c19fda13b5c9ac2892a7e2f9c6f79feb8173ecfd9b33b673327d67\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-10T22:49:27Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-10T22:49:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-10T22:49:25Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T22:49:55Z is after 2025-08-24T17:21:41Z" Dec 10 22:49:55 crc kubenswrapper[4791]: I1210 22:49:55.088465 4791 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"90250e13-6c63-443a-a5e5-6e4e341a2289\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T22:49:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T22:49:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T22:49:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T22:49:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T22:49:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ba54f95b1e00843b854db9a10fe6897592159ccacf30ed0f76002093a13eedd3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T22:49:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5aa8ffa518f3a68b60b275f8df4851b2564aa4984b61efdc9470b73072133e43\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T22:49:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://16f989503edc78d9451638e2cf6cf4b19a2cf2b8809b10b237541cb829303091\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T22:49:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://82556824d63ce6e46df14c45689ff1f81ebf729abe902061815c040b323898e9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T22:49:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://61db531940fa6f2effdae5a005b4c91bd2698bd5d1bff9588c171a65ad550eca\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T22:49:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0f65c864d885319ce2a08d61c30be24b3ffb7d23ea30c82ed7faf039cb2c006d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://0f65c864d885319ce2a08d61c30be24b3ffb7d23ea30c82ed7faf039cb2c006d\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-10T22:49:27Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-10T22:49:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e31a76d1d5b2812ba83e24e38268e25ec8bed98fc3da9adcd06bda5822b6ace1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e31a76d1d5b2812ba83e24e38268e25ec8bed98fc3da9adcd06bda5822b6ace1\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-10T22:49:30Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-10T22:49:29Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://b5b5eceae562f3986b543d41de1e977efb85a91d84c98e09159ca6d092eda691\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b5b5eceae562f3986b543d41de1e977efb85a91d84c98e09159ca6d092eda691\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-10T22:49:31Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-10T22:49:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-10T22:49:25Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T22:49:55Z is after 2025-08-24T17:21:41Z" Dec 10 22:49:55 crc kubenswrapper[4791]: I1210 22:49:55.121534 4791 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-5rb5l" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ba35653c-6e06-4cee-a4d6-137764090d18\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T22:49:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T22:49:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T22:49:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T22:49:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://dbb5c875730a98ad2daf316cfafc08fcdce21ac16148154780ac9193ed405dab\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T22:49:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fvvrs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://185eb2ff44221a514878283ec873773a6c9a9073dc64e296f982e60dee1338e4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T22:49:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fvvrs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-10T22:49:51Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-5rb5l\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T22:49:55Z is after 2025-08-24T17:21:41Z" Dec 10 22:49:55 crc kubenswrapper[4791]: I1210 22:49:55.275135 4791 generic.go:334] "Generic (PLEG): container finished" podID="ff9beb90-69b0-4732-bf37-0b81f58ecc98" containerID="9aa9c06e894bb2a2ed3738defaeff68a0a0458ead48177f36b90770944c211a0" exitCode=0 Dec 10 22:49:55 crc kubenswrapper[4791]: I1210 22:49:55.275236 4791 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-additional-cni-plugins-rw584" event={"ID":"ff9beb90-69b0-4732-bf37-0b81f58ecc98","Type":"ContainerDied","Data":"9aa9c06e894bb2a2ed3738defaeff68a0a0458ead48177f36b90770944c211a0"} Dec 10 22:49:55 crc kubenswrapper[4791]: I1210 22:49:55.293613 4791 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-5rb5l" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ba35653c-6e06-4cee-a4d6-137764090d18\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T22:49:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T22:49:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T22:49:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T22:49:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://dbb5c875730a98ad2daf316cfafc08fcdce21ac16148154780ac9193ed405dab\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T22:49:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fvvrs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://185eb2ff44221a514878283ec873773a6c9a9073dc64e296f982e60dee1338e4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T22:49:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fvvrs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-10T22:49:51Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-5rb5l\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T22:49:55Z is after 2025-08-24T17:21:41Z" Dec 10 22:49:55 crc kubenswrapper[4791]: I1210 22:49:55.314466 4791 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-10T22:49:52Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-10T22:49:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-10T22:49:52Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3734b110802828632d44263ed0a43cabc4b710833f0be120d72db0940fe783e2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T22:49:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://794803ac9be3c52509e6dcb10854d8864e1a9e406762e57c7971b07c4da1220e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T22:49:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T22:49:55Z is after 2025-08-24T17:21:41Z" Dec 10 22:49:55 crc kubenswrapper[4791]: I1210 22:49:55.328848 4791 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-vczjq" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b9f00f25-3fa1-4b63-9710-fd352224b01b\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T22:49:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T22:49:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T22:49:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T22:49:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4698cf4c6b55b41e628448d37c2eb847c2777d0efbd93afcbf367d5e78623b7e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T22:49:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lrrpv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-10T22:49:50Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-vczjq\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T22:49:55Z is after 2025-08-24T17:21:41Z" Dec 10 22:49:55 crc kubenswrapper[4791]: I1210 22:49:55.344227 4791 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-4nwkq" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"672aa28c-8169-49ed-87b8-21187d13a80c\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T22:49:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T22:49:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T22:49:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T22:49:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a2f2a62315529e148a10a33ccb241e9417fbde0fe42afb61cb7ee2438c74636f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T22:49:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xl62j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-10T22:49:51Z\\\"}}\" for pod \"openshift-multus\"/\"multus-4nwkq\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T22:49:55Z is after 2025-08-24T17:21:41Z" Dec 10 22:49:55 crc kubenswrapper[4791]: I1210 22:49:55.366215 4791 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-zhq64" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3cd47739-0fa9-4321-aff1-220f8721a0b3\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T22:49:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T22:49:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T22:49:52Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T22:49:52Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wvpw8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wvpw8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wvpw8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wvpw8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wvpw8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wvpw8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wvpw8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wvpw8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3afbd1ffcd2c4b40324aca6e52d6fc822db131c3f35a9af3a6f005c2cffdb7bf\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3afbd1ffcd2c4b40324aca6e52d6fc822db131c3f35a9af3a6f005c2cffdb7bf\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-10T22:49:53Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-10T22:49:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wvpw8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-10T22:49:52Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-zhq64\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T22:49:55Z is after 2025-08-24T17:21:41Z" Dec 10 22:49:55 crc kubenswrapper[4791]: I1210 22:49:55.384708 4791 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"698bbe9a-fbe5-436d-a835-5e80df77ba24\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T22:49:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T22:49:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T22:49:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T22:49:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T22:49:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://224d73713312376a58ec6cc2bc8fdb6ffa44c4749ae0e7691d5fe27b9afca09f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T22:49:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://372eeaffd40992c6f06efc4df6dfcb44b31de64ee8256c21508cdd7c74ecb96e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T22:49:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://76b6013b062cac5b0b4d12a94f38ce7b43f9c7ea216063455d88a956ac04d26e\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T22:49:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://78e950e122e761d950da1a3c60c06476e53920e8795f6c61db22c01e22d5136e\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T22:49:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-10T22:49:25Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T22:49:55Z is after 2025-08-24T17:21:41Z" Dec 10 22:49:55 crc kubenswrapper[4791]: I1210 22:49:55.410494 4791 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-10T22:49:50Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-10T22:49:50Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T22:49:55Z is after 2025-08-24T17:21:41Z" Dec 10 22:49:55 crc kubenswrapper[4791]: I1210 22:49:55.441979 4791 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-10T22:49:50Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-10T22:49:50Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T22:49:55Z is after 2025-08-24T17:21:41Z" Dec 10 22:49:55 crc kubenswrapper[4791]: I1210 22:49:55.474922 4791 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-tchxk" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"c0ab74e0-42a9-4f42-ac80-16030cf4ffff\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T22:49:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T22:49:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T22:49:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T22:49:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://04e0fd1b53256b7a938e8ccab93e2a4eb73dd1a59e0b9799e09c55e511baf7fe\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T22:49:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zn2qt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-10T22:49:50Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-tchxk\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T22:49:55Z is after 2025-08-24T17:21:41Z" Dec 10 22:49:55 crc kubenswrapper[4791]: I1210 22:49:55.518887 4791 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-rw584" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ff9beb90-69b0-4732-bf37-0b81f58ecc98\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T22:49:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T22:49:51Z\\\",\\\"message\\\":\\\"containers with incomplete status: [routeoverride-cni whereabouts-cni-bincopy whereabouts-cni]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T22:49:51Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T22:49:51Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fksqn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ce67e54b696bb53ffbbe48c6ce71476b635e0868993e9c427b22f904891433ee\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ce67e54b696bb53ffbbe48c6ce71476b635e0868993e9c427b22f904891433ee\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-10T22:49:52Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-10T22:49:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fksqn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://818d89bfffb8a110670152c8480684665ea8db578e9768f524fbbe6a92f27c22\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://818d89bfffb8a110670152c8480684665ea8db578e9768f524fbbe6a92f27c22\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-10T22:49:53Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-10T22:49:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fksqn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://9aa9c06e894bb2a2ed3738defaeff68a0a0458ead48177f36b90770944c211a0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://9aa9c06e894bb2a2ed3738defaeff68a0a0458ead48177f36b90770944c211a0\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-10T22:49:54Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-10T22:49:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fksqn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fksqn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fksqn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fksqn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-10T22:49:51Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-rw584\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T22:49:55Z is after 2025-08-24T17:21:41Z" Dec 10 22:49:55 crc kubenswrapper[4791]: I1210 22:49:55.558581 4791 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-10T22:49:52Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-10T22:49:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-10T22:49:52Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://20651eb0c16e2c71222f56da1ff755cb0e7b084886acbbb24f0ccaf57041ed85\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T22:49:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T22:49:55Z is after 2025-08-24T17:21:41Z" Dec 10 22:49:55 crc kubenswrapper[4791]: I1210 22:49:55.605691 4791 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"90250e13-6c63-443a-a5e5-6e4e341a2289\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T22:49:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T22:49:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T22:49:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T22:49:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T22:49:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ba54f95b1e00843b854db9a10fe6897592159ccacf30ed0f76002093a13eedd3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T22:49:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5aa8ffa518f3a68b60b275f8df4851b2564aa4984b61efdc9470b73072133e43\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T22:49:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://16f989503edc78d9451638e2cf6cf4b19a2cf2b8809b10b237541cb829303091\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T22:49:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://82556824d63ce6e46df14c45689ff1f81ebf729abe902061815c040b323898e9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T22:49:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://61db531940fa6f2effdae5a005b4c91bd2698bd5d1bff9588c171a65ad550eca\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T22:49:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0f65c864d885319ce2a08d61c30be24b3ffb7d23ea30c82ed7faf039cb2c006d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://0f65c864d885319ce2a08d61c30be24b3ffb7d23ea30c82ed7faf039cb2c006d\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-10T22:49:27Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-10T22:49:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e31a76d1d5b2812ba83e24e38268e25ec8bed98fc3da9adcd06bda5822b6ace1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e31a76d1d5b2812ba83e24e38268e25ec8bed98fc3da9adcd06bda5822b6ace1\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-10T22:49:30Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-10T22:49:29Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://b5b5eceae562f3986b543d41de1e977efb85a91d84c98e09159ca6d092eda691\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b5b5eceae562f3986b543d41de1e977efb85a91d84c98e09159ca6d092eda691\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-10T22:49:31Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-10T22:49:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-10T22:49:25Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T22:49:55Z is after 2025-08-24T17:21:41Z" Dec 10 22:49:55 crc kubenswrapper[4791]: I1210 22:49:55.644166 4791 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-10T22:49:50Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-10T22:49:50Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T22:49:55Z is after 2025-08-24T17:21:41Z" Dec 10 22:49:55 crc kubenswrapper[4791]: I1210 22:49:55.681074 4791 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-10T22:49:54Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-10T22:49:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-10T22:49:54Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://62bfaf5d7a6d1b452cdaf88086447d0c8cb7306e94434e46752be5475124d56a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T22:49:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T22:49:55Z is after 2025-08-24T17:21:41Z" Dec 10 22:49:55 crc kubenswrapper[4791]: I1210 22:49:55.725765 4791 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f8349fea-c49b-49a1-b5ee-032c5dffe021\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T22:49:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T22:49:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T22:49:25Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T22:49:25Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T22:49:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://acafe12186df99d1059a31260de0eafce35ff7771bcb8f5c6753b72665d6f60b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T22:49:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ea4c346507862a8c12756a3061d6e25e788a8d79f4621b62f81f4322185c82a8\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T22:49:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://2e6e324ec6f4f7bb69fdfab4ea3ac1aa55e480dd8e2a8984b0fe5b4c9f66636e\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T22:49:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://1b2e11ff2bf8789fa1d0d23ae648b3e6ef653106e5bc4bbb5dd6447f7cd6f05d\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d6aa7f8c5d908d90008f2df49860ac8a3480136da567e3d85f86135d779204f9\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-12-10T22:49:44Z\\\",\\\"message\\\":\\\"W1210 22:49:33.845733 1 cmd.go:257] Using insecure, self-signed certificates\\\\nI1210 22:49:33.846108 1 crypto.go:601] Generating new CA for check-endpoints-signer@1765406973 cert, and key in /tmp/serving-cert-2585701456/serving-signer.crt, /tmp/serving-cert-2585701456/serving-signer.key\\\\nI1210 22:49:34.529748 1 observer_polling.go:159] Starting file observer\\\\nW1210 22:49:34.532225 1 builder.go:272] unable to get owner reference (falling back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": dial tcp [::1]:6443: connect: connection refused\\\\nI1210 22:49:34.532599 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1210 22:49:34.533977 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-2585701456/tls.crt::/tmp/serving-cert-2585701456/tls.key\\\\\\\"\\\\nF1210 22:49:44.774710 1 cmd.go:182] error initializing delegating authentication: unable to load configmap based request-header-client-ca-file: Get \\\\\\\"https://localhost:6443/api/v1/namespaces/kube-system/configmaps/extension-apiserver-authentication\\\\\\\": net/http: TLS handshake timeout\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-10T22:49:33Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T22:49:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://0c4d036ab269f62561fe7ef1586012cb3a16e3583a08b9488a0616ec00c8e0ae\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T22:49:32Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c5856d6b26c19fda13b5c9ac2892a7e2f9c6f79feb8173ecfd9b33b673327d67\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c5856d6b26c19fda13b5c9ac2892a7e2f9c6f79feb8173ecfd9b33b673327d67\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-10T22:49:27Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-10T22:49:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-10T22:49:25Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T22:49:55Z is after 2025-08-24T17:21:41Z" Dec 10 22:49:55 crc kubenswrapper[4791]: I1210 22:49:55.727579 4791 transport.go:147] "Certificate rotation detected, shutting down client connections to start using new credentials" Dec 10 22:49:55 crc kubenswrapper[4791]: I1210 22:49:55.925312 4791 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"90250e13-6c63-443a-a5e5-6e4e341a2289\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T22:49:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T22:49:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T22:49:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T22:49:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T22:49:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ba54f95b1e00843b854db9a10fe6897592159ccacf30ed0f76002093a13eedd3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T22:49:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5aa8ffa518f3a68b60b275f8df4851b2564aa4984b61efdc9470b73072133e43\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T22:49:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://16f989503edc78d9451638e2cf6cf4b19a2cf2b8809b10b237541cb829303091\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T22:49:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://82556824d63ce6e46df14c45689ff1f81ebf729abe902061815c040b323898e9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T22:49:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://61db531940fa6f2effdae5a005b4c91bd2698bd5d1bff9588c171a65ad550eca\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T22:49:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0f65c864d885319ce2a08d61c30be24b3ffb7d23ea30c82ed7faf039cb2c006d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://0f65c864d885319ce2a08d61c30be24b3ffb7d23ea30c82ed7faf039cb2c006d\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-10T22:49:27Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-10T22:49:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e31a76d1d5b2812ba83e24e38268e25ec8bed98fc3da9adcd06bda5822b6ace1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e31a76d1d5b2812ba83e24e38268e25ec8bed98fc3da9adcd06bda5822b6ace1\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-10T22:49:30Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-10T22:49:29Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://b5b5eceae562f3986b543d41de1e977efb85a91d84c98e09159ca6d092eda691\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b5b5eceae562f3986b543d41de1e977efb85a91d84c98e09159ca6d092eda691\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-10T22:49:31Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-10T22:49:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-10T22:49:25Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T22:49:55Z is after 2025-08-24T17:21:41Z" Dec 10 22:49:55 crc kubenswrapper[4791]: I1210 22:49:55.945288 4791 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-10T22:49:50Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-10T22:49:50Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T22:49:55Z is after 2025-08-24T17:21:41Z" Dec 10 22:49:55 crc kubenswrapper[4791]: I1210 22:49:55.963531 4791 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-10T22:49:54Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-10T22:49:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-10T22:49:54Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://62bfaf5d7a6d1b452cdaf88086447d0c8cb7306e94434e46752be5475124d56a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T22:49:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T22:49:55Z is after 2025-08-24T17:21:41Z" Dec 10 22:49:55 crc kubenswrapper[4791]: I1210 22:49:55.981209 4791 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f8349fea-c49b-49a1-b5ee-032c5dffe021\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T22:49:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T22:49:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T22:49:25Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T22:49:25Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T22:49:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://acafe12186df99d1059a31260de0eafce35ff7771bcb8f5c6753b72665d6f60b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T22:49:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ea4c346507862a8c12756a3061d6e25e788a8d79f4621b62f81f4322185c82a8\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T22:49:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://2e6e324ec6f4f7bb69fdfab4ea3ac1aa55e480dd8e2a8984b0fe5b4c9f66636e\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T22:49:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://1b2e11ff2bf8789fa1d0d23ae648b3e6ef653106e5bc4bbb5dd6447f7cd6f05d\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d6aa7f8c5d908d90008f2df49860ac8a3480136da567e3d85f86135d779204f9\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-12-10T22:49:44Z\\\",\\\"message\\\":\\\"W1210 22:49:33.845733 1 cmd.go:257] Using insecure, self-signed certificates\\\\nI1210 22:49:33.846108 1 crypto.go:601] Generating new CA for check-endpoints-signer@1765406973 cert, and key in /tmp/serving-cert-2585701456/serving-signer.crt, /tmp/serving-cert-2585701456/serving-signer.key\\\\nI1210 22:49:34.529748 1 observer_polling.go:159] Starting file observer\\\\nW1210 22:49:34.532225 1 builder.go:272] unable to get owner reference (falling back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": dial tcp [::1]:6443: connect: connection refused\\\\nI1210 22:49:34.532599 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1210 22:49:34.533977 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-2585701456/tls.crt::/tmp/serving-cert-2585701456/tls.key\\\\\\\"\\\\nF1210 22:49:44.774710 1 cmd.go:182] error initializing delegating authentication: unable to load configmap based request-header-client-ca-file: Get \\\\\\\"https://localhost:6443/api/v1/namespaces/kube-system/configmaps/extension-apiserver-authentication\\\\\\\": net/http: TLS handshake timeout\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-10T22:49:33Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T22:49:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://0c4d036ab269f62561fe7ef1586012cb3a16e3583a08b9488a0616ec00c8e0ae\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T22:49:32Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c5856d6b26c19fda13b5c9ac2892a7e2f9c6f79feb8173ecfd9b33b673327d67\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c5856d6b26c19fda13b5c9ac2892a7e2f9c6f79feb8173ecfd9b33b673327d67\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-10T22:49:27Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-10T22:49:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-10T22:49:25Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T22:49:55Z is after 2025-08-24T17:21:41Z" Dec 10 22:49:56 crc kubenswrapper[4791]: I1210 22:49:56.000314 4791 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-5rb5l" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ba35653c-6e06-4cee-a4d6-137764090d18\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T22:49:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T22:49:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T22:49:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T22:49:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://dbb5c875730a98ad2daf316cfafc08fcdce21ac16148154780ac9193ed405dab\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T22:49:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fvvrs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://185eb2ff44221a514878283ec873773a6c9a9073dc64e296f982e60dee1338e4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T22:49:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fvvrs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-10T22:49:51Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-5rb5l\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T22:49:55Z is after 2025-08-24T17:21:41Z" Dec 10 22:49:56 crc kubenswrapper[4791]: I1210 22:49:56.020615 4791 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-10T22:49:52Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-10T22:49:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-10T22:49:52Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3734b110802828632d44263ed0a43cabc4b710833f0be120d72db0940fe783e2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T22:49:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://794803ac9be3c52509e6dcb10854d8864e1a9e406762e57c7971b07c4da1220e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T22:49:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T22:49:56Z is after 2025-08-24T17:21:41Z" Dec 10 22:49:56 crc kubenswrapper[4791]: I1210 22:49:56.034027 4791 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-vczjq" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b9f00f25-3fa1-4b63-9710-fd352224b01b\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T22:49:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T22:49:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T22:49:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T22:49:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4698cf4c6b55b41e628448d37c2eb847c2777d0efbd93afcbf367d5e78623b7e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T22:49:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lrrpv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-10T22:49:50Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-vczjq\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T22:49:56Z is after 2025-08-24T17:21:41Z" Dec 10 22:49:56 crc kubenswrapper[4791]: I1210 22:49:56.050568 4791 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-4nwkq" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"672aa28c-8169-49ed-87b8-21187d13a80c\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T22:49:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T22:49:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T22:49:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T22:49:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a2f2a62315529e148a10a33ccb241e9417fbde0fe42afb61cb7ee2438c74636f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T22:49:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xl62j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-10T22:49:51Z\\\"}}\" for pod \"openshift-multus\"/\"multus-4nwkq\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T22:49:56Z is after 2025-08-24T17:21:41Z" Dec 10 22:49:56 crc kubenswrapper[4791]: I1210 22:49:56.084167 4791 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-zhq64" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3cd47739-0fa9-4321-aff1-220f8721a0b3\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T22:49:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T22:49:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T22:49:52Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T22:49:52Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wvpw8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wvpw8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wvpw8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wvpw8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wvpw8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wvpw8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wvpw8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wvpw8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3afbd1ffcd2c4b40324aca6e52d6fc822db131c3f35a9af3a6f005c2cffdb7bf\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3afbd1ffcd2c4b40324aca6e52d6fc822db131c3f35a9af3a6f005c2cffdb7bf\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-10T22:49:53Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-10T22:49:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wvpw8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-10T22:49:52Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-zhq64\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T22:49:56Z is after 2025-08-24T17:21:41Z" Dec 10 22:49:56 crc kubenswrapper[4791]: I1210 22:49:56.124589 4791 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"698bbe9a-fbe5-436d-a835-5e80df77ba24\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T22:49:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T22:49:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T22:49:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T22:49:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T22:49:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://224d73713312376a58ec6cc2bc8fdb6ffa44c4749ae0e7691d5fe27b9afca09f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T22:49:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://372eeaffd40992c6f06efc4df6dfcb44b31de64ee8256c21508cdd7c74ecb96e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T22:49:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://76b6013b062cac5b0b4d12a94f38ce7b43f9c7ea216063455d88a956ac04d26e\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T22:49:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://78e950e122e761d950da1a3c60c06476e53920e8795f6c61db22c01e22d5136e\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T22:49:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-10T22:49:25Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T22:49:56Z is after 2025-08-24T17:21:41Z" Dec 10 22:49:56 crc kubenswrapper[4791]: I1210 22:49:56.165269 4791 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-10T22:49:50Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-10T22:49:50Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T22:49:56Z is after 2025-08-24T17:21:41Z" Dec 10 22:49:56 crc kubenswrapper[4791]: I1210 22:49:56.200044 4791 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-10T22:49:50Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-10T22:49:50Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T22:49:56Z is after 2025-08-24T17:21:41Z" Dec 10 22:49:56 crc kubenswrapper[4791]: I1210 22:49:56.239235 4791 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-tchxk" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"c0ab74e0-42a9-4f42-ac80-16030cf4ffff\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T22:49:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T22:49:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T22:49:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T22:49:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://04e0fd1b53256b7a938e8ccab93e2a4eb73dd1a59e0b9799e09c55e511baf7fe\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T22:49:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zn2qt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-10T22:49:50Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-tchxk\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T22:49:56Z is after 2025-08-24T17:21:41Z" Dec 10 22:49:56 crc kubenswrapper[4791]: I1210 22:49:56.283194 4791 generic.go:334] "Generic (PLEG): container finished" podID="ff9beb90-69b0-4732-bf37-0b81f58ecc98" containerID="dfd6d03336dcea7529419a30e8d70d84e2290a0fa5b0e46d09ad341c190638d6" exitCode=0 Dec 10 22:49:56 crc kubenswrapper[4791]: I1210 22:49:56.283256 4791 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-additional-cni-plugins-rw584" event={"ID":"ff9beb90-69b0-4732-bf37-0b81f58ecc98","Type":"ContainerDied","Data":"dfd6d03336dcea7529419a30e8d70d84e2290a0fa5b0e46d09ad341c190638d6"} Dec 10 22:49:56 crc kubenswrapper[4791]: I1210 22:49:56.287182 4791 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-rw584" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ff9beb90-69b0-4732-bf37-0b81f58ecc98\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T22:49:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T22:49:51Z\\\",\\\"message\\\":\\\"containers with incomplete status: [routeoverride-cni whereabouts-cni-bincopy whereabouts-cni]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T22:49:51Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T22:49:51Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fksqn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ce67e54b696bb53ffbbe48c6ce71476b635e0868993e9c427b22f904891433ee\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ce67e54b696bb53ffbbe48c6ce71476b635e0868993e9c427b22f904891433ee\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-10T22:49:52Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-10T22:49:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fksqn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://818d89bfffb8a110670152c8480684665ea8db578e9768f524fbbe6a92f27c22\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://818d89bfffb8a110670152c8480684665ea8db578e9768f524fbbe6a92f27c22\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-10T22:49:53Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-10T22:49:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fksqn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://9aa9c06e894bb2a2ed3738defaeff68a0a0458ead48177f36b90770944c211a0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://9aa9c06e894bb2a2ed3738defaeff68a0a0458ead48177f36b90770944c211a0\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-10T22:49:54Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-10T22:49:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fksqn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fksqn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fksqn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fksqn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-10T22:49:51Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-rw584\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T22:49:56Z is after 2025-08-24T17:21:41Z" Dec 10 22:49:56 crc kubenswrapper[4791]: I1210 22:49:56.327385 4791 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-10T22:49:52Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-10T22:49:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-10T22:49:52Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://20651eb0c16e2c71222f56da1ff755cb0e7b084886acbbb24f0ccaf57041ed85\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T22:49:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T22:49:56Z is after 2025-08-24T17:21:41Z" Dec 10 22:49:56 crc kubenswrapper[4791]: I1210 22:49:56.358200 4791 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-5rb5l" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ba35653c-6e06-4cee-a4d6-137764090d18\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T22:49:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T22:49:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T22:49:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T22:49:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://dbb5c875730a98ad2daf316cfafc08fcdce21ac16148154780ac9193ed405dab\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T22:49:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fvvrs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://185eb2ff44221a514878283ec873773a6c9a9073dc64e296f982e60dee1338e4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T22:49:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fvvrs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-10T22:49:51Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-5rb5l\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T22:49:56Z is after 2025-08-24T17:21:41Z" Dec 10 22:49:56 crc kubenswrapper[4791]: I1210 22:49:56.399567 4791 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-10T22:49:52Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-10T22:49:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-10T22:49:52Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3734b110802828632d44263ed0a43cabc4b710833f0be120d72db0940fe783e2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T22:49:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://794803ac9be3c52509e6dcb10854d8864e1a9e406762e57c7971b07c4da1220e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T22:49:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T22:49:56Z is after 2025-08-24T17:21:41Z" Dec 10 22:49:56 crc kubenswrapper[4791]: I1210 22:49:56.437898 4791 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-vczjq" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b9f00f25-3fa1-4b63-9710-fd352224b01b\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T22:49:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T22:49:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T22:49:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T22:49:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4698cf4c6b55b41e628448d37c2eb847c2777d0efbd93afcbf367d5e78623b7e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T22:49:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lrrpv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-10T22:49:50Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-vczjq\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T22:49:56Z is after 2025-08-24T17:21:41Z" Dec 10 22:49:56 crc kubenswrapper[4791]: I1210 22:49:56.479904 4791 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-4nwkq" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"672aa28c-8169-49ed-87b8-21187d13a80c\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T22:49:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T22:49:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T22:49:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T22:49:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a2f2a62315529e148a10a33ccb241e9417fbde0fe42afb61cb7ee2438c74636f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T22:49:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xl62j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-10T22:49:51Z\\\"}}\" for pod \"openshift-multus\"/\"multus-4nwkq\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T22:49:56Z is after 2025-08-24T17:21:41Z" Dec 10 22:49:56 crc kubenswrapper[4791]: I1210 22:49:56.527101 4791 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-zhq64" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3cd47739-0fa9-4321-aff1-220f8721a0b3\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T22:49:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T22:49:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T22:49:52Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T22:49:52Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wvpw8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wvpw8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wvpw8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wvpw8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wvpw8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wvpw8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wvpw8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wvpw8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3afbd1ffcd2c4b40324aca6e52d6fc822db131c3f35a9af3a6f005c2cffdb7bf\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3afbd1ffcd2c4b40324aca6e52d6fc822db131c3f35a9af3a6f005c2cffdb7bf\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-10T22:49:53Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-10T22:49:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wvpw8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-10T22:49:52Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-zhq64\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T22:49:56Z is after 2025-08-24T17:21:41Z" Dec 10 22:49:56 crc kubenswrapper[4791]: I1210 22:49:56.564661 4791 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"698bbe9a-fbe5-436d-a835-5e80df77ba24\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T22:49:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T22:49:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T22:49:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T22:49:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T22:49:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://224d73713312376a58ec6cc2bc8fdb6ffa44c4749ae0e7691d5fe27b9afca09f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T22:49:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://372eeaffd40992c6f06efc4df6dfcb44b31de64ee8256c21508cdd7c74ecb96e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T22:49:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://76b6013b062cac5b0b4d12a94f38ce7b43f9c7ea216063455d88a956ac04d26e\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T22:49:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://78e950e122e761d950da1a3c60c06476e53920e8795f6c61db22c01e22d5136e\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T22:49:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-10T22:49:25Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T22:49:56Z is after 2025-08-24T17:21:41Z" Dec 10 22:49:56 crc kubenswrapper[4791]: I1210 22:49:56.603834 4791 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-10T22:49:50Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-10T22:49:50Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T22:49:56Z is after 2025-08-24T17:21:41Z" Dec 10 22:49:56 crc kubenswrapper[4791]: I1210 22:49:56.639271 4791 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-10T22:49:50Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-10T22:49:50Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T22:49:56Z is after 2025-08-24T17:21:41Z" Dec 10 22:49:56 crc kubenswrapper[4791]: I1210 22:49:56.677806 4791 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-tchxk" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"c0ab74e0-42a9-4f42-ac80-16030cf4ffff\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T22:49:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T22:49:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T22:49:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T22:49:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://04e0fd1b53256b7a938e8ccab93e2a4eb73dd1a59e0b9799e09c55e511baf7fe\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T22:49:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zn2qt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-10T22:49:50Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-tchxk\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T22:49:56Z is after 2025-08-24T17:21:41Z" Dec 10 22:49:56 crc kubenswrapper[4791]: I1210 22:49:56.721220 4791 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-rw584" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ff9beb90-69b0-4732-bf37-0b81f58ecc98\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T22:49:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T22:49:51Z\\\",\\\"message\\\":\\\"containers with incomplete status: [whereabouts-cni-bincopy whereabouts-cni]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T22:49:51Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T22:49:51Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fksqn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ce67e54b696bb53ffbbe48c6ce71476b635e0868993e9c427b22f904891433ee\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ce67e54b696bb53ffbbe48c6ce71476b635e0868993e9c427b22f904891433ee\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-10T22:49:52Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-10T22:49:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fksqn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://818d89bfffb8a110670152c8480684665ea8db578e9768f524fbbe6a92f27c22\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://818d89bfffb8a110670152c8480684665ea8db578e9768f524fbbe6a92f27c22\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-10T22:49:53Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-10T22:49:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fksqn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://9aa9c06e894bb2a2ed3738defaeff68a0a0458ead48177f36b90770944c211a0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://9aa9c06e894bb2a2ed3738defaeff68a0a0458ead48177f36b90770944c211a0\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-10T22:49:54Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-10T22:49:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fksqn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://dfd6d03336dcea7529419a30e8d70d84e2290a0fa5b0e46d09ad341c190638d6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://dfd6d03336dcea7529419a30e8d70d84e2290a0fa5b0e46d09ad341c190638d6\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-10T22:49:55Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-10T22:49:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fksqn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fksqn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fksqn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-10T22:49:51Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-rw584\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T22:49:56Z is after 2025-08-24T17:21:41Z" Dec 10 22:49:56 crc kubenswrapper[4791]: I1210 22:49:56.760255 4791 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-10T22:49:52Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-10T22:49:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-10T22:49:52Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://20651eb0c16e2c71222f56da1ff755cb0e7b084886acbbb24f0ccaf57041ed85\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T22:49:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T22:49:56Z is after 2025-08-24T17:21:41Z" Dec 10 22:49:56 crc kubenswrapper[4791]: I1210 22:49:56.818635 4791 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"90250e13-6c63-443a-a5e5-6e4e341a2289\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T22:49:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T22:49:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T22:49:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T22:49:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T22:49:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ba54f95b1e00843b854db9a10fe6897592159ccacf30ed0f76002093a13eedd3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T22:49:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5aa8ffa518f3a68b60b275f8df4851b2564aa4984b61efdc9470b73072133e43\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T22:49:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://16f989503edc78d9451638e2cf6cf4b19a2cf2b8809b10b237541cb829303091\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T22:49:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://82556824d63ce6e46df14c45689ff1f81ebf729abe902061815c040b323898e9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T22:49:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://61db531940fa6f2effdae5a005b4c91bd2698bd5d1bff9588c171a65ad550eca\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T22:49:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0f65c864d885319ce2a08d61c30be24b3ffb7d23ea30c82ed7faf039cb2c006d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://0f65c864d885319ce2a08d61c30be24b3ffb7d23ea30c82ed7faf039cb2c006d\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-10T22:49:27Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-10T22:49:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e31a76d1d5b2812ba83e24e38268e25ec8bed98fc3da9adcd06bda5822b6ace1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e31a76d1d5b2812ba83e24e38268e25ec8bed98fc3da9adcd06bda5822b6ace1\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-10T22:49:30Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-10T22:49:29Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://b5b5eceae562f3986b543d41de1e977efb85a91d84c98e09159ca6d092eda691\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b5b5eceae562f3986b543d41de1e977efb85a91d84c98e09159ca6d092eda691\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-10T22:49:31Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-10T22:49:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-10T22:49:25Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T22:49:56Z is after 2025-08-24T17:21:41Z" Dec 10 22:49:56 crc kubenswrapper[4791]: I1210 22:49:56.837659 4791 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-10T22:49:50Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-10T22:49:50Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T22:49:56Z is after 2025-08-24T17:21:41Z" Dec 10 22:49:56 crc kubenswrapper[4791]: I1210 22:49:56.861276 4791 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 10 22:49:56 crc kubenswrapper[4791]: I1210 22:49:56.863488 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 22:49:56 crc kubenswrapper[4791]: I1210 22:49:56.863548 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 22:49:56 crc kubenswrapper[4791]: I1210 22:49:56.863562 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 22:49:56 crc kubenswrapper[4791]: I1210 22:49:56.863674 4791 kubelet_node_status.go:76] "Attempting to register node" node="crc" Dec 10 22:49:56 crc kubenswrapper[4791]: I1210 22:49:56.876457 4791 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-10T22:49:54Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-10T22:49:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-10T22:49:54Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://62bfaf5d7a6d1b452cdaf88086447d0c8cb7306e94434e46752be5475124d56a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T22:49:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T22:49:56Z is after 2025-08-24T17:21:41Z" Dec 10 22:49:56 crc kubenswrapper[4791]: I1210 22:49:56.884746 4791 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 10 22:49:56 crc kubenswrapper[4791]: E1210 22:49:56.884906 4791 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 10 22:49:56 crc kubenswrapper[4791]: I1210 22:49:56.884742 4791 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 10 22:49:56 crc kubenswrapper[4791]: E1210 22:49:56.885040 4791 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 10 22:49:56 crc kubenswrapper[4791]: I1210 22:49:56.884744 4791 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 10 22:49:56 crc kubenswrapper[4791]: E1210 22:49:56.885140 4791 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 10 22:49:56 crc kubenswrapper[4791]: I1210 22:49:56.930714 4791 kubelet_node_status.go:115] "Node was previously registered" node="crc" Dec 10 22:49:56 crc kubenswrapper[4791]: I1210 22:49:56.931044 4791 kubelet_node_status.go:79] "Successfully registered node" node="crc" Dec 10 22:49:56 crc kubenswrapper[4791]: I1210 22:49:56.932635 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 22:49:56 crc kubenswrapper[4791]: I1210 22:49:56.932670 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 22:49:56 crc kubenswrapper[4791]: I1210 22:49:56.932681 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 22:49:56 crc kubenswrapper[4791]: I1210 22:49:56.932696 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 22:49:56 crc kubenswrapper[4791]: I1210 22:49:56.932708 4791 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T22:49:56Z","lastTransitionTime":"2025-12-10T22:49:56Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 22:49:56 crc kubenswrapper[4791]: E1210 22:49:56.955157 4791 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-12-10T22:49:56Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-10T22:49:56Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-10T22:49:56Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-10T22:49:56Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-10T22:49:56Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-10T22:49:56Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-10T22:49:56Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-10T22:49:56Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"4fab3af6-f657-4dfc-8ef2-12b7c978c94f\\\",\\\"systemUUID\\\":\\\"442c0b68-8dce-47e0-bd68-d5210d7e0493\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T22:49:56Z is after 2025-08-24T17:21:41Z" Dec 10 22:49:56 crc kubenswrapper[4791]: I1210 22:49:56.960019 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 22:49:56 crc kubenswrapper[4791]: I1210 22:49:56.960089 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 22:49:56 crc kubenswrapper[4791]: I1210 22:49:56.960108 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 22:49:56 crc kubenswrapper[4791]: I1210 22:49:56.960133 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 22:49:56 crc kubenswrapper[4791]: I1210 22:49:56.960156 4791 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T22:49:56Z","lastTransitionTime":"2025-12-10T22:49:56Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 22:49:56 crc kubenswrapper[4791]: I1210 22:49:56.973587 4791 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f8349fea-c49b-49a1-b5ee-032c5dffe021\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T22:49:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T22:49:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T22:49:25Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T22:49:25Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T22:49:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://acafe12186df99d1059a31260de0eafce35ff7771bcb8f5c6753b72665d6f60b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T22:49:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ea4c346507862a8c12756a3061d6e25e788a8d79f4621b62f81f4322185c82a8\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T22:49:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://2e6e324ec6f4f7bb69fdfab4ea3ac1aa55e480dd8e2a8984b0fe5b4c9f66636e\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T22:49:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://1b2e11ff2bf8789fa1d0d23ae648b3e6ef653106e5bc4bbb5dd6447f7cd6f05d\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d6aa7f8c5d908d90008f2df49860ac8a3480136da567e3d85f86135d779204f9\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-12-10T22:49:44Z\\\",\\\"message\\\":\\\"W1210 22:49:33.845733 1 cmd.go:257] Using insecure, self-signed certificates\\\\nI1210 22:49:33.846108 1 crypto.go:601] Generating new CA for check-endpoints-signer@1765406973 cert, and key in /tmp/serving-cert-2585701456/serving-signer.crt, /tmp/serving-cert-2585701456/serving-signer.key\\\\nI1210 22:49:34.529748 1 observer_polling.go:159] Starting file observer\\\\nW1210 22:49:34.532225 1 builder.go:272] unable to get owner reference (falling back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": dial tcp [::1]:6443: connect: connection refused\\\\nI1210 22:49:34.532599 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1210 22:49:34.533977 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-2585701456/tls.crt::/tmp/serving-cert-2585701456/tls.key\\\\\\\"\\\\nF1210 22:49:44.774710 1 cmd.go:182] error initializing delegating authentication: unable to load configmap based request-header-client-ca-file: Get \\\\\\\"https://localhost:6443/api/v1/namespaces/kube-system/configmaps/extension-apiserver-authentication\\\\\\\": net/http: TLS handshake timeout\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-10T22:49:33Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T22:49:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://0c4d036ab269f62561fe7ef1586012cb3a16e3583a08b9488a0616ec00c8e0ae\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T22:49:32Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c5856d6b26c19fda13b5c9ac2892a7e2f9c6f79feb8173ecfd9b33b673327d67\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c5856d6b26c19fda13b5c9ac2892a7e2f9c6f79feb8173ecfd9b33b673327d67\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-10T22:49:27Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-10T22:49:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-10T22:49:25Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T22:49:56Z is after 2025-08-24T17:21:41Z" Dec 10 22:49:56 crc kubenswrapper[4791]: E1210 22:49:56.986894 4791 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-12-10T22:49:56Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-10T22:49:56Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-10T22:49:56Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-10T22:49:56Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-10T22:49:56Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-10T22:49:56Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-10T22:49:56Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-10T22:49:56Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"4fab3af6-f657-4dfc-8ef2-12b7c978c94f\\\",\\\"systemUUID\\\":\\\"442c0b68-8dce-47e0-bd68-d5210d7e0493\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T22:49:56Z is after 2025-08-24T17:21:41Z" Dec 10 22:49:56 crc kubenswrapper[4791]: I1210 22:49:56.991298 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 22:49:56 crc kubenswrapper[4791]: I1210 22:49:56.991369 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 22:49:56 crc kubenswrapper[4791]: I1210 22:49:56.991384 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 22:49:56 crc kubenswrapper[4791]: I1210 22:49:56.991402 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 22:49:56 crc kubenswrapper[4791]: I1210 22:49:56.991414 4791 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T22:49:56Z","lastTransitionTime":"2025-12-10T22:49:56Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 22:49:57 crc kubenswrapper[4791]: E1210 22:49:57.007217 4791 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-12-10T22:49:56Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-10T22:49:56Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-10T22:49:56Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-10T22:49:56Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-10T22:49:56Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-10T22:49:56Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-10T22:49:56Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-10T22:49:56Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"4fab3af6-f657-4dfc-8ef2-12b7c978c94f\\\",\\\"systemUUID\\\":\\\"442c0b68-8dce-47e0-bd68-d5210d7e0493\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T22:49:57Z is after 2025-08-24T17:21:41Z" Dec 10 22:49:57 crc kubenswrapper[4791]: I1210 22:49:57.012700 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 22:49:57 crc kubenswrapper[4791]: I1210 22:49:57.012760 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 22:49:57 crc kubenswrapper[4791]: I1210 22:49:57.012778 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 22:49:57 crc kubenswrapper[4791]: I1210 22:49:57.012802 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 22:49:57 crc kubenswrapper[4791]: I1210 22:49:57.012819 4791 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T22:49:57Z","lastTransitionTime":"2025-12-10T22:49:57Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 22:49:57 crc kubenswrapper[4791]: E1210 22:49:57.028320 4791 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-12-10T22:49:57Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-10T22:49:57Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-10T22:49:57Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-10T22:49:57Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-10T22:49:57Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-10T22:49:57Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-10T22:49:57Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-10T22:49:57Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"4fab3af6-f657-4dfc-8ef2-12b7c978c94f\\\",\\\"systemUUID\\\":\\\"442c0b68-8dce-47e0-bd68-d5210d7e0493\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T22:49:57Z is after 2025-08-24T17:21:41Z" Dec 10 22:49:57 crc kubenswrapper[4791]: I1210 22:49:57.033298 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 22:49:57 crc kubenswrapper[4791]: I1210 22:49:57.033396 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 22:49:57 crc kubenswrapper[4791]: I1210 22:49:57.033418 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 22:49:57 crc kubenswrapper[4791]: I1210 22:49:57.033485 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 22:49:57 crc kubenswrapper[4791]: I1210 22:49:57.033515 4791 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T22:49:57Z","lastTransitionTime":"2025-12-10T22:49:57Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 22:49:57 crc kubenswrapper[4791]: E1210 22:49:57.056489 4791 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-12-10T22:49:57Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-10T22:49:57Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-10T22:49:57Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-10T22:49:57Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-10T22:49:57Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-10T22:49:57Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-10T22:49:57Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-10T22:49:57Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"4fab3af6-f657-4dfc-8ef2-12b7c978c94f\\\",\\\"systemUUID\\\":\\\"442c0b68-8dce-47e0-bd68-d5210d7e0493\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T22:49:57Z is after 2025-08-24T17:21:41Z" Dec 10 22:49:57 crc kubenswrapper[4791]: E1210 22:49:57.056849 4791 kubelet_node_status.go:572] "Unable to update node status" err="update node status exceeds retry count" Dec 10 22:49:57 crc kubenswrapper[4791]: I1210 22:49:57.060174 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 22:49:57 crc kubenswrapper[4791]: I1210 22:49:57.060257 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 22:49:57 crc kubenswrapper[4791]: I1210 22:49:57.060285 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 22:49:57 crc kubenswrapper[4791]: I1210 22:49:57.060317 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 22:49:57 crc kubenswrapper[4791]: I1210 22:49:57.060381 4791 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T22:49:57Z","lastTransitionTime":"2025-12-10T22:49:57Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 22:49:57 crc kubenswrapper[4791]: I1210 22:49:57.163266 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 22:49:57 crc kubenswrapper[4791]: I1210 22:49:57.163295 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 22:49:57 crc kubenswrapper[4791]: I1210 22:49:57.163305 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 22:49:57 crc kubenswrapper[4791]: I1210 22:49:57.163318 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 22:49:57 crc kubenswrapper[4791]: I1210 22:49:57.163327 4791 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T22:49:57Z","lastTransitionTime":"2025-12-10T22:49:57Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 22:49:57 crc kubenswrapper[4791]: I1210 22:49:57.266112 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 22:49:57 crc kubenswrapper[4791]: I1210 22:49:57.266164 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 22:49:57 crc kubenswrapper[4791]: I1210 22:49:57.266180 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 22:49:57 crc kubenswrapper[4791]: I1210 22:49:57.266204 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 22:49:57 crc kubenswrapper[4791]: I1210 22:49:57.266218 4791 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T22:49:57Z","lastTransitionTime":"2025-12-10T22:49:57Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 22:49:57 crc kubenswrapper[4791]: I1210 22:49:57.289172 4791 generic.go:334] "Generic (PLEG): container finished" podID="ff9beb90-69b0-4732-bf37-0b81f58ecc98" containerID="b86738401ece6fbc3083782720c66a515850ed78d4054f34fcabab7ff525b85f" exitCode=0 Dec 10 22:49:57 crc kubenswrapper[4791]: I1210 22:49:57.289242 4791 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-additional-cni-plugins-rw584" event={"ID":"ff9beb90-69b0-4732-bf37-0b81f58ecc98","Type":"ContainerDied","Data":"b86738401ece6fbc3083782720c66a515850ed78d4054f34fcabab7ff525b85f"} Dec 10 22:49:57 crc kubenswrapper[4791]: I1210 22:49:57.295651 4791 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-zhq64" event={"ID":"3cd47739-0fa9-4321-aff1-220f8721a0b3","Type":"ContainerStarted","Data":"dd3f80e5713221f96ed50da7f2e7fd5b9fabfcd5c1656da8d8a86d818a5fdc3f"} Dec 10 22:49:57 crc kubenswrapper[4791]: I1210 22:49:57.316310 4791 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"698bbe9a-fbe5-436d-a835-5e80df77ba24\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T22:49:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T22:49:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T22:49:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T22:49:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T22:49:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://224d73713312376a58ec6cc2bc8fdb6ffa44c4749ae0e7691d5fe27b9afca09f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T22:49:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://372eeaffd40992c6f06efc4df6dfcb44b31de64ee8256c21508cdd7c74ecb96e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T22:49:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://76b6013b062cac5b0b4d12a94f38ce7b43f9c7ea216063455d88a956ac04d26e\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T22:49:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://78e950e122e761d950da1a3c60c06476e53920e8795f6c61db22c01e22d5136e\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T22:49:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-10T22:49:25Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T22:49:57Z is after 2025-08-24T17:21:41Z" Dec 10 22:49:57 crc kubenswrapper[4791]: I1210 22:49:57.337249 4791 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-10T22:49:52Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-10T22:49:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-10T22:49:52Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3734b110802828632d44263ed0a43cabc4b710833f0be120d72db0940fe783e2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T22:49:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://794803ac9be3c52509e6dcb10854d8864e1a9e406762e57c7971b07c4da1220e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T22:49:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T22:49:57Z is after 2025-08-24T17:21:41Z" Dec 10 22:49:57 crc kubenswrapper[4791]: I1210 22:49:57.354969 4791 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-vczjq" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b9f00f25-3fa1-4b63-9710-fd352224b01b\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T22:49:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T22:49:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T22:49:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T22:49:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4698cf4c6b55b41e628448d37c2eb847c2777d0efbd93afcbf367d5e78623b7e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T22:49:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lrrpv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-10T22:49:50Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-vczjq\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T22:49:57Z is after 2025-08-24T17:21:41Z" Dec 10 22:49:57 crc kubenswrapper[4791]: I1210 22:49:57.368899 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 22:49:57 crc kubenswrapper[4791]: I1210 22:49:57.368947 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 22:49:57 crc kubenswrapper[4791]: I1210 22:49:57.368959 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 22:49:57 crc kubenswrapper[4791]: I1210 22:49:57.368978 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 22:49:57 crc kubenswrapper[4791]: I1210 22:49:57.368990 4791 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T22:49:57Z","lastTransitionTime":"2025-12-10T22:49:57Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 22:49:57 crc kubenswrapper[4791]: I1210 22:49:57.374833 4791 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-4nwkq" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"672aa28c-8169-49ed-87b8-21187d13a80c\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T22:49:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T22:49:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T22:49:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T22:49:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a2f2a62315529e148a10a33ccb241e9417fbde0fe42afb61cb7ee2438c74636f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T22:49:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xl62j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-10T22:49:51Z\\\"}}\" for pod \"openshift-multus\"/\"multus-4nwkq\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T22:49:57Z is after 2025-08-24T17:21:41Z" Dec 10 22:49:57 crc kubenswrapper[4791]: I1210 22:49:57.412514 4791 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-zhq64" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3cd47739-0fa9-4321-aff1-220f8721a0b3\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T22:49:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T22:49:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T22:49:52Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T22:49:52Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wvpw8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wvpw8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wvpw8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wvpw8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wvpw8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wvpw8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wvpw8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wvpw8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3afbd1ffcd2c4b40324aca6e52d6fc822db131c3f35a9af3a6f005c2cffdb7bf\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3afbd1ffcd2c4b40324aca6e52d6fc822db131c3f35a9af3a6f005c2cffdb7bf\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-10T22:49:53Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-10T22:49:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wvpw8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-10T22:49:52Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-zhq64\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T22:49:57Z is after 2025-08-24T17:21:41Z" Dec 10 22:49:57 crc kubenswrapper[4791]: I1210 22:49:57.428894 4791 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-10T22:49:52Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-10T22:49:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-10T22:49:52Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://20651eb0c16e2c71222f56da1ff755cb0e7b084886acbbb24f0ccaf57041ed85\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T22:49:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T22:49:57Z is after 2025-08-24T17:21:41Z" Dec 10 22:49:57 crc kubenswrapper[4791]: I1210 22:49:57.441962 4791 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-10T22:49:50Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-10T22:49:50Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T22:49:57Z is after 2025-08-24T17:21:41Z" Dec 10 22:49:57 crc kubenswrapper[4791]: I1210 22:49:57.457388 4791 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-10T22:49:50Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-10T22:49:50Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T22:49:57Z is after 2025-08-24T17:21:41Z" Dec 10 22:49:57 crc kubenswrapper[4791]: I1210 22:49:57.470504 4791 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-tchxk" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"c0ab74e0-42a9-4f42-ac80-16030cf4ffff\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T22:49:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T22:49:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T22:49:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T22:49:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://04e0fd1b53256b7a938e8ccab93e2a4eb73dd1a59e0b9799e09c55e511baf7fe\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T22:49:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zn2qt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-10T22:49:50Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-tchxk\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T22:49:57Z is after 2025-08-24T17:21:41Z" Dec 10 22:49:57 crc kubenswrapper[4791]: I1210 22:49:57.472049 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 22:49:57 crc kubenswrapper[4791]: I1210 22:49:57.472114 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 22:49:57 crc kubenswrapper[4791]: I1210 22:49:57.472136 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 22:49:57 crc kubenswrapper[4791]: I1210 22:49:57.472161 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 22:49:57 crc kubenswrapper[4791]: I1210 22:49:57.472179 4791 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T22:49:57Z","lastTransitionTime":"2025-12-10T22:49:57Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 22:49:57 crc kubenswrapper[4791]: I1210 22:49:57.485483 4791 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-rw584" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ff9beb90-69b0-4732-bf37-0b81f58ecc98\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T22:49:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T22:49:51Z\\\",\\\"message\\\":\\\"containers with incomplete status: [whereabouts-cni]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T22:49:51Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T22:49:51Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fksqn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ce67e54b696bb53ffbbe48c6ce71476b635e0868993e9c427b22f904891433ee\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ce67e54b696bb53ffbbe48c6ce71476b635e0868993e9c427b22f904891433ee\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-10T22:49:52Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-10T22:49:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fksqn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://818d89bfffb8a110670152c8480684665ea8db578e9768f524fbbe6a92f27c22\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://818d89bfffb8a110670152c8480684665ea8db578e9768f524fbbe6a92f27c22\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-10T22:49:53Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-10T22:49:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fksqn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://9aa9c06e894bb2a2ed3738defaeff68a0a0458ead48177f36b90770944c211a0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://9aa9c06e894bb2a2ed3738defaeff68a0a0458ead48177f36b90770944c211a0\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-10T22:49:54Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-10T22:49:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fksqn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://dfd6d03336dcea7529419a30e8d70d84e2290a0fa5b0e46d09ad341c190638d6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://dfd6d03336dcea7529419a30e8d70d84e2290a0fa5b0e46d09ad341c190638d6\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-10T22:49:55Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-10T22:49:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fksqn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b86738401ece6fbc3083782720c66a515850ed78d4054f34fcabab7ff525b85f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b86738401ece6fbc3083782720c66a515850ed78d4054f34fcabab7ff525b85f\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-10T22:49:56Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-10T22:49:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fksqn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fksqn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-10T22:49:51Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-rw584\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T22:49:57Z is after 2025-08-24T17:21:41Z" Dec 10 22:49:57 crc kubenswrapper[4791]: I1210 22:49:57.497514 4791 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f8349fea-c49b-49a1-b5ee-032c5dffe021\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T22:49:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T22:49:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T22:49:25Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T22:49:25Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T22:49:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://acafe12186df99d1059a31260de0eafce35ff7771bcb8f5c6753b72665d6f60b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T22:49:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ea4c346507862a8c12756a3061d6e25e788a8d79f4621b62f81f4322185c82a8\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T22:49:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://2e6e324ec6f4f7bb69fdfab4ea3ac1aa55e480dd8e2a8984b0fe5b4c9f66636e\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T22:49:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://1b2e11ff2bf8789fa1d0d23ae648b3e6ef653106e5bc4bbb5dd6447f7cd6f05d\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d6aa7f8c5d908d90008f2df49860ac8a3480136da567e3d85f86135d779204f9\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-12-10T22:49:44Z\\\",\\\"message\\\":\\\"W1210 22:49:33.845733 1 cmd.go:257] Using insecure, self-signed certificates\\\\nI1210 22:49:33.846108 1 crypto.go:601] Generating new CA for check-endpoints-signer@1765406973 cert, and key in /tmp/serving-cert-2585701456/serving-signer.crt, /tmp/serving-cert-2585701456/serving-signer.key\\\\nI1210 22:49:34.529748 1 observer_polling.go:159] Starting file observer\\\\nW1210 22:49:34.532225 1 builder.go:272] unable to get owner reference (falling back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": dial tcp [::1]:6443: connect: connection refused\\\\nI1210 22:49:34.532599 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1210 22:49:34.533977 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-2585701456/tls.crt::/tmp/serving-cert-2585701456/tls.key\\\\\\\"\\\\nF1210 22:49:44.774710 1 cmd.go:182] error initializing delegating authentication: unable to load configmap based request-header-client-ca-file: Get \\\\\\\"https://localhost:6443/api/v1/namespaces/kube-system/configmaps/extension-apiserver-authentication\\\\\\\": net/http: TLS handshake timeout\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-10T22:49:33Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T22:49:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://0c4d036ab269f62561fe7ef1586012cb3a16e3583a08b9488a0616ec00c8e0ae\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T22:49:32Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c5856d6b26c19fda13b5c9ac2892a7e2f9c6f79feb8173ecfd9b33b673327d67\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c5856d6b26c19fda13b5c9ac2892a7e2f9c6f79feb8173ecfd9b33b673327d67\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-10T22:49:27Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-10T22:49:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-10T22:49:25Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T22:49:57Z is after 2025-08-24T17:21:41Z" Dec 10 22:49:57 crc kubenswrapper[4791]: I1210 22:49:57.516659 4791 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"90250e13-6c63-443a-a5e5-6e4e341a2289\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T22:49:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T22:49:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T22:49:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T22:49:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T22:49:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ba54f95b1e00843b854db9a10fe6897592159ccacf30ed0f76002093a13eedd3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T22:49:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5aa8ffa518f3a68b60b275f8df4851b2564aa4984b61efdc9470b73072133e43\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T22:49:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://16f989503edc78d9451638e2cf6cf4b19a2cf2b8809b10b237541cb829303091\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T22:49:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://82556824d63ce6e46df14c45689ff1f81ebf729abe902061815c040b323898e9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T22:49:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://61db531940fa6f2effdae5a005b4c91bd2698bd5d1bff9588c171a65ad550eca\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T22:49:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0f65c864d885319ce2a08d61c30be24b3ffb7d23ea30c82ed7faf039cb2c006d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://0f65c864d885319ce2a08d61c30be24b3ffb7d23ea30c82ed7faf039cb2c006d\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-10T22:49:27Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-10T22:49:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e31a76d1d5b2812ba83e24e38268e25ec8bed98fc3da9adcd06bda5822b6ace1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e31a76d1d5b2812ba83e24e38268e25ec8bed98fc3da9adcd06bda5822b6ace1\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-10T22:49:30Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-10T22:49:29Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://b5b5eceae562f3986b543d41de1e977efb85a91d84c98e09159ca6d092eda691\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b5b5eceae562f3986b543d41de1e977efb85a91d84c98e09159ca6d092eda691\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-10T22:49:31Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-10T22:49:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-10T22:49:25Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T22:49:57Z is after 2025-08-24T17:21:41Z" Dec 10 22:49:57 crc kubenswrapper[4791]: I1210 22:49:57.528420 4791 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-10T22:49:50Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-10T22:49:50Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T22:49:57Z is after 2025-08-24T17:21:41Z" Dec 10 22:49:57 crc kubenswrapper[4791]: I1210 22:49:57.539328 4791 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-10T22:49:54Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-10T22:49:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-10T22:49:54Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://62bfaf5d7a6d1b452cdaf88086447d0c8cb7306e94434e46752be5475124d56a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T22:49:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T22:49:57Z is after 2025-08-24T17:21:41Z" Dec 10 22:49:57 crc kubenswrapper[4791]: I1210 22:49:57.555998 4791 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-5rb5l" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ba35653c-6e06-4cee-a4d6-137764090d18\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T22:49:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T22:49:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T22:49:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T22:49:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://dbb5c875730a98ad2daf316cfafc08fcdce21ac16148154780ac9193ed405dab\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T22:49:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fvvrs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://185eb2ff44221a514878283ec873773a6c9a9073dc64e296f982e60dee1338e4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T22:49:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fvvrs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-10T22:49:51Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-5rb5l\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T22:49:57Z is after 2025-08-24T17:21:41Z" Dec 10 22:49:57 crc kubenswrapper[4791]: I1210 22:49:57.574228 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 22:49:57 crc kubenswrapper[4791]: I1210 22:49:57.574264 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 22:49:57 crc kubenswrapper[4791]: I1210 22:49:57.574273 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 22:49:57 crc kubenswrapper[4791]: I1210 22:49:57.574289 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 22:49:57 crc kubenswrapper[4791]: I1210 22:49:57.574301 4791 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T22:49:57Z","lastTransitionTime":"2025-12-10T22:49:57Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 22:49:57 crc kubenswrapper[4791]: I1210 22:49:57.676480 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 22:49:57 crc kubenswrapper[4791]: I1210 22:49:57.676546 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 22:49:57 crc kubenswrapper[4791]: I1210 22:49:57.676568 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 22:49:57 crc kubenswrapper[4791]: I1210 22:49:57.676598 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 22:49:57 crc kubenswrapper[4791]: I1210 22:49:57.676624 4791 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T22:49:57Z","lastTransitionTime":"2025-12-10T22:49:57Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 22:49:57 crc kubenswrapper[4791]: I1210 22:49:57.779939 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 22:49:57 crc kubenswrapper[4791]: I1210 22:49:57.780010 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 22:49:57 crc kubenswrapper[4791]: I1210 22:49:57.780025 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 22:49:57 crc kubenswrapper[4791]: I1210 22:49:57.780043 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 22:49:57 crc kubenswrapper[4791]: I1210 22:49:57.780054 4791 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T22:49:57Z","lastTransitionTime":"2025-12-10T22:49:57Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 22:49:57 crc kubenswrapper[4791]: I1210 22:49:57.884110 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 22:49:57 crc kubenswrapper[4791]: I1210 22:49:57.884179 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 22:49:57 crc kubenswrapper[4791]: I1210 22:49:57.884203 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 22:49:57 crc kubenswrapper[4791]: I1210 22:49:57.884229 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 22:49:57 crc kubenswrapper[4791]: I1210 22:49:57.884249 4791 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T22:49:57Z","lastTransitionTime":"2025-12-10T22:49:57Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 22:49:57 crc kubenswrapper[4791]: I1210 22:49:57.986646 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 22:49:57 crc kubenswrapper[4791]: I1210 22:49:57.986707 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 22:49:57 crc kubenswrapper[4791]: I1210 22:49:57.986723 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 22:49:57 crc kubenswrapper[4791]: I1210 22:49:57.986746 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 22:49:57 crc kubenswrapper[4791]: I1210 22:49:57.986763 4791 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T22:49:57Z","lastTransitionTime":"2025-12-10T22:49:57Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 22:49:58 crc kubenswrapper[4791]: I1210 22:49:58.089785 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 22:49:58 crc kubenswrapper[4791]: I1210 22:49:58.089852 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 22:49:58 crc kubenswrapper[4791]: I1210 22:49:58.089870 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 22:49:58 crc kubenswrapper[4791]: I1210 22:49:58.089894 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 22:49:58 crc kubenswrapper[4791]: I1210 22:49:58.089911 4791 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T22:49:58Z","lastTransitionTime":"2025-12-10T22:49:58Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 22:49:58 crc kubenswrapper[4791]: I1210 22:49:58.184369 4791 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-kube-apiserver/kube-apiserver-crc" Dec 10 22:49:58 crc kubenswrapper[4791]: I1210 22:49:58.193535 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 22:49:58 crc kubenswrapper[4791]: I1210 22:49:58.193568 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 22:49:58 crc kubenswrapper[4791]: I1210 22:49:58.193579 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 22:49:58 crc kubenswrapper[4791]: I1210 22:49:58.193595 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 22:49:58 crc kubenswrapper[4791]: I1210 22:49:58.193607 4791 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T22:49:58Z","lastTransitionTime":"2025-12-10T22:49:58Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 22:49:58 crc kubenswrapper[4791]: I1210 22:49:58.202598 4791 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-vczjq" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b9f00f25-3fa1-4b63-9710-fd352224b01b\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T22:49:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T22:49:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T22:49:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T22:49:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4698cf4c6b55b41e628448d37c2eb847c2777d0efbd93afcbf367d5e78623b7e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T22:49:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lrrpv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-10T22:49:50Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-vczjq\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T22:49:58Z is after 2025-08-24T17:21:41Z" Dec 10 22:49:58 crc kubenswrapper[4791]: I1210 22:49:58.222322 4791 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-4nwkq" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"672aa28c-8169-49ed-87b8-21187d13a80c\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T22:49:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T22:49:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T22:49:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T22:49:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a2f2a62315529e148a10a33ccb241e9417fbde0fe42afb61cb7ee2438c74636f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T22:49:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xl62j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-10T22:49:51Z\\\"}}\" for pod \"openshift-multus\"/\"multus-4nwkq\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T22:49:58Z is after 2025-08-24T17:21:41Z" Dec 10 22:49:58 crc kubenswrapper[4791]: I1210 22:49:58.253424 4791 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-zhq64" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3cd47739-0fa9-4321-aff1-220f8721a0b3\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T22:49:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T22:49:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T22:49:52Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T22:49:52Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wvpw8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wvpw8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wvpw8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wvpw8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wvpw8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wvpw8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wvpw8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wvpw8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3afbd1ffcd2c4b40324aca6e52d6fc822db131c3f35a9af3a6f005c2cffdb7bf\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3afbd1ffcd2c4b40324aca6e52d6fc822db131c3f35a9af3a6f005c2cffdb7bf\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-10T22:49:53Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-10T22:49:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wvpw8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-10T22:49:52Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-zhq64\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T22:49:58Z is after 2025-08-24T17:21:41Z" Dec 10 22:49:58 crc kubenswrapper[4791]: I1210 22:49:58.270667 4791 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"698bbe9a-fbe5-436d-a835-5e80df77ba24\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T22:49:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T22:49:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T22:49:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T22:49:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T22:49:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://224d73713312376a58ec6cc2bc8fdb6ffa44c4749ae0e7691d5fe27b9afca09f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T22:49:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://372eeaffd40992c6f06efc4df6dfcb44b31de64ee8256c21508cdd7c74ecb96e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T22:49:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://76b6013b062cac5b0b4d12a94f38ce7b43f9c7ea216063455d88a956ac04d26e\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T22:49:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://78e950e122e761d950da1a3c60c06476e53920e8795f6c61db22c01e22d5136e\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T22:49:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-10T22:49:25Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T22:49:58Z is after 2025-08-24T17:21:41Z" Dec 10 22:49:58 crc kubenswrapper[4791]: I1210 22:49:58.288256 4791 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-10T22:49:52Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-10T22:49:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-10T22:49:52Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3734b110802828632d44263ed0a43cabc4b710833f0be120d72db0940fe783e2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T22:49:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://794803ac9be3c52509e6dcb10854d8864e1a9e406762e57c7971b07c4da1220e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T22:49:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T22:49:58Z is after 2025-08-24T17:21:41Z" Dec 10 22:49:58 crc kubenswrapper[4791]: I1210 22:49:58.295817 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 22:49:58 crc kubenswrapper[4791]: I1210 22:49:58.295855 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 22:49:58 crc kubenswrapper[4791]: I1210 22:49:58.295872 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 22:49:58 crc kubenswrapper[4791]: I1210 22:49:58.295895 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 22:49:58 crc kubenswrapper[4791]: I1210 22:49:58.295913 4791 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T22:49:58Z","lastTransitionTime":"2025-12-10T22:49:58Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 22:49:58 crc kubenswrapper[4791]: I1210 22:49:58.300050 4791 generic.go:334] "Generic (PLEG): container finished" podID="ff9beb90-69b0-4732-bf37-0b81f58ecc98" containerID="18b326f572a899ddd10197a745bf9264d4ddd975690ad1cfb0ec576ccd5cf102" exitCode=0 Dec 10 22:49:58 crc kubenswrapper[4791]: I1210 22:49:58.300100 4791 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-additional-cni-plugins-rw584" event={"ID":"ff9beb90-69b0-4732-bf37-0b81f58ecc98","Type":"ContainerDied","Data":"18b326f572a899ddd10197a745bf9264d4ddd975690ad1cfb0ec576ccd5cf102"} Dec 10 22:49:58 crc kubenswrapper[4791]: I1210 22:49:58.304479 4791 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-tchxk" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"c0ab74e0-42a9-4f42-ac80-16030cf4ffff\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T22:49:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T22:49:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T22:49:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T22:49:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://04e0fd1b53256b7a938e8ccab93e2a4eb73dd1a59e0b9799e09c55e511baf7fe\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T22:49:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zn2qt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-10T22:49:50Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-tchxk\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T22:49:58Z is after 2025-08-24T17:21:41Z" Dec 10 22:49:58 crc kubenswrapper[4791]: I1210 22:49:58.330568 4791 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-rw584" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ff9beb90-69b0-4732-bf37-0b81f58ecc98\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T22:49:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T22:49:51Z\\\",\\\"message\\\":\\\"containers with incomplete status: [whereabouts-cni]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T22:49:51Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T22:49:51Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fksqn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ce67e54b696bb53ffbbe48c6ce71476b635e0868993e9c427b22f904891433ee\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ce67e54b696bb53ffbbe48c6ce71476b635e0868993e9c427b22f904891433ee\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-10T22:49:52Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-10T22:49:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fksqn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://818d89bfffb8a110670152c8480684665ea8db578e9768f524fbbe6a92f27c22\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://818d89bfffb8a110670152c8480684665ea8db578e9768f524fbbe6a92f27c22\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-10T22:49:53Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-10T22:49:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fksqn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://9aa9c06e894bb2a2ed3738defaeff68a0a0458ead48177f36b90770944c211a0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://9aa9c06e894bb2a2ed3738defaeff68a0a0458ead48177f36b90770944c211a0\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-10T22:49:54Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-10T22:49:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fksqn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://dfd6d03336dcea7529419a30e8d70d84e2290a0fa5b0e46d09ad341c190638d6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://dfd6d03336dcea7529419a30e8d70d84e2290a0fa5b0e46d09ad341c190638d6\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-10T22:49:55Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-10T22:49:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fksqn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b86738401ece6fbc3083782720c66a515850ed78d4054f34fcabab7ff525b85f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b86738401ece6fbc3083782720c66a515850ed78d4054f34fcabab7ff525b85f\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-10T22:49:56Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-10T22:49:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fksqn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fksqn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-10T22:49:51Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-rw584\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T22:49:58Z is after 2025-08-24T17:21:41Z" Dec 10 22:49:58 crc kubenswrapper[4791]: I1210 22:49:58.344904 4791 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-10T22:49:52Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-10T22:49:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-10T22:49:52Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://20651eb0c16e2c71222f56da1ff755cb0e7b084886acbbb24f0ccaf57041ed85\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T22:49:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T22:49:58Z is after 2025-08-24T17:21:41Z" Dec 10 22:49:58 crc kubenswrapper[4791]: I1210 22:49:58.356350 4791 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-10T22:49:50Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-10T22:49:50Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T22:49:58Z is after 2025-08-24T17:21:41Z" Dec 10 22:49:58 crc kubenswrapper[4791]: I1210 22:49:58.371809 4791 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-10T22:49:50Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-10T22:49:50Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T22:49:58Z is after 2025-08-24T17:21:41Z" Dec 10 22:49:58 crc kubenswrapper[4791]: I1210 22:49:58.388803 4791 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-10T22:49:50Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-10T22:49:50Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T22:49:58Z is after 2025-08-24T17:21:41Z" Dec 10 22:49:58 crc kubenswrapper[4791]: I1210 22:49:58.399410 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 22:49:58 crc kubenswrapper[4791]: I1210 22:49:58.399446 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 22:49:58 crc kubenswrapper[4791]: I1210 22:49:58.399459 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 22:49:58 crc kubenswrapper[4791]: I1210 22:49:58.399475 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 22:49:58 crc kubenswrapper[4791]: I1210 22:49:58.399486 4791 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T22:49:58Z","lastTransitionTime":"2025-12-10T22:49:58Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 22:49:58 crc kubenswrapper[4791]: I1210 22:49:58.401421 4791 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-10T22:49:54Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-10T22:49:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-10T22:49:54Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://62bfaf5d7a6d1b452cdaf88086447d0c8cb7306e94434e46752be5475124d56a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T22:49:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T22:49:58Z is after 2025-08-24T17:21:41Z" Dec 10 22:49:58 crc kubenswrapper[4791]: I1210 22:49:58.418608 4791 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f8349fea-c49b-49a1-b5ee-032c5dffe021\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T22:49:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T22:49:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T22:49:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T22:49:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T22:49:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://acafe12186df99d1059a31260de0eafce35ff7771bcb8f5c6753b72665d6f60b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T22:49:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ea4c346507862a8c12756a3061d6e25e788a8d79f4621b62f81f4322185c82a8\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T22:49:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://2e6e324ec6f4f7bb69fdfab4ea3ac1aa55e480dd8e2a8984b0fe5b4c9f66636e\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T22:49:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://1b2e11ff2bf8789fa1d0d23ae648b3e6ef653106e5bc4bbb5dd6447f7cd6f05d\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d6aa7f8c5d908d90008f2df49860ac8a3480136da567e3d85f86135d779204f9\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-12-10T22:49:44Z\\\",\\\"message\\\":\\\"W1210 22:49:33.845733 1 cmd.go:257] Using insecure, self-signed certificates\\\\nI1210 22:49:33.846108 1 crypto.go:601] Generating new CA for check-endpoints-signer@1765406973 cert, and key in /tmp/serving-cert-2585701456/serving-signer.crt, /tmp/serving-cert-2585701456/serving-signer.key\\\\nI1210 22:49:34.529748 1 observer_polling.go:159] Starting file observer\\\\nW1210 22:49:34.532225 1 builder.go:272] unable to get owner reference (falling back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": dial tcp [::1]:6443: connect: connection refused\\\\nI1210 22:49:34.532599 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1210 22:49:34.533977 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-2585701456/tls.crt::/tmp/serving-cert-2585701456/tls.key\\\\\\\"\\\\nF1210 22:49:44.774710 1 cmd.go:182] error initializing delegating authentication: unable to load configmap based request-header-client-ca-file: Get \\\\\\\"https://localhost:6443/api/v1/namespaces/kube-system/configmaps/extension-apiserver-authentication\\\\\\\": net/http: TLS handshake timeout\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-10T22:49:33Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T22:49:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://0c4d036ab269f62561fe7ef1586012cb3a16e3583a08b9488a0616ec00c8e0ae\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T22:49:32Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c5856d6b26c19fda13b5c9ac2892a7e2f9c6f79feb8173ecfd9b33b673327d67\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c5856d6b26c19fda13b5c9ac2892a7e2f9c6f79feb8173ecfd9b33b673327d67\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-10T22:49:27Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-10T22:49:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-10T22:49:25Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T22:49:58Z is after 2025-08-24T17:21:41Z" Dec 10 22:49:58 crc kubenswrapper[4791]: I1210 22:49:58.455841 4791 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"90250e13-6c63-443a-a5e5-6e4e341a2289\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T22:49:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T22:49:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T22:49:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T22:49:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T22:49:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ba54f95b1e00843b854db9a10fe6897592159ccacf30ed0f76002093a13eedd3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T22:49:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5aa8ffa518f3a68b60b275f8df4851b2564aa4984b61efdc9470b73072133e43\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T22:49:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://16f989503edc78d9451638e2cf6cf4b19a2cf2b8809b10b237541cb829303091\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T22:49:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://82556824d63ce6e46df14c45689ff1f81ebf729abe902061815c040b323898e9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T22:49:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://61db531940fa6f2effdae5a005b4c91bd2698bd5d1bff9588c171a65ad550eca\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T22:49:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0f65c864d885319ce2a08d61c30be24b3ffb7d23ea30c82ed7faf039cb2c006d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://0f65c864d885319ce2a08d61c30be24b3ffb7d23ea30c82ed7faf039cb2c006d\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-10T22:49:27Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-10T22:49:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e31a76d1d5b2812ba83e24e38268e25ec8bed98fc3da9adcd06bda5822b6ace1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e31a76d1d5b2812ba83e24e38268e25ec8bed98fc3da9adcd06bda5822b6ace1\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-10T22:49:30Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-10T22:49:29Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://b5b5eceae562f3986b543d41de1e977efb85a91d84c98e09159ca6d092eda691\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b5b5eceae562f3986b543d41de1e977efb85a91d84c98e09159ca6d092eda691\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-10T22:49:31Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-10T22:49:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-10T22:49:25Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T22:49:58Z is after 2025-08-24T17:21:41Z" Dec 10 22:49:58 crc kubenswrapper[4791]: I1210 22:49:58.467262 4791 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-5rb5l" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ba35653c-6e06-4cee-a4d6-137764090d18\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T22:49:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T22:49:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T22:49:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T22:49:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://dbb5c875730a98ad2daf316cfafc08fcdce21ac16148154780ac9193ed405dab\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T22:49:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fvvrs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://185eb2ff44221a514878283ec873773a6c9a9073dc64e296f982e60dee1338e4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T22:49:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fvvrs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-10T22:49:51Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-5rb5l\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T22:49:58Z is after 2025-08-24T17:21:41Z" Dec 10 22:49:58 crc kubenswrapper[4791]: I1210 22:49:58.479097 4791 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-10T22:49:52Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-10T22:49:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-10T22:49:52Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://20651eb0c16e2c71222f56da1ff755cb0e7b084886acbbb24f0ccaf57041ed85\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T22:49:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T22:49:58Z is after 2025-08-24T17:21:41Z" Dec 10 22:49:58 crc kubenswrapper[4791]: I1210 22:49:58.489998 4791 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-10T22:49:50Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-10T22:49:50Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T22:49:58Z is after 2025-08-24T17:21:41Z" Dec 10 22:49:58 crc kubenswrapper[4791]: I1210 22:49:58.501751 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 22:49:58 crc kubenswrapper[4791]: I1210 22:49:58.501786 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 22:49:58 crc kubenswrapper[4791]: I1210 22:49:58.501799 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 22:49:58 crc kubenswrapper[4791]: I1210 22:49:58.501816 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 22:49:58 crc kubenswrapper[4791]: I1210 22:49:58.501829 4791 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T22:49:58Z","lastTransitionTime":"2025-12-10T22:49:58Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 22:49:58 crc kubenswrapper[4791]: I1210 22:49:58.501834 4791 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-10T22:49:50Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-10T22:49:50Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T22:49:58Z is after 2025-08-24T17:21:41Z" Dec 10 22:49:58 crc kubenswrapper[4791]: I1210 22:49:58.510391 4791 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-tchxk" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"c0ab74e0-42a9-4f42-ac80-16030cf4ffff\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T22:49:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T22:49:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T22:49:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T22:49:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://04e0fd1b53256b7a938e8ccab93e2a4eb73dd1a59e0b9799e09c55e511baf7fe\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T22:49:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zn2qt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-10T22:49:50Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-tchxk\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T22:49:58Z is after 2025-08-24T17:21:41Z" Dec 10 22:49:58 crc kubenswrapper[4791]: I1210 22:49:58.521288 4791 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-rw584" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ff9beb90-69b0-4732-bf37-0b81f58ecc98\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T22:49:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T22:49:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T22:49:51Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T22:49:51Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fksqn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ce67e54b696bb53ffbbe48c6ce71476b635e0868993e9c427b22f904891433ee\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ce67e54b696bb53ffbbe48c6ce71476b635e0868993e9c427b22f904891433ee\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-10T22:49:52Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-10T22:49:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fksqn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://818d89bfffb8a110670152c8480684665ea8db578e9768f524fbbe6a92f27c22\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://818d89bfffb8a110670152c8480684665ea8db578e9768f524fbbe6a92f27c22\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-10T22:49:53Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-10T22:49:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fksqn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://9aa9c06e894bb2a2ed3738defaeff68a0a0458ead48177f36b90770944c211a0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://9aa9c06e894bb2a2ed3738defaeff68a0a0458ead48177f36b90770944c211a0\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-10T22:49:54Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-10T22:49:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fksqn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://dfd6d03336dcea7529419a30e8d70d84e2290a0fa5b0e46d09ad341c190638d6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://dfd6d03336dcea7529419a30e8d70d84e2290a0fa5b0e46d09ad341c190638d6\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-10T22:49:55Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-10T22:49:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fksqn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b86738401ece6fbc3083782720c66a515850ed78d4054f34fcabab7ff525b85f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b86738401ece6fbc3083782720c66a515850ed78d4054f34fcabab7ff525b85f\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-10T22:49:56Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-10T22:49:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fksqn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://18b326f572a899ddd10197a745bf9264d4ddd975690ad1cfb0ec576ccd5cf102\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://18b326f572a899ddd10197a745bf9264d4ddd975690ad1cfb0ec576ccd5cf102\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-10T22:49:57Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-10T22:49:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fksqn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-10T22:49:51Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-rw584\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T22:49:58Z is after 2025-08-24T17:21:41Z" Dec 10 22:49:58 crc kubenswrapper[4791]: I1210 22:49:58.532871 4791 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f8349fea-c49b-49a1-b5ee-032c5dffe021\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T22:49:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T22:49:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T22:49:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T22:49:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T22:49:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://acafe12186df99d1059a31260de0eafce35ff7771bcb8f5c6753b72665d6f60b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T22:49:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ea4c346507862a8c12756a3061d6e25e788a8d79f4621b62f81f4322185c82a8\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T22:49:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://2e6e324ec6f4f7bb69fdfab4ea3ac1aa55e480dd8e2a8984b0fe5b4c9f66636e\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T22:49:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://1b2e11ff2bf8789fa1d0d23ae648b3e6ef653106e5bc4bbb5dd6447f7cd6f05d\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d6aa7f8c5d908d90008f2df49860ac8a3480136da567e3d85f86135d779204f9\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-12-10T22:49:44Z\\\",\\\"message\\\":\\\"W1210 22:49:33.845733 1 cmd.go:257] Using insecure, self-signed certificates\\\\nI1210 22:49:33.846108 1 crypto.go:601] Generating new CA for check-endpoints-signer@1765406973 cert, and key in /tmp/serving-cert-2585701456/serving-signer.crt, /tmp/serving-cert-2585701456/serving-signer.key\\\\nI1210 22:49:34.529748 1 observer_polling.go:159] Starting file observer\\\\nW1210 22:49:34.532225 1 builder.go:272] unable to get owner reference (falling back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": dial tcp [::1]:6443: connect: connection refused\\\\nI1210 22:49:34.532599 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1210 22:49:34.533977 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-2585701456/tls.crt::/tmp/serving-cert-2585701456/tls.key\\\\\\\"\\\\nF1210 22:49:44.774710 1 cmd.go:182] error initializing delegating authentication: unable to load configmap based request-header-client-ca-file: Get \\\\\\\"https://localhost:6443/api/v1/namespaces/kube-system/configmaps/extension-apiserver-authentication\\\\\\\": net/http: TLS handshake timeout\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-10T22:49:33Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T22:49:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://0c4d036ab269f62561fe7ef1586012cb3a16e3583a08b9488a0616ec00c8e0ae\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T22:49:32Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c5856d6b26c19fda13b5c9ac2892a7e2f9c6f79feb8173ecfd9b33b673327d67\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c5856d6b26c19fda13b5c9ac2892a7e2f9c6f79feb8173ecfd9b33b673327d67\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-10T22:49:27Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-10T22:49:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-10T22:49:25Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T22:49:58Z is after 2025-08-24T17:21:41Z" Dec 10 22:49:58 crc kubenswrapper[4791]: I1210 22:49:58.548663 4791 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"90250e13-6c63-443a-a5e5-6e4e341a2289\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T22:49:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T22:49:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T22:49:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T22:49:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T22:49:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ba54f95b1e00843b854db9a10fe6897592159ccacf30ed0f76002093a13eedd3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T22:49:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5aa8ffa518f3a68b60b275f8df4851b2564aa4984b61efdc9470b73072133e43\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T22:49:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://16f989503edc78d9451638e2cf6cf4b19a2cf2b8809b10b237541cb829303091\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T22:49:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://82556824d63ce6e46df14c45689ff1f81ebf729abe902061815c040b323898e9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T22:49:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://61db531940fa6f2effdae5a005b4c91bd2698bd5d1bff9588c171a65ad550eca\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T22:49:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0f65c864d885319ce2a08d61c30be24b3ffb7d23ea30c82ed7faf039cb2c006d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://0f65c864d885319ce2a08d61c30be24b3ffb7d23ea30c82ed7faf039cb2c006d\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-10T22:49:27Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-10T22:49:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e31a76d1d5b2812ba83e24e38268e25ec8bed98fc3da9adcd06bda5822b6ace1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e31a76d1d5b2812ba83e24e38268e25ec8bed98fc3da9adcd06bda5822b6ace1\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-10T22:49:30Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-10T22:49:29Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://b5b5eceae562f3986b543d41de1e977efb85a91d84c98e09159ca6d092eda691\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b5b5eceae562f3986b543d41de1e977efb85a91d84c98e09159ca6d092eda691\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-10T22:49:31Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-10T22:49:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-10T22:49:25Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T22:49:58Z is after 2025-08-24T17:21:41Z" Dec 10 22:49:58 crc kubenswrapper[4791]: I1210 22:49:58.562452 4791 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-10T22:49:50Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-10T22:49:50Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T22:49:58Z is after 2025-08-24T17:21:41Z" Dec 10 22:49:58 crc kubenswrapper[4791]: I1210 22:49:58.574726 4791 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-10T22:49:54Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-10T22:49:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-10T22:49:54Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://62bfaf5d7a6d1b452cdaf88086447d0c8cb7306e94434e46752be5475124d56a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T22:49:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T22:49:58Z is after 2025-08-24T17:21:41Z" Dec 10 22:49:58 crc kubenswrapper[4791]: I1210 22:49:58.585788 4791 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 10 22:49:58 crc kubenswrapper[4791]: E1210 22:49:58.585939 4791 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-10 22:50:06.585919613 +0000 UTC m=+41.015537236 (durationBeforeRetry 8s). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 10 22:49:58 crc kubenswrapper[4791]: I1210 22:49:58.586516 4791 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-5rb5l" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ba35653c-6e06-4cee-a4d6-137764090d18\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T22:49:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T22:49:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T22:49:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T22:49:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://dbb5c875730a98ad2daf316cfafc08fcdce21ac16148154780ac9193ed405dab\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T22:49:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fvvrs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://185eb2ff44221a514878283ec873773a6c9a9073dc64e296f982e60dee1338e4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T22:49:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fvvrs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-10T22:49:51Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-5rb5l\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T22:49:58Z is after 2025-08-24T17:21:41Z" Dec 10 22:49:58 crc kubenswrapper[4791]: I1210 22:49:58.600112 4791 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"698bbe9a-fbe5-436d-a835-5e80df77ba24\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T22:49:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T22:49:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T22:49:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T22:49:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T22:49:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://224d73713312376a58ec6cc2bc8fdb6ffa44c4749ae0e7691d5fe27b9afca09f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T22:49:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://372eeaffd40992c6f06efc4df6dfcb44b31de64ee8256c21508cdd7c74ecb96e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T22:49:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://76b6013b062cac5b0b4d12a94f38ce7b43f9c7ea216063455d88a956ac04d26e\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T22:49:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://78e950e122e761d950da1a3c60c06476e53920e8795f6c61db22c01e22d5136e\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T22:49:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-10T22:49:25Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T22:49:58Z is after 2025-08-24T17:21:41Z" Dec 10 22:49:58 crc kubenswrapper[4791]: I1210 22:49:58.603873 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 22:49:58 crc kubenswrapper[4791]: I1210 22:49:58.603941 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 22:49:58 crc kubenswrapper[4791]: I1210 22:49:58.603951 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 22:49:58 crc kubenswrapper[4791]: I1210 22:49:58.603967 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 22:49:58 crc kubenswrapper[4791]: I1210 22:49:58.603996 4791 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T22:49:58Z","lastTransitionTime":"2025-12-10T22:49:58Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 22:49:58 crc kubenswrapper[4791]: I1210 22:49:58.641649 4791 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-10T22:49:52Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-10T22:49:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-10T22:49:52Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3734b110802828632d44263ed0a43cabc4b710833f0be120d72db0940fe783e2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T22:49:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://794803ac9be3c52509e6dcb10854d8864e1a9e406762e57c7971b07c4da1220e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T22:49:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T22:49:58Z is after 2025-08-24T17:21:41Z" Dec 10 22:49:58 crc kubenswrapper[4791]: I1210 22:49:58.681407 4791 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-vczjq" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b9f00f25-3fa1-4b63-9710-fd352224b01b\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T22:49:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T22:49:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T22:49:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T22:49:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4698cf4c6b55b41e628448d37c2eb847c2777d0efbd93afcbf367d5e78623b7e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T22:49:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lrrpv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-10T22:49:50Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-vczjq\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T22:49:58Z is after 2025-08-24T17:21:41Z" Dec 10 22:49:58 crc kubenswrapper[4791]: I1210 22:49:58.686960 4791 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nginx-conf\" (UniqueName: \"kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 10 22:49:58 crc kubenswrapper[4791]: I1210 22:49:58.687027 4791 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cqllr\" (UniqueName: \"kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr\") pod \"network-check-target-xd92c\" (UID: \"3b6479f0-333b-4a96-9adf-2099afdc2447\") " pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 10 22:49:58 crc kubenswrapper[4791]: I1210 22:49:58.687071 4791 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-s2dwl\" (UniqueName: \"kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl\") pod \"network-check-source-55646444c4-trplf\" (UID: \"9d751cbb-f2e2-430d-9754-c882a5e924a5\") " pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 10 22:49:58 crc kubenswrapper[4791]: I1210 22:49:58.687120 4791 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"networking-console-plugin-cert\" (UniqueName: \"kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 10 22:49:58 crc kubenswrapper[4791]: E1210 22:49:58.687217 4791 configmap.go:193] Couldn't get configMap openshift-network-console/networking-console-plugin: object "openshift-network-console"/"networking-console-plugin" not registered Dec 10 22:49:58 crc kubenswrapper[4791]: E1210 22:49:58.687276 4791 secret.go:188] Couldn't get secret openshift-network-console/networking-console-plugin-cert: object "openshift-network-console"/"networking-console-plugin-cert" not registered Dec 10 22:49:58 crc kubenswrapper[4791]: E1210 22:49:58.687319 4791 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Dec 10 22:49:58 crc kubenswrapper[4791]: E1210 22:49:58.687394 4791 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Dec 10 22:49:58 crc kubenswrapper[4791]: E1210 22:49:58.687435 4791 projected.go:194] Error preparing data for projected volume kube-api-access-s2dwl for pod openshift-network-diagnostics/network-check-source-55646444c4-trplf: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Dec 10 22:49:58 crc kubenswrapper[4791]: E1210 22:49:58.687325 4791 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2025-12-10 22:50:06.687299854 +0000 UTC m=+41.116917497 (durationBeforeRetry 8s). Error: MountVolume.SetUp failed for volume "nginx-conf" (UniqueName: "kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin" not registered Dec 10 22:49:58 crc kubenswrapper[4791]: E1210 22:49:58.687240 4791 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Dec 10 22:49:58 crc kubenswrapper[4791]: E1210 22:49:58.687581 4791 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2025-12-10 22:50:06.68751802 +0000 UTC m=+41.117135693 (durationBeforeRetry 8s). Error: MountVolume.SetUp failed for volume "networking-console-plugin-cert" (UniqueName: "kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin-cert" not registered Dec 10 22:49:58 crc kubenswrapper[4791]: E1210 22:49:58.687587 4791 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Dec 10 22:49:58 crc kubenswrapper[4791]: E1210 22:49:58.687622 4791 projected.go:194] Error preparing data for projected volume kube-api-access-cqllr for pod openshift-network-diagnostics/network-check-target-xd92c: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Dec 10 22:49:58 crc kubenswrapper[4791]: E1210 22:49:58.687625 4791 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl podName:9d751cbb-f2e2-430d-9754-c882a5e924a5 nodeName:}" failed. No retries permitted until 2025-12-10 22:50:06.687604902 +0000 UTC m=+41.117222565 (durationBeforeRetry 8s). Error: MountVolume.SetUp failed for volume "kube-api-access-s2dwl" (UniqueName: "kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl") pod "network-check-source-55646444c4-trplf" (UID: "9d751cbb-f2e2-430d-9754-c882a5e924a5") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Dec 10 22:49:58 crc kubenswrapper[4791]: E1210 22:49:58.687694 4791 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr podName:3b6479f0-333b-4a96-9adf-2099afdc2447 nodeName:}" failed. No retries permitted until 2025-12-10 22:50:06.687672434 +0000 UTC m=+41.117290117 (durationBeforeRetry 8s). Error: MountVolume.SetUp failed for volume "kube-api-access-cqllr" (UniqueName: "kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr") pod "network-check-target-xd92c" (UID: "3b6479f0-333b-4a96-9adf-2099afdc2447") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Dec 10 22:49:58 crc kubenswrapper[4791]: I1210 22:49:58.706639 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 22:49:58 crc kubenswrapper[4791]: I1210 22:49:58.706707 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 22:49:58 crc kubenswrapper[4791]: I1210 22:49:58.706727 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 22:49:58 crc kubenswrapper[4791]: I1210 22:49:58.706755 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 22:49:58 crc kubenswrapper[4791]: I1210 22:49:58.706774 4791 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T22:49:58Z","lastTransitionTime":"2025-12-10T22:49:58Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 22:49:58 crc kubenswrapper[4791]: I1210 22:49:58.724839 4791 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-4nwkq" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"672aa28c-8169-49ed-87b8-21187d13a80c\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T22:49:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T22:49:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T22:49:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T22:49:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a2f2a62315529e148a10a33ccb241e9417fbde0fe42afb61cb7ee2438c74636f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T22:49:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xl62j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-10T22:49:51Z\\\"}}\" for pod \"openshift-multus\"/\"multus-4nwkq\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T22:49:58Z is after 2025-08-24T17:21:41Z" Dec 10 22:49:58 crc kubenswrapper[4791]: I1210 22:49:58.773990 4791 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-zhq64" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3cd47739-0fa9-4321-aff1-220f8721a0b3\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T22:49:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T22:49:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T22:49:52Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T22:49:52Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wvpw8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wvpw8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wvpw8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wvpw8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wvpw8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wvpw8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wvpw8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wvpw8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3afbd1ffcd2c4b40324aca6e52d6fc822db131c3f35a9af3a6f005c2cffdb7bf\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3afbd1ffcd2c4b40324aca6e52d6fc822db131c3f35a9af3a6f005c2cffdb7bf\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-10T22:49:53Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-10T22:49:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wvpw8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-10T22:49:52Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-zhq64\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T22:49:58Z is after 2025-08-24T17:21:41Z" Dec 10 22:49:58 crc kubenswrapper[4791]: I1210 22:49:58.809762 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 22:49:58 crc kubenswrapper[4791]: I1210 22:49:58.809821 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 22:49:58 crc kubenswrapper[4791]: I1210 22:49:58.809845 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 22:49:58 crc kubenswrapper[4791]: I1210 22:49:58.809875 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 22:49:58 crc kubenswrapper[4791]: I1210 22:49:58.809897 4791 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T22:49:58Z","lastTransitionTime":"2025-12-10T22:49:58Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 22:49:58 crc kubenswrapper[4791]: I1210 22:49:58.884049 4791 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 10 22:49:58 crc kubenswrapper[4791]: I1210 22:49:58.884074 4791 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 10 22:49:58 crc kubenswrapper[4791]: E1210 22:49:58.884252 4791 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 10 22:49:58 crc kubenswrapper[4791]: I1210 22:49:58.884066 4791 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 10 22:49:58 crc kubenswrapper[4791]: E1210 22:49:58.884406 4791 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 10 22:49:58 crc kubenswrapper[4791]: E1210 22:49:58.884500 4791 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 10 22:49:58 crc kubenswrapper[4791]: I1210 22:49:58.912360 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 22:49:58 crc kubenswrapper[4791]: I1210 22:49:58.912416 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 22:49:58 crc kubenswrapper[4791]: I1210 22:49:58.912432 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 22:49:58 crc kubenswrapper[4791]: I1210 22:49:58.912454 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 22:49:58 crc kubenswrapper[4791]: I1210 22:49:58.912468 4791 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T22:49:58Z","lastTransitionTime":"2025-12-10T22:49:58Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 22:49:59 crc kubenswrapper[4791]: I1210 22:49:59.015625 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 22:49:59 crc kubenswrapper[4791]: I1210 22:49:59.015675 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 22:49:59 crc kubenswrapper[4791]: I1210 22:49:59.015688 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 22:49:59 crc kubenswrapper[4791]: I1210 22:49:59.015708 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 22:49:59 crc kubenswrapper[4791]: I1210 22:49:59.015721 4791 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T22:49:59Z","lastTransitionTime":"2025-12-10T22:49:59Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 22:49:59 crc kubenswrapper[4791]: I1210 22:49:59.117966 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 22:49:59 crc kubenswrapper[4791]: I1210 22:49:59.118026 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 22:49:59 crc kubenswrapper[4791]: I1210 22:49:59.118043 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 22:49:59 crc kubenswrapper[4791]: I1210 22:49:59.118069 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 22:49:59 crc kubenswrapper[4791]: I1210 22:49:59.118087 4791 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T22:49:59Z","lastTransitionTime":"2025-12-10T22:49:59Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 22:49:59 crc kubenswrapper[4791]: I1210 22:49:59.221265 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 22:49:59 crc kubenswrapper[4791]: I1210 22:49:59.221314 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 22:49:59 crc kubenswrapper[4791]: I1210 22:49:59.221331 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 22:49:59 crc kubenswrapper[4791]: I1210 22:49:59.221386 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 22:49:59 crc kubenswrapper[4791]: I1210 22:49:59.221402 4791 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T22:49:59Z","lastTransitionTime":"2025-12-10T22:49:59Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 22:49:59 crc kubenswrapper[4791]: I1210 22:49:59.312561 4791 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-zhq64" event={"ID":"3cd47739-0fa9-4321-aff1-220f8721a0b3","Type":"ContainerStarted","Data":"7a03d21e34f9d560953b7a809a4d5f374e39d3daeb8e698dbb5861b8e63c9de5"} Dec 10 22:49:59 crc kubenswrapper[4791]: I1210 22:49:59.312984 4791 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-ovn-kubernetes/ovnkube-node-zhq64" Dec 10 22:49:59 crc kubenswrapper[4791]: I1210 22:49:59.313736 4791 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-ovn-kubernetes/ovnkube-node-zhq64" Dec 10 22:49:59 crc kubenswrapper[4791]: I1210 22:49:59.320237 4791 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-additional-cni-plugins-rw584" event={"ID":"ff9beb90-69b0-4732-bf37-0b81f58ecc98","Type":"ContainerStarted","Data":"ddb9df501dc6e8e5561fbc2aefd431037a3f28f73bf8e82a6598d261a4224807"} Dec 10 22:49:59 crc kubenswrapper[4791]: I1210 22:49:59.324415 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 22:49:59 crc kubenswrapper[4791]: I1210 22:49:59.324460 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 22:49:59 crc kubenswrapper[4791]: I1210 22:49:59.324479 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 22:49:59 crc kubenswrapper[4791]: I1210 22:49:59.324503 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 22:49:59 crc kubenswrapper[4791]: I1210 22:49:59.324520 4791 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T22:49:59Z","lastTransitionTime":"2025-12-10T22:49:59Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 22:49:59 crc kubenswrapper[4791]: I1210 22:49:59.340748 4791 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f8349fea-c49b-49a1-b5ee-032c5dffe021\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T22:49:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T22:49:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T22:49:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T22:49:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T22:49:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://acafe12186df99d1059a31260de0eafce35ff7771bcb8f5c6753b72665d6f60b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T22:49:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ea4c346507862a8c12756a3061d6e25e788a8d79f4621b62f81f4322185c82a8\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T22:49:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://2e6e324ec6f4f7bb69fdfab4ea3ac1aa55e480dd8e2a8984b0fe5b4c9f66636e\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T22:49:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://1b2e11ff2bf8789fa1d0d23ae648b3e6ef653106e5bc4bbb5dd6447f7cd6f05d\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d6aa7f8c5d908d90008f2df49860ac8a3480136da567e3d85f86135d779204f9\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-12-10T22:49:44Z\\\",\\\"message\\\":\\\"W1210 22:49:33.845733 1 cmd.go:257] Using insecure, self-signed certificates\\\\nI1210 22:49:33.846108 1 crypto.go:601] Generating new CA for check-endpoints-signer@1765406973 cert, and key in /tmp/serving-cert-2585701456/serving-signer.crt, /tmp/serving-cert-2585701456/serving-signer.key\\\\nI1210 22:49:34.529748 1 observer_polling.go:159] Starting file observer\\\\nW1210 22:49:34.532225 1 builder.go:272] unable to get owner reference (falling back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": dial tcp [::1]:6443: connect: connection refused\\\\nI1210 22:49:34.532599 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1210 22:49:34.533977 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-2585701456/tls.crt::/tmp/serving-cert-2585701456/tls.key\\\\\\\"\\\\nF1210 22:49:44.774710 1 cmd.go:182] error initializing delegating authentication: unable to load configmap based request-header-client-ca-file: Get \\\\\\\"https://localhost:6443/api/v1/namespaces/kube-system/configmaps/extension-apiserver-authentication\\\\\\\": net/http: TLS handshake timeout\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-10T22:49:33Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T22:49:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://0c4d036ab269f62561fe7ef1586012cb3a16e3583a08b9488a0616ec00c8e0ae\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T22:49:32Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c5856d6b26c19fda13b5c9ac2892a7e2f9c6f79feb8173ecfd9b33b673327d67\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c5856d6b26c19fda13b5c9ac2892a7e2f9c6f79feb8173ecfd9b33b673327d67\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-10T22:49:27Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-10T22:49:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-10T22:49:25Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T22:49:59Z is after 2025-08-24T17:21:41Z" Dec 10 22:49:59 crc kubenswrapper[4791]: I1210 22:49:59.355795 4791 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-ovn-kubernetes/ovnkube-node-zhq64" Dec 10 22:49:59 crc kubenswrapper[4791]: I1210 22:49:59.356307 4791 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-ovn-kubernetes/ovnkube-node-zhq64" Dec 10 22:49:59 crc kubenswrapper[4791]: I1210 22:49:59.371263 4791 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"90250e13-6c63-443a-a5e5-6e4e341a2289\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T22:49:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T22:49:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T22:49:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T22:49:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T22:49:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ba54f95b1e00843b854db9a10fe6897592159ccacf30ed0f76002093a13eedd3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T22:49:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5aa8ffa518f3a68b60b275f8df4851b2564aa4984b61efdc9470b73072133e43\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T22:49:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://16f989503edc78d9451638e2cf6cf4b19a2cf2b8809b10b237541cb829303091\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T22:49:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://82556824d63ce6e46df14c45689ff1f81ebf729abe902061815c040b323898e9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T22:49:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://61db531940fa6f2effdae5a005b4c91bd2698bd5d1bff9588c171a65ad550eca\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T22:49:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0f65c864d885319ce2a08d61c30be24b3ffb7d23ea30c82ed7faf039cb2c006d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://0f65c864d885319ce2a08d61c30be24b3ffb7d23ea30c82ed7faf039cb2c006d\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-10T22:49:27Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-10T22:49:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e31a76d1d5b2812ba83e24e38268e25ec8bed98fc3da9adcd06bda5822b6ace1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e31a76d1d5b2812ba83e24e38268e25ec8bed98fc3da9adcd06bda5822b6ace1\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-10T22:49:30Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-10T22:49:29Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://b5b5eceae562f3986b543d41de1e977efb85a91d84c98e09159ca6d092eda691\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b5b5eceae562f3986b543d41de1e977efb85a91d84c98e09159ca6d092eda691\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-10T22:49:31Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-10T22:49:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-10T22:49:25Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T22:49:59Z is after 2025-08-24T17:21:41Z" Dec 10 22:49:59 crc kubenswrapper[4791]: I1210 22:49:59.387935 4791 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-10T22:49:50Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-10T22:49:50Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T22:49:59Z is after 2025-08-24T17:21:41Z" Dec 10 22:49:59 crc kubenswrapper[4791]: I1210 22:49:59.400930 4791 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-10T22:49:54Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-10T22:49:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-10T22:49:54Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://62bfaf5d7a6d1b452cdaf88086447d0c8cb7306e94434e46752be5475124d56a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T22:49:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T22:49:59Z is after 2025-08-24T17:21:41Z" Dec 10 22:49:59 crc kubenswrapper[4791]: I1210 22:49:59.412327 4791 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-5rb5l" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ba35653c-6e06-4cee-a4d6-137764090d18\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T22:49:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T22:49:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T22:49:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T22:49:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://dbb5c875730a98ad2daf316cfafc08fcdce21ac16148154780ac9193ed405dab\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T22:49:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fvvrs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://185eb2ff44221a514878283ec873773a6c9a9073dc64e296f982e60dee1338e4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T22:49:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fvvrs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-10T22:49:51Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-5rb5l\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T22:49:59Z is after 2025-08-24T17:21:41Z" Dec 10 22:49:59 crc kubenswrapper[4791]: I1210 22:49:59.427405 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 22:49:59 crc kubenswrapper[4791]: I1210 22:49:59.427463 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 22:49:59 crc kubenswrapper[4791]: I1210 22:49:59.427480 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 22:49:59 crc kubenswrapper[4791]: I1210 22:49:59.427507 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 22:49:59 crc kubenswrapper[4791]: I1210 22:49:59.427524 4791 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T22:49:59Z","lastTransitionTime":"2025-12-10T22:49:59Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 22:49:59 crc kubenswrapper[4791]: I1210 22:49:59.431517 4791 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"698bbe9a-fbe5-436d-a835-5e80df77ba24\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T22:49:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T22:49:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T22:49:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T22:49:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T22:49:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://224d73713312376a58ec6cc2bc8fdb6ffa44c4749ae0e7691d5fe27b9afca09f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T22:49:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://372eeaffd40992c6f06efc4df6dfcb44b31de64ee8256c21508cdd7c74ecb96e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T22:49:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://76b6013b062cac5b0b4d12a94f38ce7b43f9c7ea216063455d88a956ac04d26e\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T22:49:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://78e950e122e761d950da1a3c60c06476e53920e8795f6c61db22c01e22d5136e\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T22:49:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-10T22:49:25Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T22:49:59Z is after 2025-08-24T17:21:41Z" Dec 10 22:49:59 crc kubenswrapper[4791]: I1210 22:49:59.446315 4791 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-10T22:49:52Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-10T22:49:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-10T22:49:52Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3734b110802828632d44263ed0a43cabc4b710833f0be120d72db0940fe783e2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T22:49:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://794803ac9be3c52509e6dcb10854d8864e1a9e406762e57c7971b07c4da1220e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T22:49:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T22:49:59Z is after 2025-08-24T17:21:41Z" Dec 10 22:49:59 crc kubenswrapper[4791]: I1210 22:49:59.461438 4791 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-vczjq" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b9f00f25-3fa1-4b63-9710-fd352224b01b\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T22:49:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T22:49:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T22:49:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T22:49:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4698cf4c6b55b41e628448d37c2eb847c2777d0efbd93afcbf367d5e78623b7e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T22:49:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lrrpv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-10T22:49:50Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-vczjq\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T22:49:59Z is after 2025-08-24T17:21:41Z" Dec 10 22:49:59 crc kubenswrapper[4791]: I1210 22:49:59.477131 4791 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-4nwkq" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"672aa28c-8169-49ed-87b8-21187d13a80c\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T22:49:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T22:49:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T22:49:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T22:49:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a2f2a62315529e148a10a33ccb241e9417fbde0fe42afb61cb7ee2438c74636f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T22:49:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xl62j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-10T22:49:51Z\\\"}}\" for pod \"openshift-multus\"/\"multus-4nwkq\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T22:49:59Z is after 2025-08-24T17:21:41Z" Dec 10 22:49:59 crc kubenswrapper[4791]: I1210 22:49:59.502251 4791 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-zhq64" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3cd47739-0fa9-4321-aff1-220f8721a0b3\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T22:49:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T22:49:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T22:49:52Z\\\",\\\"message\\\":\\\"containers with unready status: [nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T22:49:52Z\\\",\\\"message\\\":\\\"containers with unready status: [nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://68795a29944e56870ffe94fa732e9dcf0bfb344c0d85d1e590f58a876129c61f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T22:49:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wvpw8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://bc721b530c08b433756e978ec0423cc18aabf6771b61eb96f903b4d57ea335ff\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T22:49:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wvpw8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://152a7618be8f38fadf4e564b3948980503cedcdb466c37344ccc20f4a3173a5f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T22:49:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wvpw8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3d18f9cb7207efb1e32de75a30d33e024f656a08f0503db95e5488bc5280971f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T22:49:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wvpw8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://937137ae9d7480cdc7c59bc4a1df0a4c10e72b3b63693af285278789953ed218\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T22:49:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wvpw8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://1a52d19b7671e8b143017c8a981a6f8d3ee3b84650e4b223b814a531ac76f364\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T22:49:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wvpw8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://7a03d21e34f9d560953b7a809a4d5f374e39d3daeb8e698dbb5861b8e63c9de5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T22:49:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wvpw8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://dd3f80e5713221f96ed50da7f2e7fd5b9fabfcd5c1656da8d8a86d818a5fdc3f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T22:49:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wvpw8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3afbd1ffcd2c4b40324aca6e52d6fc822db131c3f35a9af3a6f005c2cffdb7bf\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3afbd1ffcd2c4b40324aca6e52d6fc822db131c3f35a9af3a6f005c2cffdb7bf\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-10T22:49:53Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-10T22:49:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wvpw8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-10T22:49:52Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-zhq64\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T22:49:59Z is after 2025-08-24T17:21:41Z" Dec 10 22:49:59 crc kubenswrapper[4791]: I1210 22:49:59.517089 4791 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-10T22:49:52Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-10T22:49:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-10T22:49:52Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://20651eb0c16e2c71222f56da1ff755cb0e7b084886acbbb24f0ccaf57041ed85\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T22:49:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T22:49:59Z is after 2025-08-24T17:21:41Z" Dec 10 22:49:59 crc kubenswrapper[4791]: I1210 22:49:59.530829 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 22:49:59 crc kubenswrapper[4791]: I1210 22:49:59.530872 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 22:49:59 crc kubenswrapper[4791]: I1210 22:49:59.530885 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 22:49:59 crc kubenswrapper[4791]: I1210 22:49:59.530903 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 22:49:59 crc kubenswrapper[4791]: I1210 22:49:59.530921 4791 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T22:49:59Z","lastTransitionTime":"2025-12-10T22:49:59Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 22:49:59 crc kubenswrapper[4791]: I1210 22:49:59.537296 4791 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-10T22:49:50Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-10T22:49:50Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T22:49:59Z is after 2025-08-24T17:21:41Z" Dec 10 22:49:59 crc kubenswrapper[4791]: I1210 22:49:59.551791 4791 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-10T22:49:50Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-10T22:49:50Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T22:49:59Z is after 2025-08-24T17:21:41Z" Dec 10 22:49:59 crc kubenswrapper[4791]: I1210 22:49:59.564658 4791 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-tchxk" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"c0ab74e0-42a9-4f42-ac80-16030cf4ffff\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T22:49:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T22:49:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T22:49:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T22:49:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://04e0fd1b53256b7a938e8ccab93e2a4eb73dd1a59e0b9799e09c55e511baf7fe\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T22:49:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zn2qt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-10T22:49:50Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-tchxk\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T22:49:59Z is after 2025-08-24T17:21:41Z" Dec 10 22:49:59 crc kubenswrapper[4791]: I1210 22:49:59.581000 4791 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-rw584" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ff9beb90-69b0-4732-bf37-0b81f58ecc98\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T22:49:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T22:49:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T22:49:51Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T22:49:51Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fksqn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ce67e54b696bb53ffbbe48c6ce71476b635e0868993e9c427b22f904891433ee\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ce67e54b696bb53ffbbe48c6ce71476b635e0868993e9c427b22f904891433ee\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-10T22:49:52Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-10T22:49:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fksqn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://818d89bfffb8a110670152c8480684665ea8db578e9768f524fbbe6a92f27c22\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://818d89bfffb8a110670152c8480684665ea8db578e9768f524fbbe6a92f27c22\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-10T22:49:53Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-10T22:49:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fksqn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://9aa9c06e894bb2a2ed3738defaeff68a0a0458ead48177f36b90770944c211a0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://9aa9c06e894bb2a2ed3738defaeff68a0a0458ead48177f36b90770944c211a0\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-10T22:49:54Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-10T22:49:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fksqn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://dfd6d03336dcea7529419a30e8d70d84e2290a0fa5b0e46d09ad341c190638d6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://dfd6d03336dcea7529419a30e8d70d84e2290a0fa5b0e46d09ad341c190638d6\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-10T22:49:55Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-10T22:49:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fksqn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b86738401ece6fbc3083782720c66a515850ed78d4054f34fcabab7ff525b85f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b86738401ece6fbc3083782720c66a515850ed78d4054f34fcabab7ff525b85f\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-10T22:49:56Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-10T22:49:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fksqn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://18b326f572a899ddd10197a745bf9264d4ddd975690ad1cfb0ec576ccd5cf102\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://18b326f572a899ddd10197a745bf9264d4ddd975690ad1cfb0ec576ccd5cf102\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-10T22:49:57Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-10T22:49:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fksqn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-10T22:49:51Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-rw584\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T22:49:59Z is after 2025-08-24T17:21:41Z" Dec 10 22:49:59 crc kubenswrapper[4791]: I1210 22:49:59.592979 4791 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-5rb5l" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ba35653c-6e06-4cee-a4d6-137764090d18\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T22:49:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T22:49:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T22:49:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T22:49:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://dbb5c875730a98ad2daf316cfafc08fcdce21ac16148154780ac9193ed405dab\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T22:49:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fvvrs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://185eb2ff44221a514878283ec873773a6c9a9073dc64e296f982e60dee1338e4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T22:49:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fvvrs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-10T22:49:51Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-5rb5l\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T22:49:59Z is after 2025-08-24T17:21:41Z" Dec 10 22:49:59 crc kubenswrapper[4791]: I1210 22:49:59.611974 4791 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-10T22:49:52Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-10T22:49:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-10T22:49:52Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3734b110802828632d44263ed0a43cabc4b710833f0be120d72db0940fe783e2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T22:49:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://794803ac9be3c52509e6dcb10854d8864e1a9e406762e57c7971b07c4da1220e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T22:49:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T22:49:59Z is after 2025-08-24T17:21:41Z" Dec 10 22:49:59 crc kubenswrapper[4791]: I1210 22:49:59.624950 4791 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-vczjq" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b9f00f25-3fa1-4b63-9710-fd352224b01b\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T22:49:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T22:49:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T22:49:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T22:49:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4698cf4c6b55b41e628448d37c2eb847c2777d0efbd93afcbf367d5e78623b7e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T22:49:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lrrpv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-10T22:49:50Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-vczjq\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T22:49:59Z is after 2025-08-24T17:21:41Z" Dec 10 22:49:59 crc kubenswrapper[4791]: I1210 22:49:59.633366 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 22:49:59 crc kubenswrapper[4791]: I1210 22:49:59.633411 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 22:49:59 crc kubenswrapper[4791]: I1210 22:49:59.633429 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 22:49:59 crc kubenswrapper[4791]: I1210 22:49:59.633449 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 22:49:59 crc kubenswrapper[4791]: I1210 22:49:59.633462 4791 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T22:49:59Z","lastTransitionTime":"2025-12-10T22:49:59Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 22:49:59 crc kubenswrapper[4791]: I1210 22:49:59.641390 4791 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-4nwkq" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"672aa28c-8169-49ed-87b8-21187d13a80c\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T22:49:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T22:49:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T22:49:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T22:49:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a2f2a62315529e148a10a33ccb241e9417fbde0fe42afb61cb7ee2438c74636f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T22:49:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xl62j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-10T22:49:51Z\\\"}}\" for pod \"openshift-multus\"/\"multus-4nwkq\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T22:49:59Z is after 2025-08-24T17:21:41Z" Dec 10 22:49:59 crc kubenswrapper[4791]: I1210 22:49:59.661527 4791 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-zhq64" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3cd47739-0fa9-4321-aff1-220f8721a0b3\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T22:49:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T22:49:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T22:49:52Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T22:49:52Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://68795a29944e56870ffe94fa732e9dcf0bfb344c0d85d1e590f58a876129c61f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T22:49:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wvpw8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://bc721b530c08b433756e978ec0423cc18aabf6771b61eb96f903b4d57ea335ff\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T22:49:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wvpw8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://152a7618be8f38fadf4e564b3948980503cedcdb466c37344ccc20f4a3173a5f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T22:49:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wvpw8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3d18f9cb7207efb1e32de75a30d33e024f656a08f0503db95e5488bc5280971f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T22:49:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wvpw8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://937137ae9d7480cdc7c59bc4a1df0a4c10e72b3b63693af285278789953ed218\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T22:49:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wvpw8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://1a52d19b7671e8b143017c8a981a6f8d3ee3b84650e4b223b814a531ac76f364\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T22:49:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wvpw8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://7a03d21e34f9d560953b7a809a4d5f374e39d3daeb8e698dbb5861b8e63c9de5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T22:49:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wvpw8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://dd3f80e5713221f96ed50da7f2e7fd5b9fabfcd5c1656da8d8a86d818a5fdc3f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T22:49:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wvpw8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3afbd1ffcd2c4b40324aca6e52d6fc822db131c3f35a9af3a6f005c2cffdb7bf\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3afbd1ffcd2c4b40324aca6e52d6fc822db131c3f35a9af3a6f005c2cffdb7bf\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-10T22:49:53Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-10T22:49:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wvpw8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-10T22:49:52Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-zhq64\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T22:49:59Z is after 2025-08-24T17:21:41Z" Dec 10 22:49:59 crc kubenswrapper[4791]: I1210 22:49:59.675665 4791 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"698bbe9a-fbe5-436d-a835-5e80df77ba24\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T22:49:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T22:49:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T22:49:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T22:49:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T22:49:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://224d73713312376a58ec6cc2bc8fdb6ffa44c4749ae0e7691d5fe27b9afca09f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T22:49:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://372eeaffd40992c6f06efc4df6dfcb44b31de64ee8256c21508cdd7c74ecb96e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T22:49:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://76b6013b062cac5b0b4d12a94f38ce7b43f9c7ea216063455d88a956ac04d26e\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T22:49:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://78e950e122e761d950da1a3c60c06476e53920e8795f6c61db22c01e22d5136e\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T22:49:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-10T22:49:25Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T22:49:59Z is after 2025-08-24T17:21:41Z" Dec 10 22:49:59 crc kubenswrapper[4791]: I1210 22:49:59.691643 4791 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-10T22:49:50Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-10T22:49:50Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T22:49:59Z is after 2025-08-24T17:21:41Z" Dec 10 22:49:59 crc kubenswrapper[4791]: I1210 22:49:59.706295 4791 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-10T22:49:50Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-10T22:49:50Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T22:49:59Z is after 2025-08-24T17:21:41Z" Dec 10 22:49:59 crc kubenswrapper[4791]: I1210 22:49:59.719108 4791 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-tchxk" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"c0ab74e0-42a9-4f42-ac80-16030cf4ffff\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T22:49:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T22:49:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T22:49:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T22:49:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://04e0fd1b53256b7a938e8ccab93e2a4eb73dd1a59e0b9799e09c55e511baf7fe\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T22:49:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zn2qt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-10T22:49:50Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-tchxk\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T22:49:59Z is after 2025-08-24T17:21:41Z" Dec 10 22:49:59 crc kubenswrapper[4791]: I1210 22:49:59.735838 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 22:49:59 crc kubenswrapper[4791]: I1210 22:49:59.735882 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 22:49:59 crc kubenswrapper[4791]: I1210 22:49:59.735893 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 22:49:59 crc kubenswrapper[4791]: I1210 22:49:59.735911 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 22:49:59 crc kubenswrapper[4791]: I1210 22:49:59.735924 4791 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T22:49:59Z","lastTransitionTime":"2025-12-10T22:49:59Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 22:49:59 crc kubenswrapper[4791]: I1210 22:49:59.762687 4791 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-rw584" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ff9beb90-69b0-4732-bf37-0b81f58ecc98\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T22:49:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T22:49:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T22:49:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T22:49:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ddb9df501dc6e8e5561fbc2aefd431037a3f28f73bf8e82a6598d261a4224807\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T22:49:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fksqn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ce67e54b696bb53ffbbe48c6ce71476b635e0868993e9c427b22f904891433ee\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ce67e54b696bb53ffbbe48c6ce71476b635e0868993e9c427b22f904891433ee\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-10T22:49:52Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-10T22:49:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fksqn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://818d89bfffb8a110670152c8480684665ea8db578e9768f524fbbe6a92f27c22\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://818d89bfffb8a110670152c8480684665ea8db578e9768f524fbbe6a92f27c22\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-10T22:49:53Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-10T22:49:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fksqn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://9aa9c06e894bb2a2ed3738defaeff68a0a0458ead48177f36b90770944c211a0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://9aa9c06e894bb2a2ed3738defaeff68a0a0458ead48177f36b90770944c211a0\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-10T22:49:54Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-10T22:49:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fksqn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://dfd6d03336dcea7529419a30e8d70d84e2290a0fa5b0e46d09ad341c190638d6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://dfd6d03336dcea7529419a30e8d70d84e2290a0fa5b0e46d09ad341c190638d6\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-10T22:49:55Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-10T22:49:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fksqn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b86738401ece6fbc3083782720c66a515850ed78d4054f34fcabab7ff525b85f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b86738401ece6fbc3083782720c66a515850ed78d4054f34fcabab7ff525b85f\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-10T22:49:56Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-10T22:49:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fksqn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://18b326f572a899ddd10197a745bf9264d4ddd975690ad1cfb0ec576ccd5cf102\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://18b326f572a899ddd10197a745bf9264d4ddd975690ad1cfb0ec576ccd5cf102\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-10T22:49:57Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-10T22:49:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fksqn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-10T22:49:51Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-rw584\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T22:49:59Z is after 2025-08-24T17:21:41Z" Dec 10 22:49:59 crc kubenswrapper[4791]: I1210 22:49:59.796839 4791 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-10T22:49:52Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-10T22:49:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-10T22:49:52Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://20651eb0c16e2c71222f56da1ff755cb0e7b084886acbbb24f0ccaf57041ed85\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T22:49:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T22:49:59Z is after 2025-08-24T17:21:41Z" Dec 10 22:49:59 crc kubenswrapper[4791]: I1210 22:49:59.838330 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 22:49:59 crc kubenswrapper[4791]: I1210 22:49:59.838404 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 22:49:59 crc kubenswrapper[4791]: I1210 22:49:59.838418 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 22:49:59 crc kubenswrapper[4791]: I1210 22:49:59.838439 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 22:49:59 crc kubenswrapper[4791]: I1210 22:49:59.838455 4791 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T22:49:59Z","lastTransitionTime":"2025-12-10T22:49:59Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 22:49:59 crc kubenswrapper[4791]: I1210 22:49:59.854323 4791 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"90250e13-6c63-443a-a5e5-6e4e341a2289\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T22:49:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T22:49:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T22:49:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T22:49:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T22:49:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ba54f95b1e00843b854db9a10fe6897592159ccacf30ed0f76002093a13eedd3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T22:49:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5aa8ffa518f3a68b60b275f8df4851b2564aa4984b61efdc9470b73072133e43\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T22:49:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://16f989503edc78d9451638e2cf6cf4b19a2cf2b8809b10b237541cb829303091\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T22:49:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://82556824d63ce6e46df14c45689ff1f81ebf729abe902061815c040b323898e9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T22:49:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://61db531940fa6f2effdae5a005b4c91bd2698bd5d1bff9588c171a65ad550eca\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T22:49:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0f65c864d885319ce2a08d61c30be24b3ffb7d23ea30c82ed7faf039cb2c006d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://0f65c864d885319ce2a08d61c30be24b3ffb7d23ea30c82ed7faf039cb2c006d\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-10T22:49:27Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-10T22:49:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e31a76d1d5b2812ba83e24e38268e25ec8bed98fc3da9adcd06bda5822b6ace1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e31a76d1d5b2812ba83e24e38268e25ec8bed98fc3da9adcd06bda5822b6ace1\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-10T22:49:30Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-10T22:49:29Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://b5b5eceae562f3986b543d41de1e977efb85a91d84c98e09159ca6d092eda691\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b5b5eceae562f3986b543d41de1e977efb85a91d84c98e09159ca6d092eda691\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-10T22:49:31Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-10T22:49:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-10T22:49:25Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T22:49:59Z is after 2025-08-24T17:21:41Z" Dec 10 22:49:59 crc kubenswrapper[4791]: I1210 22:49:59.891313 4791 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-10T22:49:50Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-10T22:49:50Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T22:49:59Z is after 2025-08-24T17:21:41Z" Dec 10 22:49:59 crc kubenswrapper[4791]: I1210 22:49:59.916291 4791 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-10T22:49:54Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-10T22:49:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-10T22:49:54Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://62bfaf5d7a6d1b452cdaf88086447d0c8cb7306e94434e46752be5475124d56a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T22:49:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T22:49:59Z is after 2025-08-24T17:21:41Z" Dec 10 22:49:59 crc kubenswrapper[4791]: I1210 22:49:59.940787 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 22:49:59 crc kubenswrapper[4791]: I1210 22:49:59.940828 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 22:49:59 crc kubenswrapper[4791]: I1210 22:49:59.940840 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 22:49:59 crc kubenswrapper[4791]: I1210 22:49:59.940856 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 22:49:59 crc kubenswrapper[4791]: I1210 22:49:59.940867 4791 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T22:49:59Z","lastTransitionTime":"2025-12-10T22:49:59Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 22:49:59 crc kubenswrapper[4791]: I1210 22:49:59.963819 4791 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f8349fea-c49b-49a1-b5ee-032c5dffe021\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T22:49:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T22:49:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T22:49:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T22:49:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T22:49:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://acafe12186df99d1059a31260de0eafce35ff7771bcb8f5c6753b72665d6f60b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T22:49:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ea4c346507862a8c12756a3061d6e25e788a8d79f4621b62f81f4322185c82a8\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T22:49:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://2e6e324ec6f4f7bb69fdfab4ea3ac1aa55e480dd8e2a8984b0fe5b4c9f66636e\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T22:49:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://1b2e11ff2bf8789fa1d0d23ae648b3e6ef653106e5bc4bbb5dd6447f7cd6f05d\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d6aa7f8c5d908d90008f2df49860ac8a3480136da567e3d85f86135d779204f9\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-12-10T22:49:44Z\\\",\\\"message\\\":\\\"W1210 22:49:33.845733 1 cmd.go:257] Using insecure, self-signed certificates\\\\nI1210 22:49:33.846108 1 crypto.go:601] Generating new CA for check-endpoints-signer@1765406973 cert, and key in /tmp/serving-cert-2585701456/serving-signer.crt, /tmp/serving-cert-2585701456/serving-signer.key\\\\nI1210 22:49:34.529748 1 observer_polling.go:159] Starting file observer\\\\nW1210 22:49:34.532225 1 builder.go:272] unable to get owner reference (falling back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": dial tcp [::1]:6443: connect: connection refused\\\\nI1210 22:49:34.532599 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1210 22:49:34.533977 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-2585701456/tls.crt::/tmp/serving-cert-2585701456/tls.key\\\\\\\"\\\\nF1210 22:49:44.774710 1 cmd.go:182] error initializing delegating authentication: unable to load configmap based request-header-client-ca-file: Get \\\\\\\"https://localhost:6443/api/v1/namespaces/kube-system/configmaps/extension-apiserver-authentication\\\\\\\": net/http: TLS handshake timeout\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-10T22:49:33Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T22:49:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://0c4d036ab269f62561fe7ef1586012cb3a16e3583a08b9488a0616ec00c8e0ae\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T22:49:32Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c5856d6b26c19fda13b5c9ac2892a7e2f9c6f79feb8173ecfd9b33b673327d67\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c5856d6b26c19fda13b5c9ac2892a7e2f9c6f79feb8173ecfd9b33b673327d67\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-10T22:49:27Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-10T22:49:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-10T22:49:25Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T22:49:59Z is after 2025-08-24T17:21:41Z" Dec 10 22:50:00 crc kubenswrapper[4791]: I1210 22:50:00.043276 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 22:50:00 crc kubenswrapper[4791]: I1210 22:50:00.043321 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 22:50:00 crc kubenswrapper[4791]: I1210 22:50:00.043353 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 22:50:00 crc kubenswrapper[4791]: I1210 22:50:00.043377 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 22:50:00 crc kubenswrapper[4791]: I1210 22:50:00.043390 4791 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T22:50:00Z","lastTransitionTime":"2025-12-10T22:50:00Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 22:50:00 crc kubenswrapper[4791]: I1210 22:50:00.146467 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 22:50:00 crc kubenswrapper[4791]: I1210 22:50:00.146544 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 22:50:00 crc kubenswrapper[4791]: I1210 22:50:00.146568 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 22:50:00 crc kubenswrapper[4791]: I1210 22:50:00.146599 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 22:50:00 crc kubenswrapper[4791]: I1210 22:50:00.146664 4791 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T22:50:00Z","lastTransitionTime":"2025-12-10T22:50:00Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 22:50:00 crc kubenswrapper[4791]: I1210 22:50:00.249598 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 22:50:00 crc kubenswrapper[4791]: I1210 22:50:00.249648 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 22:50:00 crc kubenswrapper[4791]: I1210 22:50:00.249665 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 22:50:00 crc kubenswrapper[4791]: I1210 22:50:00.249689 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 22:50:00 crc kubenswrapper[4791]: I1210 22:50:00.249707 4791 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T22:50:00Z","lastTransitionTime":"2025-12-10T22:50:00Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 22:50:00 crc kubenswrapper[4791]: I1210 22:50:00.322792 4791 prober_manager.go:312] "Failed to trigger a manual run" probe="Readiness" Dec 10 22:50:00 crc kubenswrapper[4791]: I1210 22:50:00.352599 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 22:50:00 crc kubenswrapper[4791]: I1210 22:50:00.352681 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 22:50:00 crc kubenswrapper[4791]: I1210 22:50:00.352712 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 22:50:00 crc kubenswrapper[4791]: I1210 22:50:00.352743 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 22:50:00 crc kubenswrapper[4791]: I1210 22:50:00.352765 4791 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T22:50:00Z","lastTransitionTime":"2025-12-10T22:50:00Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 22:50:00 crc kubenswrapper[4791]: I1210 22:50:00.456828 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 22:50:00 crc kubenswrapper[4791]: I1210 22:50:00.456899 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 22:50:00 crc kubenswrapper[4791]: I1210 22:50:00.456925 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 22:50:00 crc kubenswrapper[4791]: I1210 22:50:00.456957 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 22:50:00 crc kubenswrapper[4791]: I1210 22:50:00.456981 4791 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T22:50:00Z","lastTransitionTime":"2025-12-10T22:50:00Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 22:50:00 crc kubenswrapper[4791]: I1210 22:50:00.560458 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 22:50:00 crc kubenswrapper[4791]: I1210 22:50:00.560520 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 22:50:00 crc kubenswrapper[4791]: I1210 22:50:00.560540 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 22:50:00 crc kubenswrapper[4791]: I1210 22:50:00.560565 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 22:50:00 crc kubenswrapper[4791]: I1210 22:50:00.560585 4791 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T22:50:00Z","lastTransitionTime":"2025-12-10T22:50:00Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 22:50:00 crc kubenswrapper[4791]: I1210 22:50:00.664060 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 22:50:00 crc kubenswrapper[4791]: I1210 22:50:00.664115 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 22:50:00 crc kubenswrapper[4791]: I1210 22:50:00.664131 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 22:50:00 crc kubenswrapper[4791]: I1210 22:50:00.664164 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 22:50:00 crc kubenswrapper[4791]: I1210 22:50:00.664181 4791 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T22:50:00Z","lastTransitionTime":"2025-12-10T22:50:00Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 22:50:00 crc kubenswrapper[4791]: I1210 22:50:00.767412 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 22:50:00 crc kubenswrapper[4791]: I1210 22:50:00.767450 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 22:50:00 crc kubenswrapper[4791]: I1210 22:50:00.767463 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 22:50:00 crc kubenswrapper[4791]: I1210 22:50:00.767479 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 22:50:00 crc kubenswrapper[4791]: I1210 22:50:00.767491 4791 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T22:50:00Z","lastTransitionTime":"2025-12-10T22:50:00Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 22:50:00 crc kubenswrapper[4791]: I1210 22:50:00.869932 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 22:50:00 crc kubenswrapper[4791]: I1210 22:50:00.869986 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 22:50:00 crc kubenswrapper[4791]: I1210 22:50:00.869995 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 22:50:00 crc kubenswrapper[4791]: I1210 22:50:00.870016 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 22:50:00 crc kubenswrapper[4791]: I1210 22:50:00.870025 4791 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T22:50:00Z","lastTransitionTime":"2025-12-10T22:50:00Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 22:50:00 crc kubenswrapper[4791]: I1210 22:50:00.884431 4791 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 10 22:50:00 crc kubenswrapper[4791]: E1210 22:50:00.884545 4791 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 10 22:50:00 crc kubenswrapper[4791]: I1210 22:50:00.884852 4791 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 10 22:50:00 crc kubenswrapper[4791]: E1210 22:50:00.884913 4791 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 10 22:50:00 crc kubenswrapper[4791]: I1210 22:50:00.884972 4791 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 10 22:50:00 crc kubenswrapper[4791]: E1210 22:50:00.885027 4791 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 10 22:50:00 crc kubenswrapper[4791]: I1210 22:50:00.972070 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 22:50:00 crc kubenswrapper[4791]: I1210 22:50:00.972101 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 22:50:00 crc kubenswrapper[4791]: I1210 22:50:00.972111 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 22:50:00 crc kubenswrapper[4791]: I1210 22:50:00.972125 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 22:50:00 crc kubenswrapper[4791]: I1210 22:50:00.972138 4791 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T22:50:00Z","lastTransitionTime":"2025-12-10T22:50:00Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 22:50:01 crc kubenswrapper[4791]: I1210 22:50:01.074726 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 22:50:01 crc kubenswrapper[4791]: I1210 22:50:01.074766 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 22:50:01 crc kubenswrapper[4791]: I1210 22:50:01.074777 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 22:50:01 crc kubenswrapper[4791]: I1210 22:50:01.074793 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 22:50:01 crc kubenswrapper[4791]: I1210 22:50:01.074803 4791 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T22:50:01Z","lastTransitionTime":"2025-12-10T22:50:01Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 22:50:01 crc kubenswrapper[4791]: I1210 22:50:01.177219 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 22:50:01 crc kubenswrapper[4791]: I1210 22:50:01.177280 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 22:50:01 crc kubenswrapper[4791]: I1210 22:50:01.177295 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 22:50:01 crc kubenswrapper[4791]: I1210 22:50:01.177316 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 22:50:01 crc kubenswrapper[4791]: I1210 22:50:01.177330 4791 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T22:50:01Z","lastTransitionTime":"2025-12-10T22:50:01Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 22:50:01 crc kubenswrapper[4791]: I1210 22:50:01.280108 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 22:50:01 crc kubenswrapper[4791]: I1210 22:50:01.280163 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 22:50:01 crc kubenswrapper[4791]: I1210 22:50:01.280175 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 22:50:01 crc kubenswrapper[4791]: I1210 22:50:01.280195 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 22:50:01 crc kubenswrapper[4791]: I1210 22:50:01.280246 4791 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T22:50:01Z","lastTransitionTime":"2025-12-10T22:50:01Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 22:50:01 crc kubenswrapper[4791]: I1210 22:50:01.325786 4791 prober_manager.go:312] "Failed to trigger a manual run" probe="Readiness" Dec 10 22:50:01 crc kubenswrapper[4791]: I1210 22:50:01.383904 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 22:50:01 crc kubenswrapper[4791]: I1210 22:50:01.383948 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 22:50:01 crc kubenswrapper[4791]: I1210 22:50:01.383988 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 22:50:01 crc kubenswrapper[4791]: I1210 22:50:01.384005 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 22:50:01 crc kubenswrapper[4791]: I1210 22:50:01.384017 4791 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T22:50:01Z","lastTransitionTime":"2025-12-10T22:50:01Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 22:50:01 crc kubenswrapper[4791]: I1210 22:50:01.485981 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 22:50:01 crc kubenswrapper[4791]: I1210 22:50:01.486012 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 22:50:01 crc kubenswrapper[4791]: I1210 22:50:01.486021 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 22:50:01 crc kubenswrapper[4791]: I1210 22:50:01.486034 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 22:50:01 crc kubenswrapper[4791]: I1210 22:50:01.486042 4791 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T22:50:01Z","lastTransitionTime":"2025-12-10T22:50:01Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 22:50:01 crc kubenswrapper[4791]: I1210 22:50:01.587892 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 22:50:01 crc kubenswrapper[4791]: I1210 22:50:01.587934 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 22:50:01 crc kubenswrapper[4791]: I1210 22:50:01.587946 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 22:50:01 crc kubenswrapper[4791]: I1210 22:50:01.587988 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 22:50:01 crc kubenswrapper[4791]: I1210 22:50:01.587998 4791 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T22:50:01Z","lastTransitionTime":"2025-12-10T22:50:01Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 22:50:01 crc kubenswrapper[4791]: I1210 22:50:01.690593 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 22:50:01 crc kubenswrapper[4791]: I1210 22:50:01.690654 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 22:50:01 crc kubenswrapper[4791]: I1210 22:50:01.690665 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 22:50:01 crc kubenswrapper[4791]: I1210 22:50:01.690680 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 22:50:01 crc kubenswrapper[4791]: I1210 22:50:01.690688 4791 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T22:50:01Z","lastTransitionTime":"2025-12-10T22:50:01Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 22:50:01 crc kubenswrapper[4791]: I1210 22:50:01.793595 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 22:50:01 crc kubenswrapper[4791]: I1210 22:50:01.793645 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 22:50:01 crc kubenswrapper[4791]: I1210 22:50:01.793665 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 22:50:01 crc kubenswrapper[4791]: I1210 22:50:01.793689 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 22:50:01 crc kubenswrapper[4791]: I1210 22:50:01.793708 4791 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T22:50:01Z","lastTransitionTime":"2025-12-10T22:50:01Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 22:50:01 crc kubenswrapper[4791]: I1210 22:50:01.896850 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 22:50:01 crc kubenswrapper[4791]: I1210 22:50:01.896896 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 22:50:01 crc kubenswrapper[4791]: I1210 22:50:01.896908 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 22:50:01 crc kubenswrapper[4791]: I1210 22:50:01.896924 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 22:50:01 crc kubenswrapper[4791]: I1210 22:50:01.896935 4791 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T22:50:01Z","lastTransitionTime":"2025-12-10T22:50:01Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 22:50:01 crc kubenswrapper[4791]: I1210 22:50:01.999092 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 22:50:01 crc kubenswrapper[4791]: I1210 22:50:01.999155 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 22:50:01 crc kubenswrapper[4791]: I1210 22:50:01.999177 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 22:50:01 crc kubenswrapper[4791]: I1210 22:50:01.999206 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 22:50:01 crc kubenswrapper[4791]: I1210 22:50:01.999227 4791 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T22:50:01Z","lastTransitionTime":"2025-12-10T22:50:01Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 22:50:02 crc kubenswrapper[4791]: I1210 22:50:02.102183 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 22:50:02 crc kubenswrapper[4791]: I1210 22:50:02.102239 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 22:50:02 crc kubenswrapper[4791]: I1210 22:50:02.102250 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 22:50:02 crc kubenswrapper[4791]: I1210 22:50:02.102268 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 22:50:02 crc kubenswrapper[4791]: I1210 22:50:02.102279 4791 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T22:50:02Z","lastTransitionTime":"2025-12-10T22:50:02Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 22:50:02 crc kubenswrapper[4791]: I1210 22:50:02.205675 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 22:50:02 crc kubenswrapper[4791]: I1210 22:50:02.205734 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 22:50:02 crc kubenswrapper[4791]: I1210 22:50:02.205751 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 22:50:02 crc kubenswrapper[4791]: I1210 22:50:02.205775 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 22:50:02 crc kubenswrapper[4791]: I1210 22:50:02.205793 4791 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T22:50:02Z","lastTransitionTime":"2025-12-10T22:50:02Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 22:50:02 crc kubenswrapper[4791]: I1210 22:50:02.309172 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 22:50:02 crc kubenswrapper[4791]: I1210 22:50:02.309227 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 22:50:02 crc kubenswrapper[4791]: I1210 22:50:02.309260 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 22:50:02 crc kubenswrapper[4791]: I1210 22:50:02.309282 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 22:50:02 crc kubenswrapper[4791]: I1210 22:50:02.309299 4791 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T22:50:02Z","lastTransitionTime":"2025-12-10T22:50:02Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 22:50:02 crc kubenswrapper[4791]: I1210 22:50:02.412024 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 22:50:02 crc kubenswrapper[4791]: I1210 22:50:02.412092 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 22:50:02 crc kubenswrapper[4791]: I1210 22:50:02.412114 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 22:50:02 crc kubenswrapper[4791]: I1210 22:50:02.412142 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 22:50:02 crc kubenswrapper[4791]: I1210 22:50:02.412164 4791 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T22:50:02Z","lastTransitionTime":"2025-12-10T22:50:02Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 22:50:02 crc kubenswrapper[4791]: I1210 22:50:02.515417 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 22:50:02 crc kubenswrapper[4791]: I1210 22:50:02.515484 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 22:50:02 crc kubenswrapper[4791]: I1210 22:50:02.515507 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 22:50:02 crc kubenswrapper[4791]: I1210 22:50:02.515535 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 22:50:02 crc kubenswrapper[4791]: I1210 22:50:02.515555 4791 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T22:50:02Z","lastTransitionTime":"2025-12-10T22:50:02Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 22:50:02 crc kubenswrapper[4791]: I1210 22:50:02.618911 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 22:50:02 crc kubenswrapper[4791]: I1210 22:50:02.618953 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 22:50:02 crc kubenswrapper[4791]: I1210 22:50:02.618969 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 22:50:02 crc kubenswrapper[4791]: I1210 22:50:02.618990 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 22:50:02 crc kubenswrapper[4791]: I1210 22:50:02.619007 4791 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T22:50:02Z","lastTransitionTime":"2025-12-10T22:50:02Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 22:50:02 crc kubenswrapper[4791]: I1210 22:50:02.722060 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 22:50:02 crc kubenswrapper[4791]: I1210 22:50:02.722107 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 22:50:02 crc kubenswrapper[4791]: I1210 22:50:02.722121 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 22:50:02 crc kubenswrapper[4791]: I1210 22:50:02.722142 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 22:50:02 crc kubenswrapper[4791]: I1210 22:50:02.722155 4791 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T22:50:02Z","lastTransitionTime":"2025-12-10T22:50:02Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 22:50:02 crc kubenswrapper[4791]: I1210 22:50:02.824444 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 22:50:02 crc kubenswrapper[4791]: I1210 22:50:02.824487 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 22:50:02 crc kubenswrapper[4791]: I1210 22:50:02.824498 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 22:50:02 crc kubenswrapper[4791]: I1210 22:50:02.824518 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 22:50:02 crc kubenswrapper[4791]: I1210 22:50:02.824530 4791 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T22:50:02Z","lastTransitionTime":"2025-12-10T22:50:02Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 22:50:02 crc kubenswrapper[4791]: I1210 22:50:02.884581 4791 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 10 22:50:02 crc kubenswrapper[4791]: I1210 22:50:02.884637 4791 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 10 22:50:02 crc kubenswrapper[4791]: I1210 22:50:02.884779 4791 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 10 22:50:02 crc kubenswrapper[4791]: E1210 22:50:02.884911 4791 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 10 22:50:02 crc kubenswrapper[4791]: E1210 22:50:02.885088 4791 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 10 22:50:02 crc kubenswrapper[4791]: E1210 22:50:02.885271 4791 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 10 22:50:02 crc kubenswrapper[4791]: I1210 22:50:02.927826 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 22:50:02 crc kubenswrapper[4791]: I1210 22:50:02.927897 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 22:50:02 crc kubenswrapper[4791]: I1210 22:50:02.927931 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 22:50:02 crc kubenswrapper[4791]: I1210 22:50:02.927964 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 22:50:02 crc kubenswrapper[4791]: I1210 22:50:02.927988 4791 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T22:50:02Z","lastTransitionTime":"2025-12-10T22:50:02Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 22:50:03 crc kubenswrapper[4791]: I1210 22:50:03.031333 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 22:50:03 crc kubenswrapper[4791]: I1210 22:50:03.031458 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 22:50:03 crc kubenswrapper[4791]: I1210 22:50:03.031488 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 22:50:03 crc kubenswrapper[4791]: I1210 22:50:03.031519 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 22:50:03 crc kubenswrapper[4791]: I1210 22:50:03.031541 4791 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T22:50:03Z","lastTransitionTime":"2025-12-10T22:50:03Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 22:50:03 crc kubenswrapper[4791]: I1210 22:50:03.134139 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 22:50:03 crc kubenswrapper[4791]: I1210 22:50:03.134193 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 22:50:03 crc kubenswrapper[4791]: I1210 22:50:03.134206 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 22:50:03 crc kubenswrapper[4791]: I1210 22:50:03.134230 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 22:50:03 crc kubenswrapper[4791]: I1210 22:50:03.134247 4791 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T22:50:03Z","lastTransitionTime":"2025-12-10T22:50:03Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 22:50:03 crc kubenswrapper[4791]: I1210 22:50:03.237460 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 22:50:03 crc kubenswrapper[4791]: I1210 22:50:03.237535 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 22:50:03 crc kubenswrapper[4791]: I1210 22:50:03.237555 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 22:50:03 crc kubenswrapper[4791]: I1210 22:50:03.237585 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 22:50:03 crc kubenswrapper[4791]: I1210 22:50:03.237610 4791 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T22:50:03Z","lastTransitionTime":"2025-12-10T22:50:03Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 22:50:03 crc kubenswrapper[4791]: I1210 22:50:03.335757 4791 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-zhq64_3cd47739-0fa9-4321-aff1-220f8721a0b3/ovnkube-controller/0.log" Dec 10 22:50:03 crc kubenswrapper[4791]: I1210 22:50:03.339428 4791 generic.go:334] "Generic (PLEG): container finished" podID="3cd47739-0fa9-4321-aff1-220f8721a0b3" containerID="7a03d21e34f9d560953b7a809a4d5f374e39d3daeb8e698dbb5861b8e63c9de5" exitCode=1 Dec 10 22:50:03 crc kubenswrapper[4791]: I1210 22:50:03.339527 4791 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-zhq64" event={"ID":"3cd47739-0fa9-4321-aff1-220f8721a0b3","Type":"ContainerDied","Data":"7a03d21e34f9d560953b7a809a4d5f374e39d3daeb8e698dbb5861b8e63c9de5"} Dec 10 22:50:03 crc kubenswrapper[4791]: I1210 22:50:03.340027 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 22:50:03 crc kubenswrapper[4791]: I1210 22:50:03.340078 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 22:50:03 crc kubenswrapper[4791]: I1210 22:50:03.340096 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 22:50:03 crc kubenswrapper[4791]: I1210 22:50:03.340121 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 22:50:03 crc kubenswrapper[4791]: I1210 22:50:03.340140 4791 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T22:50:03Z","lastTransitionTime":"2025-12-10T22:50:03Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 22:50:03 crc kubenswrapper[4791]: I1210 22:50:03.340358 4791 scope.go:117] "RemoveContainer" containerID="7a03d21e34f9d560953b7a809a4d5f374e39d3daeb8e698dbb5861b8e63c9de5" Dec 10 22:50:03 crc kubenswrapper[4791]: I1210 22:50:03.365197 4791 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"90250e13-6c63-443a-a5e5-6e4e341a2289\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T22:49:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T22:49:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T22:49:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T22:49:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T22:49:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ba54f95b1e00843b854db9a10fe6897592159ccacf30ed0f76002093a13eedd3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T22:49:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5aa8ffa518f3a68b60b275f8df4851b2564aa4984b61efdc9470b73072133e43\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T22:49:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://16f989503edc78d9451638e2cf6cf4b19a2cf2b8809b10b237541cb829303091\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T22:49:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://82556824d63ce6e46df14c45689ff1f81ebf729abe902061815c040b323898e9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T22:49:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://61db531940fa6f2effdae5a005b4c91bd2698bd5d1bff9588c171a65ad550eca\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T22:49:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0f65c864d885319ce2a08d61c30be24b3ffb7d23ea30c82ed7faf039cb2c006d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://0f65c864d885319ce2a08d61c30be24b3ffb7d23ea30c82ed7faf039cb2c006d\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-10T22:49:27Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-10T22:49:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e31a76d1d5b2812ba83e24e38268e25ec8bed98fc3da9adcd06bda5822b6ace1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e31a76d1d5b2812ba83e24e38268e25ec8bed98fc3da9adcd06bda5822b6ace1\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-10T22:49:30Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-10T22:49:29Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://b5b5eceae562f3986b543d41de1e977efb85a91d84c98e09159ca6d092eda691\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b5b5eceae562f3986b543d41de1e977efb85a91d84c98e09159ca6d092eda691\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-10T22:49:31Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-10T22:49:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-10T22:49:25Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T22:50:03Z is after 2025-08-24T17:21:41Z" Dec 10 22:50:03 crc kubenswrapper[4791]: I1210 22:50:03.378101 4791 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-s6rk8"] Dec 10 22:50:03 crc kubenswrapper[4791]: I1210 22:50:03.378782 4791 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-s6rk8" Dec 10 22:50:03 crc kubenswrapper[4791]: I1210 22:50:03.381652 4791 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ovn-kubernetes"/"ovn-control-plane-metrics-cert" Dec 10 22:50:03 crc kubenswrapper[4791]: I1210 22:50:03.382050 4791 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-10T22:49:50Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-10T22:49:50Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T22:50:03Z is after 2025-08-24T17:21:41Z" Dec 10 22:50:03 crc kubenswrapper[4791]: I1210 22:50:03.383056 4791 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ovn-kubernetes"/"ovn-kubernetes-control-plane-dockercfg-gs7dd" Dec 10 22:50:03 crc kubenswrapper[4791]: I1210 22:50:03.404798 4791 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-10T22:49:54Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-10T22:49:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-10T22:49:54Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://62bfaf5d7a6d1b452cdaf88086447d0c8cb7306e94434e46752be5475124d56a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T22:49:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T22:50:03Z is after 2025-08-24T17:21:41Z" Dec 10 22:50:03 crc kubenswrapper[4791]: I1210 22:50:03.421243 4791 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f8349fea-c49b-49a1-b5ee-032c5dffe021\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T22:49:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T22:49:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T22:49:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T22:49:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T22:49:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://acafe12186df99d1059a31260de0eafce35ff7771bcb8f5c6753b72665d6f60b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T22:49:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ea4c346507862a8c12756a3061d6e25e788a8d79f4621b62f81f4322185c82a8\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T22:49:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://2e6e324ec6f4f7bb69fdfab4ea3ac1aa55e480dd8e2a8984b0fe5b4c9f66636e\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T22:49:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://1b2e11ff2bf8789fa1d0d23ae648b3e6ef653106e5bc4bbb5dd6447f7cd6f05d\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d6aa7f8c5d908d90008f2df49860ac8a3480136da567e3d85f86135d779204f9\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-12-10T22:49:44Z\\\",\\\"message\\\":\\\"W1210 22:49:33.845733 1 cmd.go:257] Using insecure, self-signed certificates\\\\nI1210 22:49:33.846108 1 crypto.go:601] Generating new CA for check-endpoints-signer@1765406973 cert, and key in /tmp/serving-cert-2585701456/serving-signer.crt, /tmp/serving-cert-2585701456/serving-signer.key\\\\nI1210 22:49:34.529748 1 observer_polling.go:159] Starting file observer\\\\nW1210 22:49:34.532225 1 builder.go:272] unable to get owner reference (falling back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": dial tcp [::1]:6443: connect: connection refused\\\\nI1210 22:49:34.532599 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1210 22:49:34.533977 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-2585701456/tls.crt::/tmp/serving-cert-2585701456/tls.key\\\\\\\"\\\\nF1210 22:49:44.774710 1 cmd.go:182] error initializing delegating authentication: unable to load configmap based request-header-client-ca-file: Get \\\\\\\"https://localhost:6443/api/v1/namespaces/kube-system/configmaps/extension-apiserver-authentication\\\\\\\": net/http: TLS handshake timeout\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-10T22:49:33Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T22:49:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://0c4d036ab269f62561fe7ef1586012cb3a16e3583a08b9488a0616ec00c8e0ae\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T22:49:32Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c5856d6b26c19fda13b5c9ac2892a7e2f9c6f79feb8173ecfd9b33b673327d67\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c5856d6b26c19fda13b5c9ac2892a7e2f9c6f79feb8173ecfd9b33b673327d67\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-10T22:49:27Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-10T22:49:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-10T22:49:25Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T22:50:03Z is after 2025-08-24T17:21:41Z" Dec 10 22:50:03 crc kubenswrapper[4791]: I1210 22:50:03.433955 4791 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-5rb5l" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ba35653c-6e06-4cee-a4d6-137764090d18\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T22:49:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T22:49:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T22:49:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T22:49:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://dbb5c875730a98ad2daf316cfafc08fcdce21ac16148154780ac9193ed405dab\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T22:49:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fvvrs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://185eb2ff44221a514878283ec873773a6c9a9073dc64e296f982e60dee1338e4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T22:49:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fvvrs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-10T22:49:51Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-5rb5l\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T22:50:03Z is after 2025-08-24T17:21:41Z" Dec 10 22:50:03 crc kubenswrapper[4791]: I1210 22:50:03.434130 4791 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovn-control-plane-metrics-cert\" (UniqueName: \"kubernetes.io/secret/5c9d0978-a6f9-49c5-b377-6619b4e3ff52-ovn-control-plane-metrics-cert\") pod \"ovnkube-control-plane-749d76644c-s6rk8\" (UID: \"5c9d0978-a6f9-49c5-b377-6619b4e3ff52\") " pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-s6rk8" Dec 10 22:50:03 crc kubenswrapper[4791]: I1210 22:50:03.434182 4791 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-7vkgm\" (UniqueName: \"kubernetes.io/projected/5c9d0978-a6f9-49c5-b377-6619b4e3ff52-kube-api-access-7vkgm\") pod \"ovnkube-control-plane-749d76644c-s6rk8\" (UID: \"5c9d0978-a6f9-49c5-b377-6619b4e3ff52\") " pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-s6rk8" Dec 10 22:50:03 crc kubenswrapper[4791]: I1210 22:50:03.434244 4791 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/5c9d0978-a6f9-49c5-b377-6619b4e3ff52-ovnkube-config\") pod \"ovnkube-control-plane-749d76644c-s6rk8\" (UID: \"5c9d0978-a6f9-49c5-b377-6619b4e3ff52\") " pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-s6rk8" Dec 10 22:50:03 crc kubenswrapper[4791]: I1210 22:50:03.434268 4791 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/5c9d0978-a6f9-49c5-b377-6619b4e3ff52-env-overrides\") pod \"ovnkube-control-plane-749d76644c-s6rk8\" (UID: \"5c9d0978-a6f9-49c5-b377-6619b4e3ff52\") " pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-s6rk8" Dec 10 22:50:03 crc kubenswrapper[4791]: I1210 22:50:03.442560 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 22:50:03 crc kubenswrapper[4791]: I1210 22:50:03.442603 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 22:50:03 crc kubenswrapper[4791]: I1210 22:50:03.442612 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 22:50:03 crc kubenswrapper[4791]: I1210 22:50:03.442627 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 22:50:03 crc kubenswrapper[4791]: I1210 22:50:03.442636 4791 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T22:50:03Z","lastTransitionTime":"2025-12-10T22:50:03Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 22:50:03 crc kubenswrapper[4791]: I1210 22:50:03.449154 4791 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-10T22:49:52Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-10T22:49:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-10T22:49:52Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3734b110802828632d44263ed0a43cabc4b710833f0be120d72db0940fe783e2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T22:49:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://794803ac9be3c52509e6dcb10854d8864e1a9e406762e57c7971b07c4da1220e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T22:49:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T22:50:03Z is after 2025-08-24T17:21:41Z" Dec 10 22:50:03 crc kubenswrapper[4791]: I1210 22:50:03.461726 4791 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-vczjq" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b9f00f25-3fa1-4b63-9710-fd352224b01b\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T22:49:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T22:49:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T22:49:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T22:49:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4698cf4c6b55b41e628448d37c2eb847c2777d0efbd93afcbf367d5e78623b7e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T22:49:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lrrpv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-10T22:49:50Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-vczjq\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T22:50:03Z is after 2025-08-24T17:21:41Z" Dec 10 22:50:03 crc kubenswrapper[4791]: I1210 22:50:03.474243 4791 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-4nwkq" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"672aa28c-8169-49ed-87b8-21187d13a80c\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T22:49:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T22:49:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T22:49:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T22:49:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a2f2a62315529e148a10a33ccb241e9417fbde0fe42afb61cb7ee2438c74636f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T22:49:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xl62j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-10T22:49:51Z\\\"}}\" for pod \"openshift-multus\"/\"multus-4nwkq\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T22:50:03Z is after 2025-08-24T17:21:41Z" Dec 10 22:50:03 crc kubenswrapper[4791]: I1210 22:50:03.491220 4791 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-zhq64" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3cd47739-0fa9-4321-aff1-220f8721a0b3\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T22:49:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T22:49:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T22:49:52Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T22:49:52Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://68795a29944e56870ffe94fa732e9dcf0bfb344c0d85d1e590f58a876129c61f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T22:49:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wvpw8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://bc721b530c08b433756e978ec0423cc18aabf6771b61eb96f903b4d57ea335ff\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T22:49:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wvpw8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://152a7618be8f38fadf4e564b3948980503cedcdb466c37344ccc20f4a3173a5f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T22:49:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wvpw8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3d18f9cb7207efb1e32de75a30d33e024f656a08f0503db95e5488bc5280971f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T22:49:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wvpw8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://937137ae9d7480cdc7c59bc4a1df0a4c10e72b3b63693af285278789953ed218\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T22:49:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wvpw8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://1a52d19b7671e8b143017c8a981a6f8d3ee3b84650e4b223b814a531ac76f364\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T22:49:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wvpw8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://7a03d21e34f9d560953b7a809a4d5f374e39d3daeb8e698dbb5861b8e63c9de5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7a03d21e34f9d560953b7a809a4d5f374e39d3daeb8e698dbb5861b8e63c9de5\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-12-10T22:50:03Z\\\",\\\"message\\\":\\\"r removal\\\\nI1210 22:50:01.628167 6086 handler.go:190] Sending *v1.EgressIP event handler 8 for removal\\\\nI1210 22:50:01.628168 6086 handler.go:208] Removed *v1.NetworkPolicy event handler 4\\\\nI1210 22:50:01.628234 6086 handler.go:208] Removed *v1.Namespace event handler 1\\\\nI1210 22:50:01.628254 6086 handler.go:208] Removed *v1.Namespace event handler 5\\\\nI1210 22:50:01.628258 6086 handler.go:190] Sending *v1.Pod event handler 6 for removal\\\\nI1210 22:50:01.628268 6086 handler.go:190] Sending *v1.Pod event handler 3 for removal\\\\nI1210 22:50:01.628292 6086 handler.go:190] Sending *v1.Node event handler 2 for removal\\\\nI1210 22:50:01.628301 6086 handler.go:190] Sending *v1.Node event handler 7 for removal\\\\nI1210 22:50:01.628373 6086 handler.go:208] Removed *v1.Pod event handler 6\\\\nI1210 22:50:01.628392 6086 handler.go:208] Removed *v1.EgressIP event handler 8\\\\nI1210 22:50:01.628406 6086 handler.go:208] Removed *v1.Node event handler 2\\\\nI1210 22:50:01.628413 6086 handler.go:190] Sending *v1.EgressFirewall event handler 9 for removal\\\\nI1210 22:50:01.628412 6086 handler.go:208] Removed *v1.Pod event handler 3\\\\nI1210 22:50:01.628462 6086 factory.go:656] Stopping watch factory\\\\nI1210 22:50:01.628492 6086 ovnkube.go:599] Stopped ovnkube\\\\nI1210 22:50:01.628536 6086 handler.go:208] Removed *v1.EgressFirewall event handler 9\\\\nI12\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-10T22:49:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wvpw8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://dd3f80e5713221f96ed50da7f2e7fd5b9fabfcd5c1656da8d8a86d818a5fdc3f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T22:49:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wvpw8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3afbd1ffcd2c4b40324aca6e52d6fc822db131c3f35a9af3a6f005c2cffdb7bf\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3afbd1ffcd2c4b40324aca6e52d6fc822db131c3f35a9af3a6f005c2cffdb7bf\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-10T22:49:53Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-10T22:49:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wvpw8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-10T22:49:52Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-zhq64\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T22:50:03Z is after 2025-08-24T17:21:41Z" Dec 10 22:50:03 crc kubenswrapper[4791]: I1210 22:50:03.505692 4791 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"698bbe9a-fbe5-436d-a835-5e80df77ba24\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T22:49:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T22:49:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T22:49:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T22:49:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T22:49:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://224d73713312376a58ec6cc2bc8fdb6ffa44c4749ae0e7691d5fe27b9afca09f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T22:49:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://372eeaffd40992c6f06efc4df6dfcb44b31de64ee8256c21508cdd7c74ecb96e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T22:49:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://76b6013b062cac5b0b4d12a94f38ce7b43f9c7ea216063455d88a956ac04d26e\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T22:49:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://78e950e122e761d950da1a3c60c06476e53920e8795f6c61db22c01e22d5136e\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T22:49:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-10T22:49:25Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T22:50:03Z is after 2025-08-24T17:21:41Z" Dec 10 22:50:03 crc kubenswrapper[4791]: I1210 22:50:03.517816 4791 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-10T22:49:50Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-10T22:49:50Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T22:50:03Z is after 2025-08-24T17:21:41Z" Dec 10 22:50:03 crc kubenswrapper[4791]: I1210 22:50:03.531332 4791 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-10T22:49:50Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-10T22:49:50Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T22:50:03Z is after 2025-08-24T17:21:41Z" Dec 10 22:50:03 crc kubenswrapper[4791]: I1210 22:50:03.534969 4791 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovn-control-plane-metrics-cert\" (UniqueName: \"kubernetes.io/secret/5c9d0978-a6f9-49c5-b377-6619b4e3ff52-ovn-control-plane-metrics-cert\") pod \"ovnkube-control-plane-749d76644c-s6rk8\" (UID: \"5c9d0978-a6f9-49c5-b377-6619b4e3ff52\") " pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-s6rk8" Dec 10 22:50:03 crc kubenswrapper[4791]: I1210 22:50:03.535057 4791 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-7vkgm\" (UniqueName: \"kubernetes.io/projected/5c9d0978-a6f9-49c5-b377-6619b4e3ff52-kube-api-access-7vkgm\") pod \"ovnkube-control-plane-749d76644c-s6rk8\" (UID: \"5c9d0978-a6f9-49c5-b377-6619b4e3ff52\") " pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-s6rk8" Dec 10 22:50:03 crc kubenswrapper[4791]: I1210 22:50:03.535107 4791 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/5c9d0978-a6f9-49c5-b377-6619b4e3ff52-ovnkube-config\") pod \"ovnkube-control-plane-749d76644c-s6rk8\" (UID: \"5c9d0978-a6f9-49c5-b377-6619b4e3ff52\") " pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-s6rk8" Dec 10 22:50:03 crc kubenswrapper[4791]: I1210 22:50:03.535144 4791 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/5c9d0978-a6f9-49c5-b377-6619b4e3ff52-env-overrides\") pod \"ovnkube-control-plane-749d76644c-s6rk8\" (UID: \"5c9d0978-a6f9-49c5-b377-6619b4e3ff52\") " pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-s6rk8" Dec 10 22:50:03 crc kubenswrapper[4791]: I1210 22:50:03.535849 4791 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/5c9d0978-a6f9-49c5-b377-6619b4e3ff52-ovnkube-config\") pod \"ovnkube-control-plane-749d76644c-s6rk8\" (UID: \"5c9d0978-a6f9-49c5-b377-6619b4e3ff52\") " pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-s6rk8" Dec 10 22:50:03 crc kubenswrapper[4791]: I1210 22:50:03.535930 4791 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/5c9d0978-a6f9-49c5-b377-6619b4e3ff52-env-overrides\") pod \"ovnkube-control-plane-749d76644c-s6rk8\" (UID: \"5c9d0978-a6f9-49c5-b377-6619b4e3ff52\") " pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-s6rk8" Dec 10 22:50:03 crc kubenswrapper[4791]: I1210 22:50:03.543003 4791 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovn-control-plane-metrics-cert\" (UniqueName: \"kubernetes.io/secret/5c9d0978-a6f9-49c5-b377-6619b4e3ff52-ovn-control-plane-metrics-cert\") pod \"ovnkube-control-plane-749d76644c-s6rk8\" (UID: \"5c9d0978-a6f9-49c5-b377-6619b4e3ff52\") " pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-s6rk8" Dec 10 22:50:03 crc kubenswrapper[4791]: I1210 22:50:03.543763 4791 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-tchxk" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"c0ab74e0-42a9-4f42-ac80-16030cf4ffff\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T22:49:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T22:49:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T22:49:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T22:49:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://04e0fd1b53256b7a938e8ccab93e2a4eb73dd1a59e0b9799e09c55e511baf7fe\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T22:49:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zn2qt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-10T22:49:50Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-tchxk\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T22:50:03Z is after 2025-08-24T17:21:41Z" Dec 10 22:50:03 crc kubenswrapper[4791]: I1210 22:50:03.544425 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 22:50:03 crc kubenswrapper[4791]: I1210 22:50:03.544458 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 22:50:03 crc kubenswrapper[4791]: I1210 22:50:03.544470 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 22:50:03 crc kubenswrapper[4791]: I1210 22:50:03.544491 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 22:50:03 crc kubenswrapper[4791]: I1210 22:50:03.544503 4791 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T22:50:03Z","lastTransitionTime":"2025-12-10T22:50:03Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 22:50:03 crc kubenswrapper[4791]: I1210 22:50:03.553500 4791 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-7vkgm\" (UniqueName: \"kubernetes.io/projected/5c9d0978-a6f9-49c5-b377-6619b4e3ff52-kube-api-access-7vkgm\") pod \"ovnkube-control-plane-749d76644c-s6rk8\" (UID: \"5c9d0978-a6f9-49c5-b377-6619b4e3ff52\") " pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-s6rk8" Dec 10 22:50:03 crc kubenswrapper[4791]: I1210 22:50:03.562396 4791 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-rw584" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ff9beb90-69b0-4732-bf37-0b81f58ecc98\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T22:49:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T22:49:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T22:49:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T22:49:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ddb9df501dc6e8e5561fbc2aefd431037a3f28f73bf8e82a6598d261a4224807\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T22:49:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fksqn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ce67e54b696bb53ffbbe48c6ce71476b635e0868993e9c427b22f904891433ee\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ce67e54b696bb53ffbbe48c6ce71476b635e0868993e9c427b22f904891433ee\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-10T22:49:52Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-10T22:49:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fksqn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://818d89bfffb8a110670152c8480684665ea8db578e9768f524fbbe6a92f27c22\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://818d89bfffb8a110670152c8480684665ea8db578e9768f524fbbe6a92f27c22\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-10T22:49:53Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-10T22:49:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fksqn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://9aa9c06e894bb2a2ed3738defaeff68a0a0458ead48177f36b90770944c211a0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://9aa9c06e894bb2a2ed3738defaeff68a0a0458ead48177f36b90770944c211a0\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-10T22:49:54Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-10T22:49:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fksqn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://dfd6d03336dcea7529419a30e8d70d84e2290a0fa5b0e46d09ad341c190638d6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://dfd6d03336dcea7529419a30e8d70d84e2290a0fa5b0e46d09ad341c190638d6\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-10T22:49:55Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-10T22:49:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fksqn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b86738401ece6fbc3083782720c66a515850ed78d4054f34fcabab7ff525b85f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b86738401ece6fbc3083782720c66a515850ed78d4054f34fcabab7ff525b85f\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-10T22:49:56Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-10T22:49:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fksqn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://18b326f572a899ddd10197a745bf9264d4ddd975690ad1cfb0ec576ccd5cf102\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://18b326f572a899ddd10197a745bf9264d4ddd975690ad1cfb0ec576ccd5cf102\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-10T22:49:57Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-10T22:49:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fksqn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-10T22:49:51Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-rw584\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T22:50:03Z is after 2025-08-24T17:21:41Z" Dec 10 22:50:03 crc kubenswrapper[4791]: I1210 22:50:03.576980 4791 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-10T22:49:52Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-10T22:49:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-10T22:49:52Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://20651eb0c16e2c71222f56da1ff755cb0e7b084886acbbb24f0ccaf57041ed85\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T22:49:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T22:50:03Z is after 2025-08-24T17:21:41Z" Dec 10 22:50:03 crc kubenswrapper[4791]: I1210 22:50:03.587790 4791 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-tchxk" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"c0ab74e0-42a9-4f42-ac80-16030cf4ffff\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T22:49:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T22:49:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T22:49:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T22:49:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://04e0fd1b53256b7a938e8ccab93e2a4eb73dd1a59e0b9799e09c55e511baf7fe\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T22:49:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zn2qt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-10T22:49:50Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-tchxk\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T22:50:03Z is after 2025-08-24T17:21:41Z" Dec 10 22:50:03 crc kubenswrapper[4791]: I1210 22:50:03.602081 4791 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-rw584" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ff9beb90-69b0-4732-bf37-0b81f58ecc98\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T22:49:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T22:49:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T22:49:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T22:49:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ddb9df501dc6e8e5561fbc2aefd431037a3f28f73bf8e82a6598d261a4224807\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T22:49:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fksqn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ce67e54b696bb53ffbbe48c6ce71476b635e0868993e9c427b22f904891433ee\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ce67e54b696bb53ffbbe48c6ce71476b635e0868993e9c427b22f904891433ee\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-10T22:49:52Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-10T22:49:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fksqn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://818d89bfffb8a110670152c8480684665ea8db578e9768f524fbbe6a92f27c22\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://818d89bfffb8a110670152c8480684665ea8db578e9768f524fbbe6a92f27c22\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-10T22:49:53Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-10T22:49:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fksqn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://9aa9c06e894bb2a2ed3738defaeff68a0a0458ead48177f36b90770944c211a0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://9aa9c06e894bb2a2ed3738defaeff68a0a0458ead48177f36b90770944c211a0\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-10T22:49:54Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-10T22:49:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fksqn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://dfd6d03336dcea7529419a30e8d70d84e2290a0fa5b0e46d09ad341c190638d6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://dfd6d03336dcea7529419a30e8d70d84e2290a0fa5b0e46d09ad341c190638d6\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-10T22:49:55Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-10T22:49:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fksqn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b86738401ece6fbc3083782720c66a515850ed78d4054f34fcabab7ff525b85f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b86738401ece6fbc3083782720c66a515850ed78d4054f34fcabab7ff525b85f\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-10T22:49:56Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-10T22:49:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fksqn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://18b326f572a899ddd10197a745bf9264d4ddd975690ad1cfb0ec576ccd5cf102\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://18b326f572a899ddd10197a745bf9264d4ddd975690ad1cfb0ec576ccd5cf102\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-10T22:49:57Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-10T22:49:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fksqn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-10T22:49:51Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-rw584\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T22:50:03Z is after 2025-08-24T17:21:41Z" Dec 10 22:50:03 crc kubenswrapper[4791]: I1210 22:50:03.622477 4791 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-10T22:49:52Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-10T22:49:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-10T22:49:52Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://20651eb0c16e2c71222f56da1ff755cb0e7b084886acbbb24f0ccaf57041ed85\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T22:49:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T22:50:03Z is after 2025-08-24T17:21:41Z" Dec 10 22:50:03 crc kubenswrapper[4791]: I1210 22:50:03.636780 4791 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-10T22:49:50Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-10T22:49:50Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T22:50:03Z is after 2025-08-24T17:21:41Z" Dec 10 22:50:03 crc kubenswrapper[4791]: I1210 22:50:03.647694 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 22:50:03 crc kubenswrapper[4791]: I1210 22:50:03.647742 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 22:50:03 crc kubenswrapper[4791]: I1210 22:50:03.647756 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 22:50:03 crc kubenswrapper[4791]: I1210 22:50:03.647773 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 22:50:03 crc kubenswrapper[4791]: I1210 22:50:03.647786 4791 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T22:50:03Z","lastTransitionTime":"2025-12-10T22:50:03Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 22:50:03 crc kubenswrapper[4791]: I1210 22:50:03.651556 4791 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-10T22:49:50Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-10T22:49:50Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T22:50:03Z is after 2025-08-24T17:21:41Z" Dec 10 22:50:03 crc kubenswrapper[4791]: I1210 22:50:03.665245 4791 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-10T22:49:50Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-10T22:49:50Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T22:50:03Z is after 2025-08-24T17:21:41Z" Dec 10 22:50:03 crc kubenswrapper[4791]: I1210 22:50:03.679032 4791 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-10T22:49:54Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-10T22:49:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-10T22:49:54Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://62bfaf5d7a6d1b452cdaf88086447d0c8cb7306e94434e46752be5475124d56a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T22:49:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T22:50:03Z is after 2025-08-24T17:21:41Z" Dec 10 22:50:03 crc kubenswrapper[4791]: I1210 22:50:03.694523 4791 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f8349fea-c49b-49a1-b5ee-032c5dffe021\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T22:49:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T22:49:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T22:49:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T22:49:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T22:49:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://acafe12186df99d1059a31260de0eafce35ff7771bcb8f5c6753b72665d6f60b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T22:49:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ea4c346507862a8c12756a3061d6e25e788a8d79f4621b62f81f4322185c82a8\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T22:49:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://2e6e324ec6f4f7bb69fdfab4ea3ac1aa55e480dd8e2a8984b0fe5b4c9f66636e\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T22:49:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://1b2e11ff2bf8789fa1d0d23ae648b3e6ef653106e5bc4bbb5dd6447f7cd6f05d\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d6aa7f8c5d908d90008f2df49860ac8a3480136da567e3d85f86135d779204f9\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-12-10T22:49:44Z\\\",\\\"message\\\":\\\"W1210 22:49:33.845733 1 cmd.go:257] Using insecure, self-signed certificates\\\\nI1210 22:49:33.846108 1 crypto.go:601] Generating new CA for check-endpoints-signer@1765406973 cert, and key in /tmp/serving-cert-2585701456/serving-signer.crt, /tmp/serving-cert-2585701456/serving-signer.key\\\\nI1210 22:49:34.529748 1 observer_polling.go:159] Starting file observer\\\\nW1210 22:49:34.532225 1 builder.go:272] unable to get owner reference (falling back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": dial tcp [::1]:6443: connect: connection refused\\\\nI1210 22:49:34.532599 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1210 22:49:34.533977 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-2585701456/tls.crt::/tmp/serving-cert-2585701456/tls.key\\\\\\\"\\\\nF1210 22:49:44.774710 1 cmd.go:182] error initializing delegating authentication: unable to load configmap based request-header-client-ca-file: Get \\\\\\\"https://localhost:6443/api/v1/namespaces/kube-system/configmaps/extension-apiserver-authentication\\\\\\\": net/http: TLS handshake timeout\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-10T22:49:33Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T22:49:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://0c4d036ab269f62561fe7ef1586012cb3a16e3583a08b9488a0616ec00c8e0ae\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T22:49:32Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c5856d6b26c19fda13b5c9ac2892a7e2f9c6f79feb8173ecfd9b33b673327d67\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c5856d6b26c19fda13b5c9ac2892a7e2f9c6f79feb8173ecfd9b33b673327d67\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-10T22:49:27Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-10T22:49:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-10T22:49:25Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T22:50:03Z is after 2025-08-24T17:21:41Z" Dec 10 22:50:03 crc kubenswrapper[4791]: I1210 22:50:03.715821 4791 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"90250e13-6c63-443a-a5e5-6e4e341a2289\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T22:49:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T22:49:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T22:49:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T22:49:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T22:49:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ba54f95b1e00843b854db9a10fe6897592159ccacf30ed0f76002093a13eedd3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T22:49:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5aa8ffa518f3a68b60b275f8df4851b2564aa4984b61efdc9470b73072133e43\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T22:49:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://16f989503edc78d9451638e2cf6cf4b19a2cf2b8809b10b237541cb829303091\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T22:49:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://82556824d63ce6e46df14c45689ff1f81ebf729abe902061815c040b323898e9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T22:49:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://61db531940fa6f2effdae5a005b4c91bd2698bd5d1bff9588c171a65ad550eca\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T22:49:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0f65c864d885319ce2a08d61c30be24b3ffb7d23ea30c82ed7faf039cb2c006d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://0f65c864d885319ce2a08d61c30be24b3ffb7d23ea30c82ed7faf039cb2c006d\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-10T22:49:27Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-10T22:49:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e31a76d1d5b2812ba83e24e38268e25ec8bed98fc3da9adcd06bda5822b6ace1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e31a76d1d5b2812ba83e24e38268e25ec8bed98fc3da9adcd06bda5822b6ace1\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-10T22:49:30Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-10T22:49:29Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://b5b5eceae562f3986b543d41de1e977efb85a91d84c98e09159ca6d092eda691\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b5b5eceae562f3986b543d41de1e977efb85a91d84c98e09159ca6d092eda691\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-10T22:49:31Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-10T22:49:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-10T22:49:25Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T22:50:03Z is after 2025-08-24T17:21:41Z" Dec 10 22:50:03 crc kubenswrapper[4791]: I1210 22:50:03.725883 4791 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-s6rk8" Dec 10 22:50:03 crc kubenswrapper[4791]: I1210 22:50:03.732841 4791 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-5rb5l" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ba35653c-6e06-4cee-a4d6-137764090d18\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T22:49:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T22:49:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T22:49:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T22:49:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://dbb5c875730a98ad2daf316cfafc08fcdce21ac16148154780ac9193ed405dab\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T22:49:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fvvrs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://185eb2ff44221a514878283ec873773a6c9a9073dc64e296f982e60dee1338e4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T22:49:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fvvrs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-10T22:49:51Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-5rb5l\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T22:50:03Z is after 2025-08-24T17:21:41Z" Dec 10 22:50:03 crc kubenswrapper[4791]: W1210 22:50:03.739597 4791 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod5c9d0978_a6f9_49c5_b377_6619b4e3ff52.slice/crio-11613be462f4f652fb71299fa5b3bba520c7d3dc7d688b749d3642352bf96177 WatchSource:0}: Error finding container 11613be462f4f652fb71299fa5b3bba520c7d3dc7d688b749d3642352bf96177: Status 404 returned error can't find the container with id 11613be462f4f652fb71299fa5b3bba520c7d3dc7d688b749d3642352bf96177 Dec 10 22:50:03 crc kubenswrapper[4791]: I1210 22:50:03.748003 4791 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-s6rk8" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5c9d0978-a6f9-49c5-b377-6619b4e3ff52\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T22:50:03Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T22:50:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T22:50:03Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-rbac-proxy ovnkube-cluster-manager]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T22:50:03Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-rbac-proxy ovnkube-cluster-manager]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7vkgm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7vkgm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-10T22:50:03Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-s6rk8\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T22:50:03Z is after 2025-08-24T17:21:41Z" Dec 10 22:50:03 crc kubenswrapper[4791]: I1210 22:50:03.749879 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 22:50:03 crc kubenswrapper[4791]: I1210 22:50:03.749953 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 22:50:03 crc kubenswrapper[4791]: I1210 22:50:03.749971 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 22:50:03 crc kubenswrapper[4791]: I1210 22:50:03.750023 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 22:50:03 crc kubenswrapper[4791]: I1210 22:50:03.750040 4791 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T22:50:03Z","lastTransitionTime":"2025-12-10T22:50:03Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 22:50:03 crc kubenswrapper[4791]: I1210 22:50:03.764762 4791 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-vczjq" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b9f00f25-3fa1-4b63-9710-fd352224b01b\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T22:49:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T22:49:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T22:49:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T22:49:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4698cf4c6b55b41e628448d37c2eb847c2777d0efbd93afcbf367d5e78623b7e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T22:49:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lrrpv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-10T22:49:50Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-vczjq\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T22:50:03Z is after 2025-08-24T17:21:41Z" Dec 10 22:50:03 crc kubenswrapper[4791]: I1210 22:50:03.779125 4791 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-4nwkq" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"672aa28c-8169-49ed-87b8-21187d13a80c\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T22:49:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T22:49:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T22:49:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T22:49:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a2f2a62315529e148a10a33ccb241e9417fbde0fe42afb61cb7ee2438c74636f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T22:49:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xl62j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-10T22:49:51Z\\\"}}\" for pod \"openshift-multus\"/\"multus-4nwkq\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T22:50:03Z is after 2025-08-24T17:21:41Z" Dec 10 22:50:03 crc kubenswrapper[4791]: I1210 22:50:03.805693 4791 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-zhq64" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3cd47739-0fa9-4321-aff1-220f8721a0b3\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T22:49:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T22:49:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T22:49:52Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T22:49:52Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://68795a29944e56870ffe94fa732e9dcf0bfb344c0d85d1e590f58a876129c61f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T22:49:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wvpw8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://bc721b530c08b433756e978ec0423cc18aabf6771b61eb96f903b4d57ea335ff\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T22:49:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wvpw8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://152a7618be8f38fadf4e564b3948980503cedcdb466c37344ccc20f4a3173a5f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T22:49:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wvpw8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3d18f9cb7207efb1e32de75a30d33e024f656a08f0503db95e5488bc5280971f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T22:49:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wvpw8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://937137ae9d7480cdc7c59bc4a1df0a4c10e72b3b63693af285278789953ed218\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T22:49:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wvpw8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://1a52d19b7671e8b143017c8a981a6f8d3ee3b84650e4b223b814a531ac76f364\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T22:49:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wvpw8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://7a03d21e34f9d560953b7a809a4d5f374e39d3daeb8e698dbb5861b8e63c9de5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7a03d21e34f9d560953b7a809a4d5f374e39d3daeb8e698dbb5861b8e63c9de5\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-12-10T22:50:03Z\\\",\\\"message\\\":\\\"r removal\\\\nI1210 22:50:01.628167 6086 handler.go:190] Sending *v1.EgressIP event handler 8 for removal\\\\nI1210 22:50:01.628168 6086 handler.go:208] Removed *v1.NetworkPolicy event handler 4\\\\nI1210 22:50:01.628234 6086 handler.go:208] Removed *v1.Namespace event handler 1\\\\nI1210 22:50:01.628254 6086 handler.go:208] Removed *v1.Namespace event handler 5\\\\nI1210 22:50:01.628258 6086 handler.go:190] Sending *v1.Pod event handler 6 for removal\\\\nI1210 22:50:01.628268 6086 handler.go:190] Sending *v1.Pod event handler 3 for removal\\\\nI1210 22:50:01.628292 6086 handler.go:190] Sending *v1.Node event handler 2 for removal\\\\nI1210 22:50:01.628301 6086 handler.go:190] Sending *v1.Node event handler 7 for removal\\\\nI1210 22:50:01.628373 6086 handler.go:208] Removed *v1.Pod event handler 6\\\\nI1210 22:50:01.628392 6086 handler.go:208] Removed *v1.EgressIP event handler 8\\\\nI1210 22:50:01.628406 6086 handler.go:208] Removed *v1.Node event handler 2\\\\nI1210 22:50:01.628413 6086 handler.go:190] Sending *v1.EgressFirewall event handler 9 for removal\\\\nI1210 22:50:01.628412 6086 handler.go:208] Removed *v1.Pod event handler 3\\\\nI1210 22:50:01.628462 6086 factory.go:656] Stopping watch factory\\\\nI1210 22:50:01.628492 6086 ovnkube.go:599] Stopped ovnkube\\\\nI1210 22:50:01.628536 6086 handler.go:208] Removed *v1.EgressFirewall event handler 9\\\\nI12\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-10T22:49:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wvpw8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://dd3f80e5713221f96ed50da7f2e7fd5b9fabfcd5c1656da8d8a86d818a5fdc3f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T22:49:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wvpw8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3afbd1ffcd2c4b40324aca6e52d6fc822db131c3f35a9af3a6f005c2cffdb7bf\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3afbd1ffcd2c4b40324aca6e52d6fc822db131c3f35a9af3a6f005c2cffdb7bf\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-10T22:49:53Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-10T22:49:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wvpw8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-10T22:49:52Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-zhq64\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T22:50:03Z is after 2025-08-24T17:21:41Z" Dec 10 22:50:03 crc kubenswrapper[4791]: I1210 22:50:03.820272 4791 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"698bbe9a-fbe5-436d-a835-5e80df77ba24\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T22:49:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T22:49:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T22:49:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T22:49:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T22:49:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://224d73713312376a58ec6cc2bc8fdb6ffa44c4749ae0e7691d5fe27b9afca09f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T22:49:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://372eeaffd40992c6f06efc4df6dfcb44b31de64ee8256c21508cdd7c74ecb96e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T22:49:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://76b6013b062cac5b0b4d12a94f38ce7b43f9c7ea216063455d88a956ac04d26e\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T22:49:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://78e950e122e761d950da1a3c60c06476e53920e8795f6c61db22c01e22d5136e\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T22:49:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-10T22:49:25Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T22:50:03Z is after 2025-08-24T17:21:41Z" Dec 10 22:50:03 crc kubenswrapper[4791]: I1210 22:50:03.837995 4791 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-10T22:49:52Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-10T22:49:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-10T22:49:52Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3734b110802828632d44263ed0a43cabc4b710833f0be120d72db0940fe783e2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T22:49:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://794803ac9be3c52509e6dcb10854d8864e1a9e406762e57c7971b07c4da1220e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T22:49:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T22:50:03Z is after 2025-08-24T17:21:41Z" Dec 10 22:50:03 crc kubenswrapper[4791]: I1210 22:50:03.852425 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 22:50:03 crc kubenswrapper[4791]: I1210 22:50:03.852537 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 22:50:03 crc kubenswrapper[4791]: I1210 22:50:03.852550 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 22:50:03 crc kubenswrapper[4791]: I1210 22:50:03.852565 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 22:50:03 crc kubenswrapper[4791]: I1210 22:50:03.852621 4791 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T22:50:03Z","lastTransitionTime":"2025-12-10T22:50:03Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 22:50:03 crc kubenswrapper[4791]: I1210 22:50:03.955325 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 22:50:03 crc kubenswrapper[4791]: I1210 22:50:03.955412 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 22:50:03 crc kubenswrapper[4791]: I1210 22:50:03.955423 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 22:50:03 crc kubenswrapper[4791]: I1210 22:50:03.955460 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 22:50:03 crc kubenswrapper[4791]: I1210 22:50:03.955474 4791 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T22:50:03Z","lastTransitionTime":"2025-12-10T22:50:03Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 22:50:04 crc kubenswrapper[4791]: I1210 22:50:04.058749 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 22:50:04 crc kubenswrapper[4791]: I1210 22:50:04.058877 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 22:50:04 crc kubenswrapper[4791]: I1210 22:50:04.058917 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 22:50:04 crc kubenswrapper[4791]: I1210 22:50:04.058974 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 22:50:04 crc kubenswrapper[4791]: I1210 22:50:04.059004 4791 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T22:50:04Z","lastTransitionTime":"2025-12-10T22:50:04Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 22:50:04 crc kubenswrapper[4791]: I1210 22:50:04.161216 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 22:50:04 crc kubenswrapper[4791]: I1210 22:50:04.161279 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 22:50:04 crc kubenswrapper[4791]: I1210 22:50:04.161293 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 22:50:04 crc kubenswrapper[4791]: I1210 22:50:04.161311 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 22:50:04 crc kubenswrapper[4791]: I1210 22:50:04.161323 4791 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T22:50:04Z","lastTransitionTime":"2025-12-10T22:50:04Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 22:50:04 crc kubenswrapper[4791]: I1210 22:50:04.263642 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 22:50:04 crc kubenswrapper[4791]: I1210 22:50:04.263697 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 22:50:04 crc kubenswrapper[4791]: I1210 22:50:04.263713 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 22:50:04 crc kubenswrapper[4791]: I1210 22:50:04.263732 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 22:50:04 crc kubenswrapper[4791]: I1210 22:50:04.263747 4791 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T22:50:04Z","lastTransitionTime":"2025-12-10T22:50:04Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 22:50:04 crc kubenswrapper[4791]: I1210 22:50:04.348014 4791 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-s6rk8" event={"ID":"5c9d0978-a6f9-49c5-b377-6619b4e3ff52","Type":"ContainerStarted","Data":"11613be462f4f652fb71299fa5b3bba520c7d3dc7d688b749d3642352bf96177"} Dec 10 22:50:04 crc kubenswrapper[4791]: I1210 22:50:04.366207 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 22:50:04 crc kubenswrapper[4791]: I1210 22:50:04.366301 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 22:50:04 crc kubenswrapper[4791]: I1210 22:50:04.366327 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 22:50:04 crc kubenswrapper[4791]: I1210 22:50:04.366474 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 22:50:04 crc kubenswrapper[4791]: I1210 22:50:04.366509 4791 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T22:50:04Z","lastTransitionTime":"2025-12-10T22:50:04Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 22:50:04 crc kubenswrapper[4791]: I1210 22:50:04.469223 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 22:50:04 crc kubenswrapper[4791]: I1210 22:50:04.469270 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 22:50:04 crc kubenswrapper[4791]: I1210 22:50:04.469284 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 22:50:04 crc kubenswrapper[4791]: I1210 22:50:04.469303 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 22:50:04 crc kubenswrapper[4791]: I1210 22:50:04.469316 4791 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T22:50:04Z","lastTransitionTime":"2025-12-10T22:50:04Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 22:50:04 crc kubenswrapper[4791]: I1210 22:50:04.571959 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 22:50:04 crc kubenswrapper[4791]: I1210 22:50:04.572038 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 22:50:04 crc kubenswrapper[4791]: I1210 22:50:04.572057 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 22:50:04 crc kubenswrapper[4791]: I1210 22:50:04.572089 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 22:50:04 crc kubenswrapper[4791]: I1210 22:50:04.572108 4791 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T22:50:04Z","lastTransitionTime":"2025-12-10T22:50:04Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 22:50:04 crc kubenswrapper[4791]: I1210 22:50:04.676106 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 22:50:04 crc kubenswrapper[4791]: I1210 22:50:04.676210 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 22:50:04 crc kubenswrapper[4791]: I1210 22:50:04.676229 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 22:50:04 crc kubenswrapper[4791]: I1210 22:50:04.676260 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 22:50:04 crc kubenswrapper[4791]: I1210 22:50:04.676281 4791 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T22:50:04Z","lastTransitionTime":"2025-12-10T22:50:04Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 22:50:04 crc kubenswrapper[4791]: I1210 22:50:04.780073 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 22:50:04 crc kubenswrapper[4791]: I1210 22:50:04.780162 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 22:50:04 crc kubenswrapper[4791]: I1210 22:50:04.780174 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 22:50:04 crc kubenswrapper[4791]: I1210 22:50:04.780198 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 22:50:04 crc kubenswrapper[4791]: I1210 22:50:04.780211 4791 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T22:50:04Z","lastTransitionTime":"2025-12-10T22:50:04Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 22:50:04 crc kubenswrapper[4791]: I1210 22:50:04.854955 4791 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-multus/network-metrics-daemon-b6kf6"] Dec 10 22:50:04 crc kubenswrapper[4791]: I1210 22:50:04.855669 4791 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-b6kf6" Dec 10 22:50:04 crc kubenswrapper[4791]: E1210 22:50:04.855752 4791 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-b6kf6" podUID="b4527e71-bb58-4a39-a99c-fe614aa4c5f9" Dec 10 22:50:04 crc kubenswrapper[4791]: I1210 22:50:04.876547 4791 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-5rb5l" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ba35653c-6e06-4cee-a4d6-137764090d18\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T22:49:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T22:49:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T22:49:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T22:49:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://dbb5c875730a98ad2daf316cfafc08fcdce21ac16148154780ac9193ed405dab\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T22:49:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fvvrs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://185eb2ff44221a514878283ec873773a6c9a9073dc64e296f982e60dee1338e4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T22:49:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fvvrs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-10T22:49:51Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-5rb5l\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T22:50:04Z is after 2025-08-24T17:21:41Z" Dec 10 22:50:04 crc kubenswrapper[4791]: I1210 22:50:04.883009 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 22:50:04 crc kubenswrapper[4791]: I1210 22:50:04.883112 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 22:50:04 crc kubenswrapper[4791]: I1210 22:50:04.883132 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 22:50:04 crc kubenswrapper[4791]: I1210 22:50:04.883167 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 22:50:04 crc kubenswrapper[4791]: I1210 22:50:04.883184 4791 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T22:50:04Z","lastTransitionTime":"2025-12-10T22:50:04Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 22:50:04 crc kubenswrapper[4791]: I1210 22:50:04.883960 4791 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 10 22:50:04 crc kubenswrapper[4791]: I1210 22:50:04.883977 4791 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 10 22:50:04 crc kubenswrapper[4791]: I1210 22:50:04.884073 4791 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 10 22:50:04 crc kubenswrapper[4791]: E1210 22:50:04.884137 4791 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 10 22:50:04 crc kubenswrapper[4791]: E1210 22:50:04.884264 4791 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 10 22:50:04 crc kubenswrapper[4791]: E1210 22:50:04.884423 4791 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 10 22:50:04 crc kubenswrapper[4791]: I1210 22:50:04.895692 4791 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-s6rk8" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5c9d0978-a6f9-49c5-b377-6619b4e3ff52\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T22:50:03Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T22:50:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T22:50:03Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-rbac-proxy ovnkube-cluster-manager]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T22:50:03Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-rbac-proxy ovnkube-cluster-manager]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7vkgm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7vkgm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-10T22:50:03Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-s6rk8\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T22:50:04Z is after 2025-08-24T17:21:41Z" Dec 10 22:50:04 crc kubenswrapper[4791]: I1210 22:50:04.907914 4791 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-vczjq" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b9f00f25-3fa1-4b63-9710-fd352224b01b\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T22:49:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T22:49:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T22:49:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T22:49:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4698cf4c6b55b41e628448d37c2eb847c2777d0efbd93afcbf367d5e78623b7e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T22:49:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lrrpv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-10T22:49:50Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-vczjq\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T22:50:04Z is after 2025-08-24T17:21:41Z" Dec 10 22:50:04 crc kubenswrapper[4791]: I1210 22:50:04.924115 4791 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-4nwkq" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"672aa28c-8169-49ed-87b8-21187d13a80c\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T22:49:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T22:49:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T22:49:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T22:49:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a2f2a62315529e148a10a33ccb241e9417fbde0fe42afb61cb7ee2438c74636f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T22:49:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xl62j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-10T22:49:51Z\\\"}}\" for pod \"openshift-multus\"/\"multus-4nwkq\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T22:50:04Z is after 2025-08-24T17:21:41Z" Dec 10 22:50:04 crc kubenswrapper[4791]: I1210 22:50:04.949897 4791 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/b4527e71-bb58-4a39-a99c-fe614aa4c5f9-metrics-certs\") pod \"network-metrics-daemon-b6kf6\" (UID: \"b4527e71-bb58-4a39-a99c-fe614aa4c5f9\") " pod="openshift-multus/network-metrics-daemon-b6kf6" Dec 10 22:50:04 crc kubenswrapper[4791]: I1210 22:50:04.950019 4791 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-ngpr4\" (UniqueName: \"kubernetes.io/projected/b4527e71-bb58-4a39-a99c-fe614aa4c5f9-kube-api-access-ngpr4\") pod \"network-metrics-daemon-b6kf6\" (UID: \"b4527e71-bb58-4a39-a99c-fe614aa4c5f9\") " pod="openshift-multus/network-metrics-daemon-b6kf6" Dec 10 22:50:04 crc kubenswrapper[4791]: I1210 22:50:04.954882 4791 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-zhq64" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3cd47739-0fa9-4321-aff1-220f8721a0b3\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T22:49:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T22:49:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T22:49:52Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T22:49:52Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://68795a29944e56870ffe94fa732e9dcf0bfb344c0d85d1e590f58a876129c61f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T22:49:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wvpw8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://bc721b530c08b433756e978ec0423cc18aabf6771b61eb96f903b4d57ea335ff\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T22:49:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wvpw8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://152a7618be8f38fadf4e564b3948980503cedcdb466c37344ccc20f4a3173a5f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T22:49:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wvpw8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3d18f9cb7207efb1e32de75a30d33e024f656a08f0503db95e5488bc5280971f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T22:49:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wvpw8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://937137ae9d7480cdc7c59bc4a1df0a4c10e72b3b63693af285278789953ed218\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T22:49:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wvpw8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://1a52d19b7671e8b143017c8a981a6f8d3ee3b84650e4b223b814a531ac76f364\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T22:49:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wvpw8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://7a03d21e34f9d560953b7a809a4d5f374e39d3daeb8e698dbb5861b8e63c9de5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7a03d21e34f9d560953b7a809a4d5f374e39d3daeb8e698dbb5861b8e63c9de5\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-12-10T22:50:03Z\\\",\\\"message\\\":\\\"r removal\\\\nI1210 22:50:01.628167 6086 handler.go:190] Sending *v1.EgressIP event handler 8 for removal\\\\nI1210 22:50:01.628168 6086 handler.go:208] Removed *v1.NetworkPolicy event handler 4\\\\nI1210 22:50:01.628234 6086 handler.go:208] Removed *v1.Namespace event handler 1\\\\nI1210 22:50:01.628254 6086 handler.go:208] Removed *v1.Namespace event handler 5\\\\nI1210 22:50:01.628258 6086 handler.go:190] Sending *v1.Pod event handler 6 for removal\\\\nI1210 22:50:01.628268 6086 handler.go:190] Sending *v1.Pod event handler 3 for removal\\\\nI1210 22:50:01.628292 6086 handler.go:190] Sending *v1.Node event handler 2 for removal\\\\nI1210 22:50:01.628301 6086 handler.go:190] Sending *v1.Node event handler 7 for removal\\\\nI1210 22:50:01.628373 6086 handler.go:208] Removed *v1.Pod event handler 6\\\\nI1210 22:50:01.628392 6086 handler.go:208] Removed *v1.EgressIP event handler 8\\\\nI1210 22:50:01.628406 6086 handler.go:208] Removed *v1.Node event handler 2\\\\nI1210 22:50:01.628413 6086 handler.go:190] Sending *v1.EgressFirewall event handler 9 for removal\\\\nI1210 22:50:01.628412 6086 handler.go:208] Removed *v1.Pod event handler 3\\\\nI1210 22:50:01.628462 6086 factory.go:656] Stopping watch factory\\\\nI1210 22:50:01.628492 6086 ovnkube.go:599] Stopped ovnkube\\\\nI1210 22:50:01.628536 6086 handler.go:208] Removed *v1.EgressFirewall event handler 9\\\\nI12\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-10T22:49:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wvpw8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://dd3f80e5713221f96ed50da7f2e7fd5b9fabfcd5c1656da8d8a86d818a5fdc3f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T22:49:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wvpw8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3afbd1ffcd2c4b40324aca6e52d6fc822db131c3f35a9af3a6f005c2cffdb7bf\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3afbd1ffcd2c4b40324aca6e52d6fc822db131c3f35a9af3a6f005c2cffdb7bf\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-10T22:49:53Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-10T22:49:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wvpw8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-10T22:49:52Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-zhq64\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T22:50:04Z is after 2025-08-24T17:21:41Z" Dec 10 22:50:04 crc kubenswrapper[4791]: I1210 22:50:04.972199 4791 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"698bbe9a-fbe5-436d-a835-5e80df77ba24\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T22:49:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T22:49:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T22:49:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T22:49:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T22:49:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://224d73713312376a58ec6cc2bc8fdb6ffa44c4749ae0e7691d5fe27b9afca09f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T22:49:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://372eeaffd40992c6f06efc4df6dfcb44b31de64ee8256c21508cdd7c74ecb96e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T22:49:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://76b6013b062cac5b0b4d12a94f38ce7b43f9c7ea216063455d88a956ac04d26e\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T22:49:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://78e950e122e761d950da1a3c60c06476e53920e8795f6c61db22c01e22d5136e\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T22:49:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-10T22:49:25Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T22:50:04Z is after 2025-08-24T17:21:41Z" Dec 10 22:50:04 crc kubenswrapper[4791]: I1210 22:50:04.985956 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 22:50:04 crc kubenswrapper[4791]: I1210 22:50:04.986025 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 22:50:04 crc kubenswrapper[4791]: I1210 22:50:04.986077 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 22:50:04 crc kubenswrapper[4791]: I1210 22:50:04.986122 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 22:50:04 crc kubenswrapper[4791]: I1210 22:50:04.986146 4791 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T22:50:04Z","lastTransitionTime":"2025-12-10T22:50:04Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 22:50:04 crc kubenswrapper[4791]: I1210 22:50:04.991727 4791 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-10T22:49:52Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-10T22:49:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-10T22:49:52Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3734b110802828632d44263ed0a43cabc4b710833f0be120d72db0940fe783e2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T22:49:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://794803ac9be3c52509e6dcb10854d8864e1a9e406762e57c7971b07c4da1220e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T22:49:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T22:50:04Z is after 2025-08-24T17:21:41Z" Dec 10 22:50:05 crc kubenswrapper[4791]: I1210 22:50:05.005895 4791 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-tchxk" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"c0ab74e0-42a9-4f42-ac80-16030cf4ffff\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T22:49:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T22:49:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T22:49:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T22:49:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://04e0fd1b53256b7a938e8ccab93e2a4eb73dd1a59e0b9799e09c55e511baf7fe\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T22:49:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zn2qt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-10T22:49:50Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-tchxk\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T22:50:05Z is after 2025-08-24T17:21:41Z" Dec 10 22:50:05 crc kubenswrapper[4791]: I1210 22:50:05.024271 4791 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-rw584" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ff9beb90-69b0-4732-bf37-0b81f58ecc98\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T22:49:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T22:49:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T22:49:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T22:49:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ddb9df501dc6e8e5561fbc2aefd431037a3f28f73bf8e82a6598d261a4224807\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T22:49:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fksqn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ce67e54b696bb53ffbbe48c6ce71476b635e0868993e9c427b22f904891433ee\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ce67e54b696bb53ffbbe48c6ce71476b635e0868993e9c427b22f904891433ee\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-10T22:49:52Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-10T22:49:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fksqn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://818d89bfffb8a110670152c8480684665ea8db578e9768f524fbbe6a92f27c22\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://818d89bfffb8a110670152c8480684665ea8db578e9768f524fbbe6a92f27c22\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-10T22:49:53Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-10T22:49:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fksqn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://9aa9c06e894bb2a2ed3738defaeff68a0a0458ead48177f36b90770944c211a0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://9aa9c06e894bb2a2ed3738defaeff68a0a0458ead48177f36b90770944c211a0\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-10T22:49:54Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-10T22:49:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fksqn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://dfd6d03336dcea7529419a30e8d70d84e2290a0fa5b0e46d09ad341c190638d6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://dfd6d03336dcea7529419a30e8d70d84e2290a0fa5b0e46d09ad341c190638d6\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-10T22:49:55Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-10T22:49:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fksqn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b86738401ece6fbc3083782720c66a515850ed78d4054f34fcabab7ff525b85f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b86738401ece6fbc3083782720c66a515850ed78d4054f34fcabab7ff525b85f\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-10T22:49:56Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-10T22:49:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fksqn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://18b326f572a899ddd10197a745bf9264d4ddd975690ad1cfb0ec576ccd5cf102\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://18b326f572a899ddd10197a745bf9264d4ddd975690ad1cfb0ec576ccd5cf102\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-10T22:49:57Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-10T22:49:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fksqn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-10T22:49:51Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-rw584\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T22:50:05Z is after 2025-08-24T17:21:41Z" Dec 10 22:50:05 crc kubenswrapper[4791]: I1210 22:50:05.037133 4791 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-10T22:49:52Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-10T22:49:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-10T22:49:52Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://20651eb0c16e2c71222f56da1ff755cb0e7b084886acbbb24f0ccaf57041ed85\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T22:49:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T22:50:05Z is after 2025-08-24T17:21:41Z" Dec 10 22:50:05 crc kubenswrapper[4791]: I1210 22:50:05.050918 4791 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/b4527e71-bb58-4a39-a99c-fe614aa4c5f9-metrics-certs\") pod \"network-metrics-daemon-b6kf6\" (UID: \"b4527e71-bb58-4a39-a99c-fe614aa4c5f9\") " pod="openshift-multus/network-metrics-daemon-b6kf6" Dec 10 22:50:05 crc kubenswrapper[4791]: I1210 22:50:05.051001 4791 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-ngpr4\" (UniqueName: \"kubernetes.io/projected/b4527e71-bb58-4a39-a99c-fe614aa4c5f9-kube-api-access-ngpr4\") pod \"network-metrics-daemon-b6kf6\" (UID: \"b4527e71-bb58-4a39-a99c-fe614aa4c5f9\") " pod="openshift-multus/network-metrics-daemon-b6kf6" Dec 10 22:50:05 crc kubenswrapper[4791]: E1210 22:50:05.051115 4791 secret.go:188] Couldn't get secret openshift-multus/metrics-daemon-secret: object "openshift-multus"/"metrics-daemon-secret" not registered Dec 10 22:50:05 crc kubenswrapper[4791]: E1210 22:50:05.051244 4791 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/b4527e71-bb58-4a39-a99c-fe614aa4c5f9-metrics-certs podName:b4527e71-bb58-4a39-a99c-fe614aa4c5f9 nodeName:}" failed. No retries permitted until 2025-12-10 22:50:05.551215944 +0000 UTC m=+39.980833667 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "metrics-certs" (UniqueName: "kubernetes.io/secret/b4527e71-bb58-4a39-a99c-fe614aa4c5f9-metrics-certs") pod "network-metrics-daemon-b6kf6" (UID: "b4527e71-bb58-4a39-a99c-fe614aa4c5f9") : object "openshift-multus"/"metrics-daemon-secret" not registered Dec 10 22:50:05 crc kubenswrapper[4791]: I1210 22:50:05.054039 4791 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-10T22:49:50Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-10T22:49:50Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T22:50:05Z is after 2025-08-24T17:21:41Z" Dec 10 22:50:05 crc kubenswrapper[4791]: I1210 22:50:05.066700 4791 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-10T22:49:50Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-10T22:49:50Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T22:50:05Z is after 2025-08-24T17:21:41Z" Dec 10 22:50:05 crc kubenswrapper[4791]: I1210 22:50:05.073447 4791 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-ngpr4\" (UniqueName: \"kubernetes.io/projected/b4527e71-bb58-4a39-a99c-fe614aa4c5f9-kube-api-access-ngpr4\") pod \"network-metrics-daemon-b6kf6\" (UID: \"b4527e71-bb58-4a39-a99c-fe614aa4c5f9\") " pod="openshift-multus/network-metrics-daemon-b6kf6" Dec 10 22:50:05 crc kubenswrapper[4791]: I1210 22:50:05.079742 4791 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-10T22:49:50Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-10T22:49:50Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T22:50:05Z is after 2025-08-24T17:21:41Z" Dec 10 22:50:05 crc kubenswrapper[4791]: I1210 22:50:05.089072 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 22:50:05 crc kubenswrapper[4791]: I1210 22:50:05.089102 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 22:50:05 crc kubenswrapper[4791]: I1210 22:50:05.089115 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 22:50:05 crc kubenswrapper[4791]: I1210 22:50:05.089136 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 22:50:05 crc kubenswrapper[4791]: I1210 22:50:05.089149 4791 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T22:50:05Z","lastTransitionTime":"2025-12-10T22:50:05Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 22:50:05 crc kubenswrapper[4791]: I1210 22:50:05.091553 4791 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-10T22:49:54Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-10T22:49:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-10T22:49:54Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://62bfaf5d7a6d1b452cdaf88086447d0c8cb7306e94434e46752be5475124d56a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T22:49:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T22:50:05Z is after 2025-08-24T17:21:41Z" Dec 10 22:50:05 crc kubenswrapper[4791]: I1210 22:50:05.103946 4791 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-b6kf6" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b4527e71-bb58-4a39-a99c-fe614aa4c5f9\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T22:50:04Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T22:50:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T22:50:04Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T22:50:04Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ngpr4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ngpr4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-10T22:50:04Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-b6kf6\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T22:50:05Z is after 2025-08-24T17:21:41Z" Dec 10 22:50:05 crc kubenswrapper[4791]: I1210 22:50:05.117057 4791 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f8349fea-c49b-49a1-b5ee-032c5dffe021\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T22:49:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T22:49:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T22:49:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T22:49:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T22:49:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://acafe12186df99d1059a31260de0eafce35ff7771bcb8f5c6753b72665d6f60b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T22:49:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ea4c346507862a8c12756a3061d6e25e788a8d79f4621b62f81f4322185c82a8\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T22:49:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://2e6e324ec6f4f7bb69fdfab4ea3ac1aa55e480dd8e2a8984b0fe5b4c9f66636e\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T22:49:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://1b2e11ff2bf8789fa1d0d23ae648b3e6ef653106e5bc4bbb5dd6447f7cd6f05d\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d6aa7f8c5d908d90008f2df49860ac8a3480136da567e3d85f86135d779204f9\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-12-10T22:49:44Z\\\",\\\"message\\\":\\\"W1210 22:49:33.845733 1 cmd.go:257] Using insecure, self-signed certificates\\\\nI1210 22:49:33.846108 1 crypto.go:601] Generating new CA for check-endpoints-signer@1765406973 cert, and key in /tmp/serving-cert-2585701456/serving-signer.crt, /tmp/serving-cert-2585701456/serving-signer.key\\\\nI1210 22:49:34.529748 1 observer_polling.go:159] Starting file observer\\\\nW1210 22:49:34.532225 1 builder.go:272] unable to get owner reference (falling back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": dial tcp [::1]:6443: connect: connection refused\\\\nI1210 22:49:34.532599 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1210 22:49:34.533977 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-2585701456/tls.crt::/tmp/serving-cert-2585701456/tls.key\\\\\\\"\\\\nF1210 22:49:44.774710 1 cmd.go:182] error initializing delegating authentication: unable to load configmap based request-header-client-ca-file: Get \\\\\\\"https://localhost:6443/api/v1/namespaces/kube-system/configmaps/extension-apiserver-authentication\\\\\\\": net/http: TLS handshake timeout\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-10T22:49:33Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T22:49:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://0c4d036ab269f62561fe7ef1586012cb3a16e3583a08b9488a0616ec00c8e0ae\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T22:49:32Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c5856d6b26c19fda13b5c9ac2892a7e2f9c6f79feb8173ecfd9b33b673327d67\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c5856d6b26c19fda13b5c9ac2892a7e2f9c6f79feb8173ecfd9b33b673327d67\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-10T22:49:27Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-10T22:49:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-10T22:49:25Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T22:50:05Z is after 2025-08-24T17:21:41Z" Dec 10 22:50:05 crc kubenswrapper[4791]: I1210 22:50:05.137039 4791 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"90250e13-6c63-443a-a5e5-6e4e341a2289\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T22:49:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T22:49:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T22:49:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T22:49:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T22:49:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ba54f95b1e00843b854db9a10fe6897592159ccacf30ed0f76002093a13eedd3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T22:49:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5aa8ffa518f3a68b60b275f8df4851b2564aa4984b61efdc9470b73072133e43\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T22:49:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://16f989503edc78d9451638e2cf6cf4b19a2cf2b8809b10b237541cb829303091\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T22:49:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://82556824d63ce6e46df14c45689ff1f81ebf729abe902061815c040b323898e9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T22:49:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://61db531940fa6f2effdae5a005b4c91bd2698bd5d1bff9588c171a65ad550eca\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T22:49:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0f65c864d885319ce2a08d61c30be24b3ffb7d23ea30c82ed7faf039cb2c006d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://0f65c864d885319ce2a08d61c30be24b3ffb7d23ea30c82ed7faf039cb2c006d\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-10T22:49:27Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-10T22:49:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e31a76d1d5b2812ba83e24e38268e25ec8bed98fc3da9adcd06bda5822b6ace1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e31a76d1d5b2812ba83e24e38268e25ec8bed98fc3da9adcd06bda5822b6ace1\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-10T22:49:30Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-10T22:49:29Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://b5b5eceae562f3986b543d41de1e977efb85a91d84c98e09159ca6d092eda691\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b5b5eceae562f3986b543d41de1e977efb85a91d84c98e09159ca6d092eda691\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-10T22:49:31Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-10T22:49:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-10T22:49:25Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T22:50:05Z is after 2025-08-24T17:21:41Z" Dec 10 22:50:05 crc kubenswrapper[4791]: I1210 22:50:05.191314 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 22:50:05 crc kubenswrapper[4791]: I1210 22:50:05.191368 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 22:50:05 crc kubenswrapper[4791]: I1210 22:50:05.191379 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 22:50:05 crc kubenswrapper[4791]: I1210 22:50:05.191395 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 22:50:05 crc kubenswrapper[4791]: I1210 22:50:05.191405 4791 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T22:50:05Z","lastTransitionTime":"2025-12-10T22:50:05Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 22:50:05 crc kubenswrapper[4791]: I1210 22:50:05.294752 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 22:50:05 crc kubenswrapper[4791]: I1210 22:50:05.294785 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 22:50:05 crc kubenswrapper[4791]: I1210 22:50:05.294797 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 22:50:05 crc kubenswrapper[4791]: I1210 22:50:05.294812 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 22:50:05 crc kubenswrapper[4791]: I1210 22:50:05.294826 4791 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T22:50:05Z","lastTransitionTime":"2025-12-10T22:50:05Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 22:50:05 crc kubenswrapper[4791]: I1210 22:50:05.353870 4791 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-zhq64_3cd47739-0fa9-4321-aff1-220f8721a0b3/ovnkube-controller/0.log" Dec 10 22:50:05 crc kubenswrapper[4791]: I1210 22:50:05.357403 4791 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-zhq64" event={"ID":"3cd47739-0fa9-4321-aff1-220f8721a0b3","Type":"ContainerStarted","Data":"f78a47c773006a18dbc4b706edfc9d1d5d0b6ad11a023a49ea4e622eae77bba3"} Dec 10 22:50:05 crc kubenswrapper[4791]: I1210 22:50:05.357641 4791 prober_manager.go:312] "Failed to trigger a manual run" probe="Readiness" Dec 10 22:50:05 crc kubenswrapper[4791]: I1210 22:50:05.358772 4791 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-s6rk8" event={"ID":"5c9d0978-a6f9-49c5-b377-6619b4e3ff52","Type":"ContainerStarted","Data":"8e55baaf460c794ab196e237d51d12de059aa038617dc2d5366d49c529f5d5c7"} Dec 10 22:50:05 crc kubenswrapper[4791]: I1210 22:50:05.397700 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 22:50:05 crc kubenswrapper[4791]: I1210 22:50:05.397752 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 22:50:05 crc kubenswrapper[4791]: I1210 22:50:05.397767 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 22:50:05 crc kubenswrapper[4791]: I1210 22:50:05.397791 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 22:50:05 crc kubenswrapper[4791]: I1210 22:50:05.397808 4791 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T22:50:05Z","lastTransitionTime":"2025-12-10T22:50:05Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 22:50:05 crc kubenswrapper[4791]: I1210 22:50:05.399922 4791 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"90250e13-6c63-443a-a5e5-6e4e341a2289\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T22:49:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T22:49:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T22:49:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T22:49:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T22:49:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ba54f95b1e00843b854db9a10fe6897592159ccacf30ed0f76002093a13eedd3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T22:49:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5aa8ffa518f3a68b60b275f8df4851b2564aa4984b61efdc9470b73072133e43\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T22:49:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://16f989503edc78d9451638e2cf6cf4b19a2cf2b8809b10b237541cb829303091\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T22:49:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://82556824d63ce6e46df14c45689ff1f81ebf729abe902061815c040b323898e9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T22:49:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://61db531940fa6f2effdae5a005b4c91bd2698bd5d1bff9588c171a65ad550eca\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T22:49:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0f65c864d885319ce2a08d61c30be24b3ffb7d23ea30c82ed7faf039cb2c006d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://0f65c864d885319ce2a08d61c30be24b3ffb7d23ea30c82ed7faf039cb2c006d\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-10T22:49:27Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-10T22:49:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e31a76d1d5b2812ba83e24e38268e25ec8bed98fc3da9adcd06bda5822b6ace1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e31a76d1d5b2812ba83e24e38268e25ec8bed98fc3da9adcd06bda5822b6ace1\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-10T22:49:30Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-10T22:49:29Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://b5b5eceae562f3986b543d41de1e977efb85a91d84c98e09159ca6d092eda691\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b5b5eceae562f3986b543d41de1e977efb85a91d84c98e09159ca6d092eda691\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-10T22:49:31Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-10T22:49:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-10T22:49:25Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T22:50:05Z is after 2025-08-24T17:21:41Z" Dec 10 22:50:05 crc kubenswrapper[4791]: I1210 22:50:05.416904 4791 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-10T22:49:50Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-10T22:49:50Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T22:50:05Z is after 2025-08-24T17:21:41Z" Dec 10 22:50:05 crc kubenswrapper[4791]: I1210 22:50:05.431250 4791 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-10T22:49:54Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-10T22:49:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-10T22:49:54Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://62bfaf5d7a6d1b452cdaf88086447d0c8cb7306e94434e46752be5475124d56a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T22:49:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T22:50:05Z is after 2025-08-24T17:21:41Z" Dec 10 22:50:05 crc kubenswrapper[4791]: I1210 22:50:05.441615 4791 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-b6kf6" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b4527e71-bb58-4a39-a99c-fe614aa4c5f9\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T22:50:04Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T22:50:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T22:50:04Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T22:50:04Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ngpr4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ngpr4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-10T22:50:04Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-b6kf6\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T22:50:05Z is after 2025-08-24T17:21:41Z" Dec 10 22:50:05 crc kubenswrapper[4791]: I1210 22:50:05.455433 4791 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f8349fea-c49b-49a1-b5ee-032c5dffe021\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T22:49:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T22:49:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T22:49:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T22:49:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T22:49:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://acafe12186df99d1059a31260de0eafce35ff7771bcb8f5c6753b72665d6f60b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T22:49:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ea4c346507862a8c12756a3061d6e25e788a8d79f4621b62f81f4322185c82a8\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T22:49:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://2e6e324ec6f4f7bb69fdfab4ea3ac1aa55e480dd8e2a8984b0fe5b4c9f66636e\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T22:49:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://1b2e11ff2bf8789fa1d0d23ae648b3e6ef653106e5bc4bbb5dd6447f7cd6f05d\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d6aa7f8c5d908d90008f2df49860ac8a3480136da567e3d85f86135d779204f9\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-12-10T22:49:44Z\\\",\\\"message\\\":\\\"W1210 22:49:33.845733 1 cmd.go:257] Using insecure, self-signed certificates\\\\nI1210 22:49:33.846108 1 crypto.go:601] Generating new CA for check-endpoints-signer@1765406973 cert, and key in /tmp/serving-cert-2585701456/serving-signer.crt, /tmp/serving-cert-2585701456/serving-signer.key\\\\nI1210 22:49:34.529748 1 observer_polling.go:159] Starting file observer\\\\nW1210 22:49:34.532225 1 builder.go:272] unable to get owner reference (falling back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": dial tcp [::1]:6443: connect: connection refused\\\\nI1210 22:49:34.532599 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1210 22:49:34.533977 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-2585701456/tls.crt::/tmp/serving-cert-2585701456/tls.key\\\\\\\"\\\\nF1210 22:49:44.774710 1 cmd.go:182] error initializing delegating authentication: unable to load configmap based request-header-client-ca-file: Get \\\\\\\"https://localhost:6443/api/v1/namespaces/kube-system/configmaps/extension-apiserver-authentication\\\\\\\": net/http: TLS handshake timeout\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-10T22:49:33Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T22:49:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://0c4d036ab269f62561fe7ef1586012cb3a16e3583a08b9488a0616ec00c8e0ae\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T22:49:32Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c5856d6b26c19fda13b5c9ac2892a7e2f9c6f79feb8173ecfd9b33b673327d67\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c5856d6b26c19fda13b5c9ac2892a7e2f9c6f79feb8173ecfd9b33b673327d67\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-10T22:49:27Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-10T22:49:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-10T22:49:25Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T22:50:05Z is after 2025-08-24T17:21:41Z" Dec 10 22:50:05 crc kubenswrapper[4791]: I1210 22:50:05.466049 4791 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-s6rk8" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5c9d0978-a6f9-49c5-b377-6619b4e3ff52\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T22:50:03Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T22:50:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T22:50:03Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-rbac-proxy ovnkube-cluster-manager]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T22:50:03Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-rbac-proxy ovnkube-cluster-manager]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7vkgm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7vkgm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-10T22:50:03Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-s6rk8\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T22:50:05Z is after 2025-08-24T17:21:41Z" Dec 10 22:50:05 crc kubenswrapper[4791]: I1210 22:50:05.476951 4791 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-5rb5l" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ba35653c-6e06-4cee-a4d6-137764090d18\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T22:49:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T22:49:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T22:49:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T22:49:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://dbb5c875730a98ad2daf316cfafc08fcdce21ac16148154780ac9193ed405dab\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T22:49:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fvvrs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://185eb2ff44221a514878283ec873773a6c9a9073dc64e296f982e60dee1338e4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T22:49:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fvvrs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-10T22:49:51Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-5rb5l\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T22:50:05Z is after 2025-08-24T17:21:41Z" Dec 10 22:50:05 crc kubenswrapper[4791]: I1210 22:50:05.489769 4791 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-10T22:49:52Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-10T22:49:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-10T22:49:52Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3734b110802828632d44263ed0a43cabc4b710833f0be120d72db0940fe783e2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T22:49:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://794803ac9be3c52509e6dcb10854d8864e1a9e406762e57c7971b07c4da1220e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T22:49:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T22:50:05Z is after 2025-08-24T17:21:41Z" Dec 10 22:50:05 crc kubenswrapper[4791]: I1210 22:50:05.499874 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 22:50:05 crc kubenswrapper[4791]: I1210 22:50:05.499908 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 22:50:05 crc kubenswrapper[4791]: I1210 22:50:05.499919 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 22:50:05 crc kubenswrapper[4791]: I1210 22:50:05.499933 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 22:50:05 crc kubenswrapper[4791]: I1210 22:50:05.499943 4791 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T22:50:05Z","lastTransitionTime":"2025-12-10T22:50:05Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 22:50:05 crc kubenswrapper[4791]: I1210 22:50:05.505000 4791 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-vczjq" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b9f00f25-3fa1-4b63-9710-fd352224b01b\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T22:49:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T22:49:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T22:49:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T22:49:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4698cf4c6b55b41e628448d37c2eb847c2777d0efbd93afcbf367d5e78623b7e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T22:49:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lrrpv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-10T22:49:50Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-vczjq\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T22:50:05Z is after 2025-08-24T17:21:41Z" Dec 10 22:50:05 crc kubenswrapper[4791]: I1210 22:50:05.522450 4791 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-4nwkq" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"672aa28c-8169-49ed-87b8-21187d13a80c\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T22:49:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T22:49:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T22:49:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T22:49:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a2f2a62315529e148a10a33ccb241e9417fbde0fe42afb61cb7ee2438c74636f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T22:49:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xl62j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-10T22:49:51Z\\\"}}\" for pod \"openshift-multus\"/\"multus-4nwkq\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T22:50:05Z is after 2025-08-24T17:21:41Z" Dec 10 22:50:05 crc kubenswrapper[4791]: I1210 22:50:05.543329 4791 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-zhq64" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3cd47739-0fa9-4321-aff1-220f8721a0b3\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T22:49:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T22:49:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T22:49:52Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T22:49:52Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://68795a29944e56870ffe94fa732e9dcf0bfb344c0d85d1e590f58a876129c61f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T22:49:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wvpw8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://bc721b530c08b433756e978ec0423cc18aabf6771b61eb96f903b4d57ea335ff\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T22:49:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wvpw8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://152a7618be8f38fadf4e564b3948980503cedcdb466c37344ccc20f4a3173a5f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T22:49:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wvpw8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3d18f9cb7207efb1e32de75a30d33e024f656a08f0503db95e5488bc5280971f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T22:49:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wvpw8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://937137ae9d7480cdc7c59bc4a1df0a4c10e72b3b63693af285278789953ed218\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T22:49:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wvpw8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://1a52d19b7671e8b143017c8a981a6f8d3ee3b84650e4b223b814a531ac76f364\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T22:49:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wvpw8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f78a47c773006a18dbc4b706edfc9d1d5d0b6ad11a023a49ea4e622eae77bba3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7a03d21e34f9d560953b7a809a4d5f374e39d3daeb8e698dbb5861b8e63c9de5\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-12-10T22:50:03Z\\\",\\\"message\\\":\\\"r removal\\\\nI1210 22:50:01.628167 6086 handler.go:190] Sending *v1.EgressIP event handler 8 for removal\\\\nI1210 22:50:01.628168 6086 handler.go:208] Removed *v1.NetworkPolicy event handler 4\\\\nI1210 22:50:01.628234 6086 handler.go:208] Removed *v1.Namespace event handler 1\\\\nI1210 22:50:01.628254 6086 handler.go:208] Removed *v1.Namespace event handler 5\\\\nI1210 22:50:01.628258 6086 handler.go:190] Sending *v1.Pod event handler 6 for removal\\\\nI1210 22:50:01.628268 6086 handler.go:190] Sending *v1.Pod event handler 3 for removal\\\\nI1210 22:50:01.628292 6086 handler.go:190] Sending *v1.Node event handler 2 for removal\\\\nI1210 22:50:01.628301 6086 handler.go:190] Sending *v1.Node event handler 7 for removal\\\\nI1210 22:50:01.628373 6086 handler.go:208] Removed *v1.Pod event handler 6\\\\nI1210 22:50:01.628392 6086 handler.go:208] Removed *v1.EgressIP event handler 8\\\\nI1210 22:50:01.628406 6086 handler.go:208] Removed *v1.Node event handler 2\\\\nI1210 22:50:01.628413 6086 handler.go:190] Sending *v1.EgressFirewall event handler 9 for removal\\\\nI1210 22:50:01.628412 6086 handler.go:208] Removed *v1.Pod event handler 3\\\\nI1210 22:50:01.628462 6086 factory.go:656] Stopping watch factory\\\\nI1210 22:50:01.628492 6086 ovnkube.go:599] Stopped ovnkube\\\\nI1210 22:50:01.628536 6086 handler.go:208] Removed *v1.EgressFirewall event handler 9\\\\nI12\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-10T22:49:58Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T22:50:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wvpw8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://dd3f80e5713221f96ed50da7f2e7fd5b9fabfcd5c1656da8d8a86d818a5fdc3f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T22:49:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wvpw8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3afbd1ffcd2c4b40324aca6e52d6fc822db131c3f35a9af3a6f005c2cffdb7bf\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3afbd1ffcd2c4b40324aca6e52d6fc822db131c3f35a9af3a6f005c2cffdb7bf\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-10T22:49:53Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-10T22:49:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wvpw8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-10T22:49:52Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-zhq64\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T22:50:05Z is after 2025-08-24T17:21:41Z" Dec 10 22:50:05 crc kubenswrapper[4791]: I1210 22:50:05.554924 4791 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/b4527e71-bb58-4a39-a99c-fe614aa4c5f9-metrics-certs\") pod \"network-metrics-daemon-b6kf6\" (UID: \"b4527e71-bb58-4a39-a99c-fe614aa4c5f9\") " pod="openshift-multus/network-metrics-daemon-b6kf6" Dec 10 22:50:05 crc kubenswrapper[4791]: E1210 22:50:05.555081 4791 secret.go:188] Couldn't get secret openshift-multus/metrics-daemon-secret: object "openshift-multus"/"metrics-daemon-secret" not registered Dec 10 22:50:05 crc kubenswrapper[4791]: E1210 22:50:05.555359 4791 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/b4527e71-bb58-4a39-a99c-fe614aa4c5f9-metrics-certs podName:b4527e71-bb58-4a39-a99c-fe614aa4c5f9 nodeName:}" failed. No retries permitted until 2025-12-10 22:50:06.555329027 +0000 UTC m=+40.984946640 (durationBeforeRetry 1s). Error: MountVolume.SetUp failed for volume "metrics-certs" (UniqueName: "kubernetes.io/secret/b4527e71-bb58-4a39-a99c-fe614aa4c5f9-metrics-certs") pod "network-metrics-daemon-b6kf6" (UID: "b4527e71-bb58-4a39-a99c-fe614aa4c5f9") : object "openshift-multus"/"metrics-daemon-secret" not registered Dec 10 22:50:05 crc kubenswrapper[4791]: I1210 22:50:05.557404 4791 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"698bbe9a-fbe5-436d-a835-5e80df77ba24\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T22:49:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T22:49:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T22:49:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T22:49:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T22:49:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://224d73713312376a58ec6cc2bc8fdb6ffa44c4749ae0e7691d5fe27b9afca09f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T22:49:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://372eeaffd40992c6f06efc4df6dfcb44b31de64ee8256c21508cdd7c74ecb96e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T22:49:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://76b6013b062cac5b0b4d12a94f38ce7b43f9c7ea216063455d88a956ac04d26e\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T22:49:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://78e950e122e761d950da1a3c60c06476e53920e8795f6c61db22c01e22d5136e\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T22:49:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-10T22:49:25Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T22:50:05Z is after 2025-08-24T17:21:41Z" Dec 10 22:50:05 crc kubenswrapper[4791]: I1210 22:50:05.579507 4791 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-10T22:49:50Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-10T22:49:50Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T22:50:05Z is after 2025-08-24T17:21:41Z" Dec 10 22:50:05 crc kubenswrapper[4791]: I1210 22:50:05.596008 4791 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-10T22:49:50Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-10T22:49:50Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T22:50:05Z is after 2025-08-24T17:21:41Z" Dec 10 22:50:05 crc kubenswrapper[4791]: I1210 22:50:05.603046 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 22:50:05 crc kubenswrapper[4791]: I1210 22:50:05.603081 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 22:50:05 crc kubenswrapper[4791]: I1210 22:50:05.603092 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 22:50:05 crc kubenswrapper[4791]: I1210 22:50:05.603106 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 22:50:05 crc kubenswrapper[4791]: I1210 22:50:05.603116 4791 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T22:50:05Z","lastTransitionTime":"2025-12-10T22:50:05Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 22:50:05 crc kubenswrapper[4791]: I1210 22:50:05.607130 4791 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-tchxk" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"c0ab74e0-42a9-4f42-ac80-16030cf4ffff\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T22:49:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T22:49:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T22:49:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T22:49:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://04e0fd1b53256b7a938e8ccab93e2a4eb73dd1a59e0b9799e09c55e511baf7fe\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T22:49:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zn2qt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-10T22:49:50Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-tchxk\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T22:50:05Z is after 2025-08-24T17:21:41Z" Dec 10 22:50:05 crc kubenswrapper[4791]: I1210 22:50:05.620766 4791 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-rw584" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ff9beb90-69b0-4732-bf37-0b81f58ecc98\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T22:49:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T22:49:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T22:49:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T22:49:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ddb9df501dc6e8e5561fbc2aefd431037a3f28f73bf8e82a6598d261a4224807\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T22:49:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fksqn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ce67e54b696bb53ffbbe48c6ce71476b635e0868993e9c427b22f904891433ee\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ce67e54b696bb53ffbbe48c6ce71476b635e0868993e9c427b22f904891433ee\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-10T22:49:52Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-10T22:49:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fksqn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://818d89bfffb8a110670152c8480684665ea8db578e9768f524fbbe6a92f27c22\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://818d89bfffb8a110670152c8480684665ea8db578e9768f524fbbe6a92f27c22\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-10T22:49:53Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-10T22:49:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fksqn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://9aa9c06e894bb2a2ed3738defaeff68a0a0458ead48177f36b90770944c211a0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://9aa9c06e894bb2a2ed3738defaeff68a0a0458ead48177f36b90770944c211a0\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-10T22:49:54Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-10T22:49:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fksqn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://dfd6d03336dcea7529419a30e8d70d84e2290a0fa5b0e46d09ad341c190638d6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://dfd6d03336dcea7529419a30e8d70d84e2290a0fa5b0e46d09ad341c190638d6\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-10T22:49:55Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-10T22:49:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fksqn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b86738401ece6fbc3083782720c66a515850ed78d4054f34fcabab7ff525b85f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b86738401ece6fbc3083782720c66a515850ed78d4054f34fcabab7ff525b85f\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-10T22:49:56Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-10T22:49:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fksqn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://18b326f572a899ddd10197a745bf9264d4ddd975690ad1cfb0ec576ccd5cf102\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://18b326f572a899ddd10197a745bf9264d4ddd975690ad1cfb0ec576ccd5cf102\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-10T22:49:57Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-10T22:49:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fksqn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-10T22:49:51Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-rw584\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T22:50:05Z is after 2025-08-24T17:21:41Z" Dec 10 22:50:05 crc kubenswrapper[4791]: I1210 22:50:05.632019 4791 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-10T22:49:52Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-10T22:49:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-10T22:49:52Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://20651eb0c16e2c71222f56da1ff755cb0e7b084886acbbb24f0ccaf57041ed85\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T22:49:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T22:50:05Z is after 2025-08-24T17:21:41Z" Dec 10 22:50:05 crc kubenswrapper[4791]: I1210 22:50:05.705204 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 22:50:05 crc kubenswrapper[4791]: I1210 22:50:05.705239 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 22:50:05 crc kubenswrapper[4791]: I1210 22:50:05.705247 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 22:50:05 crc kubenswrapper[4791]: I1210 22:50:05.705262 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 22:50:05 crc kubenswrapper[4791]: I1210 22:50:05.705272 4791 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T22:50:05Z","lastTransitionTime":"2025-12-10T22:50:05Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 22:50:05 crc kubenswrapper[4791]: I1210 22:50:05.807167 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 22:50:05 crc kubenswrapper[4791]: I1210 22:50:05.807209 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 22:50:05 crc kubenswrapper[4791]: I1210 22:50:05.807217 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 22:50:05 crc kubenswrapper[4791]: I1210 22:50:05.807231 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 22:50:05 crc kubenswrapper[4791]: I1210 22:50:05.807240 4791 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T22:50:05Z","lastTransitionTime":"2025-12-10T22:50:05Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 22:50:05 crc kubenswrapper[4791]: I1210 22:50:05.897877 4791 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-tchxk" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"c0ab74e0-42a9-4f42-ac80-16030cf4ffff\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T22:49:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T22:49:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T22:49:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T22:49:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://04e0fd1b53256b7a938e8ccab93e2a4eb73dd1a59e0b9799e09c55e511baf7fe\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T22:49:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zn2qt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-10T22:49:50Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-tchxk\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T22:50:05Z is after 2025-08-24T17:21:41Z" Dec 10 22:50:05 crc kubenswrapper[4791]: I1210 22:50:05.909329 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 22:50:05 crc kubenswrapper[4791]: I1210 22:50:05.909385 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 22:50:05 crc kubenswrapper[4791]: I1210 22:50:05.909395 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 22:50:05 crc kubenswrapper[4791]: I1210 22:50:05.909412 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 22:50:05 crc kubenswrapper[4791]: I1210 22:50:05.909421 4791 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T22:50:05Z","lastTransitionTime":"2025-12-10T22:50:05Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 22:50:05 crc kubenswrapper[4791]: I1210 22:50:05.914216 4791 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-rw584" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ff9beb90-69b0-4732-bf37-0b81f58ecc98\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T22:49:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T22:49:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T22:49:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T22:49:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ddb9df501dc6e8e5561fbc2aefd431037a3f28f73bf8e82a6598d261a4224807\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T22:49:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fksqn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ce67e54b696bb53ffbbe48c6ce71476b635e0868993e9c427b22f904891433ee\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ce67e54b696bb53ffbbe48c6ce71476b635e0868993e9c427b22f904891433ee\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-10T22:49:52Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-10T22:49:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fksqn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://818d89bfffb8a110670152c8480684665ea8db578e9768f524fbbe6a92f27c22\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://818d89bfffb8a110670152c8480684665ea8db578e9768f524fbbe6a92f27c22\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-10T22:49:53Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-10T22:49:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fksqn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://9aa9c06e894bb2a2ed3738defaeff68a0a0458ead48177f36b90770944c211a0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://9aa9c06e894bb2a2ed3738defaeff68a0a0458ead48177f36b90770944c211a0\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-10T22:49:54Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-10T22:49:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fksqn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://dfd6d03336dcea7529419a30e8d70d84e2290a0fa5b0e46d09ad341c190638d6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://dfd6d03336dcea7529419a30e8d70d84e2290a0fa5b0e46d09ad341c190638d6\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-10T22:49:55Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-10T22:49:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fksqn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b86738401ece6fbc3083782720c66a515850ed78d4054f34fcabab7ff525b85f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b86738401ece6fbc3083782720c66a515850ed78d4054f34fcabab7ff525b85f\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-10T22:49:56Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-10T22:49:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fksqn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://18b326f572a899ddd10197a745bf9264d4ddd975690ad1cfb0ec576ccd5cf102\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://18b326f572a899ddd10197a745bf9264d4ddd975690ad1cfb0ec576ccd5cf102\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-10T22:49:57Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-10T22:49:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fksqn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-10T22:49:51Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-rw584\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T22:50:05Z is after 2025-08-24T17:21:41Z" Dec 10 22:50:05 crc kubenswrapper[4791]: I1210 22:50:05.926137 4791 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-10T22:49:52Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-10T22:49:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-10T22:49:52Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://20651eb0c16e2c71222f56da1ff755cb0e7b084886acbbb24f0ccaf57041ed85\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T22:49:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T22:50:05Z is after 2025-08-24T17:21:41Z" Dec 10 22:50:05 crc kubenswrapper[4791]: I1210 22:50:05.940453 4791 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-10T22:49:50Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-10T22:49:50Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T22:50:05Z is after 2025-08-24T17:21:41Z" Dec 10 22:50:05 crc kubenswrapper[4791]: I1210 22:50:05.960316 4791 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-10T22:49:50Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-10T22:49:50Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T22:50:05Z is after 2025-08-24T17:21:41Z" Dec 10 22:50:05 crc kubenswrapper[4791]: I1210 22:50:05.976072 4791 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-10T22:49:50Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-10T22:49:50Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T22:50:05Z is after 2025-08-24T17:21:41Z" Dec 10 22:50:05 crc kubenswrapper[4791]: I1210 22:50:05.988292 4791 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-10T22:49:54Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-10T22:49:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-10T22:49:54Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://62bfaf5d7a6d1b452cdaf88086447d0c8cb7306e94434e46752be5475124d56a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T22:49:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T22:50:05Z is after 2025-08-24T17:21:41Z" Dec 10 22:50:05 crc kubenswrapper[4791]: I1210 22:50:05.997855 4791 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-b6kf6" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b4527e71-bb58-4a39-a99c-fe614aa4c5f9\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T22:50:04Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T22:50:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T22:50:04Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T22:50:04Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ngpr4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ngpr4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-10T22:50:04Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-b6kf6\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T22:50:05Z is after 2025-08-24T17:21:41Z" Dec 10 22:50:06 crc kubenswrapper[4791]: I1210 22:50:06.009274 4791 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f8349fea-c49b-49a1-b5ee-032c5dffe021\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T22:49:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T22:49:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T22:49:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T22:49:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T22:49:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://acafe12186df99d1059a31260de0eafce35ff7771bcb8f5c6753b72665d6f60b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T22:49:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ea4c346507862a8c12756a3061d6e25e788a8d79f4621b62f81f4322185c82a8\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T22:49:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://2e6e324ec6f4f7bb69fdfab4ea3ac1aa55e480dd8e2a8984b0fe5b4c9f66636e\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T22:49:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://1b2e11ff2bf8789fa1d0d23ae648b3e6ef653106e5bc4bbb5dd6447f7cd6f05d\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d6aa7f8c5d908d90008f2df49860ac8a3480136da567e3d85f86135d779204f9\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-12-10T22:49:44Z\\\",\\\"message\\\":\\\"W1210 22:49:33.845733 1 cmd.go:257] Using insecure, self-signed certificates\\\\nI1210 22:49:33.846108 1 crypto.go:601] Generating new CA for check-endpoints-signer@1765406973 cert, and key in /tmp/serving-cert-2585701456/serving-signer.crt, /tmp/serving-cert-2585701456/serving-signer.key\\\\nI1210 22:49:34.529748 1 observer_polling.go:159] Starting file observer\\\\nW1210 22:49:34.532225 1 builder.go:272] unable to get owner reference (falling back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": dial tcp [::1]:6443: connect: connection refused\\\\nI1210 22:49:34.532599 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1210 22:49:34.533977 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-2585701456/tls.crt::/tmp/serving-cert-2585701456/tls.key\\\\\\\"\\\\nF1210 22:49:44.774710 1 cmd.go:182] error initializing delegating authentication: unable to load configmap based request-header-client-ca-file: Get \\\\\\\"https://localhost:6443/api/v1/namespaces/kube-system/configmaps/extension-apiserver-authentication\\\\\\\": net/http: TLS handshake timeout\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-10T22:49:33Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T22:49:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://0c4d036ab269f62561fe7ef1586012cb3a16e3583a08b9488a0616ec00c8e0ae\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T22:49:32Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c5856d6b26c19fda13b5c9ac2892a7e2f9c6f79feb8173ecfd9b33b673327d67\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c5856d6b26c19fda13b5c9ac2892a7e2f9c6f79feb8173ecfd9b33b673327d67\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-10T22:49:27Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-10T22:49:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-10T22:49:25Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T22:50:06Z is after 2025-08-24T17:21:41Z" Dec 10 22:50:06 crc kubenswrapper[4791]: I1210 22:50:06.012020 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 22:50:06 crc kubenswrapper[4791]: I1210 22:50:06.012052 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 22:50:06 crc kubenswrapper[4791]: I1210 22:50:06.012061 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 22:50:06 crc kubenswrapper[4791]: I1210 22:50:06.012076 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 22:50:06 crc kubenswrapper[4791]: I1210 22:50:06.012085 4791 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T22:50:06Z","lastTransitionTime":"2025-12-10T22:50:06Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 22:50:06 crc kubenswrapper[4791]: I1210 22:50:06.029456 4791 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"90250e13-6c63-443a-a5e5-6e4e341a2289\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T22:49:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T22:49:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T22:49:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T22:49:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T22:49:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ba54f95b1e00843b854db9a10fe6897592159ccacf30ed0f76002093a13eedd3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T22:49:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5aa8ffa518f3a68b60b275f8df4851b2564aa4984b61efdc9470b73072133e43\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T22:49:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://16f989503edc78d9451638e2cf6cf4b19a2cf2b8809b10b237541cb829303091\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T22:49:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://82556824d63ce6e46df14c45689ff1f81ebf729abe902061815c040b323898e9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T22:49:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://61db531940fa6f2effdae5a005b4c91bd2698bd5d1bff9588c171a65ad550eca\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T22:49:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0f65c864d885319ce2a08d61c30be24b3ffb7d23ea30c82ed7faf039cb2c006d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://0f65c864d885319ce2a08d61c30be24b3ffb7d23ea30c82ed7faf039cb2c006d\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-10T22:49:27Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-10T22:49:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e31a76d1d5b2812ba83e24e38268e25ec8bed98fc3da9adcd06bda5822b6ace1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e31a76d1d5b2812ba83e24e38268e25ec8bed98fc3da9adcd06bda5822b6ace1\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-10T22:49:30Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-10T22:49:29Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://b5b5eceae562f3986b543d41de1e977efb85a91d84c98e09159ca6d092eda691\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b5b5eceae562f3986b543d41de1e977efb85a91d84c98e09159ca6d092eda691\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-10T22:49:31Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-10T22:49:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-10T22:49:25Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T22:50:06Z is after 2025-08-24T17:21:41Z" Dec 10 22:50:06 crc kubenswrapper[4791]: I1210 22:50:06.040409 4791 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-5rb5l" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ba35653c-6e06-4cee-a4d6-137764090d18\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T22:49:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T22:49:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T22:49:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T22:49:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://dbb5c875730a98ad2daf316cfafc08fcdce21ac16148154780ac9193ed405dab\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T22:49:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fvvrs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://185eb2ff44221a514878283ec873773a6c9a9073dc64e296f982e60dee1338e4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T22:49:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fvvrs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-10T22:49:51Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-5rb5l\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T22:50:06Z is after 2025-08-24T17:21:41Z" Dec 10 22:50:06 crc kubenswrapper[4791]: I1210 22:50:06.050735 4791 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-s6rk8" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5c9d0978-a6f9-49c5-b377-6619b4e3ff52\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T22:50:03Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T22:50:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T22:50:03Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-rbac-proxy ovnkube-cluster-manager]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T22:50:03Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-rbac-proxy ovnkube-cluster-manager]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7vkgm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7vkgm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-10T22:50:03Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-s6rk8\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T22:50:06Z is after 2025-08-24T17:21:41Z" Dec 10 22:50:06 crc kubenswrapper[4791]: I1210 22:50:06.059770 4791 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-vczjq" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b9f00f25-3fa1-4b63-9710-fd352224b01b\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T22:49:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T22:49:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T22:49:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T22:49:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4698cf4c6b55b41e628448d37c2eb847c2777d0efbd93afcbf367d5e78623b7e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T22:49:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lrrpv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-10T22:49:50Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-vczjq\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T22:50:06Z is after 2025-08-24T17:21:41Z" Dec 10 22:50:06 crc kubenswrapper[4791]: I1210 22:50:06.072006 4791 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-4nwkq" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"672aa28c-8169-49ed-87b8-21187d13a80c\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T22:49:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T22:49:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T22:49:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T22:49:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a2f2a62315529e148a10a33ccb241e9417fbde0fe42afb61cb7ee2438c74636f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T22:49:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xl62j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-10T22:49:51Z\\\"}}\" for pod \"openshift-multus\"/\"multus-4nwkq\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T22:50:06Z is after 2025-08-24T17:21:41Z" Dec 10 22:50:06 crc kubenswrapper[4791]: I1210 22:50:06.087772 4791 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-zhq64" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3cd47739-0fa9-4321-aff1-220f8721a0b3\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T22:49:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T22:49:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T22:49:52Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T22:49:52Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://68795a29944e56870ffe94fa732e9dcf0bfb344c0d85d1e590f58a876129c61f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T22:49:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wvpw8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://bc721b530c08b433756e978ec0423cc18aabf6771b61eb96f903b4d57ea335ff\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T22:49:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wvpw8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://152a7618be8f38fadf4e564b3948980503cedcdb466c37344ccc20f4a3173a5f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T22:49:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wvpw8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3d18f9cb7207efb1e32de75a30d33e024f656a08f0503db95e5488bc5280971f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T22:49:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wvpw8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://937137ae9d7480cdc7c59bc4a1df0a4c10e72b3b63693af285278789953ed218\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T22:49:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wvpw8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://1a52d19b7671e8b143017c8a981a6f8d3ee3b84650e4b223b814a531ac76f364\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T22:49:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wvpw8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f78a47c773006a18dbc4b706edfc9d1d5d0b6ad11a023a49ea4e622eae77bba3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7a03d21e34f9d560953b7a809a4d5f374e39d3daeb8e698dbb5861b8e63c9de5\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-12-10T22:50:03Z\\\",\\\"message\\\":\\\"r removal\\\\nI1210 22:50:01.628167 6086 handler.go:190] Sending *v1.EgressIP event handler 8 for removal\\\\nI1210 22:50:01.628168 6086 handler.go:208] Removed *v1.NetworkPolicy event handler 4\\\\nI1210 22:50:01.628234 6086 handler.go:208] Removed *v1.Namespace event handler 1\\\\nI1210 22:50:01.628254 6086 handler.go:208] Removed *v1.Namespace event handler 5\\\\nI1210 22:50:01.628258 6086 handler.go:190] Sending *v1.Pod event handler 6 for removal\\\\nI1210 22:50:01.628268 6086 handler.go:190] Sending *v1.Pod event handler 3 for removal\\\\nI1210 22:50:01.628292 6086 handler.go:190] Sending *v1.Node event handler 2 for removal\\\\nI1210 22:50:01.628301 6086 handler.go:190] Sending *v1.Node event handler 7 for removal\\\\nI1210 22:50:01.628373 6086 handler.go:208] Removed *v1.Pod event handler 6\\\\nI1210 22:50:01.628392 6086 handler.go:208] Removed *v1.EgressIP event handler 8\\\\nI1210 22:50:01.628406 6086 handler.go:208] Removed *v1.Node event handler 2\\\\nI1210 22:50:01.628413 6086 handler.go:190] Sending *v1.EgressFirewall event handler 9 for removal\\\\nI1210 22:50:01.628412 6086 handler.go:208] Removed *v1.Pod event handler 3\\\\nI1210 22:50:01.628462 6086 factory.go:656] Stopping watch factory\\\\nI1210 22:50:01.628492 6086 ovnkube.go:599] Stopped ovnkube\\\\nI1210 22:50:01.628536 6086 handler.go:208] Removed *v1.EgressFirewall event handler 9\\\\nI12\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-10T22:49:58Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T22:50:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wvpw8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://dd3f80e5713221f96ed50da7f2e7fd5b9fabfcd5c1656da8d8a86d818a5fdc3f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T22:49:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wvpw8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3afbd1ffcd2c4b40324aca6e52d6fc822db131c3f35a9af3a6f005c2cffdb7bf\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3afbd1ffcd2c4b40324aca6e52d6fc822db131c3f35a9af3a6f005c2cffdb7bf\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-10T22:49:53Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-10T22:49:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wvpw8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-10T22:49:52Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-zhq64\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T22:50:06Z is after 2025-08-24T17:21:41Z" Dec 10 22:50:06 crc kubenswrapper[4791]: I1210 22:50:06.097648 4791 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"698bbe9a-fbe5-436d-a835-5e80df77ba24\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T22:49:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T22:49:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T22:49:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T22:49:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T22:49:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://224d73713312376a58ec6cc2bc8fdb6ffa44c4749ae0e7691d5fe27b9afca09f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T22:49:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://372eeaffd40992c6f06efc4df6dfcb44b31de64ee8256c21508cdd7c74ecb96e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T22:49:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://76b6013b062cac5b0b4d12a94f38ce7b43f9c7ea216063455d88a956ac04d26e\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T22:49:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://78e950e122e761d950da1a3c60c06476e53920e8795f6c61db22c01e22d5136e\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T22:49:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-10T22:49:25Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T22:50:06Z is after 2025-08-24T17:21:41Z" Dec 10 22:50:06 crc kubenswrapper[4791]: I1210 22:50:06.110036 4791 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-10T22:49:52Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-10T22:49:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-10T22:49:52Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3734b110802828632d44263ed0a43cabc4b710833f0be120d72db0940fe783e2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T22:49:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://794803ac9be3c52509e6dcb10854d8864e1a9e406762e57c7971b07c4da1220e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T22:49:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T22:50:06Z is after 2025-08-24T17:21:41Z" Dec 10 22:50:06 crc kubenswrapper[4791]: I1210 22:50:06.115171 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 22:50:06 crc kubenswrapper[4791]: I1210 22:50:06.115222 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 22:50:06 crc kubenswrapper[4791]: I1210 22:50:06.115238 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 22:50:06 crc kubenswrapper[4791]: I1210 22:50:06.115259 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 22:50:06 crc kubenswrapper[4791]: I1210 22:50:06.115277 4791 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T22:50:06Z","lastTransitionTime":"2025-12-10T22:50:06Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 22:50:06 crc kubenswrapper[4791]: I1210 22:50:06.218312 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 22:50:06 crc kubenswrapper[4791]: I1210 22:50:06.218388 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 22:50:06 crc kubenswrapper[4791]: I1210 22:50:06.218400 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 22:50:06 crc kubenswrapper[4791]: I1210 22:50:06.218415 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 22:50:06 crc kubenswrapper[4791]: I1210 22:50:06.218446 4791 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T22:50:06Z","lastTransitionTime":"2025-12-10T22:50:06Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 22:50:06 crc kubenswrapper[4791]: I1210 22:50:06.321561 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 22:50:06 crc kubenswrapper[4791]: I1210 22:50:06.321637 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 22:50:06 crc kubenswrapper[4791]: I1210 22:50:06.321656 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 22:50:06 crc kubenswrapper[4791]: I1210 22:50:06.322087 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 22:50:06 crc kubenswrapper[4791]: I1210 22:50:06.322142 4791 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T22:50:06Z","lastTransitionTime":"2025-12-10T22:50:06Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 22:50:06 crc kubenswrapper[4791]: I1210 22:50:06.365025 4791 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-zhq64_3cd47739-0fa9-4321-aff1-220f8721a0b3/ovnkube-controller/1.log" Dec 10 22:50:06 crc kubenswrapper[4791]: I1210 22:50:06.366117 4791 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-zhq64_3cd47739-0fa9-4321-aff1-220f8721a0b3/ovnkube-controller/0.log" Dec 10 22:50:06 crc kubenswrapper[4791]: I1210 22:50:06.370556 4791 generic.go:334] "Generic (PLEG): container finished" podID="3cd47739-0fa9-4321-aff1-220f8721a0b3" containerID="f78a47c773006a18dbc4b706edfc9d1d5d0b6ad11a023a49ea4e622eae77bba3" exitCode=1 Dec 10 22:50:06 crc kubenswrapper[4791]: I1210 22:50:06.370618 4791 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-zhq64" event={"ID":"3cd47739-0fa9-4321-aff1-220f8721a0b3","Type":"ContainerDied","Data":"f78a47c773006a18dbc4b706edfc9d1d5d0b6ad11a023a49ea4e622eae77bba3"} Dec 10 22:50:06 crc kubenswrapper[4791]: I1210 22:50:06.370731 4791 scope.go:117] "RemoveContainer" containerID="7a03d21e34f9d560953b7a809a4d5f374e39d3daeb8e698dbb5861b8e63c9de5" Dec 10 22:50:06 crc kubenswrapper[4791]: I1210 22:50:06.372637 4791 scope.go:117] "RemoveContainer" containerID="f78a47c773006a18dbc4b706edfc9d1d5d0b6ad11a023a49ea4e622eae77bba3" Dec 10 22:50:06 crc kubenswrapper[4791]: I1210 22:50:06.372993 4791 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-s6rk8" event={"ID":"5c9d0978-a6f9-49c5-b377-6619b4e3ff52","Type":"ContainerStarted","Data":"51d1b9b086c60bae9223cb51d5a998d02790a0d91b745aab10392c23e617eb9f"} Dec 10 22:50:06 crc kubenswrapper[4791]: E1210 22:50:06.373024 4791 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"ovnkube-controller\" with CrashLoopBackOff: \"back-off 10s restarting failed container=ovnkube-controller pod=ovnkube-node-zhq64_openshift-ovn-kubernetes(3cd47739-0fa9-4321-aff1-220f8721a0b3)\"" pod="openshift-ovn-kubernetes/ovnkube-node-zhq64" podUID="3cd47739-0fa9-4321-aff1-220f8721a0b3" Dec 10 22:50:06 crc kubenswrapper[4791]: I1210 22:50:06.395062 4791 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-10T22:49:52Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-10T22:49:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-10T22:49:52Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3734b110802828632d44263ed0a43cabc4b710833f0be120d72db0940fe783e2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T22:49:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://794803ac9be3c52509e6dcb10854d8864e1a9e406762e57c7971b07c4da1220e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T22:49:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T22:50:06Z is after 2025-08-24T17:21:41Z" Dec 10 22:50:06 crc kubenswrapper[4791]: I1210 22:50:06.413542 4791 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-vczjq" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b9f00f25-3fa1-4b63-9710-fd352224b01b\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T22:49:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T22:49:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T22:49:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T22:49:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4698cf4c6b55b41e628448d37c2eb847c2777d0efbd93afcbf367d5e78623b7e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T22:49:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lrrpv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-10T22:49:50Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-vczjq\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T22:50:06Z is after 2025-08-24T17:21:41Z" Dec 10 22:50:06 crc kubenswrapper[4791]: I1210 22:50:06.425184 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 22:50:06 crc kubenswrapper[4791]: I1210 22:50:06.425246 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 22:50:06 crc kubenswrapper[4791]: I1210 22:50:06.425264 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 22:50:06 crc kubenswrapper[4791]: I1210 22:50:06.425287 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 22:50:06 crc kubenswrapper[4791]: I1210 22:50:06.425306 4791 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T22:50:06Z","lastTransitionTime":"2025-12-10T22:50:06Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 22:50:06 crc kubenswrapper[4791]: I1210 22:50:06.435561 4791 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-4nwkq" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"672aa28c-8169-49ed-87b8-21187d13a80c\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T22:49:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T22:49:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T22:49:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T22:49:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a2f2a62315529e148a10a33ccb241e9417fbde0fe42afb61cb7ee2438c74636f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T22:49:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xl62j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-10T22:49:51Z\\\"}}\" for pod \"openshift-multus\"/\"multus-4nwkq\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T22:50:06Z is after 2025-08-24T17:21:41Z" Dec 10 22:50:06 crc kubenswrapper[4791]: I1210 22:50:06.468531 4791 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-zhq64" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3cd47739-0fa9-4321-aff1-220f8721a0b3\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T22:49:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T22:49:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T22:49:52Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T22:49:52Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://68795a29944e56870ffe94fa732e9dcf0bfb344c0d85d1e590f58a876129c61f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T22:49:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wvpw8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://bc721b530c08b433756e978ec0423cc18aabf6771b61eb96f903b4d57ea335ff\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T22:49:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wvpw8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://152a7618be8f38fadf4e564b3948980503cedcdb466c37344ccc20f4a3173a5f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T22:49:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wvpw8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3d18f9cb7207efb1e32de75a30d33e024f656a08f0503db95e5488bc5280971f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T22:49:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wvpw8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://937137ae9d7480cdc7c59bc4a1df0a4c10e72b3b63693af285278789953ed218\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T22:49:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wvpw8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://1a52d19b7671e8b143017c8a981a6f8d3ee3b84650e4b223b814a531ac76f364\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T22:49:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wvpw8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f78a47c773006a18dbc4b706edfc9d1d5d0b6ad11a023a49ea4e622eae77bba3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7a03d21e34f9d560953b7a809a4d5f374e39d3daeb8e698dbb5861b8e63c9de5\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-12-10T22:50:03Z\\\",\\\"message\\\":\\\"r removal\\\\nI1210 22:50:01.628167 6086 handler.go:190] Sending *v1.EgressIP event handler 8 for removal\\\\nI1210 22:50:01.628168 6086 handler.go:208] Removed *v1.NetworkPolicy event handler 4\\\\nI1210 22:50:01.628234 6086 handler.go:208] Removed *v1.Namespace event handler 1\\\\nI1210 22:50:01.628254 6086 handler.go:208] Removed *v1.Namespace event handler 5\\\\nI1210 22:50:01.628258 6086 handler.go:190] Sending *v1.Pod event handler 6 for removal\\\\nI1210 22:50:01.628268 6086 handler.go:190] Sending *v1.Pod event handler 3 for removal\\\\nI1210 22:50:01.628292 6086 handler.go:190] Sending *v1.Node event handler 2 for removal\\\\nI1210 22:50:01.628301 6086 handler.go:190] Sending *v1.Node event handler 7 for removal\\\\nI1210 22:50:01.628373 6086 handler.go:208] Removed *v1.Pod event handler 6\\\\nI1210 22:50:01.628392 6086 handler.go:208] Removed *v1.EgressIP event handler 8\\\\nI1210 22:50:01.628406 6086 handler.go:208] Removed *v1.Node event handler 2\\\\nI1210 22:50:01.628413 6086 handler.go:190] Sending *v1.EgressFirewall event handler 9 for removal\\\\nI1210 22:50:01.628412 6086 handler.go:208] Removed *v1.Pod event handler 3\\\\nI1210 22:50:01.628462 6086 factory.go:656] Stopping watch factory\\\\nI1210 22:50:01.628492 6086 ovnkube.go:599] Stopped ovnkube\\\\nI1210 22:50:01.628536 6086 handler.go:208] Removed *v1.EgressFirewall event handler 9\\\\nI12\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-10T22:49:58Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f78a47c773006a18dbc4b706edfc9d1d5d0b6ad11a023a49ea4e622eae77bba3\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-12-10T22:50:05Z\\\",\\\"message\\\":\\\"olicy controller, err: could not add Event Handler for anpInformer during admin network policy controller initialization, handler {0x1fcc6e0 0x1fcc3c0 0x1fcc360} was not added to shared informer because it has stopped already, failed to start node network controller: failed to start default node network controller: failed to set node crc annotations: Internal error occurred: failed calling webhook \\\\\\\"node.network-node-identity.openshift.io\\\\\\\": failed to call webhook: Post \\\\\\\"https://127.0.0.1:9743/node?timeout=10s\\\\\\\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T22:50:05Z is after 2025-08-24T17:21:41Z]\\\\nI1210 22:50:05.954954 6227 services_controller.go:473] Services do not match for network=default, existing lbs: []services.LB{services.LB{Name:\\\\\\\"Service_openshift-operator-lifecycle-manager/olm-operator-metrics_TCP_cluster\\\\\\\", UUID:\\\\\\\"63b1440a-0908-4cab-8799-012fa1cf0b07\\\\\\\", Protocol:\\\\\\\"tcp\\\\\\\", ExternalIDs:map[string]string{\\\\\\\"k8s.ovn.org/kind\\\\\\\":\\\\\\\"Service\\\\\\\", \\\\\\\"k8s.ovn.org/owner\\\\\\\":\\\\\\\"openshift-operator-lifecycle-manager/olm-operator-metrics\\\\\\\"}, Opts:services.LBOpts{Reject:false, EmptyLBEvents:false, AffinityTimeOut:0, SkipSNAT:false, Template:false, AddressFamily:\\\\\\\"\\\\\\\"}, Rules:[]services.LBRule{}, Templates:services.TemplateMap{}, Switches:[]string{}, Routers:[]string{}, Groups:[]string{\\\\\\\"clusterLBGroup\\\\\\\"}}}, bui\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-10T22:50:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wvpw8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://dd3f80e5713221f96ed50da7f2e7fd5b9fabfcd5c1656da8d8a86d818a5fdc3f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T22:49:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wvpw8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3afbd1ffcd2c4b40324aca6e52d6fc822db131c3f35a9af3a6f005c2cffdb7bf\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3afbd1ffcd2c4b40324aca6e52d6fc822db131c3f35a9af3a6f005c2cffdb7bf\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-10T22:49:53Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-10T22:49:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wvpw8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-10T22:49:52Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-zhq64\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T22:50:06Z is after 2025-08-24T17:21:41Z" Dec 10 22:50:06 crc kubenswrapper[4791]: I1210 22:50:06.488475 4791 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"698bbe9a-fbe5-436d-a835-5e80df77ba24\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T22:49:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T22:49:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T22:49:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T22:49:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T22:49:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://224d73713312376a58ec6cc2bc8fdb6ffa44c4749ae0e7691d5fe27b9afca09f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T22:49:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://372eeaffd40992c6f06efc4df6dfcb44b31de64ee8256c21508cdd7c74ecb96e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T22:49:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://76b6013b062cac5b0b4d12a94f38ce7b43f9c7ea216063455d88a956ac04d26e\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T22:49:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://78e950e122e761d950da1a3c60c06476e53920e8795f6c61db22c01e22d5136e\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T22:49:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-10T22:49:25Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T22:50:06Z is after 2025-08-24T17:21:41Z" Dec 10 22:50:06 crc kubenswrapper[4791]: I1210 22:50:06.506968 4791 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-10T22:49:50Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-10T22:49:50Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T22:50:06Z is after 2025-08-24T17:21:41Z" Dec 10 22:50:06 crc kubenswrapper[4791]: I1210 22:50:06.526700 4791 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-10T22:49:50Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-10T22:49:50Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T22:50:06Z is after 2025-08-24T17:21:41Z" Dec 10 22:50:06 crc kubenswrapper[4791]: I1210 22:50:06.528521 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 22:50:06 crc kubenswrapper[4791]: I1210 22:50:06.528582 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 22:50:06 crc kubenswrapper[4791]: I1210 22:50:06.528604 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 22:50:06 crc kubenswrapper[4791]: I1210 22:50:06.528625 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 22:50:06 crc kubenswrapper[4791]: I1210 22:50:06.528638 4791 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T22:50:06Z","lastTransitionTime":"2025-12-10T22:50:06Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 22:50:06 crc kubenswrapper[4791]: I1210 22:50:06.543735 4791 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-tchxk" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"c0ab74e0-42a9-4f42-ac80-16030cf4ffff\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T22:49:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T22:49:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T22:49:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T22:49:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://04e0fd1b53256b7a938e8ccab93e2a4eb73dd1a59e0b9799e09c55e511baf7fe\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T22:49:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zn2qt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-10T22:49:50Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-tchxk\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T22:50:06Z is after 2025-08-24T17:21:41Z" Dec 10 22:50:06 crc kubenswrapper[4791]: I1210 22:50:06.564452 4791 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/b4527e71-bb58-4a39-a99c-fe614aa4c5f9-metrics-certs\") pod \"network-metrics-daemon-b6kf6\" (UID: \"b4527e71-bb58-4a39-a99c-fe614aa4c5f9\") " pod="openshift-multus/network-metrics-daemon-b6kf6" Dec 10 22:50:06 crc kubenswrapper[4791]: E1210 22:50:06.564606 4791 secret.go:188] Couldn't get secret openshift-multus/metrics-daemon-secret: object "openshift-multus"/"metrics-daemon-secret" not registered Dec 10 22:50:06 crc kubenswrapper[4791]: E1210 22:50:06.564707 4791 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/b4527e71-bb58-4a39-a99c-fe614aa4c5f9-metrics-certs podName:b4527e71-bb58-4a39-a99c-fe614aa4c5f9 nodeName:}" failed. No retries permitted until 2025-12-10 22:50:08.564676333 +0000 UTC m=+42.994293986 (durationBeforeRetry 2s). Error: MountVolume.SetUp failed for volume "metrics-certs" (UniqueName: "kubernetes.io/secret/b4527e71-bb58-4a39-a99c-fe614aa4c5f9-metrics-certs") pod "network-metrics-daemon-b6kf6" (UID: "b4527e71-bb58-4a39-a99c-fe614aa4c5f9") : object "openshift-multus"/"metrics-daemon-secret" not registered Dec 10 22:50:06 crc kubenswrapper[4791]: I1210 22:50:06.570801 4791 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-rw584" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ff9beb90-69b0-4732-bf37-0b81f58ecc98\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T22:49:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T22:49:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T22:49:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T22:49:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ddb9df501dc6e8e5561fbc2aefd431037a3f28f73bf8e82a6598d261a4224807\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T22:49:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fksqn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ce67e54b696bb53ffbbe48c6ce71476b635e0868993e9c427b22f904891433ee\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ce67e54b696bb53ffbbe48c6ce71476b635e0868993e9c427b22f904891433ee\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-10T22:49:52Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-10T22:49:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fksqn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://818d89bfffb8a110670152c8480684665ea8db578e9768f524fbbe6a92f27c22\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://818d89bfffb8a110670152c8480684665ea8db578e9768f524fbbe6a92f27c22\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-10T22:49:53Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-10T22:49:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fksqn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://9aa9c06e894bb2a2ed3738defaeff68a0a0458ead48177f36b90770944c211a0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://9aa9c06e894bb2a2ed3738defaeff68a0a0458ead48177f36b90770944c211a0\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-10T22:49:54Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-10T22:49:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fksqn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://dfd6d03336dcea7529419a30e8d70d84e2290a0fa5b0e46d09ad341c190638d6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://dfd6d03336dcea7529419a30e8d70d84e2290a0fa5b0e46d09ad341c190638d6\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-10T22:49:55Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-10T22:49:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fksqn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b86738401ece6fbc3083782720c66a515850ed78d4054f34fcabab7ff525b85f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b86738401ece6fbc3083782720c66a515850ed78d4054f34fcabab7ff525b85f\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-10T22:49:56Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-10T22:49:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fksqn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://18b326f572a899ddd10197a745bf9264d4ddd975690ad1cfb0ec576ccd5cf102\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://18b326f572a899ddd10197a745bf9264d4ddd975690ad1cfb0ec576ccd5cf102\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-10T22:49:57Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-10T22:49:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fksqn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-10T22:49:51Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-rw584\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T22:50:06Z is after 2025-08-24T17:21:41Z" Dec 10 22:50:06 crc kubenswrapper[4791]: I1210 22:50:06.591835 4791 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-10T22:49:52Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-10T22:49:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-10T22:49:52Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://20651eb0c16e2c71222f56da1ff755cb0e7b084886acbbb24f0ccaf57041ed85\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T22:49:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T22:50:06Z is after 2025-08-24T17:21:41Z" Dec 10 22:50:06 crc kubenswrapper[4791]: I1210 22:50:06.626612 4791 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"90250e13-6c63-443a-a5e5-6e4e341a2289\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T22:49:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T22:49:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T22:49:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T22:49:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T22:49:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ba54f95b1e00843b854db9a10fe6897592159ccacf30ed0f76002093a13eedd3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T22:49:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5aa8ffa518f3a68b60b275f8df4851b2564aa4984b61efdc9470b73072133e43\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T22:49:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://16f989503edc78d9451638e2cf6cf4b19a2cf2b8809b10b237541cb829303091\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T22:49:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://82556824d63ce6e46df14c45689ff1f81ebf729abe902061815c040b323898e9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T22:49:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://61db531940fa6f2effdae5a005b4c91bd2698bd5d1bff9588c171a65ad550eca\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T22:49:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0f65c864d885319ce2a08d61c30be24b3ffb7d23ea30c82ed7faf039cb2c006d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://0f65c864d885319ce2a08d61c30be24b3ffb7d23ea30c82ed7faf039cb2c006d\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-10T22:49:27Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-10T22:49:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e31a76d1d5b2812ba83e24e38268e25ec8bed98fc3da9adcd06bda5822b6ace1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e31a76d1d5b2812ba83e24e38268e25ec8bed98fc3da9adcd06bda5822b6ace1\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-10T22:49:30Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-10T22:49:29Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://b5b5eceae562f3986b543d41de1e977efb85a91d84c98e09159ca6d092eda691\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b5b5eceae562f3986b543d41de1e977efb85a91d84c98e09159ca6d092eda691\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-10T22:49:31Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-10T22:49:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-10T22:49:25Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T22:50:06Z is after 2025-08-24T17:21:41Z" Dec 10 22:50:06 crc kubenswrapper[4791]: I1210 22:50:06.632001 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 22:50:06 crc kubenswrapper[4791]: I1210 22:50:06.632107 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 22:50:06 crc kubenswrapper[4791]: I1210 22:50:06.632131 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 22:50:06 crc kubenswrapper[4791]: I1210 22:50:06.632162 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 22:50:06 crc kubenswrapper[4791]: I1210 22:50:06.632182 4791 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T22:50:06Z","lastTransitionTime":"2025-12-10T22:50:06Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 22:50:06 crc kubenswrapper[4791]: I1210 22:50:06.651118 4791 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-10T22:49:50Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-10T22:49:50Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T22:50:06Z is after 2025-08-24T17:21:41Z" Dec 10 22:50:06 crc kubenswrapper[4791]: I1210 22:50:06.666023 4791 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 10 22:50:06 crc kubenswrapper[4791]: E1210 22:50:06.666241 4791 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-10 22:50:22.666212998 +0000 UTC m=+57.095830621 (durationBeforeRetry 16s). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 10 22:50:06 crc kubenswrapper[4791]: I1210 22:50:06.674518 4791 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-10T22:49:54Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-10T22:49:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-10T22:49:54Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://62bfaf5d7a6d1b452cdaf88086447d0c8cb7306e94434e46752be5475124d56a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T22:49:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T22:50:06Z is after 2025-08-24T17:21:41Z" Dec 10 22:50:06 crc kubenswrapper[4791]: I1210 22:50:06.689703 4791 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-b6kf6" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b4527e71-bb58-4a39-a99c-fe614aa4c5f9\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T22:50:04Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T22:50:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T22:50:04Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T22:50:04Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ngpr4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ngpr4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-10T22:50:04Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-b6kf6\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T22:50:06Z is after 2025-08-24T17:21:41Z" Dec 10 22:50:06 crc kubenswrapper[4791]: I1210 22:50:06.710836 4791 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f8349fea-c49b-49a1-b5ee-032c5dffe021\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T22:49:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T22:49:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T22:49:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T22:49:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T22:49:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://acafe12186df99d1059a31260de0eafce35ff7771bcb8f5c6753b72665d6f60b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T22:49:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ea4c346507862a8c12756a3061d6e25e788a8d79f4621b62f81f4322185c82a8\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T22:49:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://2e6e324ec6f4f7bb69fdfab4ea3ac1aa55e480dd8e2a8984b0fe5b4c9f66636e\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T22:49:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://1b2e11ff2bf8789fa1d0d23ae648b3e6ef653106e5bc4bbb5dd6447f7cd6f05d\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d6aa7f8c5d908d90008f2df49860ac8a3480136da567e3d85f86135d779204f9\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-12-10T22:49:44Z\\\",\\\"message\\\":\\\"W1210 22:49:33.845733 1 cmd.go:257] Using insecure, self-signed certificates\\\\nI1210 22:49:33.846108 1 crypto.go:601] Generating new CA for check-endpoints-signer@1765406973 cert, and key in /tmp/serving-cert-2585701456/serving-signer.crt, /tmp/serving-cert-2585701456/serving-signer.key\\\\nI1210 22:49:34.529748 1 observer_polling.go:159] Starting file observer\\\\nW1210 22:49:34.532225 1 builder.go:272] unable to get owner reference (falling back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": dial tcp [::1]:6443: connect: connection refused\\\\nI1210 22:49:34.532599 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1210 22:49:34.533977 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-2585701456/tls.crt::/tmp/serving-cert-2585701456/tls.key\\\\\\\"\\\\nF1210 22:49:44.774710 1 cmd.go:182] error initializing delegating authentication: unable to load configmap based request-header-client-ca-file: Get \\\\\\\"https://localhost:6443/api/v1/namespaces/kube-system/configmaps/extension-apiserver-authentication\\\\\\\": net/http: TLS handshake timeout\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-10T22:49:33Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T22:49:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://0c4d036ab269f62561fe7ef1586012cb3a16e3583a08b9488a0616ec00c8e0ae\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T22:49:32Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c5856d6b26c19fda13b5c9ac2892a7e2f9c6f79feb8173ecfd9b33b673327d67\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c5856d6b26c19fda13b5c9ac2892a7e2f9c6f79feb8173ecfd9b33b673327d67\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-10T22:49:27Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-10T22:49:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-10T22:49:25Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T22:50:06Z is after 2025-08-24T17:21:41Z" Dec 10 22:50:06 crc kubenswrapper[4791]: I1210 22:50:06.735838 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 22:50:06 crc kubenswrapper[4791]: I1210 22:50:06.735903 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 22:50:06 crc kubenswrapper[4791]: I1210 22:50:06.735921 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 22:50:06 crc kubenswrapper[4791]: I1210 22:50:06.735947 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 22:50:06 crc kubenswrapper[4791]: I1210 22:50:06.735968 4791 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T22:50:06Z","lastTransitionTime":"2025-12-10T22:50:06Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 22:50:06 crc kubenswrapper[4791]: I1210 22:50:06.754707 4791 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-s6rk8" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5c9d0978-a6f9-49c5-b377-6619b4e3ff52\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T22:50:03Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T22:50:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T22:50:03Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-rbac-proxy ovnkube-cluster-manager]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T22:50:03Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-rbac-proxy ovnkube-cluster-manager]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7vkgm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7vkgm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-10T22:50:03Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-s6rk8\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T22:50:06Z is after 2025-08-24T17:21:41Z" Dec 10 22:50:06 crc kubenswrapper[4791]: I1210 22:50:06.767477 4791 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nginx-conf\" (UniqueName: \"kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 10 22:50:06 crc kubenswrapper[4791]: I1210 22:50:06.767547 4791 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cqllr\" (UniqueName: \"kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr\") pod \"network-check-target-xd92c\" (UID: \"3b6479f0-333b-4a96-9adf-2099afdc2447\") " pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 10 22:50:06 crc kubenswrapper[4791]: I1210 22:50:06.767575 4791 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-s2dwl\" (UniqueName: \"kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl\") pod \"network-check-source-55646444c4-trplf\" (UID: \"9d751cbb-f2e2-430d-9754-c882a5e924a5\") " pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 10 22:50:06 crc kubenswrapper[4791]: I1210 22:50:06.767607 4791 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"networking-console-plugin-cert\" (UniqueName: \"kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 10 22:50:06 crc kubenswrapper[4791]: E1210 22:50:06.767664 4791 configmap.go:193] Couldn't get configMap openshift-network-console/networking-console-plugin: object "openshift-network-console"/"networking-console-plugin" not registered Dec 10 22:50:06 crc kubenswrapper[4791]: E1210 22:50:06.767710 4791 secret.go:188] Couldn't get secret openshift-network-console/networking-console-plugin-cert: object "openshift-network-console"/"networking-console-plugin-cert" not registered Dec 10 22:50:06 crc kubenswrapper[4791]: E1210 22:50:06.767728 4791 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Dec 10 22:50:06 crc kubenswrapper[4791]: E1210 22:50:06.767761 4791 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Dec 10 22:50:06 crc kubenswrapper[4791]: E1210 22:50:06.767775 4791 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2025-12-10 22:50:22.767751663 +0000 UTC m=+57.197369286 (durationBeforeRetry 16s). Error: MountVolume.SetUp failed for volume "nginx-conf" (UniqueName: "kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin" not registered Dec 10 22:50:06 crc kubenswrapper[4791]: E1210 22:50:06.767780 4791 projected.go:194] Error preparing data for projected volume kube-api-access-cqllr for pod openshift-network-diagnostics/network-check-target-xd92c: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Dec 10 22:50:06 crc kubenswrapper[4791]: E1210 22:50:06.767783 4791 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Dec 10 22:50:06 crc kubenswrapper[4791]: E1210 22:50:06.767832 4791 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Dec 10 22:50:06 crc kubenswrapper[4791]: E1210 22:50:06.767803 4791 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2025-12-10 22:50:22.767791544 +0000 UTC m=+57.197409247 (durationBeforeRetry 16s). Error: MountVolume.SetUp failed for volume "networking-console-plugin-cert" (UniqueName: "kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin-cert" not registered Dec 10 22:50:06 crc kubenswrapper[4791]: E1210 22:50:06.767851 4791 projected.go:194] Error preparing data for projected volume kube-api-access-s2dwl for pod openshift-network-diagnostics/network-check-source-55646444c4-trplf: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Dec 10 22:50:06 crc kubenswrapper[4791]: E1210 22:50:06.767876 4791 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr podName:3b6479f0-333b-4a96-9adf-2099afdc2447 nodeName:}" failed. No retries permitted until 2025-12-10 22:50:22.767853425 +0000 UTC m=+57.197471058 (durationBeforeRetry 16s). Error: MountVolume.SetUp failed for volume "kube-api-access-cqllr" (UniqueName: "kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr") pod "network-check-target-xd92c" (UID: "3b6479f0-333b-4a96-9adf-2099afdc2447") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Dec 10 22:50:06 crc kubenswrapper[4791]: E1210 22:50:06.767900 4791 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl podName:9d751cbb-f2e2-430d-9754-c882a5e924a5 nodeName:}" failed. No retries permitted until 2025-12-10 22:50:22.767886006 +0000 UTC m=+57.197503689 (durationBeforeRetry 16s). Error: MountVolume.SetUp failed for volume "kube-api-access-s2dwl" (UniqueName: "kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl") pod "network-check-source-55646444c4-trplf" (UID: "9d751cbb-f2e2-430d-9754-c882a5e924a5") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Dec 10 22:50:06 crc kubenswrapper[4791]: I1210 22:50:06.784617 4791 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-5rb5l" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ba35653c-6e06-4cee-a4d6-137764090d18\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T22:49:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T22:49:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T22:49:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T22:49:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://dbb5c875730a98ad2daf316cfafc08fcdce21ac16148154780ac9193ed405dab\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T22:49:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fvvrs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://185eb2ff44221a514878283ec873773a6c9a9073dc64e296f982e60dee1338e4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T22:49:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fvvrs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-10T22:49:51Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-5rb5l\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T22:50:06Z is after 2025-08-24T17:21:41Z" Dec 10 22:50:06 crc kubenswrapper[4791]: I1210 22:50:06.799498 4791 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-10T22:49:52Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-10T22:49:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-10T22:49:52Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://20651eb0c16e2c71222f56da1ff755cb0e7b084886acbbb24f0ccaf57041ed85\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T22:49:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T22:50:06Z is after 2025-08-24T17:21:41Z" Dec 10 22:50:06 crc kubenswrapper[4791]: I1210 22:50:06.812055 4791 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-10T22:49:50Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-10T22:49:50Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T22:50:06Z is after 2025-08-24T17:21:41Z" Dec 10 22:50:06 crc kubenswrapper[4791]: I1210 22:50:06.826564 4791 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-10T22:49:50Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-10T22:49:50Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T22:50:06Z is after 2025-08-24T17:21:41Z" Dec 10 22:50:06 crc kubenswrapper[4791]: I1210 22:50:06.838390 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 22:50:06 crc kubenswrapper[4791]: I1210 22:50:06.838440 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 22:50:06 crc kubenswrapper[4791]: I1210 22:50:06.838456 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 22:50:06 crc kubenswrapper[4791]: I1210 22:50:06.838477 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 22:50:06 crc kubenswrapper[4791]: I1210 22:50:06.838493 4791 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T22:50:06Z","lastTransitionTime":"2025-12-10T22:50:06Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 22:50:06 crc kubenswrapper[4791]: I1210 22:50:06.840933 4791 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-tchxk" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"c0ab74e0-42a9-4f42-ac80-16030cf4ffff\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T22:49:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T22:49:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T22:49:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T22:49:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://04e0fd1b53256b7a938e8ccab93e2a4eb73dd1a59e0b9799e09c55e511baf7fe\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T22:49:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zn2qt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-10T22:49:50Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-tchxk\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T22:50:06Z is after 2025-08-24T17:21:41Z" Dec 10 22:50:06 crc kubenswrapper[4791]: I1210 22:50:06.856753 4791 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-rw584" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ff9beb90-69b0-4732-bf37-0b81f58ecc98\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T22:49:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T22:49:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T22:49:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T22:49:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ddb9df501dc6e8e5561fbc2aefd431037a3f28f73bf8e82a6598d261a4224807\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T22:49:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fksqn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ce67e54b696bb53ffbbe48c6ce71476b635e0868993e9c427b22f904891433ee\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ce67e54b696bb53ffbbe48c6ce71476b635e0868993e9c427b22f904891433ee\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-10T22:49:52Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-10T22:49:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fksqn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://818d89bfffb8a110670152c8480684665ea8db578e9768f524fbbe6a92f27c22\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://818d89bfffb8a110670152c8480684665ea8db578e9768f524fbbe6a92f27c22\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-10T22:49:53Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-10T22:49:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fksqn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://9aa9c06e894bb2a2ed3738defaeff68a0a0458ead48177f36b90770944c211a0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://9aa9c06e894bb2a2ed3738defaeff68a0a0458ead48177f36b90770944c211a0\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-10T22:49:54Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-10T22:49:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fksqn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://dfd6d03336dcea7529419a30e8d70d84e2290a0fa5b0e46d09ad341c190638d6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://dfd6d03336dcea7529419a30e8d70d84e2290a0fa5b0e46d09ad341c190638d6\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-10T22:49:55Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-10T22:49:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fksqn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b86738401ece6fbc3083782720c66a515850ed78d4054f34fcabab7ff525b85f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b86738401ece6fbc3083782720c66a515850ed78d4054f34fcabab7ff525b85f\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-10T22:49:56Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-10T22:49:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fksqn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://18b326f572a899ddd10197a745bf9264d4ddd975690ad1cfb0ec576ccd5cf102\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://18b326f572a899ddd10197a745bf9264d4ddd975690ad1cfb0ec576ccd5cf102\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-10T22:49:57Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-10T22:49:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fksqn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-10T22:49:51Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-rw584\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T22:50:06Z is after 2025-08-24T17:21:41Z" Dec 10 22:50:06 crc kubenswrapper[4791]: I1210 22:50:06.876839 4791 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f8349fea-c49b-49a1-b5ee-032c5dffe021\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T22:49:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T22:49:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T22:49:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T22:49:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T22:49:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://acafe12186df99d1059a31260de0eafce35ff7771bcb8f5c6753b72665d6f60b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T22:49:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ea4c346507862a8c12756a3061d6e25e788a8d79f4621b62f81f4322185c82a8\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T22:49:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://2e6e324ec6f4f7bb69fdfab4ea3ac1aa55e480dd8e2a8984b0fe5b4c9f66636e\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T22:49:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://1b2e11ff2bf8789fa1d0d23ae648b3e6ef653106e5bc4bbb5dd6447f7cd6f05d\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d6aa7f8c5d908d90008f2df49860ac8a3480136da567e3d85f86135d779204f9\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-12-10T22:49:44Z\\\",\\\"message\\\":\\\"W1210 22:49:33.845733 1 cmd.go:257] Using insecure, self-signed certificates\\\\nI1210 22:49:33.846108 1 crypto.go:601] Generating new CA for check-endpoints-signer@1765406973 cert, and key in /tmp/serving-cert-2585701456/serving-signer.crt, /tmp/serving-cert-2585701456/serving-signer.key\\\\nI1210 22:49:34.529748 1 observer_polling.go:159] Starting file observer\\\\nW1210 22:49:34.532225 1 builder.go:272] unable to get owner reference (falling back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": dial tcp [::1]:6443: connect: connection refused\\\\nI1210 22:49:34.532599 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1210 22:49:34.533977 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-2585701456/tls.crt::/tmp/serving-cert-2585701456/tls.key\\\\\\\"\\\\nF1210 22:49:44.774710 1 cmd.go:182] error initializing delegating authentication: unable to load configmap based request-header-client-ca-file: Get \\\\\\\"https://localhost:6443/api/v1/namespaces/kube-system/configmaps/extension-apiserver-authentication\\\\\\\": net/http: TLS handshake timeout\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-10T22:49:33Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T22:49:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://0c4d036ab269f62561fe7ef1586012cb3a16e3583a08b9488a0616ec00c8e0ae\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T22:49:32Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c5856d6b26c19fda13b5c9ac2892a7e2f9c6f79feb8173ecfd9b33b673327d67\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c5856d6b26c19fda13b5c9ac2892a7e2f9c6f79feb8173ecfd9b33b673327d67\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-10T22:49:27Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-10T22:49:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-10T22:49:25Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T22:50:06Z is after 2025-08-24T17:21:41Z" Dec 10 22:50:06 crc kubenswrapper[4791]: I1210 22:50:06.884655 4791 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 10 22:50:06 crc kubenswrapper[4791]: I1210 22:50:06.884709 4791 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-b6kf6" Dec 10 22:50:06 crc kubenswrapper[4791]: I1210 22:50:06.884729 4791 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 10 22:50:06 crc kubenswrapper[4791]: E1210 22:50:06.884775 4791 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 10 22:50:06 crc kubenswrapper[4791]: E1210 22:50:06.884892 4791 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-b6kf6" podUID="b4527e71-bb58-4a39-a99c-fe614aa4c5f9" Dec 10 22:50:06 crc kubenswrapper[4791]: I1210 22:50:06.884929 4791 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 10 22:50:06 crc kubenswrapper[4791]: E1210 22:50:06.885050 4791 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 10 22:50:06 crc kubenswrapper[4791]: E1210 22:50:06.885146 4791 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 10 22:50:06 crc kubenswrapper[4791]: I1210 22:50:06.898664 4791 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"90250e13-6c63-443a-a5e5-6e4e341a2289\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T22:49:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T22:49:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T22:49:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T22:49:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T22:49:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ba54f95b1e00843b854db9a10fe6897592159ccacf30ed0f76002093a13eedd3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T22:49:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5aa8ffa518f3a68b60b275f8df4851b2564aa4984b61efdc9470b73072133e43\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T22:49:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://16f989503edc78d9451638e2cf6cf4b19a2cf2b8809b10b237541cb829303091\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T22:49:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://82556824d63ce6e46df14c45689ff1f81ebf729abe902061815c040b323898e9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T22:49:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://61db531940fa6f2effdae5a005b4c91bd2698bd5d1bff9588c171a65ad550eca\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T22:49:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0f65c864d885319ce2a08d61c30be24b3ffb7d23ea30c82ed7faf039cb2c006d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://0f65c864d885319ce2a08d61c30be24b3ffb7d23ea30c82ed7faf039cb2c006d\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-10T22:49:27Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-10T22:49:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e31a76d1d5b2812ba83e24e38268e25ec8bed98fc3da9adcd06bda5822b6ace1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e31a76d1d5b2812ba83e24e38268e25ec8bed98fc3da9adcd06bda5822b6ace1\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-10T22:49:30Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-10T22:49:29Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://b5b5eceae562f3986b543d41de1e977efb85a91d84c98e09159ca6d092eda691\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b5b5eceae562f3986b543d41de1e977efb85a91d84c98e09159ca6d092eda691\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-10T22:49:31Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-10T22:49:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-10T22:49:25Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T22:50:06Z is after 2025-08-24T17:21:41Z" Dec 10 22:50:06 crc kubenswrapper[4791]: I1210 22:50:06.912658 4791 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-10T22:49:50Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-10T22:49:50Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T22:50:06Z is after 2025-08-24T17:21:41Z" Dec 10 22:50:06 crc kubenswrapper[4791]: I1210 22:50:06.927877 4791 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-10T22:49:54Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-10T22:49:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-10T22:49:54Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://62bfaf5d7a6d1b452cdaf88086447d0c8cb7306e94434e46752be5475124d56a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T22:49:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T22:50:06Z is after 2025-08-24T17:21:41Z" Dec 10 22:50:06 crc kubenswrapper[4791]: I1210 22:50:06.940860 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 22:50:06 crc kubenswrapper[4791]: I1210 22:50:06.940941 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 22:50:06 crc kubenswrapper[4791]: I1210 22:50:06.940956 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 22:50:06 crc kubenswrapper[4791]: I1210 22:50:06.940979 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 22:50:06 crc kubenswrapper[4791]: I1210 22:50:06.941005 4791 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T22:50:06Z","lastTransitionTime":"2025-12-10T22:50:06Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 22:50:06 crc kubenswrapper[4791]: I1210 22:50:06.943057 4791 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-b6kf6" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b4527e71-bb58-4a39-a99c-fe614aa4c5f9\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T22:50:04Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T22:50:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T22:50:04Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T22:50:04Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ngpr4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ngpr4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-10T22:50:04Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-b6kf6\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T22:50:06Z is after 2025-08-24T17:21:41Z" Dec 10 22:50:06 crc kubenswrapper[4791]: I1210 22:50:06.957147 4791 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-5rb5l" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ba35653c-6e06-4cee-a4d6-137764090d18\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T22:49:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T22:49:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T22:49:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T22:49:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://dbb5c875730a98ad2daf316cfafc08fcdce21ac16148154780ac9193ed405dab\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T22:49:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fvvrs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://185eb2ff44221a514878283ec873773a6c9a9073dc64e296f982e60dee1338e4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T22:49:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fvvrs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-10T22:49:51Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-5rb5l\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T22:50:06Z is after 2025-08-24T17:21:41Z" Dec 10 22:50:06 crc kubenswrapper[4791]: I1210 22:50:06.969492 4791 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-s6rk8" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5c9d0978-a6f9-49c5-b377-6619b4e3ff52\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T22:50:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T22:50:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T22:50:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T22:50:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8e55baaf460c794ab196e237d51d12de059aa038617dc2d5366d49c529f5d5c7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T22:50:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7vkgm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://51d1b9b086c60bae9223cb51d5a998d02790a0d91b745aab10392c23e617eb9f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T22:50:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7vkgm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-10T22:50:03Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-s6rk8\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T22:50:06Z is after 2025-08-24T17:21:41Z" Dec 10 22:50:06 crc kubenswrapper[4791]: I1210 22:50:06.982127 4791 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"698bbe9a-fbe5-436d-a835-5e80df77ba24\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T22:49:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T22:49:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T22:49:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T22:49:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T22:49:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://224d73713312376a58ec6cc2bc8fdb6ffa44c4749ae0e7691d5fe27b9afca09f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T22:49:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://372eeaffd40992c6f06efc4df6dfcb44b31de64ee8256c21508cdd7c74ecb96e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T22:49:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://76b6013b062cac5b0b4d12a94f38ce7b43f9c7ea216063455d88a956ac04d26e\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T22:49:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://78e950e122e761d950da1a3c60c06476e53920e8795f6c61db22c01e22d5136e\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T22:49:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-10T22:49:25Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T22:50:06Z is after 2025-08-24T17:21:41Z" Dec 10 22:50:06 crc kubenswrapper[4791]: I1210 22:50:06.997686 4791 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-10T22:49:52Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-10T22:49:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-10T22:49:52Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3734b110802828632d44263ed0a43cabc4b710833f0be120d72db0940fe783e2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T22:49:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://794803ac9be3c52509e6dcb10854d8864e1a9e406762e57c7971b07c4da1220e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T22:49:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T22:50:06Z is after 2025-08-24T17:21:41Z" Dec 10 22:50:07 crc kubenswrapper[4791]: I1210 22:50:07.006288 4791 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-vczjq" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b9f00f25-3fa1-4b63-9710-fd352224b01b\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T22:49:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T22:49:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T22:49:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T22:49:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4698cf4c6b55b41e628448d37c2eb847c2777d0efbd93afcbf367d5e78623b7e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T22:49:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lrrpv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-10T22:49:50Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-vczjq\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T22:50:07Z is after 2025-08-24T17:21:41Z" Dec 10 22:50:07 crc kubenswrapper[4791]: I1210 22:50:07.016879 4791 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-4nwkq" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"672aa28c-8169-49ed-87b8-21187d13a80c\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T22:49:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T22:49:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T22:49:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T22:49:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a2f2a62315529e148a10a33ccb241e9417fbde0fe42afb61cb7ee2438c74636f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T22:49:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xl62j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-10T22:49:51Z\\\"}}\" for pod \"openshift-multus\"/\"multus-4nwkq\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T22:50:07Z is after 2025-08-24T17:21:41Z" Dec 10 22:50:07 crc kubenswrapper[4791]: I1210 22:50:07.034449 4791 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-zhq64" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3cd47739-0fa9-4321-aff1-220f8721a0b3\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T22:49:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T22:49:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T22:49:52Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T22:49:52Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://68795a29944e56870ffe94fa732e9dcf0bfb344c0d85d1e590f58a876129c61f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T22:49:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wvpw8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://bc721b530c08b433756e978ec0423cc18aabf6771b61eb96f903b4d57ea335ff\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T22:49:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wvpw8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://152a7618be8f38fadf4e564b3948980503cedcdb466c37344ccc20f4a3173a5f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T22:49:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wvpw8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3d18f9cb7207efb1e32de75a30d33e024f656a08f0503db95e5488bc5280971f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T22:49:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wvpw8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://937137ae9d7480cdc7c59bc4a1df0a4c10e72b3b63693af285278789953ed218\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T22:49:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wvpw8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://1a52d19b7671e8b143017c8a981a6f8d3ee3b84650e4b223b814a531ac76f364\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T22:49:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wvpw8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f78a47c773006a18dbc4b706edfc9d1d5d0b6ad11a023a49ea4e622eae77bba3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7a03d21e34f9d560953b7a809a4d5f374e39d3daeb8e698dbb5861b8e63c9de5\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-12-10T22:50:03Z\\\",\\\"message\\\":\\\"r removal\\\\nI1210 22:50:01.628167 6086 handler.go:190] Sending *v1.EgressIP event handler 8 for removal\\\\nI1210 22:50:01.628168 6086 handler.go:208] Removed *v1.NetworkPolicy event handler 4\\\\nI1210 22:50:01.628234 6086 handler.go:208] Removed *v1.Namespace event handler 1\\\\nI1210 22:50:01.628254 6086 handler.go:208] Removed *v1.Namespace event handler 5\\\\nI1210 22:50:01.628258 6086 handler.go:190] Sending *v1.Pod event handler 6 for removal\\\\nI1210 22:50:01.628268 6086 handler.go:190] Sending *v1.Pod event handler 3 for removal\\\\nI1210 22:50:01.628292 6086 handler.go:190] Sending *v1.Node event handler 2 for removal\\\\nI1210 22:50:01.628301 6086 handler.go:190] Sending *v1.Node event handler 7 for removal\\\\nI1210 22:50:01.628373 6086 handler.go:208] Removed *v1.Pod event handler 6\\\\nI1210 22:50:01.628392 6086 handler.go:208] Removed *v1.EgressIP event handler 8\\\\nI1210 22:50:01.628406 6086 handler.go:208] Removed *v1.Node event handler 2\\\\nI1210 22:50:01.628413 6086 handler.go:190] Sending *v1.EgressFirewall event handler 9 for removal\\\\nI1210 22:50:01.628412 6086 handler.go:208] Removed *v1.Pod event handler 3\\\\nI1210 22:50:01.628462 6086 factory.go:656] Stopping watch factory\\\\nI1210 22:50:01.628492 6086 ovnkube.go:599] Stopped ovnkube\\\\nI1210 22:50:01.628536 6086 handler.go:208] Removed *v1.EgressFirewall event handler 9\\\\nI12\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-10T22:49:58Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f78a47c773006a18dbc4b706edfc9d1d5d0b6ad11a023a49ea4e622eae77bba3\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-12-10T22:50:05Z\\\",\\\"message\\\":\\\"olicy controller, err: could not add Event Handler for anpInformer during admin network policy controller initialization, handler {0x1fcc6e0 0x1fcc3c0 0x1fcc360} was not added to shared informer because it has stopped already, failed to start node network controller: failed to start default node network controller: failed to set node crc annotations: Internal error occurred: failed calling webhook \\\\\\\"node.network-node-identity.openshift.io\\\\\\\": failed to call webhook: Post \\\\\\\"https://127.0.0.1:9743/node?timeout=10s\\\\\\\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T22:50:05Z is after 2025-08-24T17:21:41Z]\\\\nI1210 22:50:05.954954 6227 services_controller.go:473] Services do not match for network=default, existing lbs: []services.LB{services.LB{Name:\\\\\\\"Service_openshift-operator-lifecycle-manager/olm-operator-metrics_TCP_cluster\\\\\\\", UUID:\\\\\\\"63b1440a-0908-4cab-8799-012fa1cf0b07\\\\\\\", Protocol:\\\\\\\"tcp\\\\\\\", ExternalIDs:map[string]string{\\\\\\\"k8s.ovn.org/kind\\\\\\\":\\\\\\\"Service\\\\\\\", \\\\\\\"k8s.ovn.org/owner\\\\\\\":\\\\\\\"openshift-operator-lifecycle-manager/olm-operator-metrics\\\\\\\"}, Opts:services.LBOpts{Reject:false, EmptyLBEvents:false, AffinityTimeOut:0, SkipSNAT:false, Template:false, AddressFamily:\\\\\\\"\\\\\\\"}, Rules:[]services.LBRule{}, Templates:services.TemplateMap{}, Switches:[]string{}, Routers:[]string{}, Groups:[]string{\\\\\\\"clusterLBGroup\\\\\\\"}}}, bui\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-10T22:50:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wvpw8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://dd3f80e5713221f96ed50da7f2e7fd5b9fabfcd5c1656da8d8a86d818a5fdc3f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T22:49:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wvpw8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3afbd1ffcd2c4b40324aca6e52d6fc822db131c3f35a9af3a6f005c2cffdb7bf\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3afbd1ffcd2c4b40324aca6e52d6fc822db131c3f35a9af3a6f005c2cffdb7bf\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-10T22:49:53Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-10T22:49:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wvpw8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-10T22:49:52Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-zhq64\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T22:50:07Z is after 2025-08-24T17:21:41Z" Dec 10 22:50:07 crc kubenswrapper[4791]: I1210 22:50:07.043019 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 22:50:07 crc kubenswrapper[4791]: I1210 22:50:07.043080 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 22:50:07 crc kubenswrapper[4791]: I1210 22:50:07.043095 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 22:50:07 crc kubenswrapper[4791]: I1210 22:50:07.043117 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 22:50:07 crc kubenswrapper[4791]: I1210 22:50:07.043136 4791 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T22:50:07Z","lastTransitionTime":"2025-12-10T22:50:07Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 22:50:07 crc kubenswrapper[4791]: I1210 22:50:07.145667 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 22:50:07 crc kubenswrapper[4791]: I1210 22:50:07.145787 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 22:50:07 crc kubenswrapper[4791]: I1210 22:50:07.145809 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 22:50:07 crc kubenswrapper[4791]: I1210 22:50:07.145837 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 22:50:07 crc kubenswrapper[4791]: I1210 22:50:07.145859 4791 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T22:50:07Z","lastTransitionTime":"2025-12-10T22:50:07Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 22:50:07 crc kubenswrapper[4791]: I1210 22:50:07.249445 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 22:50:07 crc kubenswrapper[4791]: I1210 22:50:07.249515 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 22:50:07 crc kubenswrapper[4791]: I1210 22:50:07.249539 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 22:50:07 crc kubenswrapper[4791]: I1210 22:50:07.249563 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 22:50:07 crc kubenswrapper[4791]: I1210 22:50:07.249580 4791 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T22:50:07Z","lastTransitionTime":"2025-12-10T22:50:07Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 22:50:07 crc kubenswrapper[4791]: I1210 22:50:07.353293 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 22:50:07 crc kubenswrapper[4791]: I1210 22:50:07.353405 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 22:50:07 crc kubenswrapper[4791]: I1210 22:50:07.353431 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 22:50:07 crc kubenswrapper[4791]: I1210 22:50:07.353460 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 22:50:07 crc kubenswrapper[4791]: I1210 22:50:07.353481 4791 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T22:50:07Z","lastTransitionTime":"2025-12-10T22:50:07Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 22:50:07 crc kubenswrapper[4791]: I1210 22:50:07.380190 4791 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-zhq64_3cd47739-0fa9-4321-aff1-220f8721a0b3/ovnkube-controller/1.log" Dec 10 22:50:07 crc kubenswrapper[4791]: I1210 22:50:07.456052 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 22:50:07 crc kubenswrapper[4791]: I1210 22:50:07.456108 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 22:50:07 crc kubenswrapper[4791]: I1210 22:50:07.456126 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 22:50:07 crc kubenswrapper[4791]: I1210 22:50:07.456152 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 22:50:07 crc kubenswrapper[4791]: I1210 22:50:07.456176 4791 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T22:50:07Z","lastTransitionTime":"2025-12-10T22:50:07Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 22:50:07 crc kubenswrapper[4791]: I1210 22:50:07.458594 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 22:50:07 crc kubenswrapper[4791]: I1210 22:50:07.458642 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 22:50:07 crc kubenswrapper[4791]: I1210 22:50:07.458654 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 22:50:07 crc kubenswrapper[4791]: I1210 22:50:07.458672 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 22:50:07 crc kubenswrapper[4791]: I1210 22:50:07.458687 4791 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T22:50:07Z","lastTransitionTime":"2025-12-10T22:50:07Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 22:50:07 crc kubenswrapper[4791]: E1210 22:50:07.474494 4791 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-12-10T22:50:07Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-10T22:50:07Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-10T22:50:07Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-10T22:50:07Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-10T22:50:07Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-10T22:50:07Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-10T22:50:07Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-10T22:50:07Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"4fab3af6-f657-4dfc-8ef2-12b7c978c94f\\\",\\\"systemUUID\\\":\\\"442c0b68-8dce-47e0-bd68-d5210d7e0493\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T22:50:07Z is after 2025-08-24T17:21:41Z" Dec 10 22:50:07 crc kubenswrapper[4791]: I1210 22:50:07.478931 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 22:50:07 crc kubenswrapper[4791]: I1210 22:50:07.478979 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 22:50:07 crc kubenswrapper[4791]: I1210 22:50:07.478996 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 22:50:07 crc kubenswrapper[4791]: I1210 22:50:07.479021 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 22:50:07 crc kubenswrapper[4791]: I1210 22:50:07.479037 4791 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T22:50:07Z","lastTransitionTime":"2025-12-10T22:50:07Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 22:50:07 crc kubenswrapper[4791]: E1210 22:50:07.491872 4791 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-12-10T22:50:07Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-10T22:50:07Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-10T22:50:07Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-10T22:50:07Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-10T22:50:07Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-10T22:50:07Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-10T22:50:07Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-10T22:50:07Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"4fab3af6-f657-4dfc-8ef2-12b7c978c94f\\\",\\\"systemUUID\\\":\\\"442c0b68-8dce-47e0-bd68-d5210d7e0493\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T22:50:07Z is after 2025-08-24T17:21:41Z" Dec 10 22:50:07 crc kubenswrapper[4791]: I1210 22:50:07.501547 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 22:50:07 crc kubenswrapper[4791]: I1210 22:50:07.501596 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 22:50:07 crc kubenswrapper[4791]: I1210 22:50:07.501608 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 22:50:07 crc kubenswrapper[4791]: I1210 22:50:07.501624 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 22:50:07 crc kubenswrapper[4791]: I1210 22:50:07.501634 4791 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T22:50:07Z","lastTransitionTime":"2025-12-10T22:50:07Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 22:50:07 crc kubenswrapper[4791]: E1210 22:50:07.514663 4791 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-12-10T22:50:07Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-10T22:50:07Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-10T22:50:07Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-10T22:50:07Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-10T22:50:07Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-10T22:50:07Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-10T22:50:07Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-10T22:50:07Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"4fab3af6-f657-4dfc-8ef2-12b7c978c94f\\\",\\\"systemUUID\\\":\\\"442c0b68-8dce-47e0-bd68-d5210d7e0493\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T22:50:07Z is after 2025-08-24T17:21:41Z" Dec 10 22:50:07 crc kubenswrapper[4791]: I1210 22:50:07.519532 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 22:50:07 crc kubenswrapper[4791]: I1210 22:50:07.519565 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 22:50:07 crc kubenswrapper[4791]: I1210 22:50:07.519576 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 22:50:07 crc kubenswrapper[4791]: I1210 22:50:07.519593 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 22:50:07 crc kubenswrapper[4791]: I1210 22:50:07.519605 4791 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T22:50:07Z","lastTransitionTime":"2025-12-10T22:50:07Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 22:50:07 crc kubenswrapper[4791]: E1210 22:50:07.531089 4791 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-12-10T22:50:07Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-10T22:50:07Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-10T22:50:07Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-10T22:50:07Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-10T22:50:07Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-10T22:50:07Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-10T22:50:07Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-10T22:50:07Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"4fab3af6-f657-4dfc-8ef2-12b7c978c94f\\\",\\\"systemUUID\\\":\\\"442c0b68-8dce-47e0-bd68-d5210d7e0493\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T22:50:07Z is after 2025-08-24T17:21:41Z" Dec 10 22:50:07 crc kubenswrapper[4791]: I1210 22:50:07.535432 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 22:50:07 crc kubenswrapper[4791]: I1210 22:50:07.535486 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 22:50:07 crc kubenswrapper[4791]: I1210 22:50:07.535500 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 22:50:07 crc kubenswrapper[4791]: I1210 22:50:07.535518 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 22:50:07 crc kubenswrapper[4791]: I1210 22:50:07.535530 4791 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T22:50:07Z","lastTransitionTime":"2025-12-10T22:50:07Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 22:50:07 crc kubenswrapper[4791]: E1210 22:50:07.551942 4791 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-12-10T22:50:07Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-10T22:50:07Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-10T22:50:07Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-10T22:50:07Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-10T22:50:07Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-10T22:50:07Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-10T22:50:07Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-10T22:50:07Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"4fab3af6-f657-4dfc-8ef2-12b7c978c94f\\\",\\\"systemUUID\\\":\\\"442c0b68-8dce-47e0-bd68-d5210d7e0493\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T22:50:07Z is after 2025-08-24T17:21:41Z" Dec 10 22:50:07 crc kubenswrapper[4791]: E1210 22:50:07.552088 4791 kubelet_node_status.go:572] "Unable to update node status" err="update node status exceeds retry count" Dec 10 22:50:07 crc kubenswrapper[4791]: I1210 22:50:07.558889 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 22:50:07 crc kubenswrapper[4791]: I1210 22:50:07.558930 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 22:50:07 crc kubenswrapper[4791]: I1210 22:50:07.558942 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 22:50:07 crc kubenswrapper[4791]: I1210 22:50:07.558959 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 22:50:07 crc kubenswrapper[4791]: I1210 22:50:07.558971 4791 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T22:50:07Z","lastTransitionTime":"2025-12-10T22:50:07Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 22:50:07 crc kubenswrapper[4791]: I1210 22:50:07.662406 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 22:50:07 crc kubenswrapper[4791]: I1210 22:50:07.662448 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 22:50:07 crc kubenswrapper[4791]: I1210 22:50:07.662459 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 22:50:07 crc kubenswrapper[4791]: I1210 22:50:07.662476 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 22:50:07 crc kubenswrapper[4791]: I1210 22:50:07.662488 4791 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T22:50:07Z","lastTransitionTime":"2025-12-10T22:50:07Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 22:50:07 crc kubenswrapper[4791]: I1210 22:50:07.765519 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 22:50:07 crc kubenswrapper[4791]: I1210 22:50:07.765599 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 22:50:07 crc kubenswrapper[4791]: I1210 22:50:07.765629 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 22:50:07 crc kubenswrapper[4791]: I1210 22:50:07.765662 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 22:50:07 crc kubenswrapper[4791]: I1210 22:50:07.765688 4791 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T22:50:07Z","lastTransitionTime":"2025-12-10T22:50:07Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 22:50:07 crc kubenswrapper[4791]: I1210 22:50:07.868836 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 22:50:07 crc kubenswrapper[4791]: I1210 22:50:07.868884 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 22:50:07 crc kubenswrapper[4791]: I1210 22:50:07.868895 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 22:50:07 crc kubenswrapper[4791]: I1210 22:50:07.868910 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 22:50:07 crc kubenswrapper[4791]: I1210 22:50:07.868921 4791 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T22:50:07Z","lastTransitionTime":"2025-12-10T22:50:07Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 22:50:07 crc kubenswrapper[4791]: I1210 22:50:07.972241 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 22:50:07 crc kubenswrapper[4791]: I1210 22:50:07.972283 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 22:50:07 crc kubenswrapper[4791]: I1210 22:50:07.972294 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 22:50:07 crc kubenswrapper[4791]: I1210 22:50:07.972309 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 22:50:07 crc kubenswrapper[4791]: I1210 22:50:07.972320 4791 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T22:50:07Z","lastTransitionTime":"2025-12-10T22:50:07Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 22:50:08 crc kubenswrapper[4791]: I1210 22:50:08.074743 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 22:50:08 crc kubenswrapper[4791]: I1210 22:50:08.074794 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 22:50:08 crc kubenswrapper[4791]: I1210 22:50:08.074807 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 22:50:08 crc kubenswrapper[4791]: I1210 22:50:08.074824 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 22:50:08 crc kubenswrapper[4791]: I1210 22:50:08.074834 4791 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T22:50:08Z","lastTransitionTime":"2025-12-10T22:50:08Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 22:50:08 crc kubenswrapper[4791]: I1210 22:50:08.084181 4791 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-ovn-kubernetes/ovnkube-node-zhq64" Dec 10 22:50:08 crc kubenswrapper[4791]: I1210 22:50:08.085045 4791 scope.go:117] "RemoveContainer" containerID="f78a47c773006a18dbc4b706edfc9d1d5d0b6ad11a023a49ea4e622eae77bba3" Dec 10 22:50:08 crc kubenswrapper[4791]: E1210 22:50:08.085208 4791 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"ovnkube-controller\" with CrashLoopBackOff: \"back-off 10s restarting failed container=ovnkube-controller pod=ovnkube-node-zhq64_openshift-ovn-kubernetes(3cd47739-0fa9-4321-aff1-220f8721a0b3)\"" pod="openshift-ovn-kubernetes/ovnkube-node-zhq64" podUID="3cd47739-0fa9-4321-aff1-220f8721a0b3" Dec 10 22:50:08 crc kubenswrapper[4791]: I1210 22:50:08.103954 4791 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-4nwkq" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"672aa28c-8169-49ed-87b8-21187d13a80c\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T22:49:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T22:49:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T22:49:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T22:49:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a2f2a62315529e148a10a33ccb241e9417fbde0fe42afb61cb7ee2438c74636f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T22:49:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xl62j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-10T22:49:51Z\\\"}}\" for pod \"openshift-multus\"/\"multus-4nwkq\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T22:50:08Z is after 2025-08-24T17:21:41Z" Dec 10 22:50:08 crc kubenswrapper[4791]: I1210 22:50:08.125210 4791 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-zhq64" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3cd47739-0fa9-4321-aff1-220f8721a0b3\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T22:49:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T22:49:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T22:49:52Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T22:49:52Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://68795a29944e56870ffe94fa732e9dcf0bfb344c0d85d1e590f58a876129c61f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T22:49:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wvpw8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://bc721b530c08b433756e978ec0423cc18aabf6771b61eb96f903b4d57ea335ff\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T22:49:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wvpw8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://152a7618be8f38fadf4e564b3948980503cedcdb466c37344ccc20f4a3173a5f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T22:49:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wvpw8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3d18f9cb7207efb1e32de75a30d33e024f656a08f0503db95e5488bc5280971f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T22:49:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wvpw8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://937137ae9d7480cdc7c59bc4a1df0a4c10e72b3b63693af285278789953ed218\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T22:49:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wvpw8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://1a52d19b7671e8b143017c8a981a6f8d3ee3b84650e4b223b814a531ac76f364\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T22:49:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wvpw8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f78a47c773006a18dbc4b706edfc9d1d5d0b6ad11a023a49ea4e622eae77bba3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f78a47c773006a18dbc4b706edfc9d1d5d0b6ad11a023a49ea4e622eae77bba3\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-12-10T22:50:05Z\\\",\\\"message\\\":\\\"olicy controller, err: could not add Event Handler for anpInformer during admin network policy controller initialization, handler {0x1fcc6e0 0x1fcc3c0 0x1fcc360} was not added to shared informer because it has stopped already, failed to start node network controller: failed to start default node network controller: failed to set node crc annotations: Internal error occurred: failed calling webhook \\\\\\\"node.network-node-identity.openshift.io\\\\\\\": failed to call webhook: Post \\\\\\\"https://127.0.0.1:9743/node?timeout=10s\\\\\\\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T22:50:05Z is after 2025-08-24T17:21:41Z]\\\\nI1210 22:50:05.954954 6227 services_controller.go:473] Services do not match for network=default, existing lbs: []services.LB{services.LB{Name:\\\\\\\"Service_openshift-operator-lifecycle-manager/olm-operator-metrics_TCP_cluster\\\\\\\", UUID:\\\\\\\"63b1440a-0908-4cab-8799-012fa1cf0b07\\\\\\\", Protocol:\\\\\\\"tcp\\\\\\\", ExternalIDs:map[string]string{\\\\\\\"k8s.ovn.org/kind\\\\\\\":\\\\\\\"Service\\\\\\\", \\\\\\\"k8s.ovn.org/owner\\\\\\\":\\\\\\\"openshift-operator-lifecycle-manager/olm-operator-metrics\\\\\\\"}, Opts:services.LBOpts{Reject:false, EmptyLBEvents:false, AffinityTimeOut:0, SkipSNAT:false, Template:false, AddressFamily:\\\\\\\"\\\\\\\"}, Rules:[]services.LBRule{}, Templates:services.TemplateMap{}, Switches:[]string{}, Routers:[]string{}, Groups:[]string{\\\\\\\"clusterLBGroup\\\\\\\"}}}, bui\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-10T22:50:04Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 10s restarting failed container=ovnkube-controller pod=ovnkube-node-zhq64_openshift-ovn-kubernetes(3cd47739-0fa9-4321-aff1-220f8721a0b3)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wvpw8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://dd3f80e5713221f96ed50da7f2e7fd5b9fabfcd5c1656da8d8a86d818a5fdc3f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T22:49:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wvpw8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3afbd1ffcd2c4b40324aca6e52d6fc822db131c3f35a9af3a6f005c2cffdb7bf\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3afbd1ffcd2c4b40324aca6e52d6fc822db131c3f35a9af3a6f005c2cffdb7bf\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-10T22:49:53Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-10T22:49:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wvpw8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-10T22:49:52Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-zhq64\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T22:50:08Z is after 2025-08-24T17:21:41Z" Dec 10 22:50:08 crc kubenswrapper[4791]: I1210 22:50:08.143909 4791 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"698bbe9a-fbe5-436d-a835-5e80df77ba24\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T22:49:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T22:49:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T22:49:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T22:49:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T22:49:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://224d73713312376a58ec6cc2bc8fdb6ffa44c4749ae0e7691d5fe27b9afca09f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T22:49:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://372eeaffd40992c6f06efc4df6dfcb44b31de64ee8256c21508cdd7c74ecb96e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T22:49:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://76b6013b062cac5b0b4d12a94f38ce7b43f9c7ea216063455d88a956ac04d26e\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T22:49:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://78e950e122e761d950da1a3c60c06476e53920e8795f6c61db22c01e22d5136e\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T22:49:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-10T22:49:25Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T22:50:08Z is after 2025-08-24T17:21:41Z" Dec 10 22:50:08 crc kubenswrapper[4791]: I1210 22:50:08.160948 4791 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-10T22:49:52Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-10T22:49:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-10T22:49:52Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3734b110802828632d44263ed0a43cabc4b710833f0be120d72db0940fe783e2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T22:49:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://794803ac9be3c52509e6dcb10854d8864e1a9e406762e57c7971b07c4da1220e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T22:49:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T22:50:08Z is after 2025-08-24T17:21:41Z" Dec 10 22:50:08 crc kubenswrapper[4791]: I1210 22:50:08.176414 4791 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-vczjq" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b9f00f25-3fa1-4b63-9710-fd352224b01b\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T22:49:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T22:49:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T22:49:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T22:49:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4698cf4c6b55b41e628448d37c2eb847c2777d0efbd93afcbf367d5e78623b7e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T22:49:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lrrpv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-10T22:49:50Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-vczjq\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T22:50:08Z is after 2025-08-24T17:21:41Z" Dec 10 22:50:08 crc kubenswrapper[4791]: I1210 22:50:08.178303 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 22:50:08 crc kubenswrapper[4791]: I1210 22:50:08.178435 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 22:50:08 crc kubenswrapper[4791]: I1210 22:50:08.178515 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 22:50:08 crc kubenswrapper[4791]: I1210 22:50:08.178551 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 22:50:08 crc kubenswrapper[4791]: I1210 22:50:08.178616 4791 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T22:50:08Z","lastTransitionTime":"2025-12-10T22:50:08Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 22:50:08 crc kubenswrapper[4791]: I1210 22:50:08.200478 4791 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-rw584" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ff9beb90-69b0-4732-bf37-0b81f58ecc98\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T22:49:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T22:49:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T22:49:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T22:49:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ddb9df501dc6e8e5561fbc2aefd431037a3f28f73bf8e82a6598d261a4224807\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T22:49:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fksqn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ce67e54b696bb53ffbbe48c6ce71476b635e0868993e9c427b22f904891433ee\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ce67e54b696bb53ffbbe48c6ce71476b635e0868993e9c427b22f904891433ee\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-10T22:49:52Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-10T22:49:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fksqn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://818d89bfffb8a110670152c8480684665ea8db578e9768f524fbbe6a92f27c22\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://818d89bfffb8a110670152c8480684665ea8db578e9768f524fbbe6a92f27c22\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-10T22:49:53Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-10T22:49:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fksqn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://9aa9c06e894bb2a2ed3738defaeff68a0a0458ead48177f36b90770944c211a0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://9aa9c06e894bb2a2ed3738defaeff68a0a0458ead48177f36b90770944c211a0\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-10T22:49:54Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-10T22:49:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fksqn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://dfd6d03336dcea7529419a30e8d70d84e2290a0fa5b0e46d09ad341c190638d6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://dfd6d03336dcea7529419a30e8d70d84e2290a0fa5b0e46d09ad341c190638d6\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-10T22:49:55Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-10T22:49:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fksqn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b86738401ece6fbc3083782720c66a515850ed78d4054f34fcabab7ff525b85f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b86738401ece6fbc3083782720c66a515850ed78d4054f34fcabab7ff525b85f\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-10T22:49:56Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-10T22:49:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fksqn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://18b326f572a899ddd10197a745bf9264d4ddd975690ad1cfb0ec576ccd5cf102\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://18b326f572a899ddd10197a745bf9264d4ddd975690ad1cfb0ec576ccd5cf102\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-10T22:49:57Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-10T22:49:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fksqn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-10T22:49:51Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-rw584\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T22:50:08Z is after 2025-08-24T17:21:41Z" Dec 10 22:50:08 crc kubenswrapper[4791]: I1210 22:50:08.219633 4791 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-10T22:49:52Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-10T22:49:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-10T22:49:52Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://20651eb0c16e2c71222f56da1ff755cb0e7b084886acbbb24f0ccaf57041ed85\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T22:49:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T22:50:08Z is after 2025-08-24T17:21:41Z" Dec 10 22:50:08 crc kubenswrapper[4791]: I1210 22:50:08.237683 4791 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-10T22:49:50Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-10T22:49:50Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T22:50:08Z is after 2025-08-24T17:21:41Z" Dec 10 22:50:08 crc kubenswrapper[4791]: I1210 22:50:08.254678 4791 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-10T22:49:50Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-10T22:49:50Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T22:50:08Z is after 2025-08-24T17:21:41Z" Dec 10 22:50:08 crc kubenswrapper[4791]: I1210 22:50:08.269431 4791 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-tchxk" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"c0ab74e0-42a9-4f42-ac80-16030cf4ffff\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T22:49:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T22:49:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T22:49:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T22:49:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://04e0fd1b53256b7a938e8ccab93e2a4eb73dd1a59e0b9799e09c55e511baf7fe\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T22:49:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zn2qt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-10T22:49:50Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-tchxk\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T22:50:08Z is after 2025-08-24T17:21:41Z" Dec 10 22:50:08 crc kubenswrapper[4791]: I1210 22:50:08.281644 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 22:50:08 crc kubenswrapper[4791]: I1210 22:50:08.281689 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 22:50:08 crc kubenswrapper[4791]: I1210 22:50:08.281700 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 22:50:08 crc kubenswrapper[4791]: I1210 22:50:08.281718 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 22:50:08 crc kubenswrapper[4791]: I1210 22:50:08.281730 4791 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T22:50:08Z","lastTransitionTime":"2025-12-10T22:50:08Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 22:50:08 crc kubenswrapper[4791]: I1210 22:50:08.286429 4791 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-10T22:49:54Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-10T22:49:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-10T22:49:54Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://62bfaf5d7a6d1b452cdaf88086447d0c8cb7306e94434e46752be5475124d56a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T22:49:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T22:50:08Z is after 2025-08-24T17:21:41Z" Dec 10 22:50:08 crc kubenswrapper[4791]: I1210 22:50:08.301223 4791 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-b6kf6" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b4527e71-bb58-4a39-a99c-fe614aa4c5f9\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T22:50:04Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T22:50:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T22:50:04Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T22:50:04Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ngpr4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ngpr4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-10T22:50:04Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-b6kf6\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T22:50:08Z is after 2025-08-24T17:21:41Z" Dec 10 22:50:08 crc kubenswrapper[4791]: I1210 22:50:08.323062 4791 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f8349fea-c49b-49a1-b5ee-032c5dffe021\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T22:49:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T22:49:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T22:49:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T22:49:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T22:49:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://acafe12186df99d1059a31260de0eafce35ff7771bcb8f5c6753b72665d6f60b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T22:49:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ea4c346507862a8c12756a3061d6e25e788a8d79f4621b62f81f4322185c82a8\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T22:49:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://2e6e324ec6f4f7bb69fdfab4ea3ac1aa55e480dd8e2a8984b0fe5b4c9f66636e\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T22:49:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://1b2e11ff2bf8789fa1d0d23ae648b3e6ef653106e5bc4bbb5dd6447f7cd6f05d\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d6aa7f8c5d908d90008f2df49860ac8a3480136da567e3d85f86135d779204f9\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-12-10T22:49:44Z\\\",\\\"message\\\":\\\"W1210 22:49:33.845733 1 cmd.go:257] Using insecure, self-signed certificates\\\\nI1210 22:49:33.846108 1 crypto.go:601] Generating new CA for check-endpoints-signer@1765406973 cert, and key in /tmp/serving-cert-2585701456/serving-signer.crt, /tmp/serving-cert-2585701456/serving-signer.key\\\\nI1210 22:49:34.529748 1 observer_polling.go:159] Starting file observer\\\\nW1210 22:49:34.532225 1 builder.go:272] unable to get owner reference (falling back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": dial tcp [::1]:6443: connect: connection refused\\\\nI1210 22:49:34.532599 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1210 22:49:34.533977 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-2585701456/tls.crt::/tmp/serving-cert-2585701456/tls.key\\\\\\\"\\\\nF1210 22:49:44.774710 1 cmd.go:182] error initializing delegating authentication: unable to load configmap based request-header-client-ca-file: Get \\\\\\\"https://localhost:6443/api/v1/namespaces/kube-system/configmaps/extension-apiserver-authentication\\\\\\\": net/http: TLS handshake timeout\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-10T22:49:33Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T22:49:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://0c4d036ab269f62561fe7ef1586012cb3a16e3583a08b9488a0616ec00c8e0ae\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T22:49:32Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c5856d6b26c19fda13b5c9ac2892a7e2f9c6f79feb8173ecfd9b33b673327d67\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c5856d6b26c19fda13b5c9ac2892a7e2f9c6f79feb8173ecfd9b33b673327d67\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-10T22:49:27Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-10T22:49:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-10T22:49:25Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T22:50:08Z is after 2025-08-24T17:21:41Z" Dec 10 22:50:08 crc kubenswrapper[4791]: I1210 22:50:08.345200 4791 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"90250e13-6c63-443a-a5e5-6e4e341a2289\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T22:49:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T22:49:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T22:49:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T22:49:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T22:49:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ba54f95b1e00843b854db9a10fe6897592159ccacf30ed0f76002093a13eedd3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T22:49:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5aa8ffa518f3a68b60b275f8df4851b2564aa4984b61efdc9470b73072133e43\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T22:49:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://16f989503edc78d9451638e2cf6cf4b19a2cf2b8809b10b237541cb829303091\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T22:49:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://82556824d63ce6e46df14c45689ff1f81ebf729abe902061815c040b323898e9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T22:49:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://61db531940fa6f2effdae5a005b4c91bd2698bd5d1bff9588c171a65ad550eca\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T22:49:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0f65c864d885319ce2a08d61c30be24b3ffb7d23ea30c82ed7faf039cb2c006d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://0f65c864d885319ce2a08d61c30be24b3ffb7d23ea30c82ed7faf039cb2c006d\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-10T22:49:27Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-10T22:49:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e31a76d1d5b2812ba83e24e38268e25ec8bed98fc3da9adcd06bda5822b6ace1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e31a76d1d5b2812ba83e24e38268e25ec8bed98fc3da9adcd06bda5822b6ace1\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-10T22:49:30Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-10T22:49:29Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://b5b5eceae562f3986b543d41de1e977efb85a91d84c98e09159ca6d092eda691\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b5b5eceae562f3986b543d41de1e977efb85a91d84c98e09159ca6d092eda691\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-10T22:49:31Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-10T22:49:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-10T22:49:25Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T22:50:08Z is after 2025-08-24T17:21:41Z" Dec 10 22:50:08 crc kubenswrapper[4791]: I1210 22:50:08.362871 4791 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-10T22:49:50Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-10T22:49:50Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T22:50:08Z is after 2025-08-24T17:21:41Z" Dec 10 22:50:08 crc kubenswrapper[4791]: I1210 22:50:08.381213 4791 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-5rb5l" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ba35653c-6e06-4cee-a4d6-137764090d18\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T22:49:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T22:49:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T22:49:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T22:49:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://dbb5c875730a98ad2daf316cfafc08fcdce21ac16148154780ac9193ed405dab\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T22:49:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fvvrs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://185eb2ff44221a514878283ec873773a6c9a9073dc64e296f982e60dee1338e4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T22:49:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fvvrs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-10T22:49:51Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-5rb5l\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T22:50:08Z is after 2025-08-24T17:21:41Z" Dec 10 22:50:08 crc kubenswrapper[4791]: I1210 22:50:08.386660 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 22:50:08 crc kubenswrapper[4791]: I1210 22:50:08.386713 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 22:50:08 crc kubenswrapper[4791]: I1210 22:50:08.386725 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 22:50:08 crc kubenswrapper[4791]: I1210 22:50:08.386746 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 22:50:08 crc kubenswrapper[4791]: I1210 22:50:08.386756 4791 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T22:50:08Z","lastTransitionTime":"2025-12-10T22:50:08Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 22:50:08 crc kubenswrapper[4791]: I1210 22:50:08.401401 4791 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-s6rk8" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5c9d0978-a6f9-49c5-b377-6619b4e3ff52\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T22:50:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T22:50:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T22:50:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T22:50:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8e55baaf460c794ab196e237d51d12de059aa038617dc2d5366d49c529f5d5c7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T22:50:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7vkgm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://51d1b9b086c60bae9223cb51d5a998d02790a0d91b745aab10392c23e617eb9f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T22:50:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7vkgm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-10T22:50:03Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-s6rk8\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T22:50:08Z is after 2025-08-24T17:21:41Z" Dec 10 22:50:08 crc kubenswrapper[4791]: I1210 22:50:08.489881 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 22:50:08 crc kubenswrapper[4791]: I1210 22:50:08.489958 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 22:50:08 crc kubenswrapper[4791]: I1210 22:50:08.489975 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 22:50:08 crc kubenswrapper[4791]: I1210 22:50:08.490052 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 22:50:08 crc kubenswrapper[4791]: I1210 22:50:08.490072 4791 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T22:50:08Z","lastTransitionTime":"2025-12-10T22:50:08Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 22:50:08 crc kubenswrapper[4791]: I1210 22:50:08.586985 4791 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/b4527e71-bb58-4a39-a99c-fe614aa4c5f9-metrics-certs\") pod \"network-metrics-daemon-b6kf6\" (UID: \"b4527e71-bb58-4a39-a99c-fe614aa4c5f9\") " pod="openshift-multus/network-metrics-daemon-b6kf6" Dec 10 22:50:08 crc kubenswrapper[4791]: E1210 22:50:08.587393 4791 secret.go:188] Couldn't get secret openshift-multus/metrics-daemon-secret: object "openshift-multus"/"metrics-daemon-secret" not registered Dec 10 22:50:08 crc kubenswrapper[4791]: E1210 22:50:08.587581 4791 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/b4527e71-bb58-4a39-a99c-fe614aa4c5f9-metrics-certs podName:b4527e71-bb58-4a39-a99c-fe614aa4c5f9 nodeName:}" failed. No retries permitted until 2025-12-10 22:50:12.587540869 +0000 UTC m=+47.017158532 (durationBeforeRetry 4s). Error: MountVolume.SetUp failed for volume "metrics-certs" (UniqueName: "kubernetes.io/secret/b4527e71-bb58-4a39-a99c-fe614aa4c5f9-metrics-certs") pod "network-metrics-daemon-b6kf6" (UID: "b4527e71-bb58-4a39-a99c-fe614aa4c5f9") : object "openshift-multus"/"metrics-daemon-secret" not registered Dec 10 22:50:08 crc kubenswrapper[4791]: I1210 22:50:08.593260 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 22:50:08 crc kubenswrapper[4791]: I1210 22:50:08.593319 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 22:50:08 crc kubenswrapper[4791]: I1210 22:50:08.593330 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 22:50:08 crc kubenswrapper[4791]: I1210 22:50:08.593361 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 22:50:08 crc kubenswrapper[4791]: I1210 22:50:08.593371 4791 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T22:50:08Z","lastTransitionTime":"2025-12-10T22:50:08Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 22:50:08 crc kubenswrapper[4791]: I1210 22:50:08.695831 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 22:50:08 crc kubenswrapper[4791]: I1210 22:50:08.695905 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 22:50:08 crc kubenswrapper[4791]: I1210 22:50:08.695944 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 22:50:08 crc kubenswrapper[4791]: I1210 22:50:08.695974 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 22:50:08 crc kubenswrapper[4791]: I1210 22:50:08.696009 4791 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T22:50:08Z","lastTransitionTime":"2025-12-10T22:50:08Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 22:50:08 crc kubenswrapper[4791]: I1210 22:50:08.798956 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 22:50:08 crc kubenswrapper[4791]: I1210 22:50:08.799026 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 22:50:08 crc kubenswrapper[4791]: I1210 22:50:08.799045 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 22:50:08 crc kubenswrapper[4791]: I1210 22:50:08.799068 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 22:50:08 crc kubenswrapper[4791]: I1210 22:50:08.799086 4791 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T22:50:08Z","lastTransitionTime":"2025-12-10T22:50:08Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 22:50:08 crc kubenswrapper[4791]: I1210 22:50:08.884474 4791 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 10 22:50:08 crc kubenswrapper[4791]: I1210 22:50:08.884541 4791 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-b6kf6" Dec 10 22:50:08 crc kubenswrapper[4791]: I1210 22:50:08.884504 4791 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 10 22:50:08 crc kubenswrapper[4791]: I1210 22:50:08.884512 4791 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 10 22:50:08 crc kubenswrapper[4791]: E1210 22:50:08.884750 4791 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-b6kf6" podUID="b4527e71-bb58-4a39-a99c-fe614aa4c5f9" Dec 10 22:50:08 crc kubenswrapper[4791]: E1210 22:50:08.884840 4791 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 10 22:50:08 crc kubenswrapper[4791]: E1210 22:50:08.885029 4791 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 10 22:50:08 crc kubenswrapper[4791]: E1210 22:50:08.885155 4791 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 10 22:50:08 crc kubenswrapper[4791]: I1210 22:50:08.903049 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 22:50:08 crc kubenswrapper[4791]: I1210 22:50:08.903161 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 22:50:08 crc kubenswrapper[4791]: I1210 22:50:08.903185 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 22:50:08 crc kubenswrapper[4791]: I1210 22:50:08.903215 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 22:50:08 crc kubenswrapper[4791]: I1210 22:50:08.903237 4791 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T22:50:08Z","lastTransitionTime":"2025-12-10T22:50:08Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 22:50:09 crc kubenswrapper[4791]: I1210 22:50:09.005788 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 22:50:09 crc kubenswrapper[4791]: I1210 22:50:09.005848 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 22:50:09 crc kubenswrapper[4791]: I1210 22:50:09.005865 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 22:50:09 crc kubenswrapper[4791]: I1210 22:50:09.005889 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 22:50:09 crc kubenswrapper[4791]: I1210 22:50:09.005906 4791 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T22:50:09Z","lastTransitionTime":"2025-12-10T22:50:09Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 22:50:09 crc kubenswrapper[4791]: I1210 22:50:09.108540 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 22:50:09 crc kubenswrapper[4791]: I1210 22:50:09.108670 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 22:50:09 crc kubenswrapper[4791]: I1210 22:50:09.108709 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 22:50:09 crc kubenswrapper[4791]: I1210 22:50:09.108756 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 22:50:09 crc kubenswrapper[4791]: I1210 22:50:09.108793 4791 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T22:50:09Z","lastTransitionTime":"2025-12-10T22:50:09Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 22:50:09 crc kubenswrapper[4791]: I1210 22:50:09.211517 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 22:50:09 crc kubenswrapper[4791]: I1210 22:50:09.211590 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 22:50:09 crc kubenswrapper[4791]: I1210 22:50:09.211611 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 22:50:09 crc kubenswrapper[4791]: I1210 22:50:09.211637 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 22:50:09 crc kubenswrapper[4791]: I1210 22:50:09.211657 4791 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T22:50:09Z","lastTransitionTime":"2025-12-10T22:50:09Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 22:50:09 crc kubenswrapper[4791]: I1210 22:50:09.315045 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 22:50:09 crc kubenswrapper[4791]: I1210 22:50:09.315112 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 22:50:09 crc kubenswrapper[4791]: I1210 22:50:09.315137 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 22:50:09 crc kubenswrapper[4791]: I1210 22:50:09.315164 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 22:50:09 crc kubenswrapper[4791]: I1210 22:50:09.315187 4791 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T22:50:09Z","lastTransitionTime":"2025-12-10T22:50:09Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 22:50:09 crc kubenswrapper[4791]: I1210 22:50:09.417406 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 22:50:09 crc kubenswrapper[4791]: I1210 22:50:09.417503 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 22:50:09 crc kubenswrapper[4791]: I1210 22:50:09.417523 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 22:50:09 crc kubenswrapper[4791]: I1210 22:50:09.417547 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 22:50:09 crc kubenswrapper[4791]: I1210 22:50:09.417564 4791 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T22:50:09Z","lastTransitionTime":"2025-12-10T22:50:09Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 22:50:09 crc kubenswrapper[4791]: I1210 22:50:09.520511 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 22:50:09 crc kubenswrapper[4791]: I1210 22:50:09.520565 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 22:50:09 crc kubenswrapper[4791]: I1210 22:50:09.520578 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 22:50:09 crc kubenswrapper[4791]: I1210 22:50:09.520596 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 22:50:09 crc kubenswrapper[4791]: I1210 22:50:09.520610 4791 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T22:50:09Z","lastTransitionTime":"2025-12-10T22:50:09Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 22:50:09 crc kubenswrapper[4791]: I1210 22:50:09.622670 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 22:50:09 crc kubenswrapper[4791]: I1210 22:50:09.622729 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 22:50:09 crc kubenswrapper[4791]: I1210 22:50:09.622746 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 22:50:09 crc kubenswrapper[4791]: I1210 22:50:09.622769 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 22:50:09 crc kubenswrapper[4791]: I1210 22:50:09.622789 4791 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T22:50:09Z","lastTransitionTime":"2025-12-10T22:50:09Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 22:50:09 crc kubenswrapper[4791]: I1210 22:50:09.726261 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 22:50:09 crc kubenswrapper[4791]: I1210 22:50:09.726309 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 22:50:09 crc kubenswrapper[4791]: I1210 22:50:09.726320 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 22:50:09 crc kubenswrapper[4791]: I1210 22:50:09.726333 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 22:50:09 crc kubenswrapper[4791]: I1210 22:50:09.726355 4791 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T22:50:09Z","lastTransitionTime":"2025-12-10T22:50:09Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 22:50:09 crc kubenswrapper[4791]: I1210 22:50:09.829119 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 22:50:09 crc kubenswrapper[4791]: I1210 22:50:09.829160 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 22:50:09 crc kubenswrapper[4791]: I1210 22:50:09.829172 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 22:50:09 crc kubenswrapper[4791]: I1210 22:50:09.829188 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 22:50:09 crc kubenswrapper[4791]: I1210 22:50:09.829198 4791 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T22:50:09Z","lastTransitionTime":"2025-12-10T22:50:09Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 22:50:09 crc kubenswrapper[4791]: I1210 22:50:09.931993 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 22:50:09 crc kubenswrapper[4791]: I1210 22:50:09.932053 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 22:50:09 crc kubenswrapper[4791]: I1210 22:50:09.932066 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 22:50:09 crc kubenswrapper[4791]: I1210 22:50:09.932084 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 22:50:09 crc kubenswrapper[4791]: I1210 22:50:09.932096 4791 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T22:50:09Z","lastTransitionTime":"2025-12-10T22:50:09Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 22:50:10 crc kubenswrapper[4791]: I1210 22:50:10.034690 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 22:50:10 crc kubenswrapper[4791]: I1210 22:50:10.034771 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 22:50:10 crc kubenswrapper[4791]: I1210 22:50:10.034795 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 22:50:10 crc kubenswrapper[4791]: I1210 22:50:10.034821 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 22:50:10 crc kubenswrapper[4791]: I1210 22:50:10.034839 4791 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T22:50:10Z","lastTransitionTime":"2025-12-10T22:50:10Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 22:50:10 crc kubenswrapper[4791]: I1210 22:50:10.136987 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 22:50:10 crc kubenswrapper[4791]: I1210 22:50:10.137024 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 22:50:10 crc kubenswrapper[4791]: I1210 22:50:10.137032 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 22:50:10 crc kubenswrapper[4791]: I1210 22:50:10.137050 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 22:50:10 crc kubenswrapper[4791]: I1210 22:50:10.137060 4791 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T22:50:10Z","lastTransitionTime":"2025-12-10T22:50:10Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 22:50:10 crc kubenswrapper[4791]: I1210 22:50:10.240428 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 22:50:10 crc kubenswrapper[4791]: I1210 22:50:10.240516 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 22:50:10 crc kubenswrapper[4791]: I1210 22:50:10.240542 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 22:50:10 crc kubenswrapper[4791]: I1210 22:50:10.240572 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 22:50:10 crc kubenswrapper[4791]: I1210 22:50:10.240589 4791 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T22:50:10Z","lastTransitionTime":"2025-12-10T22:50:10Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 22:50:10 crc kubenswrapper[4791]: I1210 22:50:10.343327 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 22:50:10 crc kubenswrapper[4791]: I1210 22:50:10.343423 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 22:50:10 crc kubenswrapper[4791]: I1210 22:50:10.343445 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 22:50:10 crc kubenswrapper[4791]: I1210 22:50:10.343473 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 22:50:10 crc kubenswrapper[4791]: I1210 22:50:10.343495 4791 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T22:50:10Z","lastTransitionTime":"2025-12-10T22:50:10Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 22:50:10 crc kubenswrapper[4791]: I1210 22:50:10.447032 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 22:50:10 crc kubenswrapper[4791]: I1210 22:50:10.447093 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 22:50:10 crc kubenswrapper[4791]: I1210 22:50:10.447109 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 22:50:10 crc kubenswrapper[4791]: I1210 22:50:10.447140 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 22:50:10 crc kubenswrapper[4791]: I1210 22:50:10.447157 4791 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T22:50:10Z","lastTransitionTime":"2025-12-10T22:50:10Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 22:50:10 crc kubenswrapper[4791]: I1210 22:50:10.548992 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 22:50:10 crc kubenswrapper[4791]: I1210 22:50:10.549070 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 22:50:10 crc kubenswrapper[4791]: I1210 22:50:10.549088 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 22:50:10 crc kubenswrapper[4791]: I1210 22:50:10.549119 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 22:50:10 crc kubenswrapper[4791]: I1210 22:50:10.549134 4791 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T22:50:10Z","lastTransitionTime":"2025-12-10T22:50:10Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 22:50:10 crc kubenswrapper[4791]: I1210 22:50:10.651950 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 22:50:10 crc kubenswrapper[4791]: I1210 22:50:10.652021 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 22:50:10 crc kubenswrapper[4791]: I1210 22:50:10.652042 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 22:50:10 crc kubenswrapper[4791]: I1210 22:50:10.652072 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 22:50:10 crc kubenswrapper[4791]: I1210 22:50:10.652097 4791 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T22:50:10Z","lastTransitionTime":"2025-12-10T22:50:10Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 22:50:10 crc kubenswrapper[4791]: I1210 22:50:10.754855 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 22:50:10 crc kubenswrapper[4791]: I1210 22:50:10.754889 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 22:50:10 crc kubenswrapper[4791]: I1210 22:50:10.754916 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 22:50:10 crc kubenswrapper[4791]: I1210 22:50:10.754932 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 22:50:10 crc kubenswrapper[4791]: I1210 22:50:10.754949 4791 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T22:50:10Z","lastTransitionTime":"2025-12-10T22:50:10Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 22:50:10 crc kubenswrapper[4791]: I1210 22:50:10.858776 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 22:50:10 crc kubenswrapper[4791]: I1210 22:50:10.859223 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 22:50:10 crc kubenswrapper[4791]: I1210 22:50:10.859294 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 22:50:10 crc kubenswrapper[4791]: I1210 22:50:10.859396 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 22:50:10 crc kubenswrapper[4791]: I1210 22:50:10.859441 4791 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T22:50:10Z","lastTransitionTime":"2025-12-10T22:50:10Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 22:50:10 crc kubenswrapper[4791]: I1210 22:50:10.883778 4791 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 10 22:50:10 crc kubenswrapper[4791]: I1210 22:50:10.883834 4791 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 10 22:50:10 crc kubenswrapper[4791]: I1210 22:50:10.883786 4791 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 10 22:50:10 crc kubenswrapper[4791]: I1210 22:50:10.883952 4791 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-b6kf6" Dec 10 22:50:10 crc kubenswrapper[4791]: E1210 22:50:10.884041 4791 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 10 22:50:10 crc kubenswrapper[4791]: E1210 22:50:10.884235 4791 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-b6kf6" podUID="b4527e71-bb58-4a39-a99c-fe614aa4c5f9" Dec 10 22:50:10 crc kubenswrapper[4791]: E1210 22:50:10.884440 4791 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 10 22:50:10 crc kubenswrapper[4791]: E1210 22:50:10.884613 4791 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 10 22:50:10 crc kubenswrapper[4791]: I1210 22:50:10.963195 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 22:50:10 crc kubenswrapper[4791]: I1210 22:50:10.963254 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 22:50:10 crc kubenswrapper[4791]: I1210 22:50:10.963269 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 22:50:10 crc kubenswrapper[4791]: I1210 22:50:10.963290 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 22:50:10 crc kubenswrapper[4791]: I1210 22:50:10.963304 4791 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T22:50:10Z","lastTransitionTime":"2025-12-10T22:50:10Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 22:50:11 crc kubenswrapper[4791]: I1210 22:50:11.065606 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 22:50:11 crc kubenswrapper[4791]: I1210 22:50:11.065653 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 22:50:11 crc kubenswrapper[4791]: I1210 22:50:11.065668 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 22:50:11 crc kubenswrapper[4791]: I1210 22:50:11.065685 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 22:50:11 crc kubenswrapper[4791]: I1210 22:50:11.065697 4791 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T22:50:11Z","lastTransitionTime":"2025-12-10T22:50:11Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 22:50:11 crc kubenswrapper[4791]: I1210 22:50:11.168211 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 22:50:11 crc kubenswrapper[4791]: I1210 22:50:11.168285 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 22:50:11 crc kubenswrapper[4791]: I1210 22:50:11.168308 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 22:50:11 crc kubenswrapper[4791]: I1210 22:50:11.168386 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 22:50:11 crc kubenswrapper[4791]: I1210 22:50:11.168411 4791 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T22:50:11Z","lastTransitionTime":"2025-12-10T22:50:11Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 22:50:11 crc kubenswrapper[4791]: I1210 22:50:11.271517 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 22:50:11 crc kubenswrapper[4791]: I1210 22:50:11.271580 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 22:50:11 crc kubenswrapper[4791]: I1210 22:50:11.271600 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 22:50:11 crc kubenswrapper[4791]: I1210 22:50:11.271623 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 22:50:11 crc kubenswrapper[4791]: I1210 22:50:11.271641 4791 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T22:50:11Z","lastTransitionTime":"2025-12-10T22:50:11Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 22:50:11 crc kubenswrapper[4791]: I1210 22:50:11.375223 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 22:50:11 crc kubenswrapper[4791]: I1210 22:50:11.375305 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 22:50:11 crc kubenswrapper[4791]: I1210 22:50:11.375328 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 22:50:11 crc kubenswrapper[4791]: I1210 22:50:11.375390 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 22:50:11 crc kubenswrapper[4791]: I1210 22:50:11.375411 4791 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T22:50:11Z","lastTransitionTime":"2025-12-10T22:50:11Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 22:50:11 crc kubenswrapper[4791]: I1210 22:50:11.478275 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 22:50:11 crc kubenswrapper[4791]: I1210 22:50:11.478392 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 22:50:11 crc kubenswrapper[4791]: I1210 22:50:11.478416 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 22:50:11 crc kubenswrapper[4791]: I1210 22:50:11.478447 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 22:50:11 crc kubenswrapper[4791]: I1210 22:50:11.478470 4791 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T22:50:11Z","lastTransitionTime":"2025-12-10T22:50:11Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 22:50:11 crc kubenswrapper[4791]: I1210 22:50:11.581540 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 22:50:11 crc kubenswrapper[4791]: I1210 22:50:11.581612 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 22:50:11 crc kubenswrapper[4791]: I1210 22:50:11.581630 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 22:50:11 crc kubenswrapper[4791]: I1210 22:50:11.581654 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 22:50:11 crc kubenswrapper[4791]: I1210 22:50:11.581675 4791 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T22:50:11Z","lastTransitionTime":"2025-12-10T22:50:11Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 22:50:11 crc kubenswrapper[4791]: I1210 22:50:11.684188 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 22:50:11 crc kubenswrapper[4791]: I1210 22:50:11.684268 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 22:50:11 crc kubenswrapper[4791]: I1210 22:50:11.684294 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 22:50:11 crc kubenswrapper[4791]: I1210 22:50:11.684323 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 22:50:11 crc kubenswrapper[4791]: I1210 22:50:11.684379 4791 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T22:50:11Z","lastTransitionTime":"2025-12-10T22:50:11Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 22:50:11 crc kubenswrapper[4791]: I1210 22:50:11.786609 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 22:50:11 crc kubenswrapper[4791]: I1210 22:50:11.786656 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 22:50:11 crc kubenswrapper[4791]: I1210 22:50:11.786666 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 22:50:11 crc kubenswrapper[4791]: I1210 22:50:11.786681 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 22:50:11 crc kubenswrapper[4791]: I1210 22:50:11.786694 4791 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T22:50:11Z","lastTransitionTime":"2025-12-10T22:50:11Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 22:50:11 crc kubenswrapper[4791]: I1210 22:50:11.889748 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 22:50:11 crc kubenswrapper[4791]: I1210 22:50:11.889903 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 22:50:11 crc kubenswrapper[4791]: I1210 22:50:11.889956 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 22:50:11 crc kubenswrapper[4791]: I1210 22:50:11.890099 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 22:50:11 crc kubenswrapper[4791]: I1210 22:50:11.890154 4791 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T22:50:11Z","lastTransitionTime":"2025-12-10T22:50:11Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 22:50:11 crc kubenswrapper[4791]: I1210 22:50:11.993469 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 22:50:11 crc kubenswrapper[4791]: I1210 22:50:11.993507 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 22:50:11 crc kubenswrapper[4791]: I1210 22:50:11.993515 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 22:50:11 crc kubenswrapper[4791]: I1210 22:50:11.993530 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 22:50:11 crc kubenswrapper[4791]: I1210 22:50:11.993539 4791 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T22:50:11Z","lastTransitionTime":"2025-12-10T22:50:11Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 22:50:12 crc kubenswrapper[4791]: I1210 22:50:12.096585 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 22:50:12 crc kubenswrapper[4791]: I1210 22:50:12.096661 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 22:50:12 crc kubenswrapper[4791]: I1210 22:50:12.096676 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 22:50:12 crc kubenswrapper[4791]: I1210 22:50:12.096699 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 22:50:12 crc kubenswrapper[4791]: I1210 22:50:12.096721 4791 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T22:50:12Z","lastTransitionTime":"2025-12-10T22:50:12Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 22:50:12 crc kubenswrapper[4791]: I1210 22:50:12.199766 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 22:50:12 crc kubenswrapper[4791]: I1210 22:50:12.199817 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 22:50:12 crc kubenswrapper[4791]: I1210 22:50:12.199834 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 22:50:12 crc kubenswrapper[4791]: I1210 22:50:12.199872 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 22:50:12 crc kubenswrapper[4791]: I1210 22:50:12.199890 4791 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T22:50:12Z","lastTransitionTime":"2025-12-10T22:50:12Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 22:50:12 crc kubenswrapper[4791]: I1210 22:50:12.303118 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 22:50:12 crc kubenswrapper[4791]: I1210 22:50:12.303191 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 22:50:12 crc kubenswrapper[4791]: I1210 22:50:12.303209 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 22:50:12 crc kubenswrapper[4791]: I1210 22:50:12.303237 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 22:50:12 crc kubenswrapper[4791]: I1210 22:50:12.303255 4791 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T22:50:12Z","lastTransitionTime":"2025-12-10T22:50:12Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 22:50:12 crc kubenswrapper[4791]: I1210 22:50:12.405832 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 22:50:12 crc kubenswrapper[4791]: I1210 22:50:12.405965 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 22:50:12 crc kubenswrapper[4791]: I1210 22:50:12.405984 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 22:50:12 crc kubenswrapper[4791]: I1210 22:50:12.406020 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 22:50:12 crc kubenswrapper[4791]: I1210 22:50:12.406037 4791 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T22:50:12Z","lastTransitionTime":"2025-12-10T22:50:12Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 22:50:12 crc kubenswrapper[4791]: I1210 22:50:12.508958 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 22:50:12 crc kubenswrapper[4791]: I1210 22:50:12.509012 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 22:50:12 crc kubenswrapper[4791]: I1210 22:50:12.509030 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 22:50:12 crc kubenswrapper[4791]: I1210 22:50:12.509053 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 22:50:12 crc kubenswrapper[4791]: I1210 22:50:12.509071 4791 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T22:50:12Z","lastTransitionTime":"2025-12-10T22:50:12Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 22:50:12 crc kubenswrapper[4791]: I1210 22:50:12.611068 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 22:50:12 crc kubenswrapper[4791]: I1210 22:50:12.611105 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 22:50:12 crc kubenswrapper[4791]: I1210 22:50:12.611142 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 22:50:12 crc kubenswrapper[4791]: I1210 22:50:12.611212 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 22:50:12 crc kubenswrapper[4791]: I1210 22:50:12.611226 4791 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T22:50:12Z","lastTransitionTime":"2025-12-10T22:50:12Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 22:50:12 crc kubenswrapper[4791]: I1210 22:50:12.628017 4791 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/b4527e71-bb58-4a39-a99c-fe614aa4c5f9-metrics-certs\") pod \"network-metrics-daemon-b6kf6\" (UID: \"b4527e71-bb58-4a39-a99c-fe614aa4c5f9\") " pod="openshift-multus/network-metrics-daemon-b6kf6" Dec 10 22:50:12 crc kubenswrapper[4791]: E1210 22:50:12.628203 4791 secret.go:188] Couldn't get secret openshift-multus/metrics-daemon-secret: object "openshift-multus"/"metrics-daemon-secret" not registered Dec 10 22:50:12 crc kubenswrapper[4791]: E1210 22:50:12.628301 4791 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/b4527e71-bb58-4a39-a99c-fe614aa4c5f9-metrics-certs podName:b4527e71-bb58-4a39-a99c-fe614aa4c5f9 nodeName:}" failed. No retries permitted until 2025-12-10 22:50:20.628278667 +0000 UTC m=+55.057896320 (durationBeforeRetry 8s). Error: MountVolume.SetUp failed for volume "metrics-certs" (UniqueName: "kubernetes.io/secret/b4527e71-bb58-4a39-a99c-fe614aa4c5f9-metrics-certs") pod "network-metrics-daemon-b6kf6" (UID: "b4527e71-bb58-4a39-a99c-fe614aa4c5f9") : object "openshift-multus"/"metrics-daemon-secret" not registered Dec 10 22:50:12 crc kubenswrapper[4791]: I1210 22:50:12.713505 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 22:50:12 crc kubenswrapper[4791]: I1210 22:50:12.713568 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 22:50:12 crc kubenswrapper[4791]: I1210 22:50:12.713586 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 22:50:12 crc kubenswrapper[4791]: I1210 22:50:12.713611 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 22:50:12 crc kubenswrapper[4791]: I1210 22:50:12.713629 4791 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T22:50:12Z","lastTransitionTime":"2025-12-10T22:50:12Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 22:50:12 crc kubenswrapper[4791]: I1210 22:50:12.816143 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 22:50:12 crc kubenswrapper[4791]: I1210 22:50:12.816189 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 22:50:12 crc kubenswrapper[4791]: I1210 22:50:12.816201 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 22:50:12 crc kubenswrapper[4791]: I1210 22:50:12.816219 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 22:50:12 crc kubenswrapper[4791]: I1210 22:50:12.816235 4791 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T22:50:12Z","lastTransitionTime":"2025-12-10T22:50:12Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 22:50:12 crc kubenswrapper[4791]: I1210 22:50:12.884075 4791 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-b6kf6" Dec 10 22:50:12 crc kubenswrapper[4791]: E1210 22:50:12.884279 4791 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-b6kf6" podUID="b4527e71-bb58-4a39-a99c-fe614aa4c5f9" Dec 10 22:50:12 crc kubenswrapper[4791]: I1210 22:50:12.884460 4791 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 10 22:50:12 crc kubenswrapper[4791]: I1210 22:50:12.884460 4791 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 10 22:50:12 crc kubenswrapper[4791]: E1210 22:50:12.884650 4791 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 10 22:50:12 crc kubenswrapper[4791]: E1210 22:50:12.884776 4791 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 10 22:50:12 crc kubenswrapper[4791]: I1210 22:50:12.884493 4791 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 10 22:50:12 crc kubenswrapper[4791]: E1210 22:50:12.884943 4791 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 10 22:50:12 crc kubenswrapper[4791]: I1210 22:50:12.919488 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 22:50:12 crc kubenswrapper[4791]: I1210 22:50:12.919562 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 22:50:12 crc kubenswrapper[4791]: I1210 22:50:12.919585 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 22:50:12 crc kubenswrapper[4791]: I1210 22:50:12.919614 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 22:50:12 crc kubenswrapper[4791]: I1210 22:50:12.919636 4791 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T22:50:12Z","lastTransitionTime":"2025-12-10T22:50:12Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 22:50:13 crc kubenswrapper[4791]: I1210 22:50:13.023161 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 22:50:13 crc kubenswrapper[4791]: I1210 22:50:13.023229 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 22:50:13 crc kubenswrapper[4791]: I1210 22:50:13.023253 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 22:50:13 crc kubenswrapper[4791]: I1210 22:50:13.023282 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 22:50:13 crc kubenswrapper[4791]: I1210 22:50:13.023304 4791 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T22:50:13Z","lastTransitionTime":"2025-12-10T22:50:13Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 22:50:13 crc kubenswrapper[4791]: I1210 22:50:13.126226 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 22:50:13 crc kubenswrapper[4791]: I1210 22:50:13.126296 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 22:50:13 crc kubenswrapper[4791]: I1210 22:50:13.126320 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 22:50:13 crc kubenswrapper[4791]: I1210 22:50:13.126392 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 22:50:13 crc kubenswrapper[4791]: I1210 22:50:13.126416 4791 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T22:50:13Z","lastTransitionTime":"2025-12-10T22:50:13Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 22:50:13 crc kubenswrapper[4791]: I1210 22:50:13.229008 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 22:50:13 crc kubenswrapper[4791]: I1210 22:50:13.229068 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 22:50:13 crc kubenswrapper[4791]: I1210 22:50:13.229090 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 22:50:13 crc kubenswrapper[4791]: I1210 22:50:13.229122 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 22:50:13 crc kubenswrapper[4791]: I1210 22:50:13.229145 4791 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T22:50:13Z","lastTransitionTime":"2025-12-10T22:50:13Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 22:50:13 crc kubenswrapper[4791]: I1210 22:50:13.332547 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 22:50:13 crc kubenswrapper[4791]: I1210 22:50:13.332618 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 22:50:13 crc kubenswrapper[4791]: I1210 22:50:13.332641 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 22:50:13 crc kubenswrapper[4791]: I1210 22:50:13.332675 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 22:50:13 crc kubenswrapper[4791]: I1210 22:50:13.332701 4791 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T22:50:13Z","lastTransitionTime":"2025-12-10T22:50:13Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 22:50:13 crc kubenswrapper[4791]: I1210 22:50:13.435267 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 22:50:13 crc kubenswrapper[4791]: I1210 22:50:13.435324 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 22:50:13 crc kubenswrapper[4791]: I1210 22:50:13.435375 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 22:50:13 crc kubenswrapper[4791]: I1210 22:50:13.435399 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 22:50:13 crc kubenswrapper[4791]: I1210 22:50:13.435416 4791 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T22:50:13Z","lastTransitionTime":"2025-12-10T22:50:13Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 22:50:13 crc kubenswrapper[4791]: I1210 22:50:13.537791 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 22:50:13 crc kubenswrapper[4791]: I1210 22:50:13.537859 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 22:50:13 crc kubenswrapper[4791]: I1210 22:50:13.537878 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 22:50:13 crc kubenswrapper[4791]: I1210 22:50:13.537903 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 22:50:13 crc kubenswrapper[4791]: I1210 22:50:13.537921 4791 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T22:50:13Z","lastTransitionTime":"2025-12-10T22:50:13Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 22:50:13 crc kubenswrapper[4791]: I1210 22:50:13.640992 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 22:50:13 crc kubenswrapper[4791]: I1210 22:50:13.641049 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 22:50:13 crc kubenswrapper[4791]: I1210 22:50:13.641067 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 22:50:13 crc kubenswrapper[4791]: I1210 22:50:13.641092 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 22:50:13 crc kubenswrapper[4791]: I1210 22:50:13.641109 4791 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T22:50:13Z","lastTransitionTime":"2025-12-10T22:50:13Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 22:50:13 crc kubenswrapper[4791]: I1210 22:50:13.744190 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 22:50:13 crc kubenswrapper[4791]: I1210 22:50:13.744242 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 22:50:13 crc kubenswrapper[4791]: I1210 22:50:13.744261 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 22:50:13 crc kubenswrapper[4791]: I1210 22:50:13.744285 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 22:50:13 crc kubenswrapper[4791]: I1210 22:50:13.744304 4791 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T22:50:13Z","lastTransitionTime":"2025-12-10T22:50:13Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 22:50:13 crc kubenswrapper[4791]: I1210 22:50:13.847041 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 22:50:13 crc kubenswrapper[4791]: I1210 22:50:13.847133 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 22:50:13 crc kubenswrapper[4791]: I1210 22:50:13.847153 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 22:50:13 crc kubenswrapper[4791]: I1210 22:50:13.847177 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 22:50:13 crc kubenswrapper[4791]: I1210 22:50:13.847194 4791 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T22:50:13Z","lastTransitionTime":"2025-12-10T22:50:13Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 22:50:13 crc kubenswrapper[4791]: I1210 22:50:13.950578 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 22:50:13 crc kubenswrapper[4791]: I1210 22:50:13.950641 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 22:50:13 crc kubenswrapper[4791]: I1210 22:50:13.950664 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 22:50:13 crc kubenswrapper[4791]: I1210 22:50:13.950691 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 22:50:13 crc kubenswrapper[4791]: I1210 22:50:13.950711 4791 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T22:50:13Z","lastTransitionTime":"2025-12-10T22:50:13Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 22:50:14 crc kubenswrapper[4791]: I1210 22:50:14.053941 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 22:50:14 crc kubenswrapper[4791]: I1210 22:50:14.054004 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 22:50:14 crc kubenswrapper[4791]: I1210 22:50:14.054027 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 22:50:14 crc kubenswrapper[4791]: I1210 22:50:14.054057 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 22:50:14 crc kubenswrapper[4791]: I1210 22:50:14.054080 4791 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T22:50:14Z","lastTransitionTime":"2025-12-10T22:50:14Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 22:50:14 crc kubenswrapper[4791]: I1210 22:50:14.157423 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 22:50:14 crc kubenswrapper[4791]: I1210 22:50:14.157497 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 22:50:14 crc kubenswrapper[4791]: I1210 22:50:14.157521 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 22:50:14 crc kubenswrapper[4791]: I1210 22:50:14.157548 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 22:50:14 crc kubenswrapper[4791]: I1210 22:50:14.157564 4791 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T22:50:14Z","lastTransitionTime":"2025-12-10T22:50:14Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 22:50:14 crc kubenswrapper[4791]: I1210 22:50:14.260742 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 22:50:14 crc kubenswrapper[4791]: I1210 22:50:14.260806 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 22:50:14 crc kubenswrapper[4791]: I1210 22:50:14.260827 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 22:50:14 crc kubenswrapper[4791]: I1210 22:50:14.260855 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 22:50:14 crc kubenswrapper[4791]: I1210 22:50:14.260876 4791 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T22:50:14Z","lastTransitionTime":"2025-12-10T22:50:14Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 22:50:14 crc kubenswrapper[4791]: I1210 22:50:14.364474 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 22:50:14 crc kubenswrapper[4791]: I1210 22:50:14.364617 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 22:50:14 crc kubenswrapper[4791]: I1210 22:50:14.364648 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 22:50:14 crc kubenswrapper[4791]: I1210 22:50:14.364680 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 22:50:14 crc kubenswrapper[4791]: I1210 22:50:14.364703 4791 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T22:50:14Z","lastTransitionTime":"2025-12-10T22:50:14Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 22:50:14 crc kubenswrapper[4791]: I1210 22:50:14.467500 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 22:50:14 crc kubenswrapper[4791]: I1210 22:50:14.467591 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 22:50:14 crc kubenswrapper[4791]: I1210 22:50:14.467615 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 22:50:14 crc kubenswrapper[4791]: I1210 22:50:14.467651 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 22:50:14 crc kubenswrapper[4791]: I1210 22:50:14.467685 4791 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T22:50:14Z","lastTransitionTime":"2025-12-10T22:50:14Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 22:50:14 crc kubenswrapper[4791]: I1210 22:50:14.570686 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 22:50:14 crc kubenswrapper[4791]: I1210 22:50:14.570774 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 22:50:14 crc kubenswrapper[4791]: I1210 22:50:14.570801 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 22:50:14 crc kubenswrapper[4791]: I1210 22:50:14.570835 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 22:50:14 crc kubenswrapper[4791]: I1210 22:50:14.570857 4791 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T22:50:14Z","lastTransitionTime":"2025-12-10T22:50:14Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 22:50:14 crc kubenswrapper[4791]: I1210 22:50:14.673084 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 22:50:14 crc kubenswrapper[4791]: I1210 22:50:14.673126 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 22:50:14 crc kubenswrapper[4791]: I1210 22:50:14.673137 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 22:50:14 crc kubenswrapper[4791]: I1210 22:50:14.673156 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 22:50:14 crc kubenswrapper[4791]: I1210 22:50:14.673171 4791 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T22:50:14Z","lastTransitionTime":"2025-12-10T22:50:14Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 22:50:14 crc kubenswrapper[4791]: I1210 22:50:14.775682 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 22:50:14 crc kubenswrapper[4791]: I1210 22:50:14.775750 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 22:50:14 crc kubenswrapper[4791]: I1210 22:50:14.775766 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 22:50:14 crc kubenswrapper[4791]: I1210 22:50:14.775787 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 22:50:14 crc kubenswrapper[4791]: I1210 22:50:14.775805 4791 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T22:50:14Z","lastTransitionTime":"2025-12-10T22:50:14Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 22:50:14 crc kubenswrapper[4791]: I1210 22:50:14.879032 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 22:50:14 crc kubenswrapper[4791]: I1210 22:50:14.879071 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 22:50:14 crc kubenswrapper[4791]: I1210 22:50:14.879081 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 22:50:14 crc kubenswrapper[4791]: I1210 22:50:14.879096 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 22:50:14 crc kubenswrapper[4791]: I1210 22:50:14.879106 4791 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T22:50:14Z","lastTransitionTime":"2025-12-10T22:50:14Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 22:50:14 crc kubenswrapper[4791]: I1210 22:50:14.884300 4791 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 10 22:50:14 crc kubenswrapper[4791]: E1210 22:50:14.884617 4791 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 10 22:50:14 crc kubenswrapper[4791]: I1210 22:50:14.884471 4791 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 10 22:50:14 crc kubenswrapper[4791]: E1210 22:50:14.884863 4791 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 10 22:50:14 crc kubenswrapper[4791]: I1210 22:50:14.884402 4791 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-b6kf6" Dec 10 22:50:14 crc kubenswrapper[4791]: I1210 22:50:14.884484 4791 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 10 22:50:14 crc kubenswrapper[4791]: E1210 22:50:14.885120 4791 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-b6kf6" podUID="b4527e71-bb58-4a39-a99c-fe614aa4c5f9" Dec 10 22:50:14 crc kubenswrapper[4791]: E1210 22:50:14.885231 4791 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 10 22:50:14 crc kubenswrapper[4791]: I1210 22:50:14.983900 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 22:50:14 crc kubenswrapper[4791]: I1210 22:50:14.983999 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 22:50:14 crc kubenswrapper[4791]: I1210 22:50:14.984029 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 22:50:14 crc kubenswrapper[4791]: I1210 22:50:14.984063 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 22:50:14 crc kubenswrapper[4791]: I1210 22:50:14.984099 4791 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T22:50:14Z","lastTransitionTime":"2025-12-10T22:50:14Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 22:50:15 crc kubenswrapper[4791]: I1210 22:50:15.087248 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 22:50:15 crc kubenswrapper[4791]: I1210 22:50:15.087313 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 22:50:15 crc kubenswrapper[4791]: I1210 22:50:15.087330 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 22:50:15 crc kubenswrapper[4791]: I1210 22:50:15.087385 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 22:50:15 crc kubenswrapper[4791]: I1210 22:50:15.087402 4791 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T22:50:15Z","lastTransitionTime":"2025-12-10T22:50:15Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 22:50:15 crc kubenswrapper[4791]: I1210 22:50:15.190241 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 22:50:15 crc kubenswrapper[4791]: I1210 22:50:15.190298 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 22:50:15 crc kubenswrapper[4791]: I1210 22:50:15.190310 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 22:50:15 crc kubenswrapper[4791]: I1210 22:50:15.190328 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 22:50:15 crc kubenswrapper[4791]: I1210 22:50:15.190359 4791 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T22:50:15Z","lastTransitionTime":"2025-12-10T22:50:15Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 22:50:15 crc kubenswrapper[4791]: I1210 22:50:15.293318 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 22:50:15 crc kubenswrapper[4791]: I1210 22:50:15.293440 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 22:50:15 crc kubenswrapper[4791]: I1210 22:50:15.293457 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 22:50:15 crc kubenswrapper[4791]: I1210 22:50:15.293481 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 22:50:15 crc kubenswrapper[4791]: I1210 22:50:15.293498 4791 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T22:50:15Z","lastTransitionTime":"2025-12-10T22:50:15Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 22:50:15 crc kubenswrapper[4791]: I1210 22:50:15.397205 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 22:50:15 crc kubenswrapper[4791]: I1210 22:50:15.397278 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 22:50:15 crc kubenswrapper[4791]: I1210 22:50:15.397295 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 22:50:15 crc kubenswrapper[4791]: I1210 22:50:15.397320 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 22:50:15 crc kubenswrapper[4791]: I1210 22:50:15.397369 4791 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T22:50:15Z","lastTransitionTime":"2025-12-10T22:50:15Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 22:50:15 crc kubenswrapper[4791]: I1210 22:50:15.500091 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 22:50:15 crc kubenswrapper[4791]: I1210 22:50:15.500144 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 22:50:15 crc kubenswrapper[4791]: I1210 22:50:15.500156 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 22:50:15 crc kubenswrapper[4791]: I1210 22:50:15.500173 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 22:50:15 crc kubenswrapper[4791]: I1210 22:50:15.500185 4791 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T22:50:15Z","lastTransitionTime":"2025-12-10T22:50:15Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 22:50:15 crc kubenswrapper[4791]: I1210 22:50:15.603416 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 22:50:15 crc kubenswrapper[4791]: I1210 22:50:15.603492 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 22:50:15 crc kubenswrapper[4791]: I1210 22:50:15.603515 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 22:50:15 crc kubenswrapper[4791]: I1210 22:50:15.603547 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 22:50:15 crc kubenswrapper[4791]: I1210 22:50:15.603569 4791 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T22:50:15Z","lastTransitionTime":"2025-12-10T22:50:15Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 22:50:15 crc kubenswrapper[4791]: I1210 22:50:15.706702 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 22:50:15 crc kubenswrapper[4791]: I1210 22:50:15.706792 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 22:50:15 crc kubenswrapper[4791]: I1210 22:50:15.706813 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 22:50:15 crc kubenswrapper[4791]: I1210 22:50:15.706835 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 22:50:15 crc kubenswrapper[4791]: I1210 22:50:15.706854 4791 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T22:50:15Z","lastTransitionTime":"2025-12-10T22:50:15Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 22:50:15 crc kubenswrapper[4791]: I1210 22:50:15.810208 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 22:50:15 crc kubenswrapper[4791]: I1210 22:50:15.810261 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 22:50:15 crc kubenswrapper[4791]: I1210 22:50:15.810276 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 22:50:15 crc kubenswrapper[4791]: I1210 22:50:15.810300 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 22:50:15 crc kubenswrapper[4791]: I1210 22:50:15.810321 4791 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T22:50:15Z","lastTransitionTime":"2025-12-10T22:50:15Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 22:50:15 crc kubenswrapper[4791]: I1210 22:50:15.899332 4791 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-10T22:49:52Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-10T22:49:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-10T22:49:52Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://20651eb0c16e2c71222f56da1ff755cb0e7b084886acbbb24f0ccaf57041ed85\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T22:49:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T22:50:15Z is after 2025-08-24T17:21:41Z" Dec 10 22:50:15 crc kubenswrapper[4791]: I1210 22:50:15.913174 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 22:50:15 crc kubenswrapper[4791]: I1210 22:50:15.913217 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 22:50:15 crc kubenswrapper[4791]: I1210 22:50:15.913229 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 22:50:15 crc kubenswrapper[4791]: I1210 22:50:15.913252 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 22:50:15 crc kubenswrapper[4791]: I1210 22:50:15.913266 4791 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T22:50:15Z","lastTransitionTime":"2025-12-10T22:50:15Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 22:50:15 crc kubenswrapper[4791]: I1210 22:50:15.917716 4791 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-10T22:49:50Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-10T22:49:50Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T22:50:15Z is after 2025-08-24T17:21:41Z" Dec 10 22:50:15 crc kubenswrapper[4791]: I1210 22:50:15.938181 4791 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-10T22:49:50Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-10T22:49:50Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T22:50:15Z is after 2025-08-24T17:21:41Z" Dec 10 22:50:15 crc kubenswrapper[4791]: I1210 22:50:15.953319 4791 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-tchxk" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"c0ab74e0-42a9-4f42-ac80-16030cf4ffff\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T22:49:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T22:49:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T22:49:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T22:49:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://04e0fd1b53256b7a938e8ccab93e2a4eb73dd1a59e0b9799e09c55e511baf7fe\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T22:49:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zn2qt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-10T22:49:50Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-tchxk\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T22:50:15Z is after 2025-08-24T17:21:41Z" Dec 10 22:50:15 crc kubenswrapper[4791]: I1210 22:50:15.978608 4791 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-rw584" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ff9beb90-69b0-4732-bf37-0b81f58ecc98\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T22:49:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T22:49:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T22:49:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T22:49:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ddb9df501dc6e8e5561fbc2aefd431037a3f28f73bf8e82a6598d261a4224807\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T22:49:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fksqn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ce67e54b696bb53ffbbe48c6ce71476b635e0868993e9c427b22f904891433ee\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ce67e54b696bb53ffbbe48c6ce71476b635e0868993e9c427b22f904891433ee\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-10T22:49:52Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-10T22:49:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fksqn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://818d89bfffb8a110670152c8480684665ea8db578e9768f524fbbe6a92f27c22\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://818d89bfffb8a110670152c8480684665ea8db578e9768f524fbbe6a92f27c22\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-10T22:49:53Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-10T22:49:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fksqn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://9aa9c06e894bb2a2ed3738defaeff68a0a0458ead48177f36b90770944c211a0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://9aa9c06e894bb2a2ed3738defaeff68a0a0458ead48177f36b90770944c211a0\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-10T22:49:54Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-10T22:49:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fksqn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://dfd6d03336dcea7529419a30e8d70d84e2290a0fa5b0e46d09ad341c190638d6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://dfd6d03336dcea7529419a30e8d70d84e2290a0fa5b0e46d09ad341c190638d6\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-10T22:49:55Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-10T22:49:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fksqn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b86738401ece6fbc3083782720c66a515850ed78d4054f34fcabab7ff525b85f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b86738401ece6fbc3083782720c66a515850ed78d4054f34fcabab7ff525b85f\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-10T22:49:56Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-10T22:49:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fksqn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://18b326f572a899ddd10197a745bf9264d4ddd975690ad1cfb0ec576ccd5cf102\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://18b326f572a899ddd10197a745bf9264d4ddd975690ad1cfb0ec576ccd5cf102\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-10T22:49:57Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-10T22:49:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fksqn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-10T22:49:51Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-rw584\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T22:50:15Z is after 2025-08-24T17:21:41Z" Dec 10 22:50:16 crc kubenswrapper[4791]: I1210 22:50:16.005016 4791 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f8349fea-c49b-49a1-b5ee-032c5dffe021\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T22:49:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T22:49:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T22:49:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T22:49:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T22:49:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://acafe12186df99d1059a31260de0eafce35ff7771bcb8f5c6753b72665d6f60b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T22:49:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ea4c346507862a8c12756a3061d6e25e788a8d79f4621b62f81f4322185c82a8\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T22:49:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://2e6e324ec6f4f7bb69fdfab4ea3ac1aa55e480dd8e2a8984b0fe5b4c9f66636e\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T22:49:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://1b2e11ff2bf8789fa1d0d23ae648b3e6ef653106e5bc4bbb5dd6447f7cd6f05d\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d6aa7f8c5d908d90008f2df49860ac8a3480136da567e3d85f86135d779204f9\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-12-10T22:49:44Z\\\",\\\"message\\\":\\\"W1210 22:49:33.845733 1 cmd.go:257] Using insecure, self-signed certificates\\\\nI1210 22:49:33.846108 1 crypto.go:601] Generating new CA for check-endpoints-signer@1765406973 cert, and key in /tmp/serving-cert-2585701456/serving-signer.crt, /tmp/serving-cert-2585701456/serving-signer.key\\\\nI1210 22:49:34.529748 1 observer_polling.go:159] Starting file observer\\\\nW1210 22:49:34.532225 1 builder.go:272] unable to get owner reference (falling back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": dial tcp [::1]:6443: connect: connection refused\\\\nI1210 22:49:34.532599 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1210 22:49:34.533977 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-2585701456/tls.crt::/tmp/serving-cert-2585701456/tls.key\\\\\\\"\\\\nF1210 22:49:44.774710 1 cmd.go:182] error initializing delegating authentication: unable to load configmap based request-header-client-ca-file: Get \\\\\\\"https://localhost:6443/api/v1/namespaces/kube-system/configmaps/extension-apiserver-authentication\\\\\\\": net/http: TLS handshake timeout\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-10T22:49:33Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T22:49:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://0c4d036ab269f62561fe7ef1586012cb3a16e3583a08b9488a0616ec00c8e0ae\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T22:49:32Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c5856d6b26c19fda13b5c9ac2892a7e2f9c6f79feb8173ecfd9b33b673327d67\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c5856d6b26c19fda13b5c9ac2892a7e2f9c6f79feb8173ecfd9b33b673327d67\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-10T22:49:27Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-10T22:49:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-10T22:49:25Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T22:50:16Z is after 2025-08-24T17:21:41Z" Dec 10 22:50:16 crc kubenswrapper[4791]: I1210 22:50:16.015635 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 22:50:16 crc kubenswrapper[4791]: I1210 22:50:16.015797 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 22:50:16 crc kubenswrapper[4791]: I1210 22:50:16.015826 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 22:50:16 crc kubenswrapper[4791]: I1210 22:50:16.015859 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 22:50:16 crc kubenswrapper[4791]: I1210 22:50:16.015881 4791 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T22:50:16Z","lastTransitionTime":"2025-12-10T22:50:16Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 22:50:16 crc kubenswrapper[4791]: I1210 22:50:16.033830 4791 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"90250e13-6c63-443a-a5e5-6e4e341a2289\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T22:49:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T22:49:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T22:49:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T22:49:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T22:49:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ba54f95b1e00843b854db9a10fe6897592159ccacf30ed0f76002093a13eedd3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T22:49:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5aa8ffa518f3a68b60b275f8df4851b2564aa4984b61efdc9470b73072133e43\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T22:49:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://16f989503edc78d9451638e2cf6cf4b19a2cf2b8809b10b237541cb829303091\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T22:49:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://82556824d63ce6e46df14c45689ff1f81ebf729abe902061815c040b323898e9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T22:49:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://61db531940fa6f2effdae5a005b4c91bd2698bd5d1bff9588c171a65ad550eca\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T22:49:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0f65c864d885319ce2a08d61c30be24b3ffb7d23ea30c82ed7faf039cb2c006d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://0f65c864d885319ce2a08d61c30be24b3ffb7d23ea30c82ed7faf039cb2c006d\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-10T22:49:27Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-10T22:49:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e31a76d1d5b2812ba83e24e38268e25ec8bed98fc3da9adcd06bda5822b6ace1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e31a76d1d5b2812ba83e24e38268e25ec8bed98fc3da9adcd06bda5822b6ace1\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-10T22:49:30Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-10T22:49:29Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://b5b5eceae562f3986b543d41de1e977efb85a91d84c98e09159ca6d092eda691\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b5b5eceae562f3986b543d41de1e977efb85a91d84c98e09159ca6d092eda691\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-10T22:49:31Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-10T22:49:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-10T22:49:25Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T22:50:16Z is after 2025-08-24T17:21:41Z" Dec 10 22:50:16 crc kubenswrapper[4791]: I1210 22:50:16.051847 4791 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-10T22:49:50Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-10T22:49:50Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T22:50:16Z is after 2025-08-24T17:21:41Z" Dec 10 22:50:16 crc kubenswrapper[4791]: I1210 22:50:16.069269 4791 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-10T22:49:54Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-10T22:49:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-10T22:49:54Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://62bfaf5d7a6d1b452cdaf88086447d0c8cb7306e94434e46752be5475124d56a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T22:49:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T22:50:16Z is after 2025-08-24T17:21:41Z" Dec 10 22:50:16 crc kubenswrapper[4791]: I1210 22:50:16.087410 4791 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-b6kf6" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b4527e71-bb58-4a39-a99c-fe614aa4c5f9\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T22:50:04Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T22:50:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T22:50:04Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T22:50:04Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ngpr4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ngpr4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-10T22:50:04Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-b6kf6\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T22:50:16Z is after 2025-08-24T17:21:41Z" Dec 10 22:50:16 crc kubenswrapper[4791]: I1210 22:50:16.109887 4791 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-5rb5l" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ba35653c-6e06-4cee-a4d6-137764090d18\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T22:49:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T22:49:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T22:49:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T22:49:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://dbb5c875730a98ad2daf316cfafc08fcdce21ac16148154780ac9193ed405dab\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T22:49:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fvvrs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://185eb2ff44221a514878283ec873773a6c9a9073dc64e296f982e60dee1338e4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T22:49:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fvvrs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-10T22:49:51Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-5rb5l\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T22:50:16Z is after 2025-08-24T17:21:41Z" Dec 10 22:50:16 crc kubenswrapper[4791]: I1210 22:50:16.117807 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 22:50:16 crc kubenswrapper[4791]: I1210 22:50:16.117864 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 22:50:16 crc kubenswrapper[4791]: I1210 22:50:16.117885 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 22:50:16 crc kubenswrapper[4791]: I1210 22:50:16.117910 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 22:50:16 crc kubenswrapper[4791]: I1210 22:50:16.117928 4791 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T22:50:16Z","lastTransitionTime":"2025-12-10T22:50:16Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 22:50:16 crc kubenswrapper[4791]: I1210 22:50:16.128084 4791 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-s6rk8" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5c9d0978-a6f9-49c5-b377-6619b4e3ff52\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T22:50:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T22:50:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T22:50:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T22:50:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8e55baaf460c794ab196e237d51d12de059aa038617dc2d5366d49c529f5d5c7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T22:50:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7vkgm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://51d1b9b086c60bae9223cb51d5a998d02790a0d91b745aab10392c23e617eb9f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T22:50:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7vkgm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-10T22:50:03Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-s6rk8\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T22:50:16Z is after 2025-08-24T17:21:41Z" Dec 10 22:50:16 crc kubenswrapper[4791]: I1210 22:50:16.145279 4791 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"698bbe9a-fbe5-436d-a835-5e80df77ba24\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T22:49:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T22:49:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T22:49:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T22:49:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T22:49:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://224d73713312376a58ec6cc2bc8fdb6ffa44c4749ae0e7691d5fe27b9afca09f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T22:49:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://372eeaffd40992c6f06efc4df6dfcb44b31de64ee8256c21508cdd7c74ecb96e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T22:49:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://76b6013b062cac5b0b4d12a94f38ce7b43f9c7ea216063455d88a956ac04d26e\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T22:49:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://78e950e122e761d950da1a3c60c06476e53920e8795f6c61db22c01e22d5136e\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T22:49:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-10T22:49:25Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T22:50:16Z is after 2025-08-24T17:21:41Z" Dec 10 22:50:16 crc kubenswrapper[4791]: I1210 22:50:16.158695 4791 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-10T22:49:52Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-10T22:49:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-10T22:49:52Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3734b110802828632d44263ed0a43cabc4b710833f0be120d72db0940fe783e2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T22:49:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://794803ac9be3c52509e6dcb10854d8864e1a9e406762e57c7971b07c4da1220e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T22:49:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T22:50:16Z is after 2025-08-24T17:21:41Z" Dec 10 22:50:16 crc kubenswrapper[4791]: I1210 22:50:16.172836 4791 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-vczjq" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b9f00f25-3fa1-4b63-9710-fd352224b01b\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T22:49:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T22:49:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T22:49:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T22:49:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4698cf4c6b55b41e628448d37c2eb847c2777d0efbd93afcbf367d5e78623b7e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T22:49:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lrrpv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-10T22:49:50Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-vczjq\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T22:50:16Z is after 2025-08-24T17:21:41Z" Dec 10 22:50:16 crc kubenswrapper[4791]: I1210 22:50:16.190384 4791 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-4nwkq" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"672aa28c-8169-49ed-87b8-21187d13a80c\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T22:49:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T22:49:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T22:49:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T22:49:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a2f2a62315529e148a10a33ccb241e9417fbde0fe42afb61cb7ee2438c74636f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T22:49:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xl62j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-10T22:49:51Z\\\"}}\" for pod \"openshift-multus\"/\"multus-4nwkq\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T22:50:16Z is after 2025-08-24T17:21:41Z" Dec 10 22:50:16 crc kubenswrapper[4791]: I1210 22:50:16.216946 4791 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-zhq64" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3cd47739-0fa9-4321-aff1-220f8721a0b3\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T22:49:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T22:49:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T22:49:52Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T22:49:52Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://68795a29944e56870ffe94fa732e9dcf0bfb344c0d85d1e590f58a876129c61f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T22:49:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wvpw8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://bc721b530c08b433756e978ec0423cc18aabf6771b61eb96f903b4d57ea335ff\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T22:49:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wvpw8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://152a7618be8f38fadf4e564b3948980503cedcdb466c37344ccc20f4a3173a5f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T22:49:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wvpw8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3d18f9cb7207efb1e32de75a30d33e024f656a08f0503db95e5488bc5280971f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T22:49:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wvpw8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://937137ae9d7480cdc7c59bc4a1df0a4c10e72b3b63693af285278789953ed218\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T22:49:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wvpw8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://1a52d19b7671e8b143017c8a981a6f8d3ee3b84650e4b223b814a531ac76f364\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T22:49:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wvpw8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f78a47c773006a18dbc4b706edfc9d1d5d0b6ad11a023a49ea4e622eae77bba3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f78a47c773006a18dbc4b706edfc9d1d5d0b6ad11a023a49ea4e622eae77bba3\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-12-10T22:50:05Z\\\",\\\"message\\\":\\\"olicy controller, err: could not add Event Handler for anpInformer during admin network policy controller initialization, handler {0x1fcc6e0 0x1fcc3c0 0x1fcc360} was not added to shared informer because it has stopped already, failed to start node network controller: failed to start default node network controller: failed to set node crc annotations: Internal error occurred: failed calling webhook \\\\\\\"node.network-node-identity.openshift.io\\\\\\\": failed to call webhook: Post \\\\\\\"https://127.0.0.1:9743/node?timeout=10s\\\\\\\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T22:50:05Z is after 2025-08-24T17:21:41Z]\\\\nI1210 22:50:05.954954 6227 services_controller.go:473] Services do not match for network=default, existing lbs: []services.LB{services.LB{Name:\\\\\\\"Service_openshift-operator-lifecycle-manager/olm-operator-metrics_TCP_cluster\\\\\\\", UUID:\\\\\\\"63b1440a-0908-4cab-8799-012fa1cf0b07\\\\\\\", Protocol:\\\\\\\"tcp\\\\\\\", ExternalIDs:map[string]string{\\\\\\\"k8s.ovn.org/kind\\\\\\\":\\\\\\\"Service\\\\\\\", \\\\\\\"k8s.ovn.org/owner\\\\\\\":\\\\\\\"openshift-operator-lifecycle-manager/olm-operator-metrics\\\\\\\"}, Opts:services.LBOpts{Reject:false, EmptyLBEvents:false, AffinityTimeOut:0, SkipSNAT:false, Template:false, AddressFamily:\\\\\\\"\\\\\\\"}, Rules:[]services.LBRule{}, Templates:services.TemplateMap{}, Switches:[]string{}, Routers:[]string{}, Groups:[]string{\\\\\\\"clusterLBGroup\\\\\\\"}}}, bui\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-10T22:50:04Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 10s restarting failed container=ovnkube-controller pod=ovnkube-node-zhq64_openshift-ovn-kubernetes(3cd47739-0fa9-4321-aff1-220f8721a0b3)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wvpw8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://dd3f80e5713221f96ed50da7f2e7fd5b9fabfcd5c1656da8d8a86d818a5fdc3f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T22:49:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wvpw8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3afbd1ffcd2c4b40324aca6e52d6fc822db131c3f35a9af3a6f005c2cffdb7bf\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3afbd1ffcd2c4b40324aca6e52d6fc822db131c3f35a9af3a6f005c2cffdb7bf\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-10T22:49:53Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-10T22:49:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wvpw8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-10T22:49:52Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-zhq64\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T22:50:16Z is after 2025-08-24T17:21:41Z" Dec 10 22:50:16 crc kubenswrapper[4791]: I1210 22:50:16.220677 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 22:50:16 crc kubenswrapper[4791]: I1210 22:50:16.220725 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 22:50:16 crc kubenswrapper[4791]: I1210 22:50:16.220738 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 22:50:16 crc kubenswrapper[4791]: I1210 22:50:16.220756 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 22:50:16 crc kubenswrapper[4791]: I1210 22:50:16.220769 4791 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T22:50:16Z","lastTransitionTime":"2025-12-10T22:50:16Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 22:50:16 crc kubenswrapper[4791]: I1210 22:50:16.323051 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 22:50:16 crc kubenswrapper[4791]: I1210 22:50:16.323112 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 22:50:16 crc kubenswrapper[4791]: I1210 22:50:16.323129 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 22:50:16 crc kubenswrapper[4791]: I1210 22:50:16.323152 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 22:50:16 crc kubenswrapper[4791]: I1210 22:50:16.323171 4791 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T22:50:16Z","lastTransitionTime":"2025-12-10T22:50:16Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 22:50:16 crc kubenswrapper[4791]: I1210 22:50:16.426000 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 22:50:16 crc kubenswrapper[4791]: I1210 22:50:16.426062 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 22:50:16 crc kubenswrapper[4791]: I1210 22:50:16.426084 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 22:50:16 crc kubenswrapper[4791]: I1210 22:50:16.426108 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 22:50:16 crc kubenswrapper[4791]: I1210 22:50:16.426126 4791 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T22:50:16Z","lastTransitionTime":"2025-12-10T22:50:16Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 22:50:16 crc kubenswrapper[4791]: I1210 22:50:16.528818 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 22:50:16 crc kubenswrapper[4791]: I1210 22:50:16.528877 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 22:50:16 crc kubenswrapper[4791]: I1210 22:50:16.528897 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 22:50:16 crc kubenswrapper[4791]: I1210 22:50:16.528921 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 22:50:16 crc kubenswrapper[4791]: I1210 22:50:16.528938 4791 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T22:50:16Z","lastTransitionTime":"2025-12-10T22:50:16Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 22:50:16 crc kubenswrapper[4791]: I1210 22:50:16.631652 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 22:50:16 crc kubenswrapper[4791]: I1210 22:50:16.631700 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 22:50:16 crc kubenswrapper[4791]: I1210 22:50:16.631734 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 22:50:16 crc kubenswrapper[4791]: I1210 22:50:16.631753 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 22:50:16 crc kubenswrapper[4791]: I1210 22:50:16.631764 4791 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T22:50:16Z","lastTransitionTime":"2025-12-10T22:50:16Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 22:50:16 crc kubenswrapper[4791]: I1210 22:50:16.733609 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 22:50:16 crc kubenswrapper[4791]: I1210 22:50:16.733657 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 22:50:16 crc kubenswrapper[4791]: I1210 22:50:16.733667 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 22:50:16 crc kubenswrapper[4791]: I1210 22:50:16.733686 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 22:50:16 crc kubenswrapper[4791]: I1210 22:50:16.733700 4791 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T22:50:16Z","lastTransitionTime":"2025-12-10T22:50:16Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 22:50:16 crc kubenswrapper[4791]: I1210 22:50:16.836492 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 22:50:16 crc kubenswrapper[4791]: I1210 22:50:16.836529 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 22:50:16 crc kubenswrapper[4791]: I1210 22:50:16.836537 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 22:50:16 crc kubenswrapper[4791]: I1210 22:50:16.836552 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 22:50:16 crc kubenswrapper[4791]: I1210 22:50:16.836561 4791 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T22:50:16Z","lastTransitionTime":"2025-12-10T22:50:16Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 22:50:16 crc kubenswrapper[4791]: I1210 22:50:16.884723 4791 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-b6kf6" Dec 10 22:50:16 crc kubenswrapper[4791]: I1210 22:50:16.884764 4791 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 10 22:50:16 crc kubenswrapper[4791]: I1210 22:50:16.884838 4791 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 10 22:50:16 crc kubenswrapper[4791]: E1210 22:50:16.884891 4791 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-b6kf6" podUID="b4527e71-bb58-4a39-a99c-fe614aa4c5f9" Dec 10 22:50:16 crc kubenswrapper[4791]: I1210 22:50:16.884914 4791 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 10 22:50:16 crc kubenswrapper[4791]: E1210 22:50:16.885020 4791 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 10 22:50:16 crc kubenswrapper[4791]: E1210 22:50:16.885113 4791 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 10 22:50:16 crc kubenswrapper[4791]: E1210 22:50:16.885220 4791 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 10 22:50:16 crc kubenswrapper[4791]: I1210 22:50:16.939129 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 22:50:16 crc kubenswrapper[4791]: I1210 22:50:16.939171 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 22:50:16 crc kubenswrapper[4791]: I1210 22:50:16.939181 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 22:50:16 crc kubenswrapper[4791]: I1210 22:50:16.939200 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 22:50:16 crc kubenswrapper[4791]: I1210 22:50:16.939212 4791 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T22:50:16Z","lastTransitionTime":"2025-12-10T22:50:16Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 22:50:17 crc kubenswrapper[4791]: I1210 22:50:17.041897 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 22:50:17 crc kubenswrapper[4791]: I1210 22:50:17.041942 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 22:50:17 crc kubenswrapper[4791]: I1210 22:50:17.041956 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 22:50:17 crc kubenswrapper[4791]: I1210 22:50:17.041973 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 22:50:17 crc kubenswrapper[4791]: I1210 22:50:17.041986 4791 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T22:50:17Z","lastTransitionTime":"2025-12-10T22:50:17Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 22:50:17 crc kubenswrapper[4791]: I1210 22:50:17.144711 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 22:50:17 crc kubenswrapper[4791]: I1210 22:50:17.144805 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 22:50:17 crc kubenswrapper[4791]: I1210 22:50:17.144826 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 22:50:17 crc kubenswrapper[4791]: I1210 22:50:17.144851 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 22:50:17 crc kubenswrapper[4791]: I1210 22:50:17.144867 4791 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T22:50:17Z","lastTransitionTime":"2025-12-10T22:50:17Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 22:50:17 crc kubenswrapper[4791]: I1210 22:50:17.247745 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 22:50:17 crc kubenswrapper[4791]: I1210 22:50:17.247781 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 22:50:17 crc kubenswrapper[4791]: I1210 22:50:17.247789 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 22:50:17 crc kubenswrapper[4791]: I1210 22:50:17.247802 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 22:50:17 crc kubenswrapper[4791]: I1210 22:50:17.247810 4791 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T22:50:17Z","lastTransitionTime":"2025-12-10T22:50:17Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 22:50:17 crc kubenswrapper[4791]: I1210 22:50:17.351191 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 22:50:17 crc kubenswrapper[4791]: I1210 22:50:17.351254 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 22:50:17 crc kubenswrapper[4791]: I1210 22:50:17.351272 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 22:50:17 crc kubenswrapper[4791]: I1210 22:50:17.351300 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 22:50:17 crc kubenswrapper[4791]: I1210 22:50:17.351320 4791 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T22:50:17Z","lastTransitionTime":"2025-12-10T22:50:17Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 22:50:17 crc kubenswrapper[4791]: I1210 22:50:17.453971 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 22:50:17 crc kubenswrapper[4791]: I1210 22:50:17.454004 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 22:50:17 crc kubenswrapper[4791]: I1210 22:50:17.454014 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 22:50:17 crc kubenswrapper[4791]: I1210 22:50:17.454028 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 22:50:17 crc kubenswrapper[4791]: I1210 22:50:17.454040 4791 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T22:50:17Z","lastTransitionTime":"2025-12-10T22:50:17Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 22:50:17 crc kubenswrapper[4791]: I1210 22:50:17.556812 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 22:50:17 crc kubenswrapper[4791]: I1210 22:50:17.556860 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 22:50:17 crc kubenswrapper[4791]: I1210 22:50:17.556872 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 22:50:17 crc kubenswrapper[4791]: I1210 22:50:17.556889 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 22:50:17 crc kubenswrapper[4791]: I1210 22:50:17.556900 4791 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T22:50:17Z","lastTransitionTime":"2025-12-10T22:50:17Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 22:50:17 crc kubenswrapper[4791]: I1210 22:50:17.645997 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 22:50:17 crc kubenswrapper[4791]: I1210 22:50:17.646300 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 22:50:17 crc kubenswrapper[4791]: I1210 22:50:17.646315 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 22:50:17 crc kubenswrapper[4791]: I1210 22:50:17.646673 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 22:50:17 crc kubenswrapper[4791]: I1210 22:50:17.646752 4791 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T22:50:17Z","lastTransitionTime":"2025-12-10T22:50:17Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 22:50:17 crc kubenswrapper[4791]: E1210 22:50:17.665756 4791 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-12-10T22:50:17Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-10T22:50:17Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-10T22:50:17Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-10T22:50:17Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-10T22:50:17Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-10T22:50:17Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-10T22:50:17Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-10T22:50:17Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"4fab3af6-f657-4dfc-8ef2-12b7c978c94f\\\",\\\"systemUUID\\\":\\\"442c0b68-8dce-47e0-bd68-d5210d7e0493\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T22:50:17Z is after 2025-08-24T17:21:41Z" Dec 10 22:50:17 crc kubenswrapper[4791]: I1210 22:50:17.671302 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 22:50:17 crc kubenswrapper[4791]: I1210 22:50:17.671379 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 22:50:17 crc kubenswrapper[4791]: I1210 22:50:17.671400 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 22:50:17 crc kubenswrapper[4791]: I1210 22:50:17.671418 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 22:50:17 crc kubenswrapper[4791]: I1210 22:50:17.671430 4791 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T22:50:17Z","lastTransitionTime":"2025-12-10T22:50:17Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 22:50:17 crc kubenswrapper[4791]: E1210 22:50:17.689786 4791 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-12-10T22:50:17Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-10T22:50:17Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-10T22:50:17Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-10T22:50:17Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-10T22:50:17Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-10T22:50:17Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-10T22:50:17Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-10T22:50:17Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"4fab3af6-f657-4dfc-8ef2-12b7c978c94f\\\",\\\"systemUUID\\\":\\\"442c0b68-8dce-47e0-bd68-d5210d7e0493\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T22:50:17Z is after 2025-08-24T17:21:41Z" Dec 10 22:50:17 crc kubenswrapper[4791]: I1210 22:50:17.696741 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 22:50:17 crc kubenswrapper[4791]: I1210 22:50:17.696785 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 22:50:17 crc kubenswrapper[4791]: I1210 22:50:17.696797 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 22:50:17 crc kubenswrapper[4791]: I1210 22:50:17.696816 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 22:50:17 crc kubenswrapper[4791]: I1210 22:50:17.696828 4791 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T22:50:17Z","lastTransitionTime":"2025-12-10T22:50:17Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 22:50:17 crc kubenswrapper[4791]: E1210 22:50:17.708813 4791 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-12-10T22:50:17Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-10T22:50:17Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-10T22:50:17Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-10T22:50:17Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-10T22:50:17Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-10T22:50:17Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-10T22:50:17Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-10T22:50:17Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"4fab3af6-f657-4dfc-8ef2-12b7c978c94f\\\",\\\"systemUUID\\\":\\\"442c0b68-8dce-47e0-bd68-d5210d7e0493\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T22:50:17Z is after 2025-08-24T17:21:41Z" Dec 10 22:50:17 crc kubenswrapper[4791]: I1210 22:50:17.712942 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 22:50:17 crc kubenswrapper[4791]: I1210 22:50:17.713003 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 22:50:17 crc kubenswrapper[4791]: I1210 22:50:17.713028 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 22:50:17 crc kubenswrapper[4791]: I1210 22:50:17.713055 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 22:50:17 crc kubenswrapper[4791]: I1210 22:50:17.713072 4791 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T22:50:17Z","lastTransitionTime":"2025-12-10T22:50:17Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 22:50:17 crc kubenswrapper[4791]: E1210 22:50:17.729078 4791 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-12-10T22:50:17Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-10T22:50:17Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-10T22:50:17Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-10T22:50:17Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-10T22:50:17Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-10T22:50:17Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-10T22:50:17Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-10T22:50:17Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"4fab3af6-f657-4dfc-8ef2-12b7c978c94f\\\",\\\"systemUUID\\\":\\\"442c0b68-8dce-47e0-bd68-d5210d7e0493\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T22:50:17Z is after 2025-08-24T17:21:41Z" Dec 10 22:50:17 crc kubenswrapper[4791]: I1210 22:50:17.732907 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 22:50:17 crc kubenswrapper[4791]: I1210 22:50:17.733000 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 22:50:17 crc kubenswrapper[4791]: I1210 22:50:17.733019 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 22:50:17 crc kubenswrapper[4791]: I1210 22:50:17.733045 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 22:50:17 crc kubenswrapper[4791]: I1210 22:50:17.733063 4791 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T22:50:17Z","lastTransitionTime":"2025-12-10T22:50:17Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 22:50:17 crc kubenswrapper[4791]: E1210 22:50:17.745597 4791 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-12-10T22:50:17Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-10T22:50:17Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-10T22:50:17Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-10T22:50:17Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-10T22:50:17Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-10T22:50:17Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-10T22:50:17Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-10T22:50:17Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"4fab3af6-f657-4dfc-8ef2-12b7c978c94f\\\",\\\"systemUUID\\\":\\\"442c0b68-8dce-47e0-bd68-d5210d7e0493\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T22:50:17Z is after 2025-08-24T17:21:41Z" Dec 10 22:50:17 crc kubenswrapper[4791]: E1210 22:50:17.746285 4791 kubelet_node_status.go:572] "Unable to update node status" err="update node status exceeds retry count" Dec 10 22:50:17 crc kubenswrapper[4791]: I1210 22:50:17.748525 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 22:50:17 crc kubenswrapper[4791]: I1210 22:50:17.748578 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 22:50:17 crc kubenswrapper[4791]: I1210 22:50:17.748589 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 22:50:17 crc kubenswrapper[4791]: I1210 22:50:17.748608 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 22:50:17 crc kubenswrapper[4791]: I1210 22:50:17.748620 4791 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T22:50:17Z","lastTransitionTime":"2025-12-10T22:50:17Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 22:50:17 crc kubenswrapper[4791]: I1210 22:50:17.851390 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 22:50:17 crc kubenswrapper[4791]: I1210 22:50:17.851430 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 22:50:17 crc kubenswrapper[4791]: I1210 22:50:17.851442 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 22:50:17 crc kubenswrapper[4791]: I1210 22:50:17.851457 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 22:50:17 crc kubenswrapper[4791]: I1210 22:50:17.851468 4791 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T22:50:17Z","lastTransitionTime":"2025-12-10T22:50:17Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 22:50:17 crc kubenswrapper[4791]: I1210 22:50:17.953830 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 22:50:17 crc kubenswrapper[4791]: I1210 22:50:17.953877 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 22:50:17 crc kubenswrapper[4791]: I1210 22:50:17.953893 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 22:50:17 crc kubenswrapper[4791]: I1210 22:50:17.953915 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 22:50:17 crc kubenswrapper[4791]: I1210 22:50:17.953930 4791 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T22:50:17Z","lastTransitionTime":"2025-12-10T22:50:17Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 22:50:18 crc kubenswrapper[4791]: I1210 22:50:18.057872 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 22:50:18 crc kubenswrapper[4791]: I1210 22:50:18.057933 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 22:50:18 crc kubenswrapper[4791]: I1210 22:50:18.057955 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 22:50:18 crc kubenswrapper[4791]: I1210 22:50:18.057986 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 22:50:18 crc kubenswrapper[4791]: I1210 22:50:18.058048 4791 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T22:50:18Z","lastTransitionTime":"2025-12-10T22:50:18Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 22:50:18 crc kubenswrapper[4791]: I1210 22:50:18.161326 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 22:50:18 crc kubenswrapper[4791]: I1210 22:50:18.161621 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 22:50:18 crc kubenswrapper[4791]: I1210 22:50:18.161639 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 22:50:18 crc kubenswrapper[4791]: I1210 22:50:18.161665 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 22:50:18 crc kubenswrapper[4791]: I1210 22:50:18.161678 4791 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T22:50:18Z","lastTransitionTime":"2025-12-10T22:50:18Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 22:50:18 crc kubenswrapper[4791]: I1210 22:50:18.264203 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 22:50:18 crc kubenswrapper[4791]: I1210 22:50:18.264273 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 22:50:18 crc kubenswrapper[4791]: I1210 22:50:18.264297 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 22:50:18 crc kubenswrapper[4791]: I1210 22:50:18.264328 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 22:50:18 crc kubenswrapper[4791]: I1210 22:50:18.264397 4791 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T22:50:18Z","lastTransitionTime":"2025-12-10T22:50:18Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 22:50:18 crc kubenswrapper[4791]: I1210 22:50:18.367493 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 22:50:18 crc kubenswrapper[4791]: I1210 22:50:18.367561 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 22:50:18 crc kubenswrapper[4791]: I1210 22:50:18.367576 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 22:50:18 crc kubenswrapper[4791]: I1210 22:50:18.367599 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 22:50:18 crc kubenswrapper[4791]: I1210 22:50:18.367615 4791 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T22:50:18Z","lastTransitionTime":"2025-12-10T22:50:18Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 22:50:18 crc kubenswrapper[4791]: I1210 22:50:18.471002 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 22:50:18 crc kubenswrapper[4791]: I1210 22:50:18.471098 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 22:50:18 crc kubenswrapper[4791]: I1210 22:50:18.471117 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 22:50:18 crc kubenswrapper[4791]: I1210 22:50:18.471143 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 22:50:18 crc kubenswrapper[4791]: I1210 22:50:18.471161 4791 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T22:50:18Z","lastTransitionTime":"2025-12-10T22:50:18Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 22:50:18 crc kubenswrapper[4791]: I1210 22:50:18.573716 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 22:50:18 crc kubenswrapper[4791]: I1210 22:50:18.573774 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 22:50:18 crc kubenswrapper[4791]: I1210 22:50:18.573791 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 22:50:18 crc kubenswrapper[4791]: I1210 22:50:18.573815 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 22:50:18 crc kubenswrapper[4791]: I1210 22:50:18.573836 4791 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T22:50:18Z","lastTransitionTime":"2025-12-10T22:50:18Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 22:50:18 crc kubenswrapper[4791]: I1210 22:50:18.681888 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 22:50:18 crc kubenswrapper[4791]: I1210 22:50:18.681953 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 22:50:18 crc kubenswrapper[4791]: I1210 22:50:18.681980 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 22:50:18 crc kubenswrapper[4791]: I1210 22:50:18.682006 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 22:50:18 crc kubenswrapper[4791]: I1210 22:50:18.682024 4791 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T22:50:18Z","lastTransitionTime":"2025-12-10T22:50:18Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 22:50:18 crc kubenswrapper[4791]: I1210 22:50:18.785922 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 22:50:18 crc kubenswrapper[4791]: I1210 22:50:18.785991 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 22:50:18 crc kubenswrapper[4791]: I1210 22:50:18.786030 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 22:50:18 crc kubenswrapper[4791]: I1210 22:50:18.786059 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 22:50:18 crc kubenswrapper[4791]: I1210 22:50:18.786080 4791 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T22:50:18Z","lastTransitionTime":"2025-12-10T22:50:18Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 22:50:18 crc kubenswrapper[4791]: I1210 22:50:18.884315 4791 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 10 22:50:18 crc kubenswrapper[4791]: I1210 22:50:18.884505 4791 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-b6kf6" Dec 10 22:50:18 crc kubenswrapper[4791]: I1210 22:50:18.884316 4791 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 10 22:50:18 crc kubenswrapper[4791]: I1210 22:50:18.884574 4791 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 10 22:50:18 crc kubenswrapper[4791]: E1210 22:50:18.884690 4791 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 10 22:50:18 crc kubenswrapper[4791]: E1210 22:50:18.884794 4791 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 10 22:50:18 crc kubenswrapper[4791]: E1210 22:50:18.884960 4791 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-b6kf6" podUID="b4527e71-bb58-4a39-a99c-fe614aa4c5f9" Dec 10 22:50:18 crc kubenswrapper[4791]: E1210 22:50:18.885102 4791 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 10 22:50:18 crc kubenswrapper[4791]: I1210 22:50:18.889595 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 22:50:18 crc kubenswrapper[4791]: I1210 22:50:18.889655 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 22:50:18 crc kubenswrapper[4791]: I1210 22:50:18.889677 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 22:50:18 crc kubenswrapper[4791]: I1210 22:50:18.889705 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 22:50:18 crc kubenswrapper[4791]: I1210 22:50:18.889726 4791 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T22:50:18Z","lastTransitionTime":"2025-12-10T22:50:18Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 22:50:18 crc kubenswrapper[4791]: I1210 22:50:18.992966 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 22:50:18 crc kubenswrapper[4791]: I1210 22:50:18.993038 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 22:50:18 crc kubenswrapper[4791]: I1210 22:50:18.993061 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 22:50:18 crc kubenswrapper[4791]: I1210 22:50:18.993093 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 22:50:18 crc kubenswrapper[4791]: I1210 22:50:18.993117 4791 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T22:50:18Z","lastTransitionTime":"2025-12-10T22:50:18Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 22:50:19 crc kubenswrapper[4791]: I1210 22:50:19.096378 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 22:50:19 crc kubenswrapper[4791]: I1210 22:50:19.096634 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 22:50:19 crc kubenswrapper[4791]: I1210 22:50:19.096647 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 22:50:19 crc kubenswrapper[4791]: I1210 22:50:19.096666 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 22:50:19 crc kubenswrapper[4791]: I1210 22:50:19.096678 4791 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T22:50:19Z","lastTransitionTime":"2025-12-10T22:50:19Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 22:50:19 crc kubenswrapper[4791]: I1210 22:50:19.200320 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 22:50:19 crc kubenswrapper[4791]: I1210 22:50:19.200401 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 22:50:19 crc kubenswrapper[4791]: I1210 22:50:19.200415 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 22:50:19 crc kubenswrapper[4791]: I1210 22:50:19.200442 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 22:50:19 crc kubenswrapper[4791]: I1210 22:50:19.200457 4791 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T22:50:19Z","lastTransitionTime":"2025-12-10T22:50:19Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 22:50:19 crc kubenswrapper[4791]: I1210 22:50:19.302997 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 22:50:19 crc kubenswrapper[4791]: I1210 22:50:19.303077 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 22:50:19 crc kubenswrapper[4791]: I1210 22:50:19.303100 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 22:50:19 crc kubenswrapper[4791]: I1210 22:50:19.303125 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 22:50:19 crc kubenswrapper[4791]: I1210 22:50:19.303143 4791 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T22:50:19Z","lastTransitionTime":"2025-12-10T22:50:19Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 22:50:19 crc kubenswrapper[4791]: I1210 22:50:19.405334 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 22:50:19 crc kubenswrapper[4791]: I1210 22:50:19.405382 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 22:50:19 crc kubenswrapper[4791]: I1210 22:50:19.405391 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 22:50:19 crc kubenswrapper[4791]: I1210 22:50:19.405406 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 22:50:19 crc kubenswrapper[4791]: I1210 22:50:19.405415 4791 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T22:50:19Z","lastTransitionTime":"2025-12-10T22:50:19Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 22:50:19 crc kubenswrapper[4791]: I1210 22:50:19.510266 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 22:50:19 crc kubenswrapper[4791]: I1210 22:50:19.510304 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 22:50:19 crc kubenswrapper[4791]: I1210 22:50:19.510315 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 22:50:19 crc kubenswrapper[4791]: I1210 22:50:19.510331 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 22:50:19 crc kubenswrapper[4791]: I1210 22:50:19.510360 4791 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T22:50:19Z","lastTransitionTime":"2025-12-10T22:50:19Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 22:50:19 crc kubenswrapper[4791]: I1210 22:50:19.613541 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 22:50:19 crc kubenswrapper[4791]: I1210 22:50:19.613593 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 22:50:19 crc kubenswrapper[4791]: I1210 22:50:19.613609 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 22:50:19 crc kubenswrapper[4791]: I1210 22:50:19.613632 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 22:50:19 crc kubenswrapper[4791]: I1210 22:50:19.613650 4791 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T22:50:19Z","lastTransitionTime":"2025-12-10T22:50:19Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 22:50:19 crc kubenswrapper[4791]: I1210 22:50:19.716987 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 22:50:19 crc kubenswrapper[4791]: I1210 22:50:19.717049 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 22:50:19 crc kubenswrapper[4791]: I1210 22:50:19.717059 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 22:50:19 crc kubenswrapper[4791]: I1210 22:50:19.717095 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 22:50:19 crc kubenswrapper[4791]: I1210 22:50:19.717106 4791 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T22:50:19Z","lastTransitionTime":"2025-12-10T22:50:19Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 22:50:19 crc kubenswrapper[4791]: I1210 22:50:19.819984 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 22:50:19 crc kubenswrapper[4791]: I1210 22:50:19.820053 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 22:50:19 crc kubenswrapper[4791]: I1210 22:50:19.820066 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 22:50:19 crc kubenswrapper[4791]: I1210 22:50:19.820084 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 22:50:19 crc kubenswrapper[4791]: I1210 22:50:19.820098 4791 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T22:50:19Z","lastTransitionTime":"2025-12-10T22:50:19Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 22:50:19 crc kubenswrapper[4791]: I1210 22:50:19.922841 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 22:50:19 crc kubenswrapper[4791]: I1210 22:50:19.922905 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 22:50:19 crc kubenswrapper[4791]: I1210 22:50:19.922920 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 22:50:19 crc kubenswrapper[4791]: I1210 22:50:19.922943 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 22:50:19 crc kubenswrapper[4791]: I1210 22:50:19.922961 4791 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T22:50:19Z","lastTransitionTime":"2025-12-10T22:50:19Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 22:50:20 crc kubenswrapper[4791]: I1210 22:50:20.026353 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 22:50:20 crc kubenswrapper[4791]: I1210 22:50:20.026402 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 22:50:20 crc kubenswrapper[4791]: I1210 22:50:20.026413 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 22:50:20 crc kubenswrapper[4791]: I1210 22:50:20.026431 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 22:50:20 crc kubenswrapper[4791]: I1210 22:50:20.026444 4791 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T22:50:20Z","lastTransitionTime":"2025-12-10T22:50:20Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 22:50:20 crc kubenswrapper[4791]: I1210 22:50:20.129041 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 22:50:20 crc kubenswrapper[4791]: I1210 22:50:20.129090 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 22:50:20 crc kubenswrapper[4791]: I1210 22:50:20.129102 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 22:50:20 crc kubenswrapper[4791]: I1210 22:50:20.129119 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 22:50:20 crc kubenswrapper[4791]: I1210 22:50:20.129131 4791 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T22:50:20Z","lastTransitionTime":"2025-12-10T22:50:20Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 22:50:20 crc kubenswrapper[4791]: I1210 22:50:20.231826 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 22:50:20 crc kubenswrapper[4791]: I1210 22:50:20.231893 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 22:50:20 crc kubenswrapper[4791]: I1210 22:50:20.231912 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 22:50:20 crc kubenswrapper[4791]: I1210 22:50:20.231941 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 22:50:20 crc kubenswrapper[4791]: I1210 22:50:20.231958 4791 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T22:50:20Z","lastTransitionTime":"2025-12-10T22:50:20Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 22:50:20 crc kubenswrapper[4791]: I1210 22:50:20.335193 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 22:50:20 crc kubenswrapper[4791]: I1210 22:50:20.335278 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 22:50:20 crc kubenswrapper[4791]: I1210 22:50:20.335291 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 22:50:20 crc kubenswrapper[4791]: I1210 22:50:20.335310 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 22:50:20 crc kubenswrapper[4791]: I1210 22:50:20.335325 4791 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T22:50:20Z","lastTransitionTime":"2025-12-10T22:50:20Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 22:50:20 crc kubenswrapper[4791]: I1210 22:50:20.437776 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 22:50:20 crc kubenswrapper[4791]: I1210 22:50:20.437834 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 22:50:20 crc kubenswrapper[4791]: I1210 22:50:20.437852 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 22:50:20 crc kubenswrapper[4791]: I1210 22:50:20.437874 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 22:50:20 crc kubenswrapper[4791]: I1210 22:50:20.437890 4791 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T22:50:20Z","lastTransitionTime":"2025-12-10T22:50:20Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 22:50:20 crc kubenswrapper[4791]: I1210 22:50:20.540424 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 22:50:20 crc kubenswrapper[4791]: I1210 22:50:20.540486 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 22:50:20 crc kubenswrapper[4791]: I1210 22:50:20.540508 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 22:50:20 crc kubenswrapper[4791]: I1210 22:50:20.540532 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 22:50:20 crc kubenswrapper[4791]: I1210 22:50:20.540549 4791 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T22:50:20Z","lastTransitionTime":"2025-12-10T22:50:20Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 22:50:20 crc kubenswrapper[4791]: I1210 22:50:20.615279 4791 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" Dec 10 22:50:20 crc kubenswrapper[4791]: I1210 22:50:20.627837 4791 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-kube-scheduler/openshift-kube-scheduler-crc"] Dec 10 22:50:20 crc kubenswrapper[4791]: I1210 22:50:20.630651 4791 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-4nwkq" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"672aa28c-8169-49ed-87b8-21187d13a80c\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T22:49:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T22:49:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T22:49:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T22:49:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a2f2a62315529e148a10a33ccb241e9417fbde0fe42afb61cb7ee2438c74636f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T22:49:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xl62j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-10T22:49:51Z\\\"}}\" for pod \"openshift-multus\"/\"multus-4nwkq\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T22:50:20Z is after 2025-08-24T17:21:41Z" Dec 10 22:50:20 crc kubenswrapper[4791]: I1210 22:50:20.642212 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 22:50:20 crc kubenswrapper[4791]: I1210 22:50:20.642257 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 22:50:20 crc kubenswrapper[4791]: I1210 22:50:20.642272 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 22:50:20 crc kubenswrapper[4791]: I1210 22:50:20.642289 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 22:50:20 crc kubenswrapper[4791]: I1210 22:50:20.642300 4791 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T22:50:20Z","lastTransitionTime":"2025-12-10T22:50:20Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 22:50:20 crc kubenswrapper[4791]: I1210 22:50:20.651729 4791 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-zhq64" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3cd47739-0fa9-4321-aff1-220f8721a0b3\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T22:49:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T22:49:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T22:49:52Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T22:49:52Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://68795a29944e56870ffe94fa732e9dcf0bfb344c0d85d1e590f58a876129c61f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T22:49:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wvpw8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://bc721b530c08b433756e978ec0423cc18aabf6771b61eb96f903b4d57ea335ff\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T22:49:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wvpw8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://152a7618be8f38fadf4e564b3948980503cedcdb466c37344ccc20f4a3173a5f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T22:49:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wvpw8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3d18f9cb7207efb1e32de75a30d33e024f656a08f0503db95e5488bc5280971f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T22:49:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wvpw8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://937137ae9d7480cdc7c59bc4a1df0a4c10e72b3b63693af285278789953ed218\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T22:49:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wvpw8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://1a52d19b7671e8b143017c8a981a6f8d3ee3b84650e4b223b814a531ac76f364\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T22:49:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wvpw8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f78a47c773006a18dbc4b706edfc9d1d5d0b6ad11a023a49ea4e622eae77bba3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f78a47c773006a18dbc4b706edfc9d1d5d0b6ad11a023a49ea4e622eae77bba3\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-12-10T22:50:05Z\\\",\\\"message\\\":\\\"olicy controller, err: could not add Event Handler for anpInformer during admin network policy controller initialization, handler {0x1fcc6e0 0x1fcc3c0 0x1fcc360} was not added to shared informer because it has stopped already, failed to start node network controller: failed to start default node network controller: failed to set node crc annotations: Internal error occurred: failed calling webhook \\\\\\\"node.network-node-identity.openshift.io\\\\\\\": failed to call webhook: Post \\\\\\\"https://127.0.0.1:9743/node?timeout=10s\\\\\\\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T22:50:05Z is after 2025-08-24T17:21:41Z]\\\\nI1210 22:50:05.954954 6227 services_controller.go:473] Services do not match for network=default, existing lbs: []services.LB{services.LB{Name:\\\\\\\"Service_openshift-operator-lifecycle-manager/olm-operator-metrics_TCP_cluster\\\\\\\", UUID:\\\\\\\"63b1440a-0908-4cab-8799-012fa1cf0b07\\\\\\\", Protocol:\\\\\\\"tcp\\\\\\\", ExternalIDs:map[string]string{\\\\\\\"k8s.ovn.org/kind\\\\\\\":\\\\\\\"Service\\\\\\\", \\\\\\\"k8s.ovn.org/owner\\\\\\\":\\\\\\\"openshift-operator-lifecycle-manager/olm-operator-metrics\\\\\\\"}, Opts:services.LBOpts{Reject:false, EmptyLBEvents:false, AffinityTimeOut:0, SkipSNAT:false, Template:false, AddressFamily:\\\\\\\"\\\\\\\"}, Rules:[]services.LBRule{}, Templates:services.TemplateMap{}, Switches:[]string{}, Routers:[]string{}, Groups:[]string{\\\\\\\"clusterLBGroup\\\\\\\"}}}, bui\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-10T22:50:04Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 10s restarting failed container=ovnkube-controller pod=ovnkube-node-zhq64_openshift-ovn-kubernetes(3cd47739-0fa9-4321-aff1-220f8721a0b3)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wvpw8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://dd3f80e5713221f96ed50da7f2e7fd5b9fabfcd5c1656da8d8a86d818a5fdc3f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T22:49:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wvpw8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3afbd1ffcd2c4b40324aca6e52d6fc822db131c3f35a9af3a6f005c2cffdb7bf\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3afbd1ffcd2c4b40324aca6e52d6fc822db131c3f35a9af3a6f005c2cffdb7bf\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-10T22:49:53Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-10T22:49:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wvpw8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-10T22:49:52Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-zhq64\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T22:50:20Z is after 2025-08-24T17:21:41Z" Dec 10 22:50:20 crc kubenswrapper[4791]: I1210 22:50:20.665387 4791 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"698bbe9a-fbe5-436d-a835-5e80df77ba24\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T22:49:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T22:49:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T22:49:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T22:49:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T22:49:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://224d73713312376a58ec6cc2bc8fdb6ffa44c4749ae0e7691d5fe27b9afca09f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T22:49:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://372eeaffd40992c6f06efc4df6dfcb44b31de64ee8256c21508cdd7c74ecb96e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T22:49:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://76b6013b062cac5b0b4d12a94f38ce7b43f9c7ea216063455d88a956ac04d26e\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T22:49:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://78e950e122e761d950da1a3c60c06476e53920e8795f6c61db22c01e22d5136e\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T22:49:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-10T22:49:25Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T22:50:20Z is after 2025-08-24T17:21:41Z" Dec 10 22:50:20 crc kubenswrapper[4791]: I1210 22:50:20.682324 4791 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-10T22:49:52Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-10T22:49:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-10T22:49:52Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3734b110802828632d44263ed0a43cabc4b710833f0be120d72db0940fe783e2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T22:49:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://794803ac9be3c52509e6dcb10854d8864e1a9e406762e57c7971b07c4da1220e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T22:49:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T22:50:20Z is after 2025-08-24T17:21:41Z" Dec 10 22:50:20 crc kubenswrapper[4791]: I1210 22:50:20.693588 4791 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-vczjq" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b9f00f25-3fa1-4b63-9710-fd352224b01b\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T22:49:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T22:49:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T22:49:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T22:49:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4698cf4c6b55b41e628448d37c2eb847c2777d0efbd93afcbf367d5e78623b7e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T22:49:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lrrpv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-10T22:49:50Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-vczjq\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T22:50:20Z is after 2025-08-24T17:21:41Z" Dec 10 22:50:20 crc kubenswrapper[4791]: I1210 22:50:20.707473 4791 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-rw584" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ff9beb90-69b0-4732-bf37-0b81f58ecc98\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T22:49:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T22:49:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T22:49:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T22:49:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ddb9df501dc6e8e5561fbc2aefd431037a3f28f73bf8e82a6598d261a4224807\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T22:49:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fksqn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ce67e54b696bb53ffbbe48c6ce71476b635e0868993e9c427b22f904891433ee\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ce67e54b696bb53ffbbe48c6ce71476b635e0868993e9c427b22f904891433ee\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-10T22:49:52Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-10T22:49:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fksqn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://818d89bfffb8a110670152c8480684665ea8db578e9768f524fbbe6a92f27c22\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://818d89bfffb8a110670152c8480684665ea8db578e9768f524fbbe6a92f27c22\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-10T22:49:53Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-10T22:49:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fksqn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://9aa9c06e894bb2a2ed3738defaeff68a0a0458ead48177f36b90770944c211a0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://9aa9c06e894bb2a2ed3738defaeff68a0a0458ead48177f36b90770944c211a0\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-10T22:49:54Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-10T22:49:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fksqn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://dfd6d03336dcea7529419a30e8d70d84e2290a0fa5b0e46d09ad341c190638d6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://dfd6d03336dcea7529419a30e8d70d84e2290a0fa5b0e46d09ad341c190638d6\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-10T22:49:55Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-10T22:49:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fksqn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b86738401ece6fbc3083782720c66a515850ed78d4054f34fcabab7ff525b85f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b86738401ece6fbc3083782720c66a515850ed78d4054f34fcabab7ff525b85f\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-10T22:49:56Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-10T22:49:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fksqn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://18b326f572a899ddd10197a745bf9264d4ddd975690ad1cfb0ec576ccd5cf102\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://18b326f572a899ddd10197a745bf9264d4ddd975690ad1cfb0ec576ccd5cf102\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-10T22:49:57Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-10T22:49:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fksqn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-10T22:49:51Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-rw584\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T22:50:20Z is after 2025-08-24T17:21:41Z" Dec 10 22:50:20 crc kubenswrapper[4791]: I1210 22:50:20.717082 4791 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/b4527e71-bb58-4a39-a99c-fe614aa4c5f9-metrics-certs\") pod \"network-metrics-daemon-b6kf6\" (UID: \"b4527e71-bb58-4a39-a99c-fe614aa4c5f9\") " pod="openshift-multus/network-metrics-daemon-b6kf6" Dec 10 22:50:20 crc kubenswrapper[4791]: E1210 22:50:20.717249 4791 secret.go:188] Couldn't get secret openshift-multus/metrics-daemon-secret: object "openshift-multus"/"metrics-daemon-secret" not registered Dec 10 22:50:20 crc kubenswrapper[4791]: E1210 22:50:20.717312 4791 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/b4527e71-bb58-4a39-a99c-fe614aa4c5f9-metrics-certs podName:b4527e71-bb58-4a39-a99c-fe614aa4c5f9 nodeName:}" failed. No retries permitted until 2025-12-10 22:50:36.717294832 +0000 UTC m=+71.146912445 (durationBeforeRetry 16s). Error: MountVolume.SetUp failed for volume "metrics-certs" (UniqueName: "kubernetes.io/secret/b4527e71-bb58-4a39-a99c-fe614aa4c5f9-metrics-certs") pod "network-metrics-daemon-b6kf6" (UID: "b4527e71-bb58-4a39-a99c-fe614aa4c5f9") : object "openshift-multus"/"metrics-daemon-secret" not registered Dec 10 22:50:20 crc kubenswrapper[4791]: I1210 22:50:20.721931 4791 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-10T22:49:52Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-10T22:49:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-10T22:49:52Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://20651eb0c16e2c71222f56da1ff755cb0e7b084886acbbb24f0ccaf57041ed85\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T22:49:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T22:50:20Z is after 2025-08-24T17:21:41Z" Dec 10 22:50:20 crc kubenswrapper[4791]: I1210 22:50:20.735768 4791 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-10T22:49:50Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-10T22:49:50Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T22:50:20Z is after 2025-08-24T17:21:41Z" Dec 10 22:50:20 crc kubenswrapper[4791]: I1210 22:50:20.744957 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 22:50:20 crc kubenswrapper[4791]: I1210 22:50:20.744997 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 22:50:20 crc kubenswrapper[4791]: I1210 22:50:20.745008 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 22:50:20 crc kubenswrapper[4791]: I1210 22:50:20.745025 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 22:50:20 crc kubenswrapper[4791]: I1210 22:50:20.745036 4791 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T22:50:20Z","lastTransitionTime":"2025-12-10T22:50:20Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 22:50:20 crc kubenswrapper[4791]: I1210 22:50:20.748830 4791 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-10T22:49:50Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-10T22:49:50Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T22:50:20Z is after 2025-08-24T17:21:41Z" Dec 10 22:50:20 crc kubenswrapper[4791]: I1210 22:50:20.759018 4791 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-tchxk" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"c0ab74e0-42a9-4f42-ac80-16030cf4ffff\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T22:49:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T22:49:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T22:49:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T22:49:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://04e0fd1b53256b7a938e8ccab93e2a4eb73dd1a59e0b9799e09c55e511baf7fe\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T22:49:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zn2qt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-10T22:49:50Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-tchxk\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T22:50:20Z is after 2025-08-24T17:21:41Z" Dec 10 22:50:20 crc kubenswrapper[4791]: I1210 22:50:20.771977 4791 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-10T22:49:54Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-10T22:49:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-10T22:49:54Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://62bfaf5d7a6d1b452cdaf88086447d0c8cb7306e94434e46752be5475124d56a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T22:49:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T22:50:20Z is after 2025-08-24T17:21:41Z" Dec 10 22:50:20 crc kubenswrapper[4791]: I1210 22:50:20.785124 4791 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-b6kf6" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b4527e71-bb58-4a39-a99c-fe614aa4c5f9\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T22:50:04Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T22:50:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T22:50:04Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T22:50:04Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ngpr4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ngpr4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-10T22:50:04Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-b6kf6\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T22:50:20Z is after 2025-08-24T17:21:41Z" Dec 10 22:50:20 crc kubenswrapper[4791]: I1210 22:50:20.800196 4791 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f8349fea-c49b-49a1-b5ee-032c5dffe021\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T22:49:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T22:49:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T22:49:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T22:49:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T22:49:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://acafe12186df99d1059a31260de0eafce35ff7771bcb8f5c6753b72665d6f60b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T22:49:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ea4c346507862a8c12756a3061d6e25e788a8d79f4621b62f81f4322185c82a8\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T22:49:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://2e6e324ec6f4f7bb69fdfab4ea3ac1aa55e480dd8e2a8984b0fe5b4c9f66636e\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T22:49:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://1b2e11ff2bf8789fa1d0d23ae648b3e6ef653106e5bc4bbb5dd6447f7cd6f05d\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d6aa7f8c5d908d90008f2df49860ac8a3480136da567e3d85f86135d779204f9\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-12-10T22:49:44Z\\\",\\\"message\\\":\\\"W1210 22:49:33.845733 1 cmd.go:257] Using insecure, self-signed certificates\\\\nI1210 22:49:33.846108 1 crypto.go:601] Generating new CA for check-endpoints-signer@1765406973 cert, and key in /tmp/serving-cert-2585701456/serving-signer.crt, /tmp/serving-cert-2585701456/serving-signer.key\\\\nI1210 22:49:34.529748 1 observer_polling.go:159] Starting file observer\\\\nW1210 22:49:34.532225 1 builder.go:272] unable to get owner reference (falling back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": dial tcp [::1]:6443: connect: connection refused\\\\nI1210 22:49:34.532599 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1210 22:49:34.533977 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-2585701456/tls.crt::/tmp/serving-cert-2585701456/tls.key\\\\\\\"\\\\nF1210 22:49:44.774710 1 cmd.go:182] error initializing delegating authentication: unable to load configmap based request-header-client-ca-file: Get \\\\\\\"https://localhost:6443/api/v1/namespaces/kube-system/configmaps/extension-apiserver-authentication\\\\\\\": net/http: TLS handshake timeout\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-10T22:49:33Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T22:49:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://0c4d036ab269f62561fe7ef1586012cb3a16e3583a08b9488a0616ec00c8e0ae\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T22:49:32Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c5856d6b26c19fda13b5c9ac2892a7e2f9c6f79feb8173ecfd9b33b673327d67\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c5856d6b26c19fda13b5c9ac2892a7e2f9c6f79feb8173ecfd9b33b673327d67\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-10T22:49:27Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-10T22:49:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-10T22:49:25Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T22:50:20Z is after 2025-08-24T17:21:41Z" Dec 10 22:50:20 crc kubenswrapper[4791]: I1210 22:50:20.824972 4791 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"90250e13-6c63-443a-a5e5-6e4e341a2289\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T22:49:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T22:49:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T22:49:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T22:49:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T22:49:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ba54f95b1e00843b854db9a10fe6897592159ccacf30ed0f76002093a13eedd3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T22:49:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5aa8ffa518f3a68b60b275f8df4851b2564aa4984b61efdc9470b73072133e43\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T22:49:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://16f989503edc78d9451638e2cf6cf4b19a2cf2b8809b10b237541cb829303091\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T22:49:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://82556824d63ce6e46df14c45689ff1f81ebf729abe902061815c040b323898e9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T22:49:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://61db531940fa6f2effdae5a005b4c91bd2698bd5d1bff9588c171a65ad550eca\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T22:49:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0f65c864d885319ce2a08d61c30be24b3ffb7d23ea30c82ed7faf039cb2c006d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://0f65c864d885319ce2a08d61c30be24b3ffb7d23ea30c82ed7faf039cb2c006d\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-10T22:49:27Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-10T22:49:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e31a76d1d5b2812ba83e24e38268e25ec8bed98fc3da9adcd06bda5822b6ace1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e31a76d1d5b2812ba83e24e38268e25ec8bed98fc3da9adcd06bda5822b6ace1\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-10T22:49:30Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-10T22:49:29Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://b5b5eceae562f3986b543d41de1e977efb85a91d84c98e09159ca6d092eda691\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b5b5eceae562f3986b543d41de1e977efb85a91d84c98e09159ca6d092eda691\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-10T22:49:31Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-10T22:49:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-10T22:49:25Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T22:50:20Z is after 2025-08-24T17:21:41Z" Dec 10 22:50:20 crc kubenswrapper[4791]: I1210 22:50:20.841235 4791 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-10T22:49:50Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-10T22:49:50Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T22:50:20Z is after 2025-08-24T17:21:41Z" Dec 10 22:50:20 crc kubenswrapper[4791]: I1210 22:50:20.846916 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 22:50:20 crc kubenswrapper[4791]: I1210 22:50:20.846950 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 22:50:20 crc kubenswrapper[4791]: I1210 22:50:20.846960 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 22:50:20 crc kubenswrapper[4791]: I1210 22:50:20.846974 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 22:50:20 crc kubenswrapper[4791]: I1210 22:50:20.846983 4791 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T22:50:20Z","lastTransitionTime":"2025-12-10T22:50:20Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 22:50:20 crc kubenswrapper[4791]: I1210 22:50:20.856174 4791 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-5rb5l" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ba35653c-6e06-4cee-a4d6-137764090d18\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T22:49:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T22:49:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T22:49:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T22:49:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://dbb5c875730a98ad2daf316cfafc08fcdce21ac16148154780ac9193ed405dab\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T22:49:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fvvrs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://185eb2ff44221a514878283ec873773a6c9a9073dc64e296f982e60dee1338e4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T22:49:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fvvrs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-10T22:49:51Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-5rb5l\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T22:50:20Z is after 2025-08-24T17:21:41Z" Dec 10 22:50:20 crc kubenswrapper[4791]: I1210 22:50:20.869157 4791 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-s6rk8" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5c9d0978-a6f9-49c5-b377-6619b4e3ff52\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T22:50:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T22:50:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T22:50:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T22:50:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8e55baaf460c794ab196e237d51d12de059aa038617dc2d5366d49c529f5d5c7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T22:50:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7vkgm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://51d1b9b086c60bae9223cb51d5a998d02790a0d91b745aab10392c23e617eb9f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T22:50:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7vkgm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-10T22:50:03Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-s6rk8\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T22:50:20Z is after 2025-08-24T17:21:41Z" Dec 10 22:50:20 crc kubenswrapper[4791]: I1210 22:50:20.884602 4791 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 10 22:50:20 crc kubenswrapper[4791]: I1210 22:50:20.884644 4791 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 10 22:50:20 crc kubenswrapper[4791]: I1210 22:50:20.884708 4791 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-b6kf6" Dec 10 22:50:20 crc kubenswrapper[4791]: I1210 22:50:20.884699 4791 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 10 22:50:20 crc kubenswrapper[4791]: E1210 22:50:20.884718 4791 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 10 22:50:20 crc kubenswrapper[4791]: E1210 22:50:20.884797 4791 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 10 22:50:20 crc kubenswrapper[4791]: E1210 22:50:20.884897 4791 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 10 22:50:20 crc kubenswrapper[4791]: E1210 22:50:20.884960 4791 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-b6kf6" podUID="b4527e71-bb58-4a39-a99c-fe614aa4c5f9" Dec 10 22:50:20 crc kubenswrapper[4791]: I1210 22:50:20.885595 4791 scope.go:117] "RemoveContainer" containerID="f78a47c773006a18dbc4b706edfc9d1d5d0b6ad11a023a49ea4e622eae77bba3" Dec 10 22:50:20 crc kubenswrapper[4791]: I1210 22:50:20.949464 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 22:50:20 crc kubenswrapper[4791]: I1210 22:50:20.949494 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 22:50:20 crc kubenswrapper[4791]: I1210 22:50:20.949519 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 22:50:20 crc kubenswrapper[4791]: I1210 22:50:20.949536 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 22:50:20 crc kubenswrapper[4791]: I1210 22:50:20.949747 4791 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T22:50:20Z","lastTransitionTime":"2025-12-10T22:50:20Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 22:50:21 crc kubenswrapper[4791]: I1210 22:50:21.051977 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 22:50:21 crc kubenswrapper[4791]: I1210 22:50:21.052004 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 22:50:21 crc kubenswrapper[4791]: I1210 22:50:21.052016 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 22:50:21 crc kubenswrapper[4791]: I1210 22:50:21.052032 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 22:50:21 crc kubenswrapper[4791]: I1210 22:50:21.052043 4791 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T22:50:21Z","lastTransitionTime":"2025-12-10T22:50:21Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 22:50:21 crc kubenswrapper[4791]: I1210 22:50:21.154365 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 22:50:21 crc kubenswrapper[4791]: I1210 22:50:21.154402 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 22:50:21 crc kubenswrapper[4791]: I1210 22:50:21.154412 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 22:50:21 crc kubenswrapper[4791]: I1210 22:50:21.154429 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 22:50:21 crc kubenswrapper[4791]: I1210 22:50:21.154440 4791 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T22:50:21Z","lastTransitionTime":"2025-12-10T22:50:21Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 22:50:21 crc kubenswrapper[4791]: I1210 22:50:21.256835 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 22:50:21 crc kubenswrapper[4791]: I1210 22:50:21.256867 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 22:50:21 crc kubenswrapper[4791]: I1210 22:50:21.256876 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 22:50:21 crc kubenswrapper[4791]: I1210 22:50:21.256890 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 22:50:21 crc kubenswrapper[4791]: I1210 22:50:21.256900 4791 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T22:50:21Z","lastTransitionTime":"2025-12-10T22:50:21Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 22:50:21 crc kubenswrapper[4791]: I1210 22:50:21.359386 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 22:50:21 crc kubenswrapper[4791]: I1210 22:50:21.359463 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 22:50:21 crc kubenswrapper[4791]: I1210 22:50:21.359476 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 22:50:21 crc kubenswrapper[4791]: I1210 22:50:21.359493 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 22:50:21 crc kubenswrapper[4791]: I1210 22:50:21.359506 4791 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T22:50:21Z","lastTransitionTime":"2025-12-10T22:50:21Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 22:50:21 crc kubenswrapper[4791]: I1210 22:50:21.438963 4791 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-zhq64_3cd47739-0fa9-4321-aff1-220f8721a0b3/ovnkube-controller/1.log" Dec 10 22:50:21 crc kubenswrapper[4791]: I1210 22:50:21.442047 4791 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-zhq64" event={"ID":"3cd47739-0fa9-4321-aff1-220f8721a0b3","Type":"ContainerStarted","Data":"b6d739a4314c4fb536810b8623f691d6aaf1090b9c83f17b484d2b22f7d3ad01"} Dec 10 22:50:21 crc kubenswrapper[4791]: I1210 22:50:21.442618 4791 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-ovn-kubernetes/ovnkube-node-zhq64" Dec 10 22:50:21 crc kubenswrapper[4791]: I1210 22:50:21.456558 4791 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-10T22:49:52Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-10T22:49:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-10T22:49:52Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://20651eb0c16e2c71222f56da1ff755cb0e7b084886acbbb24f0ccaf57041ed85\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T22:49:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T22:50:21Z is after 2025-08-24T17:21:41Z" Dec 10 22:50:21 crc kubenswrapper[4791]: I1210 22:50:21.461735 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 22:50:21 crc kubenswrapper[4791]: I1210 22:50:21.461791 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 22:50:21 crc kubenswrapper[4791]: I1210 22:50:21.461800 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 22:50:21 crc kubenswrapper[4791]: I1210 22:50:21.461815 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 22:50:21 crc kubenswrapper[4791]: I1210 22:50:21.461826 4791 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T22:50:21Z","lastTransitionTime":"2025-12-10T22:50:21Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 22:50:21 crc kubenswrapper[4791]: I1210 22:50:21.469589 4791 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-10T22:49:50Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-10T22:49:50Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T22:50:21Z is after 2025-08-24T17:21:41Z" Dec 10 22:50:21 crc kubenswrapper[4791]: I1210 22:50:21.481719 4791 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-10T22:49:50Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-10T22:49:50Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T22:50:21Z is after 2025-08-24T17:21:41Z" Dec 10 22:50:21 crc kubenswrapper[4791]: I1210 22:50:21.491062 4791 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-tchxk" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"c0ab74e0-42a9-4f42-ac80-16030cf4ffff\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T22:49:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T22:49:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T22:49:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T22:49:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://04e0fd1b53256b7a938e8ccab93e2a4eb73dd1a59e0b9799e09c55e511baf7fe\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T22:49:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zn2qt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-10T22:49:50Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-tchxk\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T22:50:21Z is after 2025-08-24T17:21:41Z" Dec 10 22:50:21 crc kubenswrapper[4791]: I1210 22:50:21.509006 4791 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-rw584" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ff9beb90-69b0-4732-bf37-0b81f58ecc98\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T22:49:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T22:49:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T22:49:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T22:49:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ddb9df501dc6e8e5561fbc2aefd431037a3f28f73bf8e82a6598d261a4224807\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T22:49:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fksqn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ce67e54b696bb53ffbbe48c6ce71476b635e0868993e9c427b22f904891433ee\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ce67e54b696bb53ffbbe48c6ce71476b635e0868993e9c427b22f904891433ee\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-10T22:49:52Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-10T22:49:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fksqn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://818d89bfffb8a110670152c8480684665ea8db578e9768f524fbbe6a92f27c22\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://818d89bfffb8a110670152c8480684665ea8db578e9768f524fbbe6a92f27c22\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-10T22:49:53Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-10T22:49:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fksqn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://9aa9c06e894bb2a2ed3738defaeff68a0a0458ead48177f36b90770944c211a0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://9aa9c06e894bb2a2ed3738defaeff68a0a0458ead48177f36b90770944c211a0\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-10T22:49:54Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-10T22:49:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fksqn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://dfd6d03336dcea7529419a30e8d70d84e2290a0fa5b0e46d09ad341c190638d6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://dfd6d03336dcea7529419a30e8d70d84e2290a0fa5b0e46d09ad341c190638d6\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-10T22:49:55Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-10T22:49:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fksqn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b86738401ece6fbc3083782720c66a515850ed78d4054f34fcabab7ff525b85f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b86738401ece6fbc3083782720c66a515850ed78d4054f34fcabab7ff525b85f\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-10T22:49:56Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-10T22:49:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fksqn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://18b326f572a899ddd10197a745bf9264d4ddd975690ad1cfb0ec576ccd5cf102\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://18b326f572a899ddd10197a745bf9264d4ddd975690ad1cfb0ec576ccd5cf102\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-10T22:49:57Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-10T22:49:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fksqn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-10T22:49:51Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-rw584\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T22:50:21Z is after 2025-08-24T17:21:41Z" Dec 10 22:50:21 crc kubenswrapper[4791]: I1210 22:50:21.523096 4791 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f8349fea-c49b-49a1-b5ee-032c5dffe021\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T22:49:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T22:49:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T22:49:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T22:49:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T22:49:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://acafe12186df99d1059a31260de0eafce35ff7771bcb8f5c6753b72665d6f60b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T22:49:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ea4c346507862a8c12756a3061d6e25e788a8d79f4621b62f81f4322185c82a8\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T22:49:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://2e6e324ec6f4f7bb69fdfab4ea3ac1aa55e480dd8e2a8984b0fe5b4c9f66636e\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T22:49:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://1b2e11ff2bf8789fa1d0d23ae648b3e6ef653106e5bc4bbb5dd6447f7cd6f05d\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d6aa7f8c5d908d90008f2df49860ac8a3480136da567e3d85f86135d779204f9\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-12-10T22:49:44Z\\\",\\\"message\\\":\\\"W1210 22:49:33.845733 1 cmd.go:257] Using insecure, self-signed certificates\\\\nI1210 22:49:33.846108 1 crypto.go:601] Generating new CA for check-endpoints-signer@1765406973 cert, and key in /tmp/serving-cert-2585701456/serving-signer.crt, /tmp/serving-cert-2585701456/serving-signer.key\\\\nI1210 22:49:34.529748 1 observer_polling.go:159] Starting file observer\\\\nW1210 22:49:34.532225 1 builder.go:272] unable to get owner reference (falling back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": dial tcp [::1]:6443: connect: connection refused\\\\nI1210 22:49:34.532599 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1210 22:49:34.533977 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-2585701456/tls.crt::/tmp/serving-cert-2585701456/tls.key\\\\\\\"\\\\nF1210 22:49:44.774710 1 cmd.go:182] error initializing delegating authentication: unable to load configmap based request-header-client-ca-file: Get \\\\\\\"https://localhost:6443/api/v1/namespaces/kube-system/configmaps/extension-apiserver-authentication\\\\\\\": net/http: TLS handshake timeout\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-10T22:49:33Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T22:49:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://0c4d036ab269f62561fe7ef1586012cb3a16e3583a08b9488a0616ec00c8e0ae\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T22:49:32Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c5856d6b26c19fda13b5c9ac2892a7e2f9c6f79feb8173ecfd9b33b673327d67\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c5856d6b26c19fda13b5c9ac2892a7e2f9c6f79feb8173ecfd9b33b673327d67\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-10T22:49:27Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-10T22:49:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-10T22:49:25Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T22:50:21Z is after 2025-08-24T17:21:41Z" Dec 10 22:50:21 crc kubenswrapper[4791]: I1210 22:50:21.536724 4791 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"02a949d8-38e2-4be3-989f-40a015407140\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T22:49:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T22:49:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T22:50:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T22:50:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T22:49:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://cd4158bc59b30f0f1716bf43c1ae7c9916a5b7d7164d074987fc1f19ba49b778\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T22:49:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://656065fd3cbdc6a25dab5d06c7baa00af16d8f62fc5da687efe9c6cc91945de0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T22:49:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://90ed7fff816b1772ba9ecd95587a59cbed7c30ac302fff7e7c5de934c912bcf4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T22:49:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9b7a5502803eaad9f76f5ffa91b29fbb2657402065a03a4d6433c096aac7264f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"wait-for-host-port\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://9b7a5502803eaad9f76f5ffa91b29fbb2657402065a03a4d6433c096aac7264f\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-10T22:49:27Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-10T22:49:27Z\\\"}}}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-10T22:49:25Z\\\"}}\" for pod \"openshift-kube-scheduler\"/\"openshift-kube-scheduler-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T22:50:21Z is after 2025-08-24T17:21:41Z" Dec 10 22:50:21 crc kubenswrapper[4791]: I1210 22:50:21.560398 4791 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"90250e13-6c63-443a-a5e5-6e4e341a2289\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T22:49:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T22:49:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T22:49:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T22:49:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T22:49:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ba54f95b1e00843b854db9a10fe6897592159ccacf30ed0f76002093a13eedd3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T22:49:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5aa8ffa518f3a68b60b275f8df4851b2564aa4984b61efdc9470b73072133e43\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T22:49:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://16f989503edc78d9451638e2cf6cf4b19a2cf2b8809b10b237541cb829303091\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T22:49:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://82556824d63ce6e46df14c45689ff1f81ebf729abe902061815c040b323898e9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T22:49:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://61db531940fa6f2effdae5a005b4c91bd2698bd5d1bff9588c171a65ad550eca\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T22:49:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0f65c864d885319ce2a08d61c30be24b3ffb7d23ea30c82ed7faf039cb2c006d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://0f65c864d885319ce2a08d61c30be24b3ffb7d23ea30c82ed7faf039cb2c006d\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-10T22:49:27Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-10T22:49:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e31a76d1d5b2812ba83e24e38268e25ec8bed98fc3da9adcd06bda5822b6ace1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e31a76d1d5b2812ba83e24e38268e25ec8bed98fc3da9adcd06bda5822b6ace1\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-10T22:49:30Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-10T22:49:29Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://b5b5eceae562f3986b543d41de1e977efb85a91d84c98e09159ca6d092eda691\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b5b5eceae562f3986b543d41de1e977efb85a91d84c98e09159ca6d092eda691\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-10T22:49:31Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-10T22:49:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-10T22:49:25Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T22:50:21Z is after 2025-08-24T17:21:41Z" Dec 10 22:50:21 crc kubenswrapper[4791]: I1210 22:50:21.567240 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 22:50:21 crc kubenswrapper[4791]: I1210 22:50:21.567285 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 22:50:21 crc kubenswrapper[4791]: I1210 22:50:21.567305 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 22:50:21 crc kubenswrapper[4791]: I1210 22:50:21.567323 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 22:50:21 crc kubenswrapper[4791]: I1210 22:50:21.567361 4791 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T22:50:21Z","lastTransitionTime":"2025-12-10T22:50:21Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 22:50:21 crc kubenswrapper[4791]: I1210 22:50:21.578975 4791 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-10T22:49:50Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-10T22:49:50Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T22:50:21Z is after 2025-08-24T17:21:41Z" Dec 10 22:50:21 crc kubenswrapper[4791]: I1210 22:50:21.594976 4791 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-10T22:49:54Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-10T22:49:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-10T22:49:54Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://62bfaf5d7a6d1b452cdaf88086447d0c8cb7306e94434e46752be5475124d56a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T22:49:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T22:50:21Z is after 2025-08-24T17:21:41Z" Dec 10 22:50:21 crc kubenswrapper[4791]: I1210 22:50:21.609041 4791 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-b6kf6" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b4527e71-bb58-4a39-a99c-fe614aa4c5f9\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T22:50:04Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T22:50:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T22:50:04Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T22:50:04Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ngpr4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ngpr4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-10T22:50:04Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-b6kf6\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T22:50:21Z is after 2025-08-24T17:21:41Z" Dec 10 22:50:21 crc kubenswrapper[4791]: I1210 22:50:21.622476 4791 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-5rb5l" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ba35653c-6e06-4cee-a4d6-137764090d18\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T22:49:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T22:49:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T22:49:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T22:49:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://dbb5c875730a98ad2daf316cfafc08fcdce21ac16148154780ac9193ed405dab\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T22:49:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fvvrs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://185eb2ff44221a514878283ec873773a6c9a9073dc64e296f982e60dee1338e4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T22:49:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fvvrs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-10T22:49:51Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-5rb5l\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T22:50:21Z is after 2025-08-24T17:21:41Z" Dec 10 22:50:21 crc kubenswrapper[4791]: I1210 22:50:21.635203 4791 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-s6rk8" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5c9d0978-a6f9-49c5-b377-6619b4e3ff52\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T22:50:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T22:50:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T22:50:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T22:50:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8e55baaf460c794ab196e237d51d12de059aa038617dc2d5366d49c529f5d5c7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T22:50:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7vkgm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://51d1b9b086c60bae9223cb51d5a998d02790a0d91b745aab10392c23e617eb9f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T22:50:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7vkgm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-10T22:50:03Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-s6rk8\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T22:50:21Z is after 2025-08-24T17:21:41Z" Dec 10 22:50:21 crc kubenswrapper[4791]: I1210 22:50:21.645994 4791 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"698bbe9a-fbe5-436d-a835-5e80df77ba24\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T22:49:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T22:49:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T22:49:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T22:49:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T22:49:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://224d73713312376a58ec6cc2bc8fdb6ffa44c4749ae0e7691d5fe27b9afca09f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T22:49:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://372eeaffd40992c6f06efc4df6dfcb44b31de64ee8256c21508cdd7c74ecb96e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T22:49:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://76b6013b062cac5b0b4d12a94f38ce7b43f9c7ea216063455d88a956ac04d26e\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T22:49:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://78e950e122e761d950da1a3c60c06476e53920e8795f6c61db22c01e22d5136e\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T22:49:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-10T22:49:25Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T22:50:21Z is after 2025-08-24T17:21:41Z" Dec 10 22:50:21 crc kubenswrapper[4791]: I1210 22:50:21.657819 4791 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-10T22:49:52Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-10T22:49:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-10T22:49:52Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3734b110802828632d44263ed0a43cabc4b710833f0be120d72db0940fe783e2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T22:49:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://794803ac9be3c52509e6dcb10854d8864e1a9e406762e57c7971b07c4da1220e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T22:49:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T22:50:21Z is after 2025-08-24T17:21:41Z" Dec 10 22:50:21 crc kubenswrapper[4791]: I1210 22:50:21.667166 4791 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-vczjq" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b9f00f25-3fa1-4b63-9710-fd352224b01b\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T22:49:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T22:49:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T22:49:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T22:49:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4698cf4c6b55b41e628448d37c2eb847c2777d0efbd93afcbf367d5e78623b7e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T22:49:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lrrpv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-10T22:49:50Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-vczjq\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T22:50:21Z is after 2025-08-24T17:21:41Z" Dec 10 22:50:21 crc kubenswrapper[4791]: I1210 22:50:21.670104 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 22:50:21 crc kubenswrapper[4791]: I1210 22:50:21.670146 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 22:50:21 crc kubenswrapper[4791]: I1210 22:50:21.670158 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 22:50:21 crc kubenswrapper[4791]: I1210 22:50:21.670176 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 22:50:21 crc kubenswrapper[4791]: I1210 22:50:21.670187 4791 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T22:50:21Z","lastTransitionTime":"2025-12-10T22:50:21Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 22:50:21 crc kubenswrapper[4791]: I1210 22:50:21.679855 4791 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-4nwkq" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"672aa28c-8169-49ed-87b8-21187d13a80c\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T22:49:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T22:49:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T22:49:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T22:49:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a2f2a62315529e148a10a33ccb241e9417fbde0fe42afb61cb7ee2438c74636f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T22:49:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xl62j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-10T22:49:51Z\\\"}}\" for pod \"openshift-multus\"/\"multus-4nwkq\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T22:50:21Z is after 2025-08-24T17:21:41Z" Dec 10 22:50:21 crc kubenswrapper[4791]: I1210 22:50:21.696785 4791 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-zhq64" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3cd47739-0fa9-4321-aff1-220f8721a0b3\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T22:49:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T22:49:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T22:49:52Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T22:49:52Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://68795a29944e56870ffe94fa732e9dcf0bfb344c0d85d1e590f58a876129c61f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T22:49:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wvpw8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://bc721b530c08b433756e978ec0423cc18aabf6771b61eb96f903b4d57ea335ff\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T22:49:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wvpw8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://152a7618be8f38fadf4e564b3948980503cedcdb466c37344ccc20f4a3173a5f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T22:49:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wvpw8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3d18f9cb7207efb1e32de75a30d33e024f656a08f0503db95e5488bc5280971f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T22:49:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wvpw8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://937137ae9d7480cdc7c59bc4a1df0a4c10e72b3b63693af285278789953ed218\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T22:49:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wvpw8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://1a52d19b7671e8b143017c8a981a6f8d3ee3b84650e4b223b814a531ac76f364\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T22:49:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wvpw8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b6d739a4314c4fb536810b8623f691d6aaf1090b9c83f17b484d2b22f7d3ad01\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f78a47c773006a18dbc4b706edfc9d1d5d0b6ad11a023a49ea4e622eae77bba3\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-12-10T22:50:05Z\\\",\\\"message\\\":\\\"olicy controller, err: could not add Event Handler for anpInformer during admin network policy controller initialization, handler {0x1fcc6e0 0x1fcc3c0 0x1fcc360} was not added to shared informer because it has stopped already, failed to start node network controller: failed to start default node network controller: failed to set node crc annotations: Internal error occurred: failed calling webhook \\\\\\\"node.network-node-identity.openshift.io\\\\\\\": failed to call webhook: Post \\\\\\\"https://127.0.0.1:9743/node?timeout=10s\\\\\\\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T22:50:05Z is after 2025-08-24T17:21:41Z]\\\\nI1210 22:50:05.954954 6227 services_controller.go:473] Services do not match for network=default, existing lbs: []services.LB{services.LB{Name:\\\\\\\"Service_openshift-operator-lifecycle-manager/olm-operator-metrics_TCP_cluster\\\\\\\", UUID:\\\\\\\"63b1440a-0908-4cab-8799-012fa1cf0b07\\\\\\\", Protocol:\\\\\\\"tcp\\\\\\\", ExternalIDs:map[string]string{\\\\\\\"k8s.ovn.org/kind\\\\\\\":\\\\\\\"Service\\\\\\\", \\\\\\\"k8s.ovn.org/owner\\\\\\\":\\\\\\\"openshift-operator-lifecycle-manager/olm-operator-metrics\\\\\\\"}, Opts:services.LBOpts{Reject:false, EmptyLBEvents:false, AffinityTimeOut:0, SkipSNAT:false, Template:false, AddressFamily:\\\\\\\"\\\\\\\"}, Rules:[]services.LBRule{}, Templates:services.TemplateMap{}, Switches:[]string{}, Routers:[]string{}, Groups:[]string{\\\\\\\"clusterLBGroup\\\\\\\"}}}, bui\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-10T22:50:04Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":2,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T22:50:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wvpw8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://dd3f80e5713221f96ed50da7f2e7fd5b9fabfcd5c1656da8d8a86d818a5fdc3f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T22:49:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wvpw8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3afbd1ffcd2c4b40324aca6e52d6fc822db131c3f35a9af3a6f005c2cffdb7bf\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3afbd1ffcd2c4b40324aca6e52d6fc822db131c3f35a9af3a6f005c2cffdb7bf\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-10T22:49:53Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-10T22:49:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wvpw8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-10T22:49:52Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-zhq64\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T22:50:21Z is after 2025-08-24T17:21:41Z" Dec 10 22:50:21 crc kubenswrapper[4791]: I1210 22:50:21.772453 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 22:50:21 crc kubenswrapper[4791]: I1210 22:50:21.772507 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 22:50:21 crc kubenswrapper[4791]: I1210 22:50:21.772524 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 22:50:21 crc kubenswrapper[4791]: I1210 22:50:21.772544 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 22:50:21 crc kubenswrapper[4791]: I1210 22:50:21.772559 4791 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T22:50:21Z","lastTransitionTime":"2025-12-10T22:50:21Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 22:50:21 crc kubenswrapper[4791]: I1210 22:50:21.874690 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 22:50:21 crc kubenswrapper[4791]: I1210 22:50:21.874732 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 22:50:21 crc kubenswrapper[4791]: I1210 22:50:21.874743 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 22:50:21 crc kubenswrapper[4791]: I1210 22:50:21.874761 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 22:50:21 crc kubenswrapper[4791]: I1210 22:50:21.874773 4791 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T22:50:21Z","lastTransitionTime":"2025-12-10T22:50:21Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 22:50:21 crc kubenswrapper[4791]: I1210 22:50:21.978335 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 22:50:21 crc kubenswrapper[4791]: I1210 22:50:21.978447 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 22:50:21 crc kubenswrapper[4791]: I1210 22:50:21.978468 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 22:50:21 crc kubenswrapper[4791]: I1210 22:50:21.978491 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 22:50:21 crc kubenswrapper[4791]: I1210 22:50:21.978508 4791 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T22:50:21Z","lastTransitionTime":"2025-12-10T22:50:21Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 22:50:22 crc kubenswrapper[4791]: I1210 22:50:22.081007 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 22:50:22 crc kubenswrapper[4791]: I1210 22:50:22.081079 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 22:50:22 crc kubenswrapper[4791]: I1210 22:50:22.081109 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 22:50:22 crc kubenswrapper[4791]: I1210 22:50:22.081143 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 22:50:22 crc kubenswrapper[4791]: I1210 22:50:22.081167 4791 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T22:50:22Z","lastTransitionTime":"2025-12-10T22:50:22Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 22:50:22 crc kubenswrapper[4791]: I1210 22:50:22.184215 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 22:50:22 crc kubenswrapper[4791]: I1210 22:50:22.184256 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 22:50:22 crc kubenswrapper[4791]: I1210 22:50:22.184271 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 22:50:22 crc kubenswrapper[4791]: I1210 22:50:22.184291 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 22:50:22 crc kubenswrapper[4791]: I1210 22:50:22.184306 4791 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T22:50:22Z","lastTransitionTime":"2025-12-10T22:50:22Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 22:50:22 crc kubenswrapper[4791]: I1210 22:50:22.287089 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 22:50:22 crc kubenswrapper[4791]: I1210 22:50:22.287358 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 22:50:22 crc kubenswrapper[4791]: I1210 22:50:22.287436 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 22:50:22 crc kubenswrapper[4791]: I1210 22:50:22.287503 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 22:50:22 crc kubenswrapper[4791]: I1210 22:50:22.287581 4791 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T22:50:22Z","lastTransitionTime":"2025-12-10T22:50:22Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 22:50:22 crc kubenswrapper[4791]: I1210 22:50:22.390268 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 22:50:22 crc kubenswrapper[4791]: I1210 22:50:22.390307 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 22:50:22 crc kubenswrapper[4791]: I1210 22:50:22.390322 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 22:50:22 crc kubenswrapper[4791]: I1210 22:50:22.390348 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 22:50:22 crc kubenswrapper[4791]: I1210 22:50:22.390358 4791 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T22:50:22Z","lastTransitionTime":"2025-12-10T22:50:22Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 22:50:22 crc kubenswrapper[4791]: I1210 22:50:22.446042 4791 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-zhq64_3cd47739-0fa9-4321-aff1-220f8721a0b3/ovnkube-controller/2.log" Dec 10 22:50:22 crc kubenswrapper[4791]: I1210 22:50:22.446917 4791 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-zhq64_3cd47739-0fa9-4321-aff1-220f8721a0b3/ovnkube-controller/1.log" Dec 10 22:50:22 crc kubenswrapper[4791]: I1210 22:50:22.450257 4791 generic.go:334] "Generic (PLEG): container finished" podID="3cd47739-0fa9-4321-aff1-220f8721a0b3" containerID="b6d739a4314c4fb536810b8623f691d6aaf1090b9c83f17b484d2b22f7d3ad01" exitCode=1 Dec 10 22:50:22 crc kubenswrapper[4791]: I1210 22:50:22.450322 4791 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-zhq64" event={"ID":"3cd47739-0fa9-4321-aff1-220f8721a0b3","Type":"ContainerDied","Data":"b6d739a4314c4fb536810b8623f691d6aaf1090b9c83f17b484d2b22f7d3ad01"} Dec 10 22:50:22 crc kubenswrapper[4791]: I1210 22:50:22.450419 4791 scope.go:117] "RemoveContainer" containerID="f78a47c773006a18dbc4b706edfc9d1d5d0b6ad11a023a49ea4e622eae77bba3" Dec 10 22:50:22 crc kubenswrapper[4791]: I1210 22:50:22.450881 4791 scope.go:117] "RemoveContainer" containerID="b6d739a4314c4fb536810b8623f691d6aaf1090b9c83f17b484d2b22f7d3ad01" Dec 10 22:50:22 crc kubenswrapper[4791]: E1210 22:50:22.451044 4791 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"ovnkube-controller\" with CrashLoopBackOff: \"back-off 20s restarting failed container=ovnkube-controller pod=ovnkube-node-zhq64_openshift-ovn-kubernetes(3cd47739-0fa9-4321-aff1-220f8721a0b3)\"" pod="openshift-ovn-kubernetes/ovnkube-node-zhq64" podUID="3cd47739-0fa9-4321-aff1-220f8721a0b3" Dec 10 22:50:22 crc kubenswrapper[4791]: I1210 22:50:22.484765 4791 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-10T22:49:50Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-10T22:49:50Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T22:50:22Z is after 2025-08-24T17:21:41Z" Dec 10 22:50:22 crc kubenswrapper[4791]: I1210 22:50:22.492527 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 22:50:22 crc kubenswrapper[4791]: I1210 22:50:22.492567 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 22:50:22 crc kubenswrapper[4791]: I1210 22:50:22.492578 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 22:50:22 crc kubenswrapper[4791]: I1210 22:50:22.492593 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 22:50:22 crc kubenswrapper[4791]: I1210 22:50:22.492606 4791 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T22:50:22Z","lastTransitionTime":"2025-12-10T22:50:22Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 22:50:22 crc kubenswrapper[4791]: I1210 22:50:22.503567 4791 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-10T22:49:54Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-10T22:49:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-10T22:49:54Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://62bfaf5d7a6d1b452cdaf88086447d0c8cb7306e94434e46752be5475124d56a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T22:49:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T22:50:22Z is after 2025-08-24T17:21:41Z" Dec 10 22:50:22 crc kubenswrapper[4791]: I1210 22:50:22.515206 4791 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-b6kf6" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b4527e71-bb58-4a39-a99c-fe614aa4c5f9\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T22:50:04Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T22:50:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T22:50:04Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T22:50:04Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ngpr4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ngpr4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-10T22:50:04Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-b6kf6\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T22:50:22Z is after 2025-08-24T17:21:41Z" Dec 10 22:50:22 crc kubenswrapper[4791]: I1210 22:50:22.529526 4791 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f8349fea-c49b-49a1-b5ee-032c5dffe021\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T22:49:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T22:49:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T22:49:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T22:49:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T22:49:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://acafe12186df99d1059a31260de0eafce35ff7771bcb8f5c6753b72665d6f60b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T22:49:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ea4c346507862a8c12756a3061d6e25e788a8d79f4621b62f81f4322185c82a8\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T22:49:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://2e6e324ec6f4f7bb69fdfab4ea3ac1aa55e480dd8e2a8984b0fe5b4c9f66636e\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T22:49:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://1b2e11ff2bf8789fa1d0d23ae648b3e6ef653106e5bc4bbb5dd6447f7cd6f05d\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d6aa7f8c5d908d90008f2df49860ac8a3480136da567e3d85f86135d779204f9\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-12-10T22:49:44Z\\\",\\\"message\\\":\\\"W1210 22:49:33.845733 1 cmd.go:257] Using insecure, self-signed certificates\\\\nI1210 22:49:33.846108 1 crypto.go:601] Generating new CA for check-endpoints-signer@1765406973 cert, and key in /tmp/serving-cert-2585701456/serving-signer.crt, /tmp/serving-cert-2585701456/serving-signer.key\\\\nI1210 22:49:34.529748 1 observer_polling.go:159] Starting file observer\\\\nW1210 22:49:34.532225 1 builder.go:272] unable to get owner reference (falling back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": dial tcp [::1]:6443: connect: connection refused\\\\nI1210 22:49:34.532599 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1210 22:49:34.533977 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-2585701456/tls.crt::/tmp/serving-cert-2585701456/tls.key\\\\\\\"\\\\nF1210 22:49:44.774710 1 cmd.go:182] error initializing delegating authentication: unable to load configmap based request-header-client-ca-file: Get \\\\\\\"https://localhost:6443/api/v1/namespaces/kube-system/configmaps/extension-apiserver-authentication\\\\\\\": net/http: TLS handshake timeout\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-10T22:49:33Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T22:49:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://0c4d036ab269f62561fe7ef1586012cb3a16e3583a08b9488a0616ec00c8e0ae\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T22:49:32Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c5856d6b26c19fda13b5c9ac2892a7e2f9c6f79feb8173ecfd9b33b673327d67\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c5856d6b26c19fda13b5c9ac2892a7e2f9c6f79feb8173ecfd9b33b673327d67\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-10T22:49:27Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-10T22:49:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-10T22:49:25Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T22:50:22Z is after 2025-08-24T17:21:41Z" Dec 10 22:50:22 crc kubenswrapper[4791]: I1210 22:50:22.540612 4791 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"02a949d8-38e2-4be3-989f-40a015407140\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T22:49:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T22:49:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T22:50:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T22:50:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T22:49:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://cd4158bc59b30f0f1716bf43c1ae7c9916a5b7d7164d074987fc1f19ba49b778\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T22:49:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://656065fd3cbdc6a25dab5d06c7baa00af16d8f62fc5da687efe9c6cc91945de0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T22:49:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://90ed7fff816b1772ba9ecd95587a59cbed7c30ac302fff7e7c5de934c912bcf4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T22:49:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9b7a5502803eaad9f76f5ffa91b29fbb2657402065a03a4d6433c096aac7264f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"wait-for-host-port\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://9b7a5502803eaad9f76f5ffa91b29fbb2657402065a03a4d6433c096aac7264f\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-10T22:49:27Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-10T22:49:27Z\\\"}}}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-10T22:49:25Z\\\"}}\" for pod \"openshift-kube-scheduler\"/\"openshift-kube-scheduler-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T22:50:22Z is after 2025-08-24T17:21:41Z" Dec 10 22:50:22 crc kubenswrapper[4791]: I1210 22:50:22.562244 4791 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"90250e13-6c63-443a-a5e5-6e4e341a2289\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T22:49:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T22:49:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T22:49:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T22:49:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T22:49:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ba54f95b1e00843b854db9a10fe6897592159ccacf30ed0f76002093a13eedd3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T22:49:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5aa8ffa518f3a68b60b275f8df4851b2564aa4984b61efdc9470b73072133e43\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T22:49:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://16f989503edc78d9451638e2cf6cf4b19a2cf2b8809b10b237541cb829303091\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T22:49:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://82556824d63ce6e46df14c45689ff1f81ebf729abe902061815c040b323898e9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T22:49:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://61db531940fa6f2effdae5a005b4c91bd2698bd5d1bff9588c171a65ad550eca\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T22:49:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0f65c864d885319ce2a08d61c30be24b3ffb7d23ea30c82ed7faf039cb2c006d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://0f65c864d885319ce2a08d61c30be24b3ffb7d23ea30c82ed7faf039cb2c006d\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-10T22:49:27Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-10T22:49:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e31a76d1d5b2812ba83e24e38268e25ec8bed98fc3da9adcd06bda5822b6ace1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e31a76d1d5b2812ba83e24e38268e25ec8bed98fc3da9adcd06bda5822b6ace1\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-10T22:49:30Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-10T22:49:29Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://b5b5eceae562f3986b543d41de1e977efb85a91d84c98e09159ca6d092eda691\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b5b5eceae562f3986b543d41de1e977efb85a91d84c98e09159ca6d092eda691\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-10T22:49:31Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-10T22:49:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-10T22:49:25Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T22:50:22Z is after 2025-08-24T17:21:41Z" Dec 10 22:50:22 crc kubenswrapper[4791]: I1210 22:50:22.575480 4791 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-5rb5l" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ba35653c-6e06-4cee-a4d6-137764090d18\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T22:49:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T22:49:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T22:49:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T22:49:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://dbb5c875730a98ad2daf316cfafc08fcdce21ac16148154780ac9193ed405dab\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T22:49:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fvvrs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://185eb2ff44221a514878283ec873773a6c9a9073dc64e296f982e60dee1338e4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T22:49:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fvvrs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-10T22:49:51Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-5rb5l\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T22:50:22Z is after 2025-08-24T17:21:41Z" Dec 10 22:50:22 crc kubenswrapper[4791]: I1210 22:50:22.586776 4791 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-s6rk8" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5c9d0978-a6f9-49c5-b377-6619b4e3ff52\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T22:50:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T22:50:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T22:50:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T22:50:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8e55baaf460c794ab196e237d51d12de059aa038617dc2d5366d49c529f5d5c7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T22:50:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7vkgm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://51d1b9b086c60bae9223cb51d5a998d02790a0d91b745aab10392c23e617eb9f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T22:50:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7vkgm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-10T22:50:03Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-s6rk8\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T22:50:22Z is after 2025-08-24T17:21:41Z" Dec 10 22:50:22 crc kubenswrapper[4791]: I1210 22:50:22.594251 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 22:50:22 crc kubenswrapper[4791]: I1210 22:50:22.594276 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 22:50:22 crc kubenswrapper[4791]: I1210 22:50:22.594286 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 22:50:22 crc kubenswrapper[4791]: I1210 22:50:22.594301 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 22:50:22 crc kubenswrapper[4791]: I1210 22:50:22.594311 4791 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T22:50:22Z","lastTransitionTime":"2025-12-10T22:50:22Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 22:50:22 crc kubenswrapper[4791]: I1210 22:50:22.597547 4791 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-vczjq" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b9f00f25-3fa1-4b63-9710-fd352224b01b\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T22:49:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T22:49:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T22:49:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T22:49:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4698cf4c6b55b41e628448d37c2eb847c2777d0efbd93afcbf367d5e78623b7e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T22:49:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lrrpv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-10T22:49:50Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-vczjq\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T22:50:22Z is after 2025-08-24T17:21:41Z" Dec 10 22:50:22 crc kubenswrapper[4791]: I1210 22:50:22.610002 4791 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-4nwkq" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"672aa28c-8169-49ed-87b8-21187d13a80c\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T22:49:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T22:49:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T22:49:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T22:49:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a2f2a62315529e148a10a33ccb241e9417fbde0fe42afb61cb7ee2438c74636f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T22:49:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xl62j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-10T22:49:51Z\\\"}}\" for pod \"openshift-multus\"/\"multus-4nwkq\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T22:50:22Z is after 2025-08-24T17:21:41Z" Dec 10 22:50:22 crc kubenswrapper[4791]: I1210 22:50:22.629211 4791 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-zhq64" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3cd47739-0fa9-4321-aff1-220f8721a0b3\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T22:49:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T22:49:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T22:49:52Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T22:49:52Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://68795a29944e56870ffe94fa732e9dcf0bfb344c0d85d1e590f58a876129c61f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T22:49:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wvpw8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://bc721b530c08b433756e978ec0423cc18aabf6771b61eb96f903b4d57ea335ff\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T22:49:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wvpw8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://152a7618be8f38fadf4e564b3948980503cedcdb466c37344ccc20f4a3173a5f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T22:49:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wvpw8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3d18f9cb7207efb1e32de75a30d33e024f656a08f0503db95e5488bc5280971f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T22:49:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wvpw8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://937137ae9d7480cdc7c59bc4a1df0a4c10e72b3b63693af285278789953ed218\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T22:49:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wvpw8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://1a52d19b7671e8b143017c8a981a6f8d3ee3b84650e4b223b814a531ac76f364\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T22:49:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wvpw8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b6d739a4314c4fb536810b8623f691d6aaf1090b9c83f17b484d2b22f7d3ad01\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f78a47c773006a18dbc4b706edfc9d1d5d0b6ad11a023a49ea4e622eae77bba3\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-12-10T22:50:05Z\\\",\\\"message\\\":\\\"olicy controller, err: could not add Event Handler for anpInformer during admin network policy controller initialization, handler {0x1fcc6e0 0x1fcc3c0 0x1fcc360} was not added to shared informer because it has stopped already, failed to start node network controller: failed to start default node network controller: failed to set node crc annotations: Internal error occurred: failed calling webhook \\\\\\\"node.network-node-identity.openshift.io\\\\\\\": failed to call webhook: Post \\\\\\\"https://127.0.0.1:9743/node?timeout=10s\\\\\\\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T22:50:05Z is after 2025-08-24T17:21:41Z]\\\\nI1210 22:50:05.954954 6227 services_controller.go:473] Services do not match for network=default, existing lbs: []services.LB{services.LB{Name:\\\\\\\"Service_openshift-operator-lifecycle-manager/olm-operator-metrics_TCP_cluster\\\\\\\", UUID:\\\\\\\"63b1440a-0908-4cab-8799-012fa1cf0b07\\\\\\\", Protocol:\\\\\\\"tcp\\\\\\\", ExternalIDs:map[string]string{\\\\\\\"k8s.ovn.org/kind\\\\\\\":\\\\\\\"Service\\\\\\\", \\\\\\\"k8s.ovn.org/owner\\\\\\\":\\\\\\\"openshift-operator-lifecycle-manager/olm-operator-metrics\\\\\\\"}, Opts:services.LBOpts{Reject:false, EmptyLBEvents:false, AffinityTimeOut:0, SkipSNAT:false, Template:false, AddressFamily:\\\\\\\"\\\\\\\"}, Rules:[]services.LBRule{}, Templates:services.TemplateMap{}, Switches:[]string{}, Routers:[]string{}, Groups:[]string{\\\\\\\"clusterLBGroup\\\\\\\"}}}, bui\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-10T22:50:04Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":2,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b6d739a4314c4fb536810b8623f691d6aaf1090b9c83f17b484d2b22f7d3ad01\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-12-10T22:50:21Z\\\",\\\"message\\\":\\\"llocateLoadBalancerNodePorts:nil,LoadBalancerClass:nil,InternalTrafficPolicy:*Cluster,TrafficDistribution:nil,},Status:ServiceStatus{LoadBalancer:LoadBalancerStatus{Ingress:[]LoadBalancerIngress{},},Conditions:[]Condition{},},}\\\\nI1210 22:50:21.614710 6424 metrics.go:553] Stopping metrics server at address \\\\\\\"127.0.0.1:29103\\\\\\\"\\\\nI1210 22:50:21.614731 6424 lb_config.go:1031] Cluster endpoints for openshift-kube-scheduler/scheduler for network=default are: map[]\\\\nI1210 22:50:21.614619 6424 services_controller.go:360] Finished syncing service ovn-kubernetes-control-plane on namespace openshift-ovn-kubernetes for network=default : 7.02µs\\\\nI1210 22:50:21.614757 6424 services_controller.go:356] Processing sync for service openshift-route-controller-manager/route-controller-manager for network=default\\\\nI1210 22:50:21.614615 6424 services_controller.go:444] Built service openshift-machine-config-operator/machine-config-daemon LB per-node configs for network=default: []services.lbConfig(nil)\\\\nF1210 22:50:21.614772 6424 ovnkube.go:137] failed to run ovnkube: [failed to start network controller: failed to start default network controller: unable to create admin network policy controller, err: could not add Event Handler for anpInformer during admin network policy controller initialization, handler {0x1fcc6e0 0x1fcc3c0 0x1fcc360} was not added to\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-10T22:50:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wvpw8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://dd3f80e5713221f96ed50da7f2e7fd5b9fabfcd5c1656da8d8a86d818a5fdc3f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T22:49:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wvpw8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3afbd1ffcd2c4b40324aca6e52d6fc822db131c3f35a9af3a6f005c2cffdb7bf\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3afbd1ffcd2c4b40324aca6e52d6fc822db131c3f35a9af3a6f005c2cffdb7bf\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-10T22:49:53Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-10T22:49:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wvpw8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-10T22:49:52Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-zhq64\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T22:50:22Z is after 2025-08-24T17:21:41Z" Dec 10 22:50:22 crc kubenswrapper[4791]: I1210 22:50:22.643084 4791 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"698bbe9a-fbe5-436d-a835-5e80df77ba24\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T22:49:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T22:49:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T22:49:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T22:49:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T22:49:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://224d73713312376a58ec6cc2bc8fdb6ffa44c4749ae0e7691d5fe27b9afca09f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T22:49:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://372eeaffd40992c6f06efc4df6dfcb44b31de64ee8256c21508cdd7c74ecb96e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T22:49:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://76b6013b062cac5b0b4d12a94f38ce7b43f9c7ea216063455d88a956ac04d26e\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T22:49:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://78e950e122e761d950da1a3c60c06476e53920e8795f6c61db22c01e22d5136e\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T22:49:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-10T22:49:25Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T22:50:22Z is after 2025-08-24T17:21:41Z" Dec 10 22:50:22 crc kubenswrapper[4791]: I1210 22:50:22.654364 4791 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-10T22:49:52Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-10T22:49:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-10T22:49:52Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3734b110802828632d44263ed0a43cabc4b710833f0be120d72db0940fe783e2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T22:49:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://794803ac9be3c52509e6dcb10854d8864e1a9e406762e57c7971b07c4da1220e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T22:49:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T22:50:22Z is after 2025-08-24T17:21:41Z" Dec 10 22:50:22 crc kubenswrapper[4791]: I1210 22:50:22.667240 4791 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-tchxk" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"c0ab74e0-42a9-4f42-ac80-16030cf4ffff\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T22:49:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T22:49:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T22:49:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T22:49:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://04e0fd1b53256b7a938e8ccab93e2a4eb73dd1a59e0b9799e09c55e511baf7fe\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T22:49:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zn2qt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-10T22:49:50Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-tchxk\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T22:50:22Z is after 2025-08-24T17:21:41Z" Dec 10 22:50:22 crc kubenswrapper[4791]: I1210 22:50:22.682746 4791 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-rw584" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ff9beb90-69b0-4732-bf37-0b81f58ecc98\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T22:49:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T22:49:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T22:49:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T22:49:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ddb9df501dc6e8e5561fbc2aefd431037a3f28f73bf8e82a6598d261a4224807\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T22:49:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fksqn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ce67e54b696bb53ffbbe48c6ce71476b635e0868993e9c427b22f904891433ee\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ce67e54b696bb53ffbbe48c6ce71476b635e0868993e9c427b22f904891433ee\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-10T22:49:52Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-10T22:49:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fksqn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://818d89bfffb8a110670152c8480684665ea8db578e9768f524fbbe6a92f27c22\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://818d89bfffb8a110670152c8480684665ea8db578e9768f524fbbe6a92f27c22\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-10T22:49:53Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-10T22:49:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fksqn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://9aa9c06e894bb2a2ed3738defaeff68a0a0458ead48177f36b90770944c211a0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://9aa9c06e894bb2a2ed3738defaeff68a0a0458ead48177f36b90770944c211a0\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-10T22:49:54Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-10T22:49:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fksqn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://dfd6d03336dcea7529419a30e8d70d84e2290a0fa5b0e46d09ad341c190638d6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://dfd6d03336dcea7529419a30e8d70d84e2290a0fa5b0e46d09ad341c190638d6\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-10T22:49:55Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-10T22:49:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fksqn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b86738401ece6fbc3083782720c66a515850ed78d4054f34fcabab7ff525b85f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b86738401ece6fbc3083782720c66a515850ed78d4054f34fcabab7ff525b85f\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-10T22:49:56Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-10T22:49:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fksqn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://18b326f572a899ddd10197a745bf9264d4ddd975690ad1cfb0ec576ccd5cf102\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://18b326f572a899ddd10197a745bf9264d4ddd975690ad1cfb0ec576ccd5cf102\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-10T22:49:57Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-10T22:49:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fksqn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-10T22:49:51Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-rw584\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T22:50:22Z is after 2025-08-24T17:21:41Z" Dec 10 22:50:22 crc kubenswrapper[4791]: I1210 22:50:22.696884 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 22:50:22 crc kubenswrapper[4791]: I1210 22:50:22.696935 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 22:50:22 crc kubenswrapper[4791]: I1210 22:50:22.696951 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 22:50:22 crc kubenswrapper[4791]: I1210 22:50:22.696971 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 22:50:22 crc kubenswrapper[4791]: I1210 22:50:22.696987 4791 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T22:50:22Z","lastTransitionTime":"2025-12-10T22:50:22Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 22:50:22 crc kubenswrapper[4791]: I1210 22:50:22.704803 4791 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-10T22:49:52Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-10T22:49:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-10T22:49:52Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://20651eb0c16e2c71222f56da1ff755cb0e7b084886acbbb24f0ccaf57041ed85\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T22:49:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T22:50:22Z is after 2025-08-24T17:21:41Z" Dec 10 22:50:22 crc kubenswrapper[4791]: I1210 22:50:22.719034 4791 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-10T22:49:50Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-10T22:49:50Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T22:50:22Z is after 2025-08-24T17:21:41Z" Dec 10 22:50:22 crc kubenswrapper[4791]: I1210 22:50:22.730873 4791 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-10T22:49:50Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-10T22:49:50Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T22:50:22Z is after 2025-08-24T17:21:41Z" Dec 10 22:50:22 crc kubenswrapper[4791]: I1210 22:50:22.735528 4791 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 10 22:50:22 crc kubenswrapper[4791]: E1210 22:50:22.735897 4791 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-10 22:50:54.73585705 +0000 UTC m=+89.165474693 (durationBeforeRetry 32s). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 10 22:50:22 crc kubenswrapper[4791]: I1210 22:50:22.799584 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 22:50:22 crc kubenswrapper[4791]: I1210 22:50:22.799673 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 22:50:22 crc kubenswrapper[4791]: I1210 22:50:22.799687 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 22:50:22 crc kubenswrapper[4791]: I1210 22:50:22.799705 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 22:50:22 crc kubenswrapper[4791]: I1210 22:50:22.799719 4791 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T22:50:22Z","lastTransitionTime":"2025-12-10T22:50:22Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 22:50:22 crc kubenswrapper[4791]: I1210 22:50:22.836482 4791 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"networking-console-plugin-cert\" (UniqueName: \"kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 10 22:50:22 crc kubenswrapper[4791]: I1210 22:50:22.836568 4791 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nginx-conf\" (UniqueName: \"kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 10 22:50:22 crc kubenswrapper[4791]: I1210 22:50:22.836625 4791 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cqllr\" (UniqueName: \"kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr\") pod \"network-check-target-xd92c\" (UID: \"3b6479f0-333b-4a96-9adf-2099afdc2447\") " pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 10 22:50:22 crc kubenswrapper[4791]: I1210 22:50:22.836651 4791 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-s2dwl\" (UniqueName: \"kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl\") pod \"network-check-source-55646444c4-trplf\" (UID: \"9d751cbb-f2e2-430d-9754-c882a5e924a5\") " pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 10 22:50:22 crc kubenswrapper[4791]: E1210 22:50:22.836713 4791 secret.go:188] Couldn't get secret openshift-network-console/networking-console-plugin-cert: object "openshift-network-console"/"networking-console-plugin-cert" not registered Dec 10 22:50:22 crc kubenswrapper[4791]: E1210 22:50:22.836792 4791 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Dec 10 22:50:22 crc kubenswrapper[4791]: E1210 22:50:22.836814 4791 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Dec 10 22:50:22 crc kubenswrapper[4791]: E1210 22:50:22.836821 4791 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2025-12-10 22:50:54.83679854 +0000 UTC m=+89.266416213 (durationBeforeRetry 32s). Error: MountVolume.SetUp failed for volume "networking-console-plugin-cert" (UniqueName: "kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin-cert" not registered Dec 10 22:50:22 crc kubenswrapper[4791]: E1210 22:50:22.836826 4791 projected.go:194] Error preparing data for projected volume kube-api-access-s2dwl for pod openshift-network-diagnostics/network-check-source-55646444c4-trplf: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Dec 10 22:50:22 crc kubenswrapper[4791]: E1210 22:50:22.836859 4791 configmap.go:193] Couldn't get configMap openshift-network-console/networking-console-plugin: object "openshift-network-console"/"networking-console-plugin" not registered Dec 10 22:50:22 crc kubenswrapper[4791]: E1210 22:50:22.836868 4791 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl podName:9d751cbb-f2e2-430d-9754-c882a5e924a5 nodeName:}" failed. No retries permitted until 2025-12-10 22:50:54.836860162 +0000 UTC m=+89.266477895 (durationBeforeRetry 32s). Error: MountVolume.SetUp failed for volume "kube-api-access-s2dwl" (UniqueName: "kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl") pod "network-check-source-55646444c4-trplf" (UID: "9d751cbb-f2e2-430d-9754-c882a5e924a5") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Dec 10 22:50:22 crc kubenswrapper[4791]: E1210 22:50:22.836907 4791 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2025-12-10 22:50:54.836891263 +0000 UTC m=+89.266508976 (durationBeforeRetry 32s). Error: MountVolume.SetUp failed for volume "nginx-conf" (UniqueName: "kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin" not registered Dec 10 22:50:22 crc kubenswrapper[4791]: E1210 22:50:22.837014 4791 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Dec 10 22:50:22 crc kubenswrapper[4791]: E1210 22:50:22.837031 4791 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Dec 10 22:50:22 crc kubenswrapper[4791]: E1210 22:50:22.837042 4791 projected.go:194] Error preparing data for projected volume kube-api-access-cqllr for pod openshift-network-diagnostics/network-check-target-xd92c: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Dec 10 22:50:22 crc kubenswrapper[4791]: E1210 22:50:22.837079 4791 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr podName:3b6479f0-333b-4a96-9adf-2099afdc2447 nodeName:}" failed. No retries permitted until 2025-12-10 22:50:54.837070357 +0000 UTC m=+89.266688060 (durationBeforeRetry 32s). Error: MountVolume.SetUp failed for volume "kube-api-access-cqllr" (UniqueName: "kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr") pod "network-check-target-xd92c" (UID: "3b6479f0-333b-4a96-9adf-2099afdc2447") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Dec 10 22:50:22 crc kubenswrapper[4791]: I1210 22:50:22.884661 4791 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 10 22:50:22 crc kubenswrapper[4791]: I1210 22:50:22.884766 4791 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 10 22:50:22 crc kubenswrapper[4791]: E1210 22:50:22.884830 4791 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 10 22:50:22 crc kubenswrapper[4791]: I1210 22:50:22.884862 4791 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 10 22:50:22 crc kubenswrapper[4791]: I1210 22:50:22.884889 4791 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-b6kf6" Dec 10 22:50:22 crc kubenswrapper[4791]: E1210 22:50:22.885045 4791 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 10 22:50:22 crc kubenswrapper[4791]: E1210 22:50:22.885378 4791 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-b6kf6" podUID="b4527e71-bb58-4a39-a99c-fe614aa4c5f9" Dec 10 22:50:22 crc kubenswrapper[4791]: E1210 22:50:22.885237 4791 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 10 22:50:22 crc kubenswrapper[4791]: I1210 22:50:22.902560 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 22:50:22 crc kubenswrapper[4791]: I1210 22:50:22.902617 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 22:50:22 crc kubenswrapper[4791]: I1210 22:50:22.902629 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 22:50:22 crc kubenswrapper[4791]: I1210 22:50:22.902649 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 22:50:22 crc kubenswrapper[4791]: I1210 22:50:22.902661 4791 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T22:50:22Z","lastTransitionTime":"2025-12-10T22:50:22Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 22:50:23 crc kubenswrapper[4791]: I1210 22:50:23.005010 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 22:50:23 crc kubenswrapper[4791]: I1210 22:50:23.005069 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 22:50:23 crc kubenswrapper[4791]: I1210 22:50:23.005093 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 22:50:23 crc kubenswrapper[4791]: I1210 22:50:23.005116 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 22:50:23 crc kubenswrapper[4791]: I1210 22:50:23.005132 4791 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T22:50:23Z","lastTransitionTime":"2025-12-10T22:50:23Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 22:50:23 crc kubenswrapper[4791]: I1210 22:50:23.108430 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 22:50:23 crc kubenswrapper[4791]: I1210 22:50:23.108509 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 22:50:23 crc kubenswrapper[4791]: I1210 22:50:23.108534 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 22:50:23 crc kubenswrapper[4791]: I1210 22:50:23.108567 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 22:50:23 crc kubenswrapper[4791]: I1210 22:50:23.108590 4791 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T22:50:23Z","lastTransitionTime":"2025-12-10T22:50:23Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 22:50:23 crc kubenswrapper[4791]: I1210 22:50:23.213163 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 22:50:23 crc kubenswrapper[4791]: I1210 22:50:23.213254 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 22:50:23 crc kubenswrapper[4791]: I1210 22:50:23.213277 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 22:50:23 crc kubenswrapper[4791]: I1210 22:50:23.213311 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 22:50:23 crc kubenswrapper[4791]: I1210 22:50:23.213370 4791 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T22:50:23Z","lastTransitionTime":"2025-12-10T22:50:23Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 22:50:23 crc kubenswrapper[4791]: I1210 22:50:23.316297 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 22:50:23 crc kubenswrapper[4791]: I1210 22:50:23.316333 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 22:50:23 crc kubenswrapper[4791]: I1210 22:50:23.316373 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 22:50:23 crc kubenswrapper[4791]: I1210 22:50:23.316390 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 22:50:23 crc kubenswrapper[4791]: I1210 22:50:23.316402 4791 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T22:50:23Z","lastTransitionTime":"2025-12-10T22:50:23Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 22:50:23 crc kubenswrapper[4791]: I1210 22:50:23.420182 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 22:50:23 crc kubenswrapper[4791]: I1210 22:50:23.420248 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 22:50:23 crc kubenswrapper[4791]: I1210 22:50:23.420266 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 22:50:23 crc kubenswrapper[4791]: I1210 22:50:23.420291 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 22:50:23 crc kubenswrapper[4791]: I1210 22:50:23.420308 4791 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T22:50:23Z","lastTransitionTime":"2025-12-10T22:50:23Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 22:50:23 crc kubenswrapper[4791]: I1210 22:50:23.456718 4791 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-zhq64_3cd47739-0fa9-4321-aff1-220f8721a0b3/ovnkube-controller/2.log" Dec 10 22:50:23 crc kubenswrapper[4791]: I1210 22:50:23.462841 4791 scope.go:117] "RemoveContainer" containerID="b6d739a4314c4fb536810b8623f691d6aaf1090b9c83f17b484d2b22f7d3ad01" Dec 10 22:50:23 crc kubenswrapper[4791]: E1210 22:50:23.463120 4791 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"ovnkube-controller\" with CrashLoopBackOff: \"back-off 20s restarting failed container=ovnkube-controller pod=ovnkube-node-zhq64_openshift-ovn-kubernetes(3cd47739-0fa9-4321-aff1-220f8721a0b3)\"" pod="openshift-ovn-kubernetes/ovnkube-node-zhq64" podUID="3cd47739-0fa9-4321-aff1-220f8721a0b3" Dec 10 22:50:23 crc kubenswrapper[4791]: I1210 22:50:23.485284 4791 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-5rb5l" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ba35653c-6e06-4cee-a4d6-137764090d18\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T22:49:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T22:49:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T22:49:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T22:49:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://dbb5c875730a98ad2daf316cfafc08fcdce21ac16148154780ac9193ed405dab\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T22:49:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fvvrs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://185eb2ff44221a514878283ec873773a6c9a9073dc64e296f982e60dee1338e4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T22:49:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fvvrs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-10T22:49:51Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-5rb5l\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T22:50:23Z is after 2025-08-24T17:21:41Z" Dec 10 22:50:23 crc kubenswrapper[4791]: I1210 22:50:23.506765 4791 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-s6rk8" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5c9d0978-a6f9-49c5-b377-6619b4e3ff52\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T22:50:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T22:50:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T22:50:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T22:50:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8e55baaf460c794ab196e237d51d12de059aa038617dc2d5366d49c529f5d5c7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T22:50:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7vkgm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://51d1b9b086c60bae9223cb51d5a998d02790a0d91b745aab10392c23e617eb9f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T22:50:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7vkgm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-10T22:50:03Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-s6rk8\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T22:50:23Z is after 2025-08-24T17:21:41Z" Dec 10 22:50:23 crc kubenswrapper[4791]: I1210 22:50:23.522825 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 22:50:23 crc kubenswrapper[4791]: I1210 22:50:23.523175 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 22:50:23 crc kubenswrapper[4791]: I1210 22:50:23.523469 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 22:50:23 crc kubenswrapper[4791]: I1210 22:50:23.523709 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 22:50:23 crc kubenswrapper[4791]: I1210 22:50:23.523914 4791 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T22:50:23Z","lastTransitionTime":"2025-12-10T22:50:23Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 22:50:23 crc kubenswrapper[4791]: I1210 22:50:23.524794 4791 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-vczjq" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b9f00f25-3fa1-4b63-9710-fd352224b01b\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T22:49:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T22:49:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T22:49:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T22:49:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4698cf4c6b55b41e628448d37c2eb847c2777d0efbd93afcbf367d5e78623b7e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T22:49:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lrrpv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-10T22:49:50Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-vczjq\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T22:50:23Z is after 2025-08-24T17:21:41Z" Dec 10 22:50:23 crc kubenswrapper[4791]: I1210 22:50:23.547233 4791 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-4nwkq" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"672aa28c-8169-49ed-87b8-21187d13a80c\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T22:49:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T22:49:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T22:49:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T22:49:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a2f2a62315529e148a10a33ccb241e9417fbde0fe42afb61cb7ee2438c74636f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T22:49:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xl62j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-10T22:49:51Z\\\"}}\" for pod \"openshift-multus\"/\"multus-4nwkq\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T22:50:23Z is after 2025-08-24T17:21:41Z" Dec 10 22:50:23 crc kubenswrapper[4791]: I1210 22:50:23.576579 4791 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-zhq64" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3cd47739-0fa9-4321-aff1-220f8721a0b3\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T22:49:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T22:49:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T22:49:52Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T22:49:52Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://68795a29944e56870ffe94fa732e9dcf0bfb344c0d85d1e590f58a876129c61f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T22:49:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wvpw8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://bc721b530c08b433756e978ec0423cc18aabf6771b61eb96f903b4d57ea335ff\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T22:49:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wvpw8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://152a7618be8f38fadf4e564b3948980503cedcdb466c37344ccc20f4a3173a5f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T22:49:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wvpw8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3d18f9cb7207efb1e32de75a30d33e024f656a08f0503db95e5488bc5280971f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T22:49:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wvpw8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://937137ae9d7480cdc7c59bc4a1df0a4c10e72b3b63693af285278789953ed218\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T22:49:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wvpw8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://1a52d19b7671e8b143017c8a981a6f8d3ee3b84650e4b223b814a531ac76f364\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T22:49:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wvpw8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b6d739a4314c4fb536810b8623f691d6aaf1090b9c83f17b484d2b22f7d3ad01\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b6d739a4314c4fb536810b8623f691d6aaf1090b9c83f17b484d2b22f7d3ad01\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-12-10T22:50:21Z\\\",\\\"message\\\":\\\"llocateLoadBalancerNodePorts:nil,LoadBalancerClass:nil,InternalTrafficPolicy:*Cluster,TrafficDistribution:nil,},Status:ServiceStatus{LoadBalancer:LoadBalancerStatus{Ingress:[]LoadBalancerIngress{},},Conditions:[]Condition{},},}\\\\nI1210 22:50:21.614710 6424 metrics.go:553] Stopping metrics server at address \\\\\\\"127.0.0.1:29103\\\\\\\"\\\\nI1210 22:50:21.614731 6424 lb_config.go:1031] Cluster endpoints for openshift-kube-scheduler/scheduler for network=default are: map[]\\\\nI1210 22:50:21.614619 6424 services_controller.go:360] Finished syncing service ovn-kubernetes-control-plane on namespace openshift-ovn-kubernetes for network=default : 7.02µs\\\\nI1210 22:50:21.614757 6424 services_controller.go:356] Processing sync for service openshift-route-controller-manager/route-controller-manager for network=default\\\\nI1210 22:50:21.614615 6424 services_controller.go:444] Built service openshift-machine-config-operator/machine-config-daemon LB per-node configs for network=default: []services.lbConfig(nil)\\\\nF1210 22:50:21.614772 6424 ovnkube.go:137] failed to run ovnkube: [failed to start network controller: failed to start default network controller: unable to create admin network policy controller, err: could not add Event Handler for anpInformer during admin network policy controller initialization, handler {0x1fcc6e0 0x1fcc3c0 0x1fcc360} was not added to\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-10T22:50:20Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":2,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 20s restarting failed container=ovnkube-controller pod=ovnkube-node-zhq64_openshift-ovn-kubernetes(3cd47739-0fa9-4321-aff1-220f8721a0b3)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wvpw8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://dd3f80e5713221f96ed50da7f2e7fd5b9fabfcd5c1656da8d8a86d818a5fdc3f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T22:49:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wvpw8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3afbd1ffcd2c4b40324aca6e52d6fc822db131c3f35a9af3a6f005c2cffdb7bf\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3afbd1ffcd2c4b40324aca6e52d6fc822db131c3f35a9af3a6f005c2cffdb7bf\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-10T22:49:53Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-10T22:49:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wvpw8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-10T22:49:52Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-zhq64\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T22:50:23Z is after 2025-08-24T17:21:41Z" Dec 10 22:50:23 crc kubenswrapper[4791]: I1210 22:50:23.596184 4791 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"698bbe9a-fbe5-436d-a835-5e80df77ba24\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T22:49:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T22:49:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T22:49:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T22:49:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T22:49:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://224d73713312376a58ec6cc2bc8fdb6ffa44c4749ae0e7691d5fe27b9afca09f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T22:49:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://372eeaffd40992c6f06efc4df6dfcb44b31de64ee8256c21508cdd7c74ecb96e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T22:49:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://76b6013b062cac5b0b4d12a94f38ce7b43f9c7ea216063455d88a956ac04d26e\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T22:49:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://78e950e122e761d950da1a3c60c06476e53920e8795f6c61db22c01e22d5136e\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T22:49:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-10T22:49:25Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T22:50:23Z is after 2025-08-24T17:21:41Z" Dec 10 22:50:23 crc kubenswrapper[4791]: I1210 22:50:23.617857 4791 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-10T22:49:52Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-10T22:49:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-10T22:49:52Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3734b110802828632d44263ed0a43cabc4b710833f0be120d72db0940fe783e2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T22:49:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://794803ac9be3c52509e6dcb10854d8864e1a9e406762e57c7971b07c4da1220e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T22:49:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T22:50:23Z is after 2025-08-24T17:21:41Z" Dec 10 22:50:23 crc kubenswrapper[4791]: I1210 22:50:23.626895 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 22:50:23 crc kubenswrapper[4791]: I1210 22:50:23.626950 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 22:50:23 crc kubenswrapper[4791]: I1210 22:50:23.626970 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 22:50:23 crc kubenswrapper[4791]: I1210 22:50:23.626993 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 22:50:23 crc kubenswrapper[4791]: I1210 22:50:23.627009 4791 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T22:50:23Z","lastTransitionTime":"2025-12-10T22:50:23Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 22:50:23 crc kubenswrapper[4791]: I1210 22:50:23.634658 4791 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-tchxk" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"c0ab74e0-42a9-4f42-ac80-16030cf4ffff\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T22:49:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T22:49:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T22:49:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T22:49:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://04e0fd1b53256b7a938e8ccab93e2a4eb73dd1a59e0b9799e09c55e511baf7fe\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T22:49:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zn2qt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-10T22:49:50Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-tchxk\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T22:50:23Z is after 2025-08-24T17:21:41Z" Dec 10 22:50:23 crc kubenswrapper[4791]: I1210 22:50:23.658102 4791 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-rw584" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ff9beb90-69b0-4732-bf37-0b81f58ecc98\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T22:49:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T22:49:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T22:49:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T22:49:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ddb9df501dc6e8e5561fbc2aefd431037a3f28f73bf8e82a6598d261a4224807\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T22:49:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fksqn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ce67e54b696bb53ffbbe48c6ce71476b635e0868993e9c427b22f904891433ee\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ce67e54b696bb53ffbbe48c6ce71476b635e0868993e9c427b22f904891433ee\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-10T22:49:52Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-10T22:49:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fksqn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://818d89bfffb8a110670152c8480684665ea8db578e9768f524fbbe6a92f27c22\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://818d89bfffb8a110670152c8480684665ea8db578e9768f524fbbe6a92f27c22\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-10T22:49:53Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-10T22:49:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fksqn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://9aa9c06e894bb2a2ed3738defaeff68a0a0458ead48177f36b90770944c211a0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://9aa9c06e894bb2a2ed3738defaeff68a0a0458ead48177f36b90770944c211a0\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-10T22:49:54Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-10T22:49:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fksqn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://dfd6d03336dcea7529419a30e8d70d84e2290a0fa5b0e46d09ad341c190638d6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://dfd6d03336dcea7529419a30e8d70d84e2290a0fa5b0e46d09ad341c190638d6\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-10T22:49:55Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-10T22:49:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fksqn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b86738401ece6fbc3083782720c66a515850ed78d4054f34fcabab7ff525b85f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b86738401ece6fbc3083782720c66a515850ed78d4054f34fcabab7ff525b85f\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-10T22:49:56Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-10T22:49:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fksqn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://18b326f572a899ddd10197a745bf9264d4ddd975690ad1cfb0ec576ccd5cf102\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://18b326f572a899ddd10197a745bf9264d4ddd975690ad1cfb0ec576ccd5cf102\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-10T22:49:57Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-10T22:49:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fksqn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-10T22:49:51Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-rw584\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T22:50:23Z is after 2025-08-24T17:21:41Z" Dec 10 22:50:23 crc kubenswrapper[4791]: I1210 22:50:23.678105 4791 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-10T22:49:52Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-10T22:49:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-10T22:49:52Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://20651eb0c16e2c71222f56da1ff755cb0e7b084886acbbb24f0ccaf57041ed85\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T22:49:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T22:50:23Z is after 2025-08-24T17:21:41Z" Dec 10 22:50:23 crc kubenswrapper[4791]: I1210 22:50:23.696302 4791 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-10T22:49:50Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-10T22:49:50Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T22:50:23Z is after 2025-08-24T17:21:41Z" Dec 10 22:50:23 crc kubenswrapper[4791]: I1210 22:50:23.715125 4791 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-10T22:49:50Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-10T22:49:50Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T22:50:23Z is after 2025-08-24T17:21:41Z" Dec 10 22:50:23 crc kubenswrapper[4791]: I1210 22:50:23.730012 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 22:50:23 crc kubenswrapper[4791]: I1210 22:50:23.730052 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 22:50:23 crc kubenswrapper[4791]: I1210 22:50:23.730066 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 22:50:23 crc kubenswrapper[4791]: I1210 22:50:23.730081 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 22:50:23 crc kubenswrapper[4791]: I1210 22:50:23.730093 4791 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T22:50:23Z","lastTransitionTime":"2025-12-10T22:50:23Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 22:50:23 crc kubenswrapper[4791]: I1210 22:50:23.734716 4791 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-10T22:49:50Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-10T22:49:50Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T22:50:23Z is after 2025-08-24T17:21:41Z" Dec 10 22:50:23 crc kubenswrapper[4791]: I1210 22:50:23.750307 4791 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-10T22:49:54Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-10T22:49:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-10T22:49:54Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://62bfaf5d7a6d1b452cdaf88086447d0c8cb7306e94434e46752be5475124d56a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T22:49:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T22:50:23Z is after 2025-08-24T17:21:41Z" Dec 10 22:50:23 crc kubenswrapper[4791]: I1210 22:50:23.763837 4791 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-b6kf6" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b4527e71-bb58-4a39-a99c-fe614aa4c5f9\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T22:50:04Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T22:50:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T22:50:04Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T22:50:04Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ngpr4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ngpr4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-10T22:50:04Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-b6kf6\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T22:50:23Z is after 2025-08-24T17:21:41Z" Dec 10 22:50:23 crc kubenswrapper[4791]: I1210 22:50:23.779741 4791 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f8349fea-c49b-49a1-b5ee-032c5dffe021\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T22:49:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T22:49:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T22:49:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T22:49:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T22:49:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://acafe12186df99d1059a31260de0eafce35ff7771bcb8f5c6753b72665d6f60b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T22:49:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ea4c346507862a8c12756a3061d6e25e788a8d79f4621b62f81f4322185c82a8\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T22:49:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://2e6e324ec6f4f7bb69fdfab4ea3ac1aa55e480dd8e2a8984b0fe5b4c9f66636e\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T22:49:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://1b2e11ff2bf8789fa1d0d23ae648b3e6ef653106e5bc4bbb5dd6447f7cd6f05d\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d6aa7f8c5d908d90008f2df49860ac8a3480136da567e3d85f86135d779204f9\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-12-10T22:49:44Z\\\",\\\"message\\\":\\\"W1210 22:49:33.845733 1 cmd.go:257] Using insecure, self-signed certificates\\\\nI1210 22:49:33.846108 1 crypto.go:601] Generating new CA for check-endpoints-signer@1765406973 cert, and key in /tmp/serving-cert-2585701456/serving-signer.crt, /tmp/serving-cert-2585701456/serving-signer.key\\\\nI1210 22:49:34.529748 1 observer_polling.go:159] Starting file observer\\\\nW1210 22:49:34.532225 1 builder.go:272] unable to get owner reference (falling back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": dial tcp [::1]:6443: connect: connection refused\\\\nI1210 22:49:34.532599 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1210 22:49:34.533977 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-2585701456/tls.crt::/tmp/serving-cert-2585701456/tls.key\\\\\\\"\\\\nF1210 22:49:44.774710 1 cmd.go:182] error initializing delegating authentication: unable to load configmap based request-header-client-ca-file: Get \\\\\\\"https://localhost:6443/api/v1/namespaces/kube-system/configmaps/extension-apiserver-authentication\\\\\\\": net/http: TLS handshake timeout\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-10T22:49:33Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T22:49:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://0c4d036ab269f62561fe7ef1586012cb3a16e3583a08b9488a0616ec00c8e0ae\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T22:49:32Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c5856d6b26c19fda13b5c9ac2892a7e2f9c6f79feb8173ecfd9b33b673327d67\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c5856d6b26c19fda13b5c9ac2892a7e2f9c6f79feb8173ecfd9b33b673327d67\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-10T22:49:27Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-10T22:49:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-10T22:49:25Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T22:50:23Z is after 2025-08-24T17:21:41Z" Dec 10 22:50:23 crc kubenswrapper[4791]: I1210 22:50:23.794275 4791 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"02a949d8-38e2-4be3-989f-40a015407140\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T22:49:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T22:49:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T22:50:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T22:50:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T22:49:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://cd4158bc59b30f0f1716bf43c1ae7c9916a5b7d7164d074987fc1f19ba49b778\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T22:49:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://656065fd3cbdc6a25dab5d06c7baa00af16d8f62fc5da687efe9c6cc91945de0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T22:49:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://90ed7fff816b1772ba9ecd95587a59cbed7c30ac302fff7e7c5de934c912bcf4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T22:49:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9b7a5502803eaad9f76f5ffa91b29fbb2657402065a03a4d6433c096aac7264f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"wait-for-host-port\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://9b7a5502803eaad9f76f5ffa91b29fbb2657402065a03a4d6433c096aac7264f\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-10T22:49:27Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-10T22:49:27Z\\\"}}}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-10T22:49:25Z\\\"}}\" for pod \"openshift-kube-scheduler\"/\"openshift-kube-scheduler-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T22:50:23Z is after 2025-08-24T17:21:41Z" Dec 10 22:50:23 crc kubenswrapper[4791]: I1210 22:50:23.820200 4791 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"90250e13-6c63-443a-a5e5-6e4e341a2289\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T22:49:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T22:49:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T22:49:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T22:49:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T22:49:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ba54f95b1e00843b854db9a10fe6897592159ccacf30ed0f76002093a13eedd3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T22:49:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5aa8ffa518f3a68b60b275f8df4851b2564aa4984b61efdc9470b73072133e43\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T22:49:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://16f989503edc78d9451638e2cf6cf4b19a2cf2b8809b10b237541cb829303091\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T22:49:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://82556824d63ce6e46df14c45689ff1f81ebf729abe902061815c040b323898e9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T22:49:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://61db531940fa6f2effdae5a005b4c91bd2698bd5d1bff9588c171a65ad550eca\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T22:49:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0f65c864d885319ce2a08d61c30be24b3ffb7d23ea30c82ed7faf039cb2c006d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://0f65c864d885319ce2a08d61c30be24b3ffb7d23ea30c82ed7faf039cb2c006d\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-10T22:49:27Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-10T22:49:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e31a76d1d5b2812ba83e24e38268e25ec8bed98fc3da9adcd06bda5822b6ace1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e31a76d1d5b2812ba83e24e38268e25ec8bed98fc3da9adcd06bda5822b6ace1\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-10T22:49:30Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-10T22:49:29Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://b5b5eceae562f3986b543d41de1e977efb85a91d84c98e09159ca6d092eda691\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b5b5eceae562f3986b543d41de1e977efb85a91d84c98e09159ca6d092eda691\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-10T22:49:31Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-10T22:49:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-10T22:49:25Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T22:50:23Z is after 2025-08-24T17:21:41Z" Dec 10 22:50:23 crc kubenswrapper[4791]: I1210 22:50:23.832970 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 22:50:23 crc kubenswrapper[4791]: I1210 22:50:23.833232 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 22:50:23 crc kubenswrapper[4791]: I1210 22:50:23.833357 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 22:50:23 crc kubenswrapper[4791]: I1210 22:50:23.833482 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 22:50:23 crc kubenswrapper[4791]: I1210 22:50:23.833573 4791 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T22:50:23Z","lastTransitionTime":"2025-12-10T22:50:23Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 22:50:23 crc kubenswrapper[4791]: I1210 22:50:23.936185 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 22:50:23 crc kubenswrapper[4791]: I1210 22:50:23.936567 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 22:50:23 crc kubenswrapper[4791]: I1210 22:50:23.936718 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 22:50:23 crc kubenswrapper[4791]: I1210 22:50:23.936860 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 22:50:23 crc kubenswrapper[4791]: I1210 22:50:23.936989 4791 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T22:50:23Z","lastTransitionTime":"2025-12-10T22:50:23Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 22:50:24 crc kubenswrapper[4791]: I1210 22:50:24.039851 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 22:50:24 crc kubenswrapper[4791]: I1210 22:50:24.039911 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 22:50:24 crc kubenswrapper[4791]: I1210 22:50:24.039930 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 22:50:24 crc kubenswrapper[4791]: I1210 22:50:24.039953 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 22:50:24 crc kubenswrapper[4791]: I1210 22:50:24.039970 4791 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T22:50:24Z","lastTransitionTime":"2025-12-10T22:50:24Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 22:50:24 crc kubenswrapper[4791]: I1210 22:50:24.142657 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 22:50:24 crc kubenswrapper[4791]: I1210 22:50:24.142707 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 22:50:24 crc kubenswrapper[4791]: I1210 22:50:24.142725 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 22:50:24 crc kubenswrapper[4791]: I1210 22:50:24.142746 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 22:50:24 crc kubenswrapper[4791]: I1210 22:50:24.142762 4791 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T22:50:24Z","lastTransitionTime":"2025-12-10T22:50:24Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 22:50:24 crc kubenswrapper[4791]: I1210 22:50:24.246117 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 22:50:24 crc kubenswrapper[4791]: I1210 22:50:24.246168 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 22:50:24 crc kubenswrapper[4791]: I1210 22:50:24.246185 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 22:50:24 crc kubenswrapper[4791]: I1210 22:50:24.246206 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 22:50:24 crc kubenswrapper[4791]: I1210 22:50:24.246221 4791 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T22:50:24Z","lastTransitionTime":"2025-12-10T22:50:24Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 22:50:24 crc kubenswrapper[4791]: I1210 22:50:24.348587 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 22:50:24 crc kubenswrapper[4791]: I1210 22:50:24.348626 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 22:50:24 crc kubenswrapper[4791]: I1210 22:50:24.348638 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 22:50:24 crc kubenswrapper[4791]: I1210 22:50:24.348656 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 22:50:24 crc kubenswrapper[4791]: I1210 22:50:24.348667 4791 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T22:50:24Z","lastTransitionTime":"2025-12-10T22:50:24Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 22:50:24 crc kubenswrapper[4791]: I1210 22:50:24.452640 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 22:50:24 crc kubenswrapper[4791]: I1210 22:50:24.452702 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 22:50:24 crc kubenswrapper[4791]: I1210 22:50:24.452721 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 22:50:24 crc kubenswrapper[4791]: I1210 22:50:24.452745 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 22:50:24 crc kubenswrapper[4791]: I1210 22:50:24.452763 4791 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T22:50:24Z","lastTransitionTime":"2025-12-10T22:50:24Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 22:50:24 crc kubenswrapper[4791]: I1210 22:50:24.556240 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 22:50:24 crc kubenswrapper[4791]: I1210 22:50:24.556309 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 22:50:24 crc kubenswrapper[4791]: I1210 22:50:24.556332 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 22:50:24 crc kubenswrapper[4791]: I1210 22:50:24.556395 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 22:50:24 crc kubenswrapper[4791]: I1210 22:50:24.556418 4791 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T22:50:24Z","lastTransitionTime":"2025-12-10T22:50:24Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 22:50:24 crc kubenswrapper[4791]: I1210 22:50:24.659117 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 22:50:24 crc kubenswrapper[4791]: I1210 22:50:24.659200 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 22:50:24 crc kubenswrapper[4791]: I1210 22:50:24.659226 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 22:50:24 crc kubenswrapper[4791]: I1210 22:50:24.659262 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 22:50:24 crc kubenswrapper[4791]: I1210 22:50:24.659287 4791 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T22:50:24Z","lastTransitionTime":"2025-12-10T22:50:24Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 22:50:24 crc kubenswrapper[4791]: I1210 22:50:24.762926 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 22:50:24 crc kubenswrapper[4791]: I1210 22:50:24.762970 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 22:50:24 crc kubenswrapper[4791]: I1210 22:50:24.762981 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 22:50:24 crc kubenswrapper[4791]: I1210 22:50:24.762997 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 22:50:24 crc kubenswrapper[4791]: I1210 22:50:24.763008 4791 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T22:50:24Z","lastTransitionTime":"2025-12-10T22:50:24Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 22:50:24 crc kubenswrapper[4791]: I1210 22:50:24.865045 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 22:50:24 crc kubenswrapper[4791]: I1210 22:50:24.865107 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 22:50:24 crc kubenswrapper[4791]: I1210 22:50:24.865122 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 22:50:24 crc kubenswrapper[4791]: I1210 22:50:24.865164 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 22:50:24 crc kubenswrapper[4791]: I1210 22:50:24.865178 4791 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T22:50:24Z","lastTransitionTime":"2025-12-10T22:50:24Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 22:50:24 crc kubenswrapper[4791]: I1210 22:50:24.884507 4791 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-b6kf6" Dec 10 22:50:24 crc kubenswrapper[4791]: I1210 22:50:24.884554 4791 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 10 22:50:24 crc kubenswrapper[4791]: I1210 22:50:24.884590 4791 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 10 22:50:24 crc kubenswrapper[4791]: I1210 22:50:24.884554 4791 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 10 22:50:24 crc kubenswrapper[4791]: E1210 22:50:24.884695 4791 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-b6kf6" podUID="b4527e71-bb58-4a39-a99c-fe614aa4c5f9" Dec 10 22:50:24 crc kubenswrapper[4791]: E1210 22:50:24.884871 4791 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 10 22:50:24 crc kubenswrapper[4791]: E1210 22:50:24.885005 4791 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 10 22:50:24 crc kubenswrapper[4791]: E1210 22:50:24.885078 4791 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 10 22:50:24 crc kubenswrapper[4791]: I1210 22:50:24.967215 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 22:50:24 crc kubenswrapper[4791]: I1210 22:50:24.967256 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 22:50:24 crc kubenswrapper[4791]: I1210 22:50:24.967264 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 22:50:24 crc kubenswrapper[4791]: I1210 22:50:24.967276 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 22:50:24 crc kubenswrapper[4791]: I1210 22:50:24.967285 4791 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T22:50:24Z","lastTransitionTime":"2025-12-10T22:50:24Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 22:50:25 crc kubenswrapper[4791]: I1210 22:50:25.070210 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 22:50:25 crc kubenswrapper[4791]: I1210 22:50:25.070262 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 22:50:25 crc kubenswrapper[4791]: I1210 22:50:25.070273 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 22:50:25 crc kubenswrapper[4791]: I1210 22:50:25.070288 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 22:50:25 crc kubenswrapper[4791]: I1210 22:50:25.070301 4791 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T22:50:25Z","lastTransitionTime":"2025-12-10T22:50:25Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 22:50:25 crc kubenswrapper[4791]: I1210 22:50:25.173548 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 22:50:25 crc kubenswrapper[4791]: I1210 22:50:25.173598 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 22:50:25 crc kubenswrapper[4791]: I1210 22:50:25.173612 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 22:50:25 crc kubenswrapper[4791]: I1210 22:50:25.173632 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 22:50:25 crc kubenswrapper[4791]: I1210 22:50:25.173645 4791 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T22:50:25Z","lastTransitionTime":"2025-12-10T22:50:25Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 22:50:25 crc kubenswrapper[4791]: I1210 22:50:25.276119 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 22:50:25 crc kubenswrapper[4791]: I1210 22:50:25.276191 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 22:50:25 crc kubenswrapper[4791]: I1210 22:50:25.276215 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 22:50:25 crc kubenswrapper[4791]: I1210 22:50:25.276244 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 22:50:25 crc kubenswrapper[4791]: I1210 22:50:25.276267 4791 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T22:50:25Z","lastTransitionTime":"2025-12-10T22:50:25Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 22:50:25 crc kubenswrapper[4791]: I1210 22:50:25.379387 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 22:50:25 crc kubenswrapper[4791]: I1210 22:50:25.379450 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 22:50:25 crc kubenswrapper[4791]: I1210 22:50:25.379461 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 22:50:25 crc kubenswrapper[4791]: I1210 22:50:25.379482 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 22:50:25 crc kubenswrapper[4791]: I1210 22:50:25.379495 4791 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T22:50:25Z","lastTransitionTime":"2025-12-10T22:50:25Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 22:50:25 crc kubenswrapper[4791]: I1210 22:50:25.482288 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 22:50:25 crc kubenswrapper[4791]: I1210 22:50:25.482324 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 22:50:25 crc kubenswrapper[4791]: I1210 22:50:25.482332 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 22:50:25 crc kubenswrapper[4791]: I1210 22:50:25.482360 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 22:50:25 crc kubenswrapper[4791]: I1210 22:50:25.482374 4791 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T22:50:25Z","lastTransitionTime":"2025-12-10T22:50:25Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 22:50:25 crc kubenswrapper[4791]: I1210 22:50:25.585205 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 22:50:25 crc kubenswrapper[4791]: I1210 22:50:25.585259 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 22:50:25 crc kubenswrapper[4791]: I1210 22:50:25.585272 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 22:50:25 crc kubenswrapper[4791]: I1210 22:50:25.585289 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 22:50:25 crc kubenswrapper[4791]: I1210 22:50:25.585302 4791 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T22:50:25Z","lastTransitionTime":"2025-12-10T22:50:25Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 22:50:25 crc kubenswrapper[4791]: I1210 22:50:25.687675 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 22:50:25 crc kubenswrapper[4791]: I1210 22:50:25.687743 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 22:50:25 crc kubenswrapper[4791]: I1210 22:50:25.687762 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 22:50:25 crc kubenswrapper[4791]: I1210 22:50:25.687784 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 22:50:25 crc kubenswrapper[4791]: I1210 22:50:25.687801 4791 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T22:50:25Z","lastTransitionTime":"2025-12-10T22:50:25Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 22:50:25 crc kubenswrapper[4791]: I1210 22:50:25.790743 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 22:50:25 crc kubenswrapper[4791]: I1210 22:50:25.790814 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 22:50:25 crc kubenswrapper[4791]: I1210 22:50:25.790832 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 22:50:25 crc kubenswrapper[4791]: I1210 22:50:25.790880 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 22:50:25 crc kubenswrapper[4791]: I1210 22:50:25.790898 4791 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T22:50:25Z","lastTransitionTime":"2025-12-10T22:50:25Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 22:50:25 crc kubenswrapper[4791]: I1210 22:50:25.892948 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 22:50:25 crc kubenswrapper[4791]: I1210 22:50:25.892986 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 22:50:25 crc kubenswrapper[4791]: I1210 22:50:25.892995 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 22:50:25 crc kubenswrapper[4791]: I1210 22:50:25.893009 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 22:50:25 crc kubenswrapper[4791]: I1210 22:50:25.893017 4791 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T22:50:25Z","lastTransitionTime":"2025-12-10T22:50:25Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 22:50:25 crc kubenswrapper[4791]: I1210 22:50:25.900733 4791 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-b6kf6" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b4527e71-bb58-4a39-a99c-fe614aa4c5f9\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T22:50:04Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T22:50:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T22:50:04Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T22:50:04Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ngpr4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ngpr4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-10T22:50:04Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-b6kf6\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T22:50:25Z is after 2025-08-24T17:21:41Z" Dec 10 22:50:25 crc kubenswrapper[4791]: I1210 22:50:25.914765 4791 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f8349fea-c49b-49a1-b5ee-032c5dffe021\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T22:49:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T22:49:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T22:49:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T22:49:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T22:49:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://acafe12186df99d1059a31260de0eafce35ff7771bcb8f5c6753b72665d6f60b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T22:49:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ea4c346507862a8c12756a3061d6e25e788a8d79f4621b62f81f4322185c82a8\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T22:49:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://2e6e324ec6f4f7bb69fdfab4ea3ac1aa55e480dd8e2a8984b0fe5b4c9f66636e\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T22:49:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://1b2e11ff2bf8789fa1d0d23ae648b3e6ef653106e5bc4bbb5dd6447f7cd6f05d\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d6aa7f8c5d908d90008f2df49860ac8a3480136da567e3d85f86135d779204f9\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-12-10T22:49:44Z\\\",\\\"message\\\":\\\"W1210 22:49:33.845733 1 cmd.go:257] Using insecure, self-signed certificates\\\\nI1210 22:49:33.846108 1 crypto.go:601] Generating new CA for check-endpoints-signer@1765406973 cert, and key in /tmp/serving-cert-2585701456/serving-signer.crt, /tmp/serving-cert-2585701456/serving-signer.key\\\\nI1210 22:49:34.529748 1 observer_polling.go:159] Starting file observer\\\\nW1210 22:49:34.532225 1 builder.go:272] unable to get owner reference (falling back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": dial tcp [::1]:6443: connect: connection refused\\\\nI1210 22:49:34.532599 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1210 22:49:34.533977 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-2585701456/tls.crt::/tmp/serving-cert-2585701456/tls.key\\\\\\\"\\\\nF1210 22:49:44.774710 1 cmd.go:182] error initializing delegating authentication: unable to load configmap based request-header-client-ca-file: Get \\\\\\\"https://localhost:6443/api/v1/namespaces/kube-system/configmaps/extension-apiserver-authentication\\\\\\\": net/http: TLS handshake timeout\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-10T22:49:33Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T22:49:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://0c4d036ab269f62561fe7ef1586012cb3a16e3583a08b9488a0616ec00c8e0ae\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T22:49:32Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c5856d6b26c19fda13b5c9ac2892a7e2f9c6f79feb8173ecfd9b33b673327d67\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c5856d6b26c19fda13b5c9ac2892a7e2f9c6f79feb8173ecfd9b33b673327d67\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-10T22:49:27Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-10T22:49:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-10T22:49:25Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T22:50:25Z is after 2025-08-24T17:21:41Z" Dec 10 22:50:25 crc kubenswrapper[4791]: I1210 22:50:25.930568 4791 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"02a949d8-38e2-4be3-989f-40a015407140\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T22:49:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T22:49:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T22:50:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T22:50:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T22:49:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://cd4158bc59b30f0f1716bf43c1ae7c9916a5b7d7164d074987fc1f19ba49b778\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T22:49:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://656065fd3cbdc6a25dab5d06c7baa00af16d8f62fc5da687efe9c6cc91945de0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T22:49:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://90ed7fff816b1772ba9ecd95587a59cbed7c30ac302fff7e7c5de934c912bcf4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T22:49:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9b7a5502803eaad9f76f5ffa91b29fbb2657402065a03a4d6433c096aac7264f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"wait-for-host-port\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://9b7a5502803eaad9f76f5ffa91b29fbb2657402065a03a4d6433c096aac7264f\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-10T22:49:27Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-10T22:49:27Z\\\"}}}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-10T22:49:25Z\\\"}}\" for pod \"openshift-kube-scheduler\"/\"openshift-kube-scheduler-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T22:50:25Z is after 2025-08-24T17:21:41Z" Dec 10 22:50:25 crc kubenswrapper[4791]: I1210 22:50:25.954103 4791 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"90250e13-6c63-443a-a5e5-6e4e341a2289\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T22:49:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T22:49:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T22:49:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T22:49:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T22:49:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ba54f95b1e00843b854db9a10fe6897592159ccacf30ed0f76002093a13eedd3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T22:49:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5aa8ffa518f3a68b60b275f8df4851b2564aa4984b61efdc9470b73072133e43\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T22:49:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://16f989503edc78d9451638e2cf6cf4b19a2cf2b8809b10b237541cb829303091\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T22:49:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://82556824d63ce6e46df14c45689ff1f81ebf729abe902061815c040b323898e9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T22:49:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://61db531940fa6f2effdae5a005b4c91bd2698bd5d1bff9588c171a65ad550eca\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T22:49:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0f65c864d885319ce2a08d61c30be24b3ffb7d23ea30c82ed7faf039cb2c006d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://0f65c864d885319ce2a08d61c30be24b3ffb7d23ea30c82ed7faf039cb2c006d\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-10T22:49:27Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-10T22:49:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e31a76d1d5b2812ba83e24e38268e25ec8bed98fc3da9adcd06bda5822b6ace1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e31a76d1d5b2812ba83e24e38268e25ec8bed98fc3da9adcd06bda5822b6ace1\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-10T22:49:30Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-10T22:49:29Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://b5b5eceae562f3986b543d41de1e977efb85a91d84c98e09159ca6d092eda691\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b5b5eceae562f3986b543d41de1e977efb85a91d84c98e09159ca6d092eda691\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-10T22:49:31Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-10T22:49:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-10T22:49:25Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T22:50:25Z is after 2025-08-24T17:21:41Z" Dec 10 22:50:25 crc kubenswrapper[4791]: I1210 22:50:25.971878 4791 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-10T22:49:50Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-10T22:49:50Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T22:50:25Z is after 2025-08-24T17:21:41Z" Dec 10 22:50:25 crc kubenswrapper[4791]: I1210 22:50:25.985582 4791 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-10T22:49:54Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-10T22:49:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-10T22:49:54Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://62bfaf5d7a6d1b452cdaf88086447d0c8cb7306e94434e46752be5475124d56a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T22:49:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T22:50:25Z is after 2025-08-24T17:21:41Z" Dec 10 22:50:25 crc kubenswrapper[4791]: I1210 22:50:25.997018 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 22:50:25 crc kubenswrapper[4791]: I1210 22:50:25.997068 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 22:50:25 crc kubenswrapper[4791]: I1210 22:50:25.997081 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 22:50:25 crc kubenswrapper[4791]: I1210 22:50:25.997098 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 22:50:25 crc kubenswrapper[4791]: I1210 22:50:25.997109 4791 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T22:50:25Z","lastTransitionTime":"2025-12-10T22:50:25Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 22:50:25 crc kubenswrapper[4791]: I1210 22:50:25.998662 4791 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-5rb5l" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ba35653c-6e06-4cee-a4d6-137764090d18\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T22:49:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T22:49:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T22:49:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T22:49:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://dbb5c875730a98ad2daf316cfafc08fcdce21ac16148154780ac9193ed405dab\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T22:49:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fvvrs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://185eb2ff44221a514878283ec873773a6c9a9073dc64e296f982e60dee1338e4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T22:49:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fvvrs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-10T22:49:51Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-5rb5l\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T22:50:25Z is after 2025-08-24T17:21:41Z" Dec 10 22:50:26 crc kubenswrapper[4791]: I1210 22:50:26.010504 4791 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-s6rk8" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5c9d0978-a6f9-49c5-b377-6619b4e3ff52\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T22:50:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T22:50:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T22:50:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T22:50:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8e55baaf460c794ab196e237d51d12de059aa038617dc2d5366d49c529f5d5c7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T22:50:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7vkgm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://51d1b9b086c60bae9223cb51d5a998d02790a0d91b745aab10392c23e617eb9f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T22:50:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7vkgm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-10T22:50:03Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-s6rk8\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T22:50:26Z is after 2025-08-24T17:21:41Z" Dec 10 22:50:26 crc kubenswrapper[4791]: I1210 22:50:26.028626 4791 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-zhq64" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3cd47739-0fa9-4321-aff1-220f8721a0b3\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T22:49:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T22:49:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T22:49:52Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T22:49:52Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://68795a29944e56870ffe94fa732e9dcf0bfb344c0d85d1e590f58a876129c61f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T22:49:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wvpw8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://bc721b530c08b433756e978ec0423cc18aabf6771b61eb96f903b4d57ea335ff\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T22:49:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wvpw8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://152a7618be8f38fadf4e564b3948980503cedcdb466c37344ccc20f4a3173a5f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T22:49:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wvpw8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3d18f9cb7207efb1e32de75a30d33e024f656a08f0503db95e5488bc5280971f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T22:49:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wvpw8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://937137ae9d7480cdc7c59bc4a1df0a4c10e72b3b63693af285278789953ed218\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T22:49:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wvpw8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://1a52d19b7671e8b143017c8a981a6f8d3ee3b84650e4b223b814a531ac76f364\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T22:49:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wvpw8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b6d739a4314c4fb536810b8623f691d6aaf1090b9c83f17b484d2b22f7d3ad01\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b6d739a4314c4fb536810b8623f691d6aaf1090b9c83f17b484d2b22f7d3ad01\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-12-10T22:50:21Z\\\",\\\"message\\\":\\\"llocateLoadBalancerNodePorts:nil,LoadBalancerClass:nil,InternalTrafficPolicy:*Cluster,TrafficDistribution:nil,},Status:ServiceStatus{LoadBalancer:LoadBalancerStatus{Ingress:[]LoadBalancerIngress{},},Conditions:[]Condition{},},}\\\\nI1210 22:50:21.614710 6424 metrics.go:553] Stopping metrics server at address \\\\\\\"127.0.0.1:29103\\\\\\\"\\\\nI1210 22:50:21.614731 6424 lb_config.go:1031] Cluster endpoints for openshift-kube-scheduler/scheduler for network=default are: map[]\\\\nI1210 22:50:21.614619 6424 services_controller.go:360] Finished syncing service ovn-kubernetes-control-plane on namespace openshift-ovn-kubernetes for network=default : 7.02µs\\\\nI1210 22:50:21.614757 6424 services_controller.go:356] Processing sync for service openshift-route-controller-manager/route-controller-manager for network=default\\\\nI1210 22:50:21.614615 6424 services_controller.go:444] Built service openshift-machine-config-operator/machine-config-daemon LB per-node configs for network=default: []services.lbConfig(nil)\\\\nF1210 22:50:21.614772 6424 ovnkube.go:137] failed to run ovnkube: [failed to start network controller: failed to start default network controller: unable to create admin network policy controller, err: could not add Event Handler for anpInformer during admin network policy controller initialization, handler {0x1fcc6e0 0x1fcc3c0 0x1fcc360} was not added to\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-10T22:50:20Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":2,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 20s restarting failed container=ovnkube-controller pod=ovnkube-node-zhq64_openshift-ovn-kubernetes(3cd47739-0fa9-4321-aff1-220f8721a0b3)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wvpw8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://dd3f80e5713221f96ed50da7f2e7fd5b9fabfcd5c1656da8d8a86d818a5fdc3f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T22:49:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wvpw8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3afbd1ffcd2c4b40324aca6e52d6fc822db131c3f35a9af3a6f005c2cffdb7bf\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3afbd1ffcd2c4b40324aca6e52d6fc822db131c3f35a9af3a6f005c2cffdb7bf\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-10T22:49:53Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-10T22:49:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wvpw8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-10T22:49:52Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-zhq64\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T22:50:26Z is after 2025-08-24T17:21:41Z" Dec 10 22:50:26 crc kubenswrapper[4791]: I1210 22:50:26.040972 4791 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"698bbe9a-fbe5-436d-a835-5e80df77ba24\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T22:49:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T22:49:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T22:49:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T22:49:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T22:49:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://224d73713312376a58ec6cc2bc8fdb6ffa44c4749ae0e7691d5fe27b9afca09f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T22:49:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://372eeaffd40992c6f06efc4df6dfcb44b31de64ee8256c21508cdd7c74ecb96e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T22:49:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://76b6013b062cac5b0b4d12a94f38ce7b43f9c7ea216063455d88a956ac04d26e\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T22:49:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://78e950e122e761d950da1a3c60c06476e53920e8795f6c61db22c01e22d5136e\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T22:49:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-10T22:49:25Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T22:50:26Z is after 2025-08-24T17:21:41Z" Dec 10 22:50:26 crc kubenswrapper[4791]: I1210 22:50:26.053192 4791 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-10T22:49:52Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-10T22:49:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-10T22:49:52Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3734b110802828632d44263ed0a43cabc4b710833f0be120d72db0940fe783e2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T22:49:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://794803ac9be3c52509e6dcb10854d8864e1a9e406762e57c7971b07c4da1220e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T22:49:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T22:50:26Z is after 2025-08-24T17:21:41Z" Dec 10 22:50:26 crc kubenswrapper[4791]: I1210 22:50:26.063789 4791 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-vczjq" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b9f00f25-3fa1-4b63-9710-fd352224b01b\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T22:49:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T22:49:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T22:49:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T22:49:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4698cf4c6b55b41e628448d37c2eb847c2777d0efbd93afcbf367d5e78623b7e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T22:49:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lrrpv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-10T22:49:50Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-vczjq\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T22:50:26Z is after 2025-08-24T17:21:41Z" Dec 10 22:50:26 crc kubenswrapper[4791]: I1210 22:50:26.076979 4791 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-4nwkq" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"672aa28c-8169-49ed-87b8-21187d13a80c\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T22:49:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T22:49:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T22:49:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T22:49:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a2f2a62315529e148a10a33ccb241e9417fbde0fe42afb61cb7ee2438c74636f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T22:49:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xl62j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-10T22:49:51Z\\\"}}\" for pod \"openshift-multus\"/\"multus-4nwkq\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T22:50:26Z is after 2025-08-24T17:21:41Z" Dec 10 22:50:26 crc kubenswrapper[4791]: I1210 22:50:26.089699 4791 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-10T22:49:52Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-10T22:49:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-10T22:49:52Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://20651eb0c16e2c71222f56da1ff755cb0e7b084886acbbb24f0ccaf57041ed85\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T22:49:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T22:50:26Z is after 2025-08-24T17:21:41Z" Dec 10 22:50:26 crc kubenswrapper[4791]: I1210 22:50:26.099963 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 22:50:26 crc kubenswrapper[4791]: I1210 22:50:26.100004 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 22:50:26 crc kubenswrapper[4791]: I1210 22:50:26.100018 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 22:50:26 crc kubenswrapper[4791]: I1210 22:50:26.100034 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 22:50:26 crc kubenswrapper[4791]: I1210 22:50:26.100047 4791 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T22:50:26Z","lastTransitionTime":"2025-12-10T22:50:26Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 22:50:26 crc kubenswrapper[4791]: I1210 22:50:26.103275 4791 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-10T22:49:50Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-10T22:49:50Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T22:50:26Z is after 2025-08-24T17:21:41Z" Dec 10 22:50:26 crc kubenswrapper[4791]: I1210 22:50:26.114658 4791 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-10T22:49:50Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-10T22:49:50Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T22:50:26Z is after 2025-08-24T17:21:41Z" Dec 10 22:50:26 crc kubenswrapper[4791]: I1210 22:50:26.126477 4791 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-tchxk" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"c0ab74e0-42a9-4f42-ac80-16030cf4ffff\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T22:49:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T22:49:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T22:49:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T22:49:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://04e0fd1b53256b7a938e8ccab93e2a4eb73dd1a59e0b9799e09c55e511baf7fe\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T22:49:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zn2qt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-10T22:49:50Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-tchxk\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T22:50:26Z is after 2025-08-24T17:21:41Z" Dec 10 22:50:26 crc kubenswrapper[4791]: I1210 22:50:26.141668 4791 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-rw584" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ff9beb90-69b0-4732-bf37-0b81f58ecc98\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T22:49:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T22:49:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T22:49:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T22:49:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ddb9df501dc6e8e5561fbc2aefd431037a3f28f73bf8e82a6598d261a4224807\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T22:49:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fksqn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ce67e54b696bb53ffbbe48c6ce71476b635e0868993e9c427b22f904891433ee\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ce67e54b696bb53ffbbe48c6ce71476b635e0868993e9c427b22f904891433ee\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-10T22:49:52Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-10T22:49:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fksqn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://818d89bfffb8a110670152c8480684665ea8db578e9768f524fbbe6a92f27c22\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://818d89bfffb8a110670152c8480684665ea8db578e9768f524fbbe6a92f27c22\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-10T22:49:53Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-10T22:49:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fksqn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://9aa9c06e894bb2a2ed3738defaeff68a0a0458ead48177f36b90770944c211a0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://9aa9c06e894bb2a2ed3738defaeff68a0a0458ead48177f36b90770944c211a0\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-10T22:49:54Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-10T22:49:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fksqn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://dfd6d03336dcea7529419a30e8d70d84e2290a0fa5b0e46d09ad341c190638d6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://dfd6d03336dcea7529419a30e8d70d84e2290a0fa5b0e46d09ad341c190638d6\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-10T22:49:55Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-10T22:49:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fksqn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b86738401ece6fbc3083782720c66a515850ed78d4054f34fcabab7ff525b85f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b86738401ece6fbc3083782720c66a515850ed78d4054f34fcabab7ff525b85f\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-10T22:49:56Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-10T22:49:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fksqn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://18b326f572a899ddd10197a745bf9264d4ddd975690ad1cfb0ec576ccd5cf102\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://18b326f572a899ddd10197a745bf9264d4ddd975690ad1cfb0ec576ccd5cf102\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-10T22:49:57Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-10T22:49:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fksqn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-10T22:49:51Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-rw584\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T22:50:26Z is after 2025-08-24T17:21:41Z" Dec 10 22:50:26 crc kubenswrapper[4791]: I1210 22:50:26.201998 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 22:50:26 crc kubenswrapper[4791]: I1210 22:50:26.202026 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 22:50:26 crc kubenswrapper[4791]: I1210 22:50:26.202034 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 22:50:26 crc kubenswrapper[4791]: I1210 22:50:26.202046 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 22:50:26 crc kubenswrapper[4791]: I1210 22:50:26.202059 4791 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T22:50:26Z","lastTransitionTime":"2025-12-10T22:50:26Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 22:50:26 crc kubenswrapper[4791]: I1210 22:50:26.304780 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 22:50:26 crc kubenswrapper[4791]: I1210 22:50:26.305010 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 22:50:26 crc kubenswrapper[4791]: I1210 22:50:26.305090 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 22:50:26 crc kubenswrapper[4791]: I1210 22:50:26.305174 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 22:50:26 crc kubenswrapper[4791]: I1210 22:50:26.305254 4791 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T22:50:26Z","lastTransitionTime":"2025-12-10T22:50:26Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 22:50:26 crc kubenswrapper[4791]: I1210 22:50:26.408520 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 22:50:26 crc kubenswrapper[4791]: I1210 22:50:26.408948 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 22:50:26 crc kubenswrapper[4791]: I1210 22:50:26.409178 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 22:50:26 crc kubenswrapper[4791]: I1210 22:50:26.409437 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 22:50:26 crc kubenswrapper[4791]: I1210 22:50:26.409663 4791 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T22:50:26Z","lastTransitionTime":"2025-12-10T22:50:26Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 22:50:26 crc kubenswrapper[4791]: I1210 22:50:26.513271 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 22:50:26 crc kubenswrapper[4791]: I1210 22:50:26.513678 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 22:50:26 crc kubenswrapper[4791]: I1210 22:50:26.513915 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 22:50:26 crc kubenswrapper[4791]: I1210 22:50:26.514106 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 22:50:26 crc kubenswrapper[4791]: I1210 22:50:26.514285 4791 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T22:50:26Z","lastTransitionTime":"2025-12-10T22:50:26Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 22:50:26 crc kubenswrapper[4791]: I1210 22:50:26.617681 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 22:50:26 crc kubenswrapper[4791]: I1210 22:50:26.617717 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 22:50:26 crc kubenswrapper[4791]: I1210 22:50:26.617727 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 22:50:26 crc kubenswrapper[4791]: I1210 22:50:26.617741 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 22:50:26 crc kubenswrapper[4791]: I1210 22:50:26.617752 4791 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T22:50:26Z","lastTransitionTime":"2025-12-10T22:50:26Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 22:50:26 crc kubenswrapper[4791]: I1210 22:50:26.720830 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 22:50:26 crc kubenswrapper[4791]: I1210 22:50:26.720886 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 22:50:26 crc kubenswrapper[4791]: I1210 22:50:26.720902 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 22:50:26 crc kubenswrapper[4791]: I1210 22:50:26.720923 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 22:50:26 crc kubenswrapper[4791]: I1210 22:50:26.720937 4791 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T22:50:26Z","lastTransitionTime":"2025-12-10T22:50:26Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 22:50:26 crc kubenswrapper[4791]: I1210 22:50:26.824262 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 22:50:26 crc kubenswrapper[4791]: I1210 22:50:26.824304 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 22:50:26 crc kubenswrapper[4791]: I1210 22:50:26.824313 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 22:50:26 crc kubenswrapper[4791]: I1210 22:50:26.824326 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 22:50:26 crc kubenswrapper[4791]: I1210 22:50:26.824351 4791 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T22:50:26Z","lastTransitionTime":"2025-12-10T22:50:26Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 22:50:26 crc kubenswrapper[4791]: I1210 22:50:26.884476 4791 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 10 22:50:26 crc kubenswrapper[4791]: I1210 22:50:26.884569 4791 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 10 22:50:26 crc kubenswrapper[4791]: I1210 22:50:26.884573 4791 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 10 22:50:26 crc kubenswrapper[4791]: E1210 22:50:26.884917 4791 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 10 22:50:26 crc kubenswrapper[4791]: E1210 22:50:26.885071 4791 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 10 22:50:26 crc kubenswrapper[4791]: E1210 22:50:26.885189 4791 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 10 22:50:26 crc kubenswrapper[4791]: I1210 22:50:26.884733 4791 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-b6kf6" Dec 10 22:50:26 crc kubenswrapper[4791]: E1210 22:50:26.885454 4791 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-b6kf6" podUID="b4527e71-bb58-4a39-a99c-fe614aa4c5f9" Dec 10 22:50:26 crc kubenswrapper[4791]: I1210 22:50:26.926816 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 22:50:26 crc kubenswrapper[4791]: I1210 22:50:26.926858 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 22:50:26 crc kubenswrapper[4791]: I1210 22:50:26.926873 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 22:50:26 crc kubenswrapper[4791]: I1210 22:50:26.926896 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 22:50:26 crc kubenswrapper[4791]: I1210 22:50:26.926910 4791 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T22:50:26Z","lastTransitionTime":"2025-12-10T22:50:26Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 22:50:27 crc kubenswrapper[4791]: I1210 22:50:27.029635 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 22:50:27 crc kubenswrapper[4791]: I1210 22:50:27.029711 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 22:50:27 crc kubenswrapper[4791]: I1210 22:50:27.029735 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 22:50:27 crc kubenswrapper[4791]: I1210 22:50:27.029765 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 22:50:27 crc kubenswrapper[4791]: I1210 22:50:27.029786 4791 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T22:50:27Z","lastTransitionTime":"2025-12-10T22:50:27Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 22:50:27 crc kubenswrapper[4791]: I1210 22:50:27.132289 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 22:50:27 crc kubenswrapper[4791]: I1210 22:50:27.132370 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 22:50:27 crc kubenswrapper[4791]: I1210 22:50:27.132392 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 22:50:27 crc kubenswrapper[4791]: I1210 22:50:27.132419 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 22:50:27 crc kubenswrapper[4791]: I1210 22:50:27.132443 4791 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T22:50:27Z","lastTransitionTime":"2025-12-10T22:50:27Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 22:50:27 crc kubenswrapper[4791]: I1210 22:50:27.234913 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 22:50:27 crc kubenswrapper[4791]: I1210 22:50:27.234977 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 22:50:27 crc kubenswrapper[4791]: I1210 22:50:27.234998 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 22:50:27 crc kubenswrapper[4791]: I1210 22:50:27.235032 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 22:50:27 crc kubenswrapper[4791]: I1210 22:50:27.235072 4791 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T22:50:27Z","lastTransitionTime":"2025-12-10T22:50:27Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 22:50:27 crc kubenswrapper[4791]: I1210 22:50:27.338807 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 22:50:27 crc kubenswrapper[4791]: I1210 22:50:27.338941 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 22:50:27 crc kubenswrapper[4791]: I1210 22:50:27.338976 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 22:50:27 crc kubenswrapper[4791]: I1210 22:50:27.339009 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 22:50:27 crc kubenswrapper[4791]: I1210 22:50:27.339031 4791 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T22:50:27Z","lastTransitionTime":"2025-12-10T22:50:27Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 22:50:27 crc kubenswrapper[4791]: I1210 22:50:27.441675 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 22:50:27 crc kubenswrapper[4791]: I1210 22:50:27.441737 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 22:50:27 crc kubenswrapper[4791]: I1210 22:50:27.441754 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 22:50:27 crc kubenswrapper[4791]: I1210 22:50:27.441776 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 22:50:27 crc kubenswrapper[4791]: I1210 22:50:27.441794 4791 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T22:50:27Z","lastTransitionTime":"2025-12-10T22:50:27Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 22:50:27 crc kubenswrapper[4791]: I1210 22:50:27.544820 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 22:50:27 crc kubenswrapper[4791]: I1210 22:50:27.544895 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 22:50:27 crc kubenswrapper[4791]: I1210 22:50:27.544919 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 22:50:27 crc kubenswrapper[4791]: I1210 22:50:27.544951 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 22:50:27 crc kubenswrapper[4791]: I1210 22:50:27.544976 4791 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T22:50:27Z","lastTransitionTime":"2025-12-10T22:50:27Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 22:50:27 crc kubenswrapper[4791]: I1210 22:50:27.647243 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 22:50:27 crc kubenswrapper[4791]: I1210 22:50:27.647319 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 22:50:27 crc kubenswrapper[4791]: I1210 22:50:27.647376 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 22:50:27 crc kubenswrapper[4791]: I1210 22:50:27.647407 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 22:50:27 crc kubenswrapper[4791]: I1210 22:50:27.647428 4791 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T22:50:27Z","lastTransitionTime":"2025-12-10T22:50:27Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 22:50:27 crc kubenswrapper[4791]: I1210 22:50:27.750112 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 22:50:27 crc kubenswrapper[4791]: I1210 22:50:27.750164 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 22:50:27 crc kubenswrapper[4791]: I1210 22:50:27.750175 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 22:50:27 crc kubenswrapper[4791]: I1210 22:50:27.750192 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 22:50:27 crc kubenswrapper[4791]: I1210 22:50:27.750200 4791 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T22:50:27Z","lastTransitionTime":"2025-12-10T22:50:27Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 22:50:27 crc kubenswrapper[4791]: I1210 22:50:27.852736 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 22:50:27 crc kubenswrapper[4791]: I1210 22:50:27.852818 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 22:50:27 crc kubenswrapper[4791]: I1210 22:50:27.852842 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 22:50:27 crc kubenswrapper[4791]: I1210 22:50:27.852864 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 22:50:27 crc kubenswrapper[4791]: I1210 22:50:27.852901 4791 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T22:50:27Z","lastTransitionTime":"2025-12-10T22:50:27Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 22:50:27 crc kubenswrapper[4791]: I1210 22:50:27.955925 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 22:50:27 crc kubenswrapper[4791]: I1210 22:50:27.955977 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 22:50:27 crc kubenswrapper[4791]: I1210 22:50:27.955988 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 22:50:27 crc kubenswrapper[4791]: I1210 22:50:27.956007 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 22:50:27 crc kubenswrapper[4791]: I1210 22:50:27.956019 4791 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T22:50:27Z","lastTransitionTime":"2025-12-10T22:50:27Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 22:50:28 crc kubenswrapper[4791]: I1210 22:50:28.047735 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 22:50:28 crc kubenswrapper[4791]: I1210 22:50:28.047817 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 22:50:28 crc kubenswrapper[4791]: I1210 22:50:28.047837 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 22:50:28 crc kubenswrapper[4791]: I1210 22:50:28.047882 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 22:50:28 crc kubenswrapper[4791]: I1210 22:50:28.047902 4791 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T22:50:28Z","lastTransitionTime":"2025-12-10T22:50:28Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 22:50:28 crc kubenswrapper[4791]: E1210 22:50:28.067028 4791 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-12-10T22:50:28Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-10T22:50:28Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-10T22:50:28Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-10T22:50:28Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-10T22:50:28Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-10T22:50:28Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-10T22:50:28Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-10T22:50:28Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"4fab3af6-f657-4dfc-8ef2-12b7c978c94f\\\",\\\"systemUUID\\\":\\\"442c0b68-8dce-47e0-bd68-d5210d7e0493\\\"},\\\"runtimeHandlers\\\":[{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":false},\\\"name\\\":\\\"runc\\\"},{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":true},\\\"name\\\":\\\"crun\\\"},{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":true},\\\"name\\\":\\\"\\\"}]}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T22:50:28Z is after 2025-08-24T17:21:41Z" Dec 10 22:50:28 crc kubenswrapper[4791]: I1210 22:50:28.072149 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 22:50:28 crc kubenswrapper[4791]: I1210 22:50:28.072273 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 22:50:28 crc kubenswrapper[4791]: I1210 22:50:28.072294 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 22:50:28 crc kubenswrapper[4791]: I1210 22:50:28.072323 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 22:50:28 crc kubenswrapper[4791]: I1210 22:50:28.072367 4791 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T22:50:28Z","lastTransitionTime":"2025-12-10T22:50:28Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 22:50:28 crc kubenswrapper[4791]: E1210 22:50:28.089215 4791 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-12-10T22:50:28Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-10T22:50:28Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-10T22:50:28Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-10T22:50:28Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-10T22:50:28Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-10T22:50:28Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-10T22:50:28Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-10T22:50:28Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"4fab3af6-f657-4dfc-8ef2-12b7c978c94f\\\",\\\"systemUUID\\\":\\\"442c0b68-8dce-47e0-bd68-d5210d7e0493\\\"},\\\"runtimeHandlers\\\":[{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":false},\\\"name\\\":\\\"runc\\\"},{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":true},\\\"name\\\":\\\"crun\\\"},{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":true},\\\"name\\\":\\\"\\\"}]}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T22:50:28Z is after 2025-08-24T17:21:41Z" Dec 10 22:50:28 crc kubenswrapper[4791]: I1210 22:50:28.093584 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 22:50:28 crc kubenswrapper[4791]: I1210 22:50:28.093660 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 22:50:28 crc kubenswrapper[4791]: I1210 22:50:28.093681 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 22:50:28 crc kubenswrapper[4791]: I1210 22:50:28.093708 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 22:50:28 crc kubenswrapper[4791]: I1210 22:50:28.093731 4791 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T22:50:28Z","lastTransitionTime":"2025-12-10T22:50:28Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 22:50:28 crc kubenswrapper[4791]: E1210 22:50:28.113580 4791 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-12-10T22:50:28Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-10T22:50:28Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-10T22:50:28Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-10T22:50:28Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-10T22:50:28Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-10T22:50:28Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-10T22:50:28Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-10T22:50:28Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"4fab3af6-f657-4dfc-8ef2-12b7c978c94f\\\",\\\"systemUUID\\\":\\\"442c0b68-8dce-47e0-bd68-d5210d7e0493\\\"},\\\"runtimeHandlers\\\":[{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":false},\\\"name\\\":\\\"runc\\\"},{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":true},\\\"name\\\":\\\"crun\\\"},{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":true},\\\"name\\\":\\\"\\\"}]}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T22:50:28Z is after 2025-08-24T17:21:41Z" Dec 10 22:50:28 crc kubenswrapper[4791]: I1210 22:50:28.117253 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 22:50:28 crc kubenswrapper[4791]: I1210 22:50:28.117311 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 22:50:28 crc kubenswrapper[4791]: I1210 22:50:28.117333 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 22:50:28 crc kubenswrapper[4791]: I1210 22:50:28.117389 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 22:50:28 crc kubenswrapper[4791]: I1210 22:50:28.117412 4791 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T22:50:28Z","lastTransitionTime":"2025-12-10T22:50:28Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 22:50:28 crc kubenswrapper[4791]: E1210 22:50:28.132397 4791 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-12-10T22:50:28Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-10T22:50:28Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-10T22:50:28Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-10T22:50:28Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-10T22:50:28Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-10T22:50:28Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-10T22:50:28Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-10T22:50:28Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"4fab3af6-f657-4dfc-8ef2-12b7c978c94f\\\",\\\"systemUUID\\\":\\\"442c0b68-8dce-47e0-bd68-d5210d7e0493\\\"},\\\"runtimeHandlers\\\":[{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":false},\\\"name\\\":\\\"runc\\\"},{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":true},\\\"name\\\":\\\"crun\\\"},{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":true},\\\"name\\\":\\\"\\\"}]}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T22:50:28Z is after 2025-08-24T17:21:41Z" Dec 10 22:50:28 crc kubenswrapper[4791]: I1210 22:50:28.141245 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 22:50:28 crc kubenswrapper[4791]: I1210 22:50:28.141283 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 22:50:28 crc kubenswrapper[4791]: I1210 22:50:28.141301 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 22:50:28 crc kubenswrapper[4791]: I1210 22:50:28.141318 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 22:50:28 crc kubenswrapper[4791]: I1210 22:50:28.141332 4791 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T22:50:28Z","lastTransitionTime":"2025-12-10T22:50:28Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 22:50:28 crc kubenswrapper[4791]: E1210 22:50:28.156461 4791 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-12-10T22:50:28Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-10T22:50:28Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-10T22:50:28Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-10T22:50:28Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-10T22:50:28Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-10T22:50:28Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-10T22:50:28Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-10T22:50:28Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"4fab3af6-f657-4dfc-8ef2-12b7c978c94f\\\",\\\"systemUUID\\\":\\\"442c0b68-8dce-47e0-bd68-d5210d7e0493\\\"},\\\"runtimeHandlers\\\":[{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":false},\\\"name\\\":\\\"runc\\\"},{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":true},\\\"name\\\":\\\"crun\\\"},{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":true},\\\"name\\\":\\\"\\\"}]}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T22:50:28Z is after 2025-08-24T17:21:41Z" Dec 10 22:50:28 crc kubenswrapper[4791]: E1210 22:50:28.156628 4791 kubelet_node_status.go:572] "Unable to update node status" err="update node status exceeds retry count" Dec 10 22:50:28 crc kubenswrapper[4791]: I1210 22:50:28.158247 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 22:50:28 crc kubenswrapper[4791]: I1210 22:50:28.158326 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 22:50:28 crc kubenswrapper[4791]: I1210 22:50:28.158378 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 22:50:28 crc kubenswrapper[4791]: I1210 22:50:28.158405 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 22:50:28 crc kubenswrapper[4791]: I1210 22:50:28.158423 4791 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T22:50:28Z","lastTransitionTime":"2025-12-10T22:50:28Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 22:50:28 crc kubenswrapper[4791]: I1210 22:50:28.260981 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 22:50:28 crc kubenswrapper[4791]: I1210 22:50:28.261042 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 22:50:28 crc kubenswrapper[4791]: I1210 22:50:28.261061 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 22:50:28 crc kubenswrapper[4791]: I1210 22:50:28.261085 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 22:50:28 crc kubenswrapper[4791]: I1210 22:50:28.261102 4791 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T22:50:28Z","lastTransitionTime":"2025-12-10T22:50:28Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 22:50:28 crc kubenswrapper[4791]: I1210 22:50:28.363803 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 22:50:28 crc kubenswrapper[4791]: I1210 22:50:28.363838 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 22:50:28 crc kubenswrapper[4791]: I1210 22:50:28.363849 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 22:50:28 crc kubenswrapper[4791]: I1210 22:50:28.363866 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 22:50:28 crc kubenswrapper[4791]: I1210 22:50:28.363879 4791 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T22:50:28Z","lastTransitionTime":"2025-12-10T22:50:28Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 22:50:28 crc kubenswrapper[4791]: I1210 22:50:28.466579 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 22:50:28 crc kubenswrapper[4791]: I1210 22:50:28.466644 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 22:50:28 crc kubenswrapper[4791]: I1210 22:50:28.466668 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 22:50:28 crc kubenswrapper[4791]: I1210 22:50:28.466699 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 22:50:28 crc kubenswrapper[4791]: I1210 22:50:28.466718 4791 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T22:50:28Z","lastTransitionTime":"2025-12-10T22:50:28Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 22:50:28 crc kubenswrapper[4791]: I1210 22:50:28.569512 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 22:50:28 crc kubenswrapper[4791]: I1210 22:50:28.569570 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 22:50:28 crc kubenswrapper[4791]: I1210 22:50:28.569586 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 22:50:28 crc kubenswrapper[4791]: I1210 22:50:28.569608 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 22:50:28 crc kubenswrapper[4791]: I1210 22:50:28.569625 4791 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T22:50:28Z","lastTransitionTime":"2025-12-10T22:50:28Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 22:50:28 crc kubenswrapper[4791]: I1210 22:50:28.672857 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 22:50:28 crc kubenswrapper[4791]: I1210 22:50:28.672949 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 22:50:28 crc kubenswrapper[4791]: I1210 22:50:28.672969 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 22:50:28 crc kubenswrapper[4791]: I1210 22:50:28.672993 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 22:50:28 crc kubenswrapper[4791]: I1210 22:50:28.673010 4791 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T22:50:28Z","lastTransitionTime":"2025-12-10T22:50:28Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 22:50:28 crc kubenswrapper[4791]: I1210 22:50:28.775541 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 22:50:28 crc kubenswrapper[4791]: I1210 22:50:28.775597 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 22:50:28 crc kubenswrapper[4791]: I1210 22:50:28.775610 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 22:50:28 crc kubenswrapper[4791]: I1210 22:50:28.775628 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 22:50:28 crc kubenswrapper[4791]: I1210 22:50:28.775671 4791 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T22:50:28Z","lastTransitionTime":"2025-12-10T22:50:28Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 22:50:28 crc kubenswrapper[4791]: I1210 22:50:28.878786 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 22:50:28 crc kubenswrapper[4791]: I1210 22:50:28.878852 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 22:50:28 crc kubenswrapper[4791]: I1210 22:50:28.878869 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 22:50:28 crc kubenswrapper[4791]: I1210 22:50:28.878894 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 22:50:28 crc kubenswrapper[4791]: I1210 22:50:28.878913 4791 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T22:50:28Z","lastTransitionTime":"2025-12-10T22:50:28Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 22:50:28 crc kubenswrapper[4791]: I1210 22:50:28.884218 4791 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-b6kf6" Dec 10 22:50:28 crc kubenswrapper[4791]: I1210 22:50:28.884269 4791 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 10 22:50:28 crc kubenswrapper[4791]: I1210 22:50:28.884218 4791 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 10 22:50:28 crc kubenswrapper[4791]: I1210 22:50:28.884384 4791 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 10 22:50:28 crc kubenswrapper[4791]: E1210 22:50:28.884485 4791 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-b6kf6" podUID="b4527e71-bb58-4a39-a99c-fe614aa4c5f9" Dec 10 22:50:28 crc kubenswrapper[4791]: E1210 22:50:28.884610 4791 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 10 22:50:28 crc kubenswrapper[4791]: E1210 22:50:28.884760 4791 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 10 22:50:28 crc kubenswrapper[4791]: E1210 22:50:28.884852 4791 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 10 22:50:28 crc kubenswrapper[4791]: I1210 22:50:28.981484 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 22:50:28 crc kubenswrapper[4791]: I1210 22:50:28.981542 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 22:50:28 crc kubenswrapper[4791]: I1210 22:50:28.981551 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 22:50:28 crc kubenswrapper[4791]: I1210 22:50:28.981565 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 22:50:28 crc kubenswrapper[4791]: I1210 22:50:28.981574 4791 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T22:50:28Z","lastTransitionTime":"2025-12-10T22:50:28Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 22:50:29 crc kubenswrapper[4791]: I1210 22:50:29.084143 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 22:50:29 crc kubenswrapper[4791]: I1210 22:50:29.084201 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 22:50:29 crc kubenswrapper[4791]: I1210 22:50:29.084215 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 22:50:29 crc kubenswrapper[4791]: I1210 22:50:29.084232 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 22:50:29 crc kubenswrapper[4791]: I1210 22:50:29.084243 4791 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T22:50:29Z","lastTransitionTime":"2025-12-10T22:50:29Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 22:50:29 crc kubenswrapper[4791]: I1210 22:50:29.187613 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 22:50:29 crc kubenswrapper[4791]: I1210 22:50:29.187674 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 22:50:29 crc kubenswrapper[4791]: I1210 22:50:29.187687 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 22:50:29 crc kubenswrapper[4791]: I1210 22:50:29.187710 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 22:50:29 crc kubenswrapper[4791]: I1210 22:50:29.187725 4791 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T22:50:29Z","lastTransitionTime":"2025-12-10T22:50:29Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 22:50:29 crc kubenswrapper[4791]: I1210 22:50:29.290546 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 22:50:29 crc kubenswrapper[4791]: I1210 22:50:29.290591 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 22:50:29 crc kubenswrapper[4791]: I1210 22:50:29.290602 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 22:50:29 crc kubenswrapper[4791]: I1210 22:50:29.290643 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 22:50:29 crc kubenswrapper[4791]: I1210 22:50:29.290661 4791 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T22:50:29Z","lastTransitionTime":"2025-12-10T22:50:29Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 22:50:29 crc kubenswrapper[4791]: I1210 22:50:29.393630 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 22:50:29 crc kubenswrapper[4791]: I1210 22:50:29.393700 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 22:50:29 crc kubenswrapper[4791]: I1210 22:50:29.393715 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 22:50:29 crc kubenswrapper[4791]: I1210 22:50:29.393742 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 22:50:29 crc kubenswrapper[4791]: I1210 22:50:29.393760 4791 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T22:50:29Z","lastTransitionTime":"2025-12-10T22:50:29Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 22:50:29 crc kubenswrapper[4791]: I1210 22:50:29.498251 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 22:50:29 crc kubenswrapper[4791]: I1210 22:50:29.498296 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 22:50:29 crc kubenswrapper[4791]: I1210 22:50:29.498307 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 22:50:29 crc kubenswrapper[4791]: I1210 22:50:29.498327 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 22:50:29 crc kubenswrapper[4791]: I1210 22:50:29.498351 4791 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T22:50:29Z","lastTransitionTime":"2025-12-10T22:50:29Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 22:50:29 crc kubenswrapper[4791]: I1210 22:50:29.600756 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 22:50:29 crc kubenswrapper[4791]: I1210 22:50:29.600798 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 22:50:29 crc kubenswrapper[4791]: I1210 22:50:29.600809 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 22:50:29 crc kubenswrapper[4791]: I1210 22:50:29.600998 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 22:50:29 crc kubenswrapper[4791]: I1210 22:50:29.601011 4791 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T22:50:29Z","lastTransitionTime":"2025-12-10T22:50:29Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 22:50:29 crc kubenswrapper[4791]: I1210 22:50:29.703443 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 22:50:29 crc kubenswrapper[4791]: I1210 22:50:29.703484 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 22:50:29 crc kubenswrapper[4791]: I1210 22:50:29.703494 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 22:50:29 crc kubenswrapper[4791]: I1210 22:50:29.703514 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 22:50:29 crc kubenswrapper[4791]: I1210 22:50:29.703527 4791 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T22:50:29Z","lastTransitionTime":"2025-12-10T22:50:29Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 22:50:29 crc kubenswrapper[4791]: I1210 22:50:29.806637 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 22:50:29 crc kubenswrapper[4791]: I1210 22:50:29.806701 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 22:50:29 crc kubenswrapper[4791]: I1210 22:50:29.806710 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 22:50:29 crc kubenswrapper[4791]: I1210 22:50:29.806742 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 22:50:29 crc kubenswrapper[4791]: I1210 22:50:29.806756 4791 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T22:50:29Z","lastTransitionTime":"2025-12-10T22:50:29Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 22:50:29 crc kubenswrapper[4791]: I1210 22:50:29.909572 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 22:50:29 crc kubenswrapper[4791]: I1210 22:50:29.909626 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 22:50:29 crc kubenswrapper[4791]: I1210 22:50:29.909638 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 22:50:29 crc kubenswrapper[4791]: I1210 22:50:29.909654 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 22:50:29 crc kubenswrapper[4791]: I1210 22:50:29.909665 4791 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T22:50:29Z","lastTransitionTime":"2025-12-10T22:50:29Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 22:50:30 crc kubenswrapper[4791]: I1210 22:50:30.012687 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 22:50:30 crc kubenswrapper[4791]: I1210 22:50:30.012753 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 22:50:30 crc kubenswrapper[4791]: I1210 22:50:30.012770 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 22:50:30 crc kubenswrapper[4791]: I1210 22:50:30.012801 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 22:50:30 crc kubenswrapper[4791]: I1210 22:50:30.012823 4791 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T22:50:30Z","lastTransitionTime":"2025-12-10T22:50:30Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 22:50:30 crc kubenswrapper[4791]: I1210 22:50:30.115025 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 22:50:30 crc kubenswrapper[4791]: I1210 22:50:30.115063 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 22:50:30 crc kubenswrapper[4791]: I1210 22:50:30.115074 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 22:50:30 crc kubenswrapper[4791]: I1210 22:50:30.115089 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 22:50:30 crc kubenswrapper[4791]: I1210 22:50:30.115099 4791 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T22:50:30Z","lastTransitionTime":"2025-12-10T22:50:30Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 22:50:30 crc kubenswrapper[4791]: I1210 22:50:30.217551 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 22:50:30 crc kubenswrapper[4791]: I1210 22:50:30.217600 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 22:50:30 crc kubenswrapper[4791]: I1210 22:50:30.217608 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 22:50:30 crc kubenswrapper[4791]: I1210 22:50:30.217631 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 22:50:30 crc kubenswrapper[4791]: I1210 22:50:30.217641 4791 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T22:50:30Z","lastTransitionTime":"2025-12-10T22:50:30Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 22:50:30 crc kubenswrapper[4791]: I1210 22:50:30.320124 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 22:50:30 crc kubenswrapper[4791]: I1210 22:50:30.320166 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 22:50:30 crc kubenswrapper[4791]: I1210 22:50:30.320175 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 22:50:30 crc kubenswrapper[4791]: I1210 22:50:30.320191 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 22:50:30 crc kubenswrapper[4791]: I1210 22:50:30.320202 4791 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T22:50:30Z","lastTransitionTime":"2025-12-10T22:50:30Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 22:50:30 crc kubenswrapper[4791]: I1210 22:50:30.423001 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 22:50:30 crc kubenswrapper[4791]: I1210 22:50:30.423046 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 22:50:30 crc kubenswrapper[4791]: I1210 22:50:30.423057 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 22:50:30 crc kubenswrapper[4791]: I1210 22:50:30.423074 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 22:50:30 crc kubenswrapper[4791]: I1210 22:50:30.423086 4791 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T22:50:30Z","lastTransitionTime":"2025-12-10T22:50:30Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 22:50:30 crc kubenswrapper[4791]: I1210 22:50:30.525835 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 22:50:30 crc kubenswrapper[4791]: I1210 22:50:30.525888 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 22:50:30 crc kubenswrapper[4791]: I1210 22:50:30.525900 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 22:50:30 crc kubenswrapper[4791]: I1210 22:50:30.525917 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 22:50:30 crc kubenswrapper[4791]: I1210 22:50:30.525928 4791 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T22:50:30Z","lastTransitionTime":"2025-12-10T22:50:30Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 22:50:30 crc kubenswrapper[4791]: I1210 22:50:30.627760 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 22:50:30 crc kubenswrapper[4791]: I1210 22:50:30.627792 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 22:50:30 crc kubenswrapper[4791]: I1210 22:50:30.627800 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 22:50:30 crc kubenswrapper[4791]: I1210 22:50:30.627813 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 22:50:30 crc kubenswrapper[4791]: I1210 22:50:30.627822 4791 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T22:50:30Z","lastTransitionTime":"2025-12-10T22:50:30Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 22:50:30 crc kubenswrapper[4791]: I1210 22:50:30.729813 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 22:50:30 crc kubenswrapper[4791]: I1210 22:50:30.729854 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 22:50:30 crc kubenswrapper[4791]: I1210 22:50:30.729863 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 22:50:30 crc kubenswrapper[4791]: I1210 22:50:30.729878 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 22:50:30 crc kubenswrapper[4791]: I1210 22:50:30.729888 4791 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T22:50:30Z","lastTransitionTime":"2025-12-10T22:50:30Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 22:50:30 crc kubenswrapper[4791]: I1210 22:50:30.831918 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 22:50:30 crc kubenswrapper[4791]: I1210 22:50:30.831968 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 22:50:30 crc kubenswrapper[4791]: I1210 22:50:30.831984 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 22:50:30 crc kubenswrapper[4791]: I1210 22:50:30.832006 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 22:50:30 crc kubenswrapper[4791]: I1210 22:50:30.832020 4791 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T22:50:30Z","lastTransitionTime":"2025-12-10T22:50:30Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 22:50:30 crc kubenswrapper[4791]: I1210 22:50:30.883759 4791 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 10 22:50:30 crc kubenswrapper[4791]: I1210 22:50:30.883811 4791 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 10 22:50:30 crc kubenswrapper[4791]: I1210 22:50:30.883759 4791 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 10 22:50:30 crc kubenswrapper[4791]: I1210 22:50:30.883770 4791 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-b6kf6" Dec 10 22:50:30 crc kubenswrapper[4791]: E1210 22:50:30.883874 4791 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 10 22:50:30 crc kubenswrapper[4791]: E1210 22:50:30.884127 4791 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 10 22:50:30 crc kubenswrapper[4791]: E1210 22:50:30.884180 4791 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-b6kf6" podUID="b4527e71-bb58-4a39-a99c-fe614aa4c5f9" Dec 10 22:50:30 crc kubenswrapper[4791]: E1210 22:50:30.884404 4791 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 10 22:50:30 crc kubenswrapper[4791]: I1210 22:50:30.933936 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 22:50:30 crc kubenswrapper[4791]: I1210 22:50:30.933973 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 22:50:30 crc kubenswrapper[4791]: I1210 22:50:30.933981 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 22:50:30 crc kubenswrapper[4791]: I1210 22:50:30.933995 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 22:50:30 crc kubenswrapper[4791]: I1210 22:50:30.934005 4791 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T22:50:30Z","lastTransitionTime":"2025-12-10T22:50:30Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 22:50:31 crc kubenswrapper[4791]: I1210 22:50:31.035974 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 22:50:31 crc kubenswrapper[4791]: I1210 22:50:31.036020 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 22:50:31 crc kubenswrapper[4791]: I1210 22:50:31.036032 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 22:50:31 crc kubenswrapper[4791]: I1210 22:50:31.036051 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 22:50:31 crc kubenswrapper[4791]: I1210 22:50:31.036064 4791 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T22:50:31Z","lastTransitionTime":"2025-12-10T22:50:31Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 22:50:31 crc kubenswrapper[4791]: I1210 22:50:31.138664 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 22:50:31 crc kubenswrapper[4791]: I1210 22:50:31.138716 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 22:50:31 crc kubenswrapper[4791]: I1210 22:50:31.138731 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 22:50:31 crc kubenswrapper[4791]: I1210 22:50:31.138758 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 22:50:31 crc kubenswrapper[4791]: I1210 22:50:31.138782 4791 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T22:50:31Z","lastTransitionTime":"2025-12-10T22:50:31Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 22:50:31 crc kubenswrapper[4791]: I1210 22:50:31.241066 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 22:50:31 crc kubenswrapper[4791]: I1210 22:50:31.241111 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 22:50:31 crc kubenswrapper[4791]: I1210 22:50:31.241144 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 22:50:31 crc kubenswrapper[4791]: I1210 22:50:31.241162 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 22:50:31 crc kubenswrapper[4791]: I1210 22:50:31.241172 4791 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T22:50:31Z","lastTransitionTime":"2025-12-10T22:50:31Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 22:50:31 crc kubenswrapper[4791]: I1210 22:50:31.343937 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 22:50:31 crc kubenswrapper[4791]: I1210 22:50:31.344007 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 22:50:31 crc kubenswrapper[4791]: I1210 22:50:31.344020 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 22:50:31 crc kubenswrapper[4791]: I1210 22:50:31.344037 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 22:50:31 crc kubenswrapper[4791]: I1210 22:50:31.344050 4791 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T22:50:31Z","lastTransitionTime":"2025-12-10T22:50:31Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 22:50:31 crc kubenswrapper[4791]: I1210 22:50:31.446577 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 22:50:31 crc kubenswrapper[4791]: I1210 22:50:31.446636 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 22:50:31 crc kubenswrapper[4791]: I1210 22:50:31.446654 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 22:50:31 crc kubenswrapper[4791]: I1210 22:50:31.446675 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 22:50:31 crc kubenswrapper[4791]: I1210 22:50:31.446690 4791 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T22:50:31Z","lastTransitionTime":"2025-12-10T22:50:31Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 22:50:31 crc kubenswrapper[4791]: I1210 22:50:31.549711 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 22:50:31 crc kubenswrapper[4791]: I1210 22:50:31.549784 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 22:50:31 crc kubenswrapper[4791]: I1210 22:50:31.549797 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 22:50:31 crc kubenswrapper[4791]: I1210 22:50:31.549815 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 22:50:31 crc kubenswrapper[4791]: I1210 22:50:31.549828 4791 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T22:50:31Z","lastTransitionTime":"2025-12-10T22:50:31Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 22:50:31 crc kubenswrapper[4791]: I1210 22:50:31.652699 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 22:50:31 crc kubenswrapper[4791]: I1210 22:50:31.652740 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 22:50:31 crc kubenswrapper[4791]: I1210 22:50:31.652748 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 22:50:31 crc kubenswrapper[4791]: I1210 22:50:31.652762 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 22:50:31 crc kubenswrapper[4791]: I1210 22:50:31.652772 4791 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T22:50:31Z","lastTransitionTime":"2025-12-10T22:50:31Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 22:50:31 crc kubenswrapper[4791]: I1210 22:50:31.755095 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 22:50:31 crc kubenswrapper[4791]: I1210 22:50:31.755122 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 22:50:31 crc kubenswrapper[4791]: I1210 22:50:31.755130 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 22:50:31 crc kubenswrapper[4791]: I1210 22:50:31.755143 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 22:50:31 crc kubenswrapper[4791]: I1210 22:50:31.755153 4791 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T22:50:31Z","lastTransitionTime":"2025-12-10T22:50:31Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 22:50:31 crc kubenswrapper[4791]: I1210 22:50:31.857579 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 22:50:31 crc kubenswrapper[4791]: I1210 22:50:31.857632 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 22:50:31 crc kubenswrapper[4791]: I1210 22:50:31.857643 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 22:50:31 crc kubenswrapper[4791]: I1210 22:50:31.857660 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 22:50:31 crc kubenswrapper[4791]: I1210 22:50:31.857673 4791 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T22:50:31Z","lastTransitionTime":"2025-12-10T22:50:31Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 22:50:31 crc kubenswrapper[4791]: I1210 22:50:31.960057 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 22:50:31 crc kubenswrapper[4791]: I1210 22:50:31.960091 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 22:50:31 crc kubenswrapper[4791]: I1210 22:50:31.960099 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 22:50:31 crc kubenswrapper[4791]: I1210 22:50:31.960111 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 22:50:31 crc kubenswrapper[4791]: I1210 22:50:31.960120 4791 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T22:50:31Z","lastTransitionTime":"2025-12-10T22:50:31Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 22:50:32 crc kubenswrapper[4791]: I1210 22:50:32.062755 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 22:50:32 crc kubenswrapper[4791]: I1210 22:50:32.062800 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 22:50:32 crc kubenswrapper[4791]: I1210 22:50:32.062814 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 22:50:32 crc kubenswrapper[4791]: I1210 22:50:32.062832 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 22:50:32 crc kubenswrapper[4791]: I1210 22:50:32.062846 4791 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T22:50:32Z","lastTransitionTime":"2025-12-10T22:50:32Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 22:50:32 crc kubenswrapper[4791]: I1210 22:50:32.165484 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 22:50:32 crc kubenswrapper[4791]: I1210 22:50:32.165530 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 22:50:32 crc kubenswrapper[4791]: I1210 22:50:32.165541 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 22:50:32 crc kubenswrapper[4791]: I1210 22:50:32.165555 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 22:50:32 crc kubenswrapper[4791]: I1210 22:50:32.165566 4791 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T22:50:32Z","lastTransitionTime":"2025-12-10T22:50:32Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 22:50:32 crc kubenswrapper[4791]: I1210 22:50:32.268013 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 22:50:32 crc kubenswrapper[4791]: I1210 22:50:32.268064 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 22:50:32 crc kubenswrapper[4791]: I1210 22:50:32.268080 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 22:50:32 crc kubenswrapper[4791]: I1210 22:50:32.268104 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 22:50:32 crc kubenswrapper[4791]: I1210 22:50:32.268118 4791 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T22:50:32Z","lastTransitionTime":"2025-12-10T22:50:32Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 22:50:32 crc kubenswrapper[4791]: I1210 22:50:32.370441 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 22:50:32 crc kubenswrapper[4791]: I1210 22:50:32.370500 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 22:50:32 crc kubenswrapper[4791]: I1210 22:50:32.370509 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 22:50:32 crc kubenswrapper[4791]: I1210 22:50:32.370523 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 22:50:32 crc kubenswrapper[4791]: I1210 22:50:32.370532 4791 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T22:50:32Z","lastTransitionTime":"2025-12-10T22:50:32Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 22:50:32 crc kubenswrapper[4791]: I1210 22:50:32.472904 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 22:50:32 crc kubenswrapper[4791]: I1210 22:50:32.472963 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 22:50:32 crc kubenswrapper[4791]: I1210 22:50:32.472980 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 22:50:32 crc kubenswrapper[4791]: I1210 22:50:32.472999 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 22:50:32 crc kubenswrapper[4791]: I1210 22:50:32.473011 4791 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T22:50:32Z","lastTransitionTime":"2025-12-10T22:50:32Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 22:50:32 crc kubenswrapper[4791]: I1210 22:50:32.575277 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 22:50:32 crc kubenswrapper[4791]: I1210 22:50:32.575313 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 22:50:32 crc kubenswrapper[4791]: I1210 22:50:32.575326 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 22:50:32 crc kubenswrapper[4791]: I1210 22:50:32.575365 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 22:50:32 crc kubenswrapper[4791]: I1210 22:50:32.575384 4791 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T22:50:32Z","lastTransitionTime":"2025-12-10T22:50:32Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 22:50:32 crc kubenswrapper[4791]: I1210 22:50:32.677842 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 22:50:32 crc kubenswrapper[4791]: I1210 22:50:32.677939 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 22:50:32 crc kubenswrapper[4791]: I1210 22:50:32.677950 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 22:50:32 crc kubenswrapper[4791]: I1210 22:50:32.677989 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 22:50:32 crc kubenswrapper[4791]: I1210 22:50:32.678007 4791 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T22:50:32Z","lastTransitionTime":"2025-12-10T22:50:32Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 22:50:32 crc kubenswrapper[4791]: I1210 22:50:32.780805 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 22:50:32 crc kubenswrapper[4791]: I1210 22:50:32.780843 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 22:50:32 crc kubenswrapper[4791]: I1210 22:50:32.780854 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 22:50:32 crc kubenswrapper[4791]: I1210 22:50:32.780873 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 22:50:32 crc kubenswrapper[4791]: I1210 22:50:32.780884 4791 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T22:50:32Z","lastTransitionTime":"2025-12-10T22:50:32Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 22:50:32 crc kubenswrapper[4791]: I1210 22:50:32.882952 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 22:50:32 crc kubenswrapper[4791]: I1210 22:50:32.882991 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 22:50:32 crc kubenswrapper[4791]: I1210 22:50:32.883003 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 22:50:32 crc kubenswrapper[4791]: I1210 22:50:32.883019 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 22:50:32 crc kubenswrapper[4791]: I1210 22:50:32.883031 4791 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T22:50:32Z","lastTransitionTime":"2025-12-10T22:50:32Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 22:50:32 crc kubenswrapper[4791]: I1210 22:50:32.884203 4791 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 10 22:50:32 crc kubenswrapper[4791]: I1210 22:50:32.884227 4791 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 10 22:50:32 crc kubenswrapper[4791]: I1210 22:50:32.884269 4791 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-b6kf6" Dec 10 22:50:32 crc kubenswrapper[4791]: E1210 22:50:32.884305 4791 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 10 22:50:32 crc kubenswrapper[4791]: I1210 22:50:32.884203 4791 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 10 22:50:32 crc kubenswrapper[4791]: E1210 22:50:32.884498 4791 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 10 22:50:32 crc kubenswrapper[4791]: E1210 22:50:32.884548 4791 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-b6kf6" podUID="b4527e71-bb58-4a39-a99c-fe614aa4c5f9" Dec 10 22:50:32 crc kubenswrapper[4791]: E1210 22:50:32.884654 4791 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 10 22:50:32 crc kubenswrapper[4791]: I1210 22:50:32.985398 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 22:50:32 crc kubenswrapper[4791]: I1210 22:50:32.985446 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 22:50:32 crc kubenswrapper[4791]: I1210 22:50:32.985458 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 22:50:32 crc kubenswrapper[4791]: I1210 22:50:32.985475 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 22:50:32 crc kubenswrapper[4791]: I1210 22:50:32.985488 4791 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T22:50:32Z","lastTransitionTime":"2025-12-10T22:50:32Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 22:50:33 crc kubenswrapper[4791]: I1210 22:50:33.088570 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 22:50:33 crc kubenswrapper[4791]: I1210 22:50:33.088610 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 22:50:33 crc kubenswrapper[4791]: I1210 22:50:33.088625 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 22:50:33 crc kubenswrapper[4791]: I1210 22:50:33.088649 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 22:50:33 crc kubenswrapper[4791]: I1210 22:50:33.088665 4791 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T22:50:33Z","lastTransitionTime":"2025-12-10T22:50:33Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 22:50:33 crc kubenswrapper[4791]: I1210 22:50:33.190874 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 22:50:33 crc kubenswrapper[4791]: I1210 22:50:33.190943 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 22:50:33 crc kubenswrapper[4791]: I1210 22:50:33.190962 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 22:50:33 crc kubenswrapper[4791]: I1210 22:50:33.190988 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 22:50:33 crc kubenswrapper[4791]: I1210 22:50:33.191005 4791 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T22:50:33Z","lastTransitionTime":"2025-12-10T22:50:33Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 22:50:33 crc kubenswrapper[4791]: I1210 22:50:33.293094 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 22:50:33 crc kubenswrapper[4791]: I1210 22:50:33.293129 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 22:50:33 crc kubenswrapper[4791]: I1210 22:50:33.293165 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 22:50:33 crc kubenswrapper[4791]: I1210 22:50:33.293183 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 22:50:33 crc kubenswrapper[4791]: I1210 22:50:33.293194 4791 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T22:50:33Z","lastTransitionTime":"2025-12-10T22:50:33Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 22:50:33 crc kubenswrapper[4791]: I1210 22:50:33.395313 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 22:50:33 crc kubenswrapper[4791]: I1210 22:50:33.395414 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 22:50:33 crc kubenswrapper[4791]: I1210 22:50:33.395438 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 22:50:33 crc kubenswrapper[4791]: I1210 22:50:33.395462 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 22:50:33 crc kubenswrapper[4791]: I1210 22:50:33.395480 4791 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T22:50:33Z","lastTransitionTime":"2025-12-10T22:50:33Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 22:50:33 crc kubenswrapper[4791]: I1210 22:50:33.497909 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 22:50:33 crc kubenswrapper[4791]: I1210 22:50:33.497972 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 22:50:33 crc kubenswrapper[4791]: I1210 22:50:33.497996 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 22:50:33 crc kubenswrapper[4791]: I1210 22:50:33.498023 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 22:50:33 crc kubenswrapper[4791]: I1210 22:50:33.498046 4791 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T22:50:33Z","lastTransitionTime":"2025-12-10T22:50:33Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 22:50:33 crc kubenswrapper[4791]: I1210 22:50:33.600804 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 22:50:33 crc kubenswrapper[4791]: I1210 22:50:33.600842 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 22:50:33 crc kubenswrapper[4791]: I1210 22:50:33.600853 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 22:50:33 crc kubenswrapper[4791]: I1210 22:50:33.600868 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 22:50:33 crc kubenswrapper[4791]: I1210 22:50:33.600877 4791 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T22:50:33Z","lastTransitionTime":"2025-12-10T22:50:33Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 22:50:33 crc kubenswrapper[4791]: I1210 22:50:33.703329 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 22:50:33 crc kubenswrapper[4791]: I1210 22:50:33.703380 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 22:50:33 crc kubenswrapper[4791]: I1210 22:50:33.703388 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 22:50:33 crc kubenswrapper[4791]: I1210 22:50:33.703401 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 22:50:33 crc kubenswrapper[4791]: I1210 22:50:33.703411 4791 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T22:50:33Z","lastTransitionTime":"2025-12-10T22:50:33Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 22:50:33 crc kubenswrapper[4791]: I1210 22:50:33.805542 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 22:50:33 crc kubenswrapper[4791]: I1210 22:50:33.805583 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 22:50:33 crc kubenswrapper[4791]: I1210 22:50:33.805596 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 22:50:33 crc kubenswrapper[4791]: I1210 22:50:33.805611 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 22:50:33 crc kubenswrapper[4791]: I1210 22:50:33.805621 4791 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T22:50:33Z","lastTransitionTime":"2025-12-10T22:50:33Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 22:50:33 crc kubenswrapper[4791]: I1210 22:50:33.908630 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 22:50:33 crc kubenswrapper[4791]: I1210 22:50:33.908676 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 22:50:33 crc kubenswrapper[4791]: I1210 22:50:33.908689 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 22:50:33 crc kubenswrapper[4791]: I1210 22:50:33.908708 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 22:50:33 crc kubenswrapper[4791]: I1210 22:50:33.908732 4791 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T22:50:33Z","lastTransitionTime":"2025-12-10T22:50:33Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 22:50:34 crc kubenswrapper[4791]: I1210 22:50:34.011470 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 22:50:34 crc kubenswrapper[4791]: I1210 22:50:34.011523 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 22:50:34 crc kubenswrapper[4791]: I1210 22:50:34.011536 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 22:50:34 crc kubenswrapper[4791]: I1210 22:50:34.011558 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 22:50:34 crc kubenswrapper[4791]: I1210 22:50:34.011571 4791 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T22:50:34Z","lastTransitionTime":"2025-12-10T22:50:34Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 22:50:34 crc kubenswrapper[4791]: I1210 22:50:34.113229 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 22:50:34 crc kubenswrapper[4791]: I1210 22:50:34.113263 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 22:50:34 crc kubenswrapper[4791]: I1210 22:50:34.113274 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 22:50:34 crc kubenswrapper[4791]: I1210 22:50:34.113290 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 22:50:34 crc kubenswrapper[4791]: I1210 22:50:34.113301 4791 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T22:50:34Z","lastTransitionTime":"2025-12-10T22:50:34Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 22:50:34 crc kubenswrapper[4791]: I1210 22:50:34.215854 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 22:50:34 crc kubenswrapper[4791]: I1210 22:50:34.215889 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 22:50:34 crc kubenswrapper[4791]: I1210 22:50:34.215897 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 22:50:34 crc kubenswrapper[4791]: I1210 22:50:34.215912 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 22:50:34 crc kubenswrapper[4791]: I1210 22:50:34.215929 4791 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T22:50:34Z","lastTransitionTime":"2025-12-10T22:50:34Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 22:50:34 crc kubenswrapper[4791]: I1210 22:50:34.319167 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 22:50:34 crc kubenswrapper[4791]: I1210 22:50:34.319234 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 22:50:34 crc kubenswrapper[4791]: I1210 22:50:34.319251 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 22:50:34 crc kubenswrapper[4791]: I1210 22:50:34.319277 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 22:50:34 crc kubenswrapper[4791]: I1210 22:50:34.319295 4791 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T22:50:34Z","lastTransitionTime":"2025-12-10T22:50:34Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 22:50:34 crc kubenswrapper[4791]: I1210 22:50:34.421857 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 22:50:34 crc kubenswrapper[4791]: I1210 22:50:34.421904 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 22:50:34 crc kubenswrapper[4791]: I1210 22:50:34.421916 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 22:50:34 crc kubenswrapper[4791]: I1210 22:50:34.421933 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 22:50:34 crc kubenswrapper[4791]: I1210 22:50:34.421945 4791 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T22:50:34Z","lastTransitionTime":"2025-12-10T22:50:34Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 22:50:34 crc kubenswrapper[4791]: I1210 22:50:34.523667 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 22:50:34 crc kubenswrapper[4791]: I1210 22:50:34.523737 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 22:50:34 crc kubenswrapper[4791]: I1210 22:50:34.523754 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 22:50:34 crc kubenswrapper[4791]: I1210 22:50:34.523774 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 22:50:34 crc kubenswrapper[4791]: I1210 22:50:34.523808 4791 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T22:50:34Z","lastTransitionTime":"2025-12-10T22:50:34Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 22:50:34 crc kubenswrapper[4791]: I1210 22:50:34.626182 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 22:50:34 crc kubenswrapper[4791]: I1210 22:50:34.626239 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 22:50:34 crc kubenswrapper[4791]: I1210 22:50:34.626256 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 22:50:34 crc kubenswrapper[4791]: I1210 22:50:34.626282 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 22:50:34 crc kubenswrapper[4791]: I1210 22:50:34.626298 4791 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T22:50:34Z","lastTransitionTime":"2025-12-10T22:50:34Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 22:50:34 crc kubenswrapper[4791]: I1210 22:50:34.728192 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 22:50:34 crc kubenswrapper[4791]: I1210 22:50:34.728234 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 22:50:34 crc kubenswrapper[4791]: I1210 22:50:34.728244 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 22:50:34 crc kubenswrapper[4791]: I1210 22:50:34.728261 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 22:50:34 crc kubenswrapper[4791]: I1210 22:50:34.728271 4791 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T22:50:34Z","lastTransitionTime":"2025-12-10T22:50:34Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 22:50:34 crc kubenswrapper[4791]: I1210 22:50:34.830963 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 22:50:34 crc kubenswrapper[4791]: I1210 22:50:34.831024 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 22:50:34 crc kubenswrapper[4791]: I1210 22:50:34.831037 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 22:50:34 crc kubenswrapper[4791]: I1210 22:50:34.831056 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 22:50:34 crc kubenswrapper[4791]: I1210 22:50:34.831069 4791 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T22:50:34Z","lastTransitionTime":"2025-12-10T22:50:34Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 22:50:34 crc kubenswrapper[4791]: I1210 22:50:34.883959 4791 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-b6kf6" Dec 10 22:50:34 crc kubenswrapper[4791]: I1210 22:50:34.883999 4791 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 10 22:50:34 crc kubenswrapper[4791]: I1210 22:50:34.884045 4791 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 10 22:50:34 crc kubenswrapper[4791]: I1210 22:50:34.884074 4791 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 10 22:50:34 crc kubenswrapper[4791]: E1210 22:50:34.884111 4791 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-b6kf6" podUID="b4527e71-bb58-4a39-a99c-fe614aa4c5f9" Dec 10 22:50:34 crc kubenswrapper[4791]: E1210 22:50:34.884259 4791 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 10 22:50:34 crc kubenswrapper[4791]: E1210 22:50:34.884307 4791 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 10 22:50:34 crc kubenswrapper[4791]: E1210 22:50:34.884392 4791 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 10 22:50:34 crc kubenswrapper[4791]: I1210 22:50:34.933183 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 22:50:34 crc kubenswrapper[4791]: I1210 22:50:34.933226 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 22:50:34 crc kubenswrapper[4791]: I1210 22:50:34.933238 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 22:50:34 crc kubenswrapper[4791]: I1210 22:50:34.933269 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 22:50:34 crc kubenswrapper[4791]: I1210 22:50:34.933280 4791 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T22:50:34Z","lastTransitionTime":"2025-12-10T22:50:34Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 22:50:35 crc kubenswrapper[4791]: I1210 22:50:35.035909 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 22:50:35 crc kubenswrapper[4791]: I1210 22:50:35.035978 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 22:50:35 crc kubenswrapper[4791]: I1210 22:50:35.035996 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 22:50:35 crc kubenswrapper[4791]: I1210 22:50:35.036013 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 22:50:35 crc kubenswrapper[4791]: I1210 22:50:35.036024 4791 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T22:50:35Z","lastTransitionTime":"2025-12-10T22:50:35Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 22:50:35 crc kubenswrapper[4791]: I1210 22:50:35.139129 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 22:50:35 crc kubenswrapper[4791]: I1210 22:50:35.139191 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 22:50:35 crc kubenswrapper[4791]: I1210 22:50:35.139209 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 22:50:35 crc kubenswrapper[4791]: I1210 22:50:35.139232 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 22:50:35 crc kubenswrapper[4791]: I1210 22:50:35.139249 4791 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T22:50:35Z","lastTransitionTime":"2025-12-10T22:50:35Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 22:50:35 crc kubenswrapper[4791]: I1210 22:50:35.241579 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 22:50:35 crc kubenswrapper[4791]: I1210 22:50:35.241635 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 22:50:35 crc kubenswrapper[4791]: I1210 22:50:35.241647 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 22:50:35 crc kubenswrapper[4791]: I1210 22:50:35.241668 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 22:50:35 crc kubenswrapper[4791]: I1210 22:50:35.241682 4791 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T22:50:35Z","lastTransitionTime":"2025-12-10T22:50:35Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 22:50:35 crc kubenswrapper[4791]: I1210 22:50:35.343613 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 22:50:35 crc kubenswrapper[4791]: I1210 22:50:35.343667 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 22:50:35 crc kubenswrapper[4791]: I1210 22:50:35.343682 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 22:50:35 crc kubenswrapper[4791]: I1210 22:50:35.343698 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 22:50:35 crc kubenswrapper[4791]: I1210 22:50:35.343710 4791 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T22:50:35Z","lastTransitionTime":"2025-12-10T22:50:35Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 22:50:35 crc kubenswrapper[4791]: I1210 22:50:35.447163 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 22:50:35 crc kubenswrapper[4791]: I1210 22:50:35.447222 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 22:50:35 crc kubenswrapper[4791]: I1210 22:50:35.447231 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 22:50:35 crc kubenswrapper[4791]: I1210 22:50:35.447249 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 22:50:35 crc kubenswrapper[4791]: I1210 22:50:35.447262 4791 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T22:50:35Z","lastTransitionTime":"2025-12-10T22:50:35Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 22:50:35 crc kubenswrapper[4791]: I1210 22:50:35.549913 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 22:50:35 crc kubenswrapper[4791]: I1210 22:50:35.550000 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 22:50:35 crc kubenswrapper[4791]: I1210 22:50:35.550011 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 22:50:35 crc kubenswrapper[4791]: I1210 22:50:35.550031 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 22:50:35 crc kubenswrapper[4791]: I1210 22:50:35.550043 4791 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T22:50:35Z","lastTransitionTime":"2025-12-10T22:50:35Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 22:50:35 crc kubenswrapper[4791]: I1210 22:50:35.653108 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 22:50:35 crc kubenswrapper[4791]: I1210 22:50:35.653415 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 22:50:35 crc kubenswrapper[4791]: I1210 22:50:35.653446 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 22:50:35 crc kubenswrapper[4791]: I1210 22:50:35.653475 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 22:50:35 crc kubenswrapper[4791]: I1210 22:50:35.653497 4791 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T22:50:35Z","lastTransitionTime":"2025-12-10T22:50:35Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 22:50:35 crc kubenswrapper[4791]: I1210 22:50:35.756084 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 22:50:35 crc kubenswrapper[4791]: I1210 22:50:35.756315 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 22:50:35 crc kubenswrapper[4791]: I1210 22:50:35.756326 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 22:50:35 crc kubenswrapper[4791]: I1210 22:50:35.756356 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 22:50:35 crc kubenswrapper[4791]: I1210 22:50:35.756368 4791 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T22:50:35Z","lastTransitionTime":"2025-12-10T22:50:35Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 22:50:35 crc kubenswrapper[4791]: I1210 22:50:35.858951 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 22:50:35 crc kubenswrapper[4791]: I1210 22:50:35.858990 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 22:50:35 crc kubenswrapper[4791]: I1210 22:50:35.858998 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 22:50:35 crc kubenswrapper[4791]: I1210 22:50:35.859016 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 22:50:35 crc kubenswrapper[4791]: I1210 22:50:35.859027 4791 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T22:50:35Z","lastTransitionTime":"2025-12-10T22:50:35Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 22:50:35 crc kubenswrapper[4791]: I1210 22:50:35.884685 4791 scope.go:117] "RemoveContainer" containerID="b6d739a4314c4fb536810b8623f691d6aaf1090b9c83f17b484d2b22f7d3ad01" Dec 10 22:50:35 crc kubenswrapper[4791]: E1210 22:50:35.884846 4791 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"ovnkube-controller\" with CrashLoopBackOff: \"back-off 20s restarting failed container=ovnkube-controller pod=ovnkube-node-zhq64_openshift-ovn-kubernetes(3cd47739-0fa9-4321-aff1-220f8721a0b3)\"" pod="openshift-ovn-kubernetes/ovnkube-node-zhq64" podUID="3cd47739-0fa9-4321-aff1-220f8721a0b3" Dec 10 22:50:35 crc kubenswrapper[4791]: I1210 22:50:35.903747 4791 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-zhq64" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3cd47739-0fa9-4321-aff1-220f8721a0b3\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T22:49:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T22:49:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T22:49:52Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T22:49:52Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://68795a29944e56870ffe94fa732e9dcf0bfb344c0d85d1e590f58a876129c61f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T22:49:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wvpw8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://bc721b530c08b433756e978ec0423cc18aabf6771b61eb96f903b4d57ea335ff\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T22:49:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wvpw8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://152a7618be8f38fadf4e564b3948980503cedcdb466c37344ccc20f4a3173a5f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T22:49:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wvpw8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3d18f9cb7207efb1e32de75a30d33e024f656a08f0503db95e5488bc5280971f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T22:49:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wvpw8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://937137ae9d7480cdc7c59bc4a1df0a4c10e72b3b63693af285278789953ed218\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T22:49:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wvpw8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://1a52d19b7671e8b143017c8a981a6f8d3ee3b84650e4b223b814a531ac76f364\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T22:49:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wvpw8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b6d739a4314c4fb536810b8623f691d6aaf1090b9c83f17b484d2b22f7d3ad01\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b6d739a4314c4fb536810b8623f691d6aaf1090b9c83f17b484d2b22f7d3ad01\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-12-10T22:50:21Z\\\",\\\"message\\\":\\\"llocateLoadBalancerNodePorts:nil,LoadBalancerClass:nil,InternalTrafficPolicy:*Cluster,TrafficDistribution:nil,},Status:ServiceStatus{LoadBalancer:LoadBalancerStatus{Ingress:[]LoadBalancerIngress{},},Conditions:[]Condition{},},}\\\\nI1210 22:50:21.614710 6424 metrics.go:553] Stopping metrics server at address \\\\\\\"127.0.0.1:29103\\\\\\\"\\\\nI1210 22:50:21.614731 6424 lb_config.go:1031] Cluster endpoints for openshift-kube-scheduler/scheduler for network=default are: map[]\\\\nI1210 22:50:21.614619 6424 services_controller.go:360] Finished syncing service ovn-kubernetes-control-plane on namespace openshift-ovn-kubernetes for network=default : 7.02µs\\\\nI1210 22:50:21.614757 6424 services_controller.go:356] Processing sync for service openshift-route-controller-manager/route-controller-manager for network=default\\\\nI1210 22:50:21.614615 6424 services_controller.go:444] Built service openshift-machine-config-operator/machine-config-daemon LB per-node configs for network=default: []services.lbConfig(nil)\\\\nF1210 22:50:21.614772 6424 ovnkube.go:137] failed to run ovnkube: [failed to start network controller: failed to start default network controller: unable to create admin network policy controller, err: could not add Event Handler for anpInformer during admin network policy controller initialization, handler {0x1fcc6e0 0x1fcc3c0 0x1fcc360} was not added to\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-10T22:50:20Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":2,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 20s restarting failed container=ovnkube-controller pod=ovnkube-node-zhq64_openshift-ovn-kubernetes(3cd47739-0fa9-4321-aff1-220f8721a0b3)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wvpw8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://dd3f80e5713221f96ed50da7f2e7fd5b9fabfcd5c1656da8d8a86d818a5fdc3f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T22:49:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wvpw8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3afbd1ffcd2c4b40324aca6e52d6fc822db131c3f35a9af3a6f005c2cffdb7bf\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3afbd1ffcd2c4b40324aca6e52d6fc822db131c3f35a9af3a6f005c2cffdb7bf\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-10T22:49:53Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-10T22:49:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wvpw8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-10T22:49:52Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-zhq64\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T22:50:35Z is after 2025-08-24T17:21:41Z" Dec 10 22:50:35 crc kubenswrapper[4791]: I1210 22:50:35.916624 4791 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"698bbe9a-fbe5-436d-a835-5e80df77ba24\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T22:49:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T22:49:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T22:49:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T22:49:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T22:49:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://224d73713312376a58ec6cc2bc8fdb6ffa44c4749ae0e7691d5fe27b9afca09f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T22:49:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://372eeaffd40992c6f06efc4df6dfcb44b31de64ee8256c21508cdd7c74ecb96e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T22:49:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://76b6013b062cac5b0b4d12a94f38ce7b43f9c7ea216063455d88a956ac04d26e\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T22:49:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://78e950e122e761d950da1a3c60c06476e53920e8795f6c61db22c01e22d5136e\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T22:49:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-10T22:49:25Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T22:50:35Z is after 2025-08-24T17:21:41Z" Dec 10 22:50:35 crc kubenswrapper[4791]: I1210 22:50:35.929185 4791 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-10T22:49:52Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-10T22:49:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-10T22:49:52Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3734b110802828632d44263ed0a43cabc4b710833f0be120d72db0940fe783e2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T22:49:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://794803ac9be3c52509e6dcb10854d8864e1a9e406762e57c7971b07c4da1220e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T22:49:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T22:50:35Z is after 2025-08-24T17:21:41Z" Dec 10 22:50:35 crc kubenswrapper[4791]: I1210 22:50:35.938525 4791 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-vczjq" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b9f00f25-3fa1-4b63-9710-fd352224b01b\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T22:49:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T22:49:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T22:49:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T22:49:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4698cf4c6b55b41e628448d37c2eb847c2777d0efbd93afcbf367d5e78623b7e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T22:49:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lrrpv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-10T22:49:50Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-vczjq\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T22:50:35Z is after 2025-08-24T17:21:41Z" Dec 10 22:50:35 crc kubenswrapper[4791]: I1210 22:50:35.951869 4791 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-4nwkq" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"672aa28c-8169-49ed-87b8-21187d13a80c\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T22:49:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T22:49:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T22:49:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T22:49:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a2f2a62315529e148a10a33ccb241e9417fbde0fe42afb61cb7ee2438c74636f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T22:49:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xl62j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-10T22:49:51Z\\\"}}\" for pod \"openshift-multus\"/\"multus-4nwkq\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T22:50:35Z is after 2025-08-24T17:21:41Z" Dec 10 22:50:35 crc kubenswrapper[4791]: I1210 22:50:35.961650 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 22:50:35 crc kubenswrapper[4791]: I1210 22:50:35.961684 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 22:50:35 crc kubenswrapper[4791]: I1210 22:50:35.961693 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 22:50:35 crc kubenswrapper[4791]: I1210 22:50:35.961707 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 22:50:35 crc kubenswrapper[4791]: I1210 22:50:35.961717 4791 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T22:50:35Z","lastTransitionTime":"2025-12-10T22:50:35Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 22:50:35 crc kubenswrapper[4791]: I1210 22:50:35.969779 4791 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-10T22:49:52Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-10T22:49:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-10T22:49:52Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://20651eb0c16e2c71222f56da1ff755cb0e7b084886acbbb24f0ccaf57041ed85\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T22:49:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T22:50:35Z is after 2025-08-24T17:21:41Z" Dec 10 22:50:35 crc kubenswrapper[4791]: I1210 22:50:35.982170 4791 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-10T22:49:50Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-10T22:49:50Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T22:50:35Z is after 2025-08-24T17:21:41Z" Dec 10 22:50:35 crc kubenswrapper[4791]: I1210 22:50:35.992623 4791 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-10T22:49:50Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-10T22:49:50Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T22:50:35Z is after 2025-08-24T17:21:41Z" Dec 10 22:50:36 crc kubenswrapper[4791]: I1210 22:50:36.002467 4791 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-tchxk" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"c0ab74e0-42a9-4f42-ac80-16030cf4ffff\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T22:49:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T22:49:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T22:49:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T22:49:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://04e0fd1b53256b7a938e8ccab93e2a4eb73dd1a59e0b9799e09c55e511baf7fe\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T22:49:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zn2qt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-10T22:49:50Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-tchxk\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T22:50:36Z is after 2025-08-24T17:21:41Z" Dec 10 22:50:36 crc kubenswrapper[4791]: I1210 22:50:36.014805 4791 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-rw584" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ff9beb90-69b0-4732-bf37-0b81f58ecc98\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T22:49:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T22:49:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T22:49:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T22:49:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ddb9df501dc6e8e5561fbc2aefd431037a3f28f73bf8e82a6598d261a4224807\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T22:49:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fksqn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ce67e54b696bb53ffbbe48c6ce71476b635e0868993e9c427b22f904891433ee\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ce67e54b696bb53ffbbe48c6ce71476b635e0868993e9c427b22f904891433ee\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-10T22:49:52Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-10T22:49:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fksqn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://818d89bfffb8a110670152c8480684665ea8db578e9768f524fbbe6a92f27c22\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://818d89bfffb8a110670152c8480684665ea8db578e9768f524fbbe6a92f27c22\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-10T22:49:53Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-10T22:49:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fksqn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://9aa9c06e894bb2a2ed3738defaeff68a0a0458ead48177f36b90770944c211a0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://9aa9c06e894bb2a2ed3738defaeff68a0a0458ead48177f36b90770944c211a0\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-10T22:49:54Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-10T22:49:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fksqn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://dfd6d03336dcea7529419a30e8d70d84e2290a0fa5b0e46d09ad341c190638d6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://dfd6d03336dcea7529419a30e8d70d84e2290a0fa5b0e46d09ad341c190638d6\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-10T22:49:55Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-10T22:49:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fksqn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b86738401ece6fbc3083782720c66a515850ed78d4054f34fcabab7ff525b85f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b86738401ece6fbc3083782720c66a515850ed78d4054f34fcabab7ff525b85f\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-10T22:49:56Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-10T22:49:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fksqn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://18b326f572a899ddd10197a745bf9264d4ddd975690ad1cfb0ec576ccd5cf102\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://18b326f572a899ddd10197a745bf9264d4ddd975690ad1cfb0ec576ccd5cf102\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-10T22:49:57Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-10T22:49:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fksqn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-10T22:49:51Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-rw584\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T22:50:36Z is after 2025-08-24T17:21:41Z" Dec 10 22:50:36 crc kubenswrapper[4791]: I1210 22:50:36.024663 4791 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-b6kf6" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b4527e71-bb58-4a39-a99c-fe614aa4c5f9\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T22:50:04Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T22:50:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T22:50:04Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T22:50:04Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ngpr4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ngpr4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-10T22:50:04Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-b6kf6\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T22:50:36Z is after 2025-08-24T17:21:41Z" Dec 10 22:50:36 crc kubenswrapper[4791]: I1210 22:50:36.035698 4791 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f8349fea-c49b-49a1-b5ee-032c5dffe021\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T22:49:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T22:49:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T22:49:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T22:49:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T22:49:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://acafe12186df99d1059a31260de0eafce35ff7771bcb8f5c6753b72665d6f60b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T22:49:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ea4c346507862a8c12756a3061d6e25e788a8d79f4621b62f81f4322185c82a8\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T22:49:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://2e6e324ec6f4f7bb69fdfab4ea3ac1aa55e480dd8e2a8984b0fe5b4c9f66636e\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T22:49:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://1b2e11ff2bf8789fa1d0d23ae648b3e6ef653106e5bc4bbb5dd6447f7cd6f05d\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d6aa7f8c5d908d90008f2df49860ac8a3480136da567e3d85f86135d779204f9\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-12-10T22:49:44Z\\\",\\\"message\\\":\\\"W1210 22:49:33.845733 1 cmd.go:257] Using insecure, self-signed certificates\\\\nI1210 22:49:33.846108 1 crypto.go:601] Generating new CA for check-endpoints-signer@1765406973 cert, and key in /tmp/serving-cert-2585701456/serving-signer.crt, /tmp/serving-cert-2585701456/serving-signer.key\\\\nI1210 22:49:34.529748 1 observer_polling.go:159] Starting file observer\\\\nW1210 22:49:34.532225 1 builder.go:272] unable to get owner reference (falling back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": dial tcp [::1]:6443: connect: connection refused\\\\nI1210 22:49:34.532599 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1210 22:49:34.533977 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-2585701456/tls.crt::/tmp/serving-cert-2585701456/tls.key\\\\\\\"\\\\nF1210 22:49:44.774710 1 cmd.go:182] error initializing delegating authentication: unable to load configmap based request-header-client-ca-file: Get \\\\\\\"https://localhost:6443/api/v1/namespaces/kube-system/configmaps/extension-apiserver-authentication\\\\\\\": net/http: TLS handshake timeout\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-10T22:49:33Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T22:49:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://0c4d036ab269f62561fe7ef1586012cb3a16e3583a08b9488a0616ec00c8e0ae\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T22:49:32Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c5856d6b26c19fda13b5c9ac2892a7e2f9c6f79feb8173ecfd9b33b673327d67\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c5856d6b26c19fda13b5c9ac2892a7e2f9c6f79feb8173ecfd9b33b673327d67\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-10T22:49:27Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-10T22:49:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-10T22:49:25Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T22:50:36Z is after 2025-08-24T17:21:41Z" Dec 10 22:50:36 crc kubenswrapper[4791]: I1210 22:50:36.045687 4791 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"02a949d8-38e2-4be3-989f-40a015407140\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T22:49:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T22:49:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T22:50:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T22:50:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T22:49:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://cd4158bc59b30f0f1716bf43c1ae7c9916a5b7d7164d074987fc1f19ba49b778\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T22:49:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://656065fd3cbdc6a25dab5d06c7baa00af16d8f62fc5da687efe9c6cc91945de0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T22:49:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://90ed7fff816b1772ba9ecd95587a59cbed7c30ac302fff7e7c5de934c912bcf4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T22:49:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9b7a5502803eaad9f76f5ffa91b29fbb2657402065a03a4d6433c096aac7264f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"wait-for-host-port\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://9b7a5502803eaad9f76f5ffa91b29fbb2657402065a03a4d6433c096aac7264f\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-10T22:49:27Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-10T22:49:27Z\\\"}}}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-10T22:49:25Z\\\"}}\" for pod \"openshift-kube-scheduler\"/\"openshift-kube-scheduler-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T22:50:36Z is after 2025-08-24T17:21:41Z" Dec 10 22:50:36 crc kubenswrapper[4791]: I1210 22:50:36.064282 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 22:50:36 crc kubenswrapper[4791]: I1210 22:50:36.064311 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 22:50:36 crc kubenswrapper[4791]: I1210 22:50:36.064321 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 22:50:36 crc kubenswrapper[4791]: I1210 22:50:36.064352 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 22:50:36 crc kubenswrapper[4791]: I1210 22:50:36.064362 4791 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T22:50:36Z","lastTransitionTime":"2025-12-10T22:50:36Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 22:50:36 crc kubenswrapper[4791]: I1210 22:50:36.072605 4791 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"90250e13-6c63-443a-a5e5-6e4e341a2289\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T22:49:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T22:49:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T22:49:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T22:49:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T22:49:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ba54f95b1e00843b854db9a10fe6897592159ccacf30ed0f76002093a13eedd3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T22:49:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5aa8ffa518f3a68b60b275f8df4851b2564aa4984b61efdc9470b73072133e43\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T22:49:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://16f989503edc78d9451638e2cf6cf4b19a2cf2b8809b10b237541cb829303091\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T22:49:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://82556824d63ce6e46df14c45689ff1f81ebf729abe902061815c040b323898e9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T22:49:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://61db531940fa6f2effdae5a005b4c91bd2698bd5d1bff9588c171a65ad550eca\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T22:49:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0f65c864d885319ce2a08d61c30be24b3ffb7d23ea30c82ed7faf039cb2c006d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://0f65c864d885319ce2a08d61c30be24b3ffb7d23ea30c82ed7faf039cb2c006d\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-10T22:49:27Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-10T22:49:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e31a76d1d5b2812ba83e24e38268e25ec8bed98fc3da9adcd06bda5822b6ace1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e31a76d1d5b2812ba83e24e38268e25ec8bed98fc3da9adcd06bda5822b6ace1\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-10T22:49:30Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-10T22:49:29Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://b5b5eceae562f3986b543d41de1e977efb85a91d84c98e09159ca6d092eda691\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b5b5eceae562f3986b543d41de1e977efb85a91d84c98e09159ca6d092eda691\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-10T22:49:31Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-10T22:49:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-10T22:49:25Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T22:50:36Z is after 2025-08-24T17:21:41Z" Dec 10 22:50:36 crc kubenswrapper[4791]: I1210 22:50:36.084800 4791 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-10T22:49:50Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-10T22:49:50Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T22:50:36Z is after 2025-08-24T17:21:41Z" Dec 10 22:50:36 crc kubenswrapper[4791]: I1210 22:50:36.095311 4791 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-10T22:49:54Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-10T22:49:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-10T22:49:54Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://62bfaf5d7a6d1b452cdaf88086447d0c8cb7306e94434e46752be5475124d56a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T22:49:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T22:50:36Z is after 2025-08-24T17:21:41Z" Dec 10 22:50:36 crc kubenswrapper[4791]: I1210 22:50:36.106386 4791 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-5rb5l" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ba35653c-6e06-4cee-a4d6-137764090d18\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T22:49:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T22:49:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T22:49:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T22:49:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://dbb5c875730a98ad2daf316cfafc08fcdce21ac16148154780ac9193ed405dab\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T22:49:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fvvrs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://185eb2ff44221a514878283ec873773a6c9a9073dc64e296f982e60dee1338e4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T22:49:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fvvrs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-10T22:49:51Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-5rb5l\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T22:50:36Z is after 2025-08-24T17:21:41Z" Dec 10 22:50:36 crc kubenswrapper[4791]: I1210 22:50:36.117768 4791 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-s6rk8" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5c9d0978-a6f9-49c5-b377-6619b4e3ff52\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T22:50:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T22:50:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T22:50:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T22:50:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8e55baaf460c794ab196e237d51d12de059aa038617dc2d5366d49c529f5d5c7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T22:50:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7vkgm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://51d1b9b086c60bae9223cb51d5a998d02790a0d91b745aab10392c23e617eb9f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T22:50:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7vkgm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-10T22:50:03Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-s6rk8\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T22:50:36Z is after 2025-08-24T17:21:41Z" Dec 10 22:50:36 crc kubenswrapper[4791]: I1210 22:50:36.166917 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 22:50:36 crc kubenswrapper[4791]: I1210 22:50:36.166969 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 22:50:36 crc kubenswrapper[4791]: I1210 22:50:36.166986 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 22:50:36 crc kubenswrapper[4791]: I1210 22:50:36.167006 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 22:50:36 crc kubenswrapper[4791]: I1210 22:50:36.167023 4791 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T22:50:36Z","lastTransitionTime":"2025-12-10T22:50:36Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 22:50:36 crc kubenswrapper[4791]: I1210 22:50:36.269905 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 22:50:36 crc kubenswrapper[4791]: I1210 22:50:36.269970 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 22:50:36 crc kubenswrapper[4791]: I1210 22:50:36.269991 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 22:50:36 crc kubenswrapper[4791]: I1210 22:50:36.270014 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 22:50:36 crc kubenswrapper[4791]: I1210 22:50:36.270031 4791 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T22:50:36Z","lastTransitionTime":"2025-12-10T22:50:36Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 22:50:36 crc kubenswrapper[4791]: I1210 22:50:36.372140 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 22:50:36 crc kubenswrapper[4791]: I1210 22:50:36.372186 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 22:50:36 crc kubenswrapper[4791]: I1210 22:50:36.372200 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 22:50:36 crc kubenswrapper[4791]: I1210 22:50:36.372217 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 22:50:36 crc kubenswrapper[4791]: I1210 22:50:36.372228 4791 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T22:50:36Z","lastTransitionTime":"2025-12-10T22:50:36Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 22:50:36 crc kubenswrapper[4791]: I1210 22:50:36.474979 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 22:50:36 crc kubenswrapper[4791]: I1210 22:50:36.475024 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 22:50:36 crc kubenswrapper[4791]: I1210 22:50:36.475042 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 22:50:36 crc kubenswrapper[4791]: I1210 22:50:36.475065 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 22:50:36 crc kubenswrapper[4791]: I1210 22:50:36.475082 4791 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T22:50:36Z","lastTransitionTime":"2025-12-10T22:50:36Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 22:50:36 crc kubenswrapper[4791]: I1210 22:50:36.577849 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 22:50:36 crc kubenswrapper[4791]: I1210 22:50:36.577941 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 22:50:36 crc kubenswrapper[4791]: I1210 22:50:36.577956 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 22:50:36 crc kubenswrapper[4791]: I1210 22:50:36.577976 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 22:50:36 crc kubenswrapper[4791]: I1210 22:50:36.577990 4791 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T22:50:36Z","lastTransitionTime":"2025-12-10T22:50:36Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 22:50:36 crc kubenswrapper[4791]: I1210 22:50:36.680482 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 22:50:36 crc kubenswrapper[4791]: I1210 22:50:36.680516 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 22:50:36 crc kubenswrapper[4791]: I1210 22:50:36.680526 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 22:50:36 crc kubenswrapper[4791]: I1210 22:50:36.680540 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 22:50:36 crc kubenswrapper[4791]: I1210 22:50:36.680549 4791 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T22:50:36Z","lastTransitionTime":"2025-12-10T22:50:36Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 22:50:36 crc kubenswrapper[4791]: I1210 22:50:36.782232 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 22:50:36 crc kubenswrapper[4791]: I1210 22:50:36.782291 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 22:50:36 crc kubenswrapper[4791]: I1210 22:50:36.782304 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 22:50:36 crc kubenswrapper[4791]: I1210 22:50:36.782322 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 22:50:36 crc kubenswrapper[4791]: I1210 22:50:36.782351 4791 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T22:50:36Z","lastTransitionTime":"2025-12-10T22:50:36Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 22:50:36 crc kubenswrapper[4791]: I1210 22:50:36.791829 4791 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/b4527e71-bb58-4a39-a99c-fe614aa4c5f9-metrics-certs\") pod \"network-metrics-daemon-b6kf6\" (UID: \"b4527e71-bb58-4a39-a99c-fe614aa4c5f9\") " pod="openshift-multus/network-metrics-daemon-b6kf6" Dec 10 22:50:36 crc kubenswrapper[4791]: E1210 22:50:36.791953 4791 secret.go:188] Couldn't get secret openshift-multus/metrics-daemon-secret: object "openshift-multus"/"metrics-daemon-secret" not registered Dec 10 22:50:36 crc kubenswrapper[4791]: E1210 22:50:36.792009 4791 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/b4527e71-bb58-4a39-a99c-fe614aa4c5f9-metrics-certs podName:b4527e71-bb58-4a39-a99c-fe614aa4c5f9 nodeName:}" failed. No retries permitted until 2025-12-10 22:51:08.791991656 +0000 UTC m=+103.221609269 (durationBeforeRetry 32s). Error: MountVolume.SetUp failed for volume "metrics-certs" (UniqueName: "kubernetes.io/secret/b4527e71-bb58-4a39-a99c-fe614aa4c5f9-metrics-certs") pod "network-metrics-daemon-b6kf6" (UID: "b4527e71-bb58-4a39-a99c-fe614aa4c5f9") : object "openshift-multus"/"metrics-daemon-secret" not registered Dec 10 22:50:36 crc kubenswrapper[4791]: I1210 22:50:36.883641 4791 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 10 22:50:36 crc kubenswrapper[4791]: I1210 22:50:36.883704 4791 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 10 22:50:36 crc kubenswrapper[4791]: I1210 22:50:36.883793 4791 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-b6kf6" Dec 10 22:50:36 crc kubenswrapper[4791]: E1210 22:50:36.883785 4791 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 10 22:50:36 crc kubenswrapper[4791]: E1210 22:50:36.884030 4791 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-b6kf6" podUID="b4527e71-bb58-4a39-a99c-fe614aa4c5f9" Dec 10 22:50:36 crc kubenswrapper[4791]: I1210 22:50:36.884085 4791 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 10 22:50:36 crc kubenswrapper[4791]: E1210 22:50:36.884152 4791 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 10 22:50:36 crc kubenswrapper[4791]: E1210 22:50:36.884261 4791 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 10 22:50:36 crc kubenswrapper[4791]: I1210 22:50:36.885193 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 22:50:36 crc kubenswrapper[4791]: I1210 22:50:36.885226 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 22:50:36 crc kubenswrapper[4791]: I1210 22:50:36.885236 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 22:50:36 crc kubenswrapper[4791]: I1210 22:50:36.885250 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 22:50:36 crc kubenswrapper[4791]: I1210 22:50:36.885259 4791 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T22:50:36Z","lastTransitionTime":"2025-12-10T22:50:36Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 22:50:36 crc kubenswrapper[4791]: I1210 22:50:36.987783 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 22:50:36 crc kubenswrapper[4791]: I1210 22:50:36.987820 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 22:50:36 crc kubenswrapper[4791]: I1210 22:50:36.987829 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 22:50:36 crc kubenswrapper[4791]: I1210 22:50:36.987878 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 22:50:36 crc kubenswrapper[4791]: I1210 22:50:36.987889 4791 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T22:50:36Z","lastTransitionTime":"2025-12-10T22:50:36Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 22:50:37 crc kubenswrapper[4791]: I1210 22:50:37.089797 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 22:50:37 crc kubenswrapper[4791]: I1210 22:50:37.089850 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 22:50:37 crc kubenswrapper[4791]: I1210 22:50:37.089861 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 22:50:37 crc kubenswrapper[4791]: I1210 22:50:37.089916 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 22:50:37 crc kubenswrapper[4791]: I1210 22:50:37.089930 4791 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T22:50:37Z","lastTransitionTime":"2025-12-10T22:50:37Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 22:50:37 crc kubenswrapper[4791]: I1210 22:50:37.191504 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 22:50:37 crc kubenswrapper[4791]: I1210 22:50:37.191567 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 22:50:37 crc kubenswrapper[4791]: I1210 22:50:37.191582 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 22:50:37 crc kubenswrapper[4791]: I1210 22:50:37.191599 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 22:50:37 crc kubenswrapper[4791]: I1210 22:50:37.191629 4791 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T22:50:37Z","lastTransitionTime":"2025-12-10T22:50:37Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 22:50:37 crc kubenswrapper[4791]: I1210 22:50:37.294378 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 22:50:37 crc kubenswrapper[4791]: I1210 22:50:37.294419 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 22:50:37 crc kubenswrapper[4791]: I1210 22:50:37.294430 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 22:50:37 crc kubenswrapper[4791]: I1210 22:50:37.294445 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 22:50:37 crc kubenswrapper[4791]: I1210 22:50:37.294457 4791 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T22:50:37Z","lastTransitionTime":"2025-12-10T22:50:37Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 22:50:37 crc kubenswrapper[4791]: I1210 22:50:37.400244 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 22:50:37 crc kubenswrapper[4791]: I1210 22:50:37.400278 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 22:50:37 crc kubenswrapper[4791]: I1210 22:50:37.400287 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 22:50:37 crc kubenswrapper[4791]: I1210 22:50:37.400300 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 22:50:37 crc kubenswrapper[4791]: I1210 22:50:37.400309 4791 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T22:50:37Z","lastTransitionTime":"2025-12-10T22:50:37Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 22:50:37 crc kubenswrapper[4791]: I1210 22:50:37.502428 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 22:50:37 crc kubenswrapper[4791]: I1210 22:50:37.502750 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 22:50:37 crc kubenswrapper[4791]: I1210 22:50:37.502920 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 22:50:37 crc kubenswrapper[4791]: I1210 22:50:37.503104 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 22:50:37 crc kubenswrapper[4791]: I1210 22:50:37.503303 4791 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T22:50:37Z","lastTransitionTime":"2025-12-10T22:50:37Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 22:50:37 crc kubenswrapper[4791]: I1210 22:50:37.608977 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 22:50:37 crc kubenswrapper[4791]: I1210 22:50:37.609030 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 22:50:37 crc kubenswrapper[4791]: I1210 22:50:37.609041 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 22:50:37 crc kubenswrapper[4791]: I1210 22:50:37.609060 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 22:50:37 crc kubenswrapper[4791]: I1210 22:50:37.609072 4791 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T22:50:37Z","lastTransitionTime":"2025-12-10T22:50:37Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 22:50:37 crc kubenswrapper[4791]: I1210 22:50:37.711997 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 22:50:37 crc kubenswrapper[4791]: I1210 22:50:37.712053 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 22:50:37 crc kubenswrapper[4791]: I1210 22:50:37.712069 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 22:50:37 crc kubenswrapper[4791]: I1210 22:50:37.712091 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 22:50:37 crc kubenswrapper[4791]: I1210 22:50:37.712107 4791 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T22:50:37Z","lastTransitionTime":"2025-12-10T22:50:37Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 22:50:37 crc kubenswrapper[4791]: I1210 22:50:37.814448 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 22:50:37 crc kubenswrapper[4791]: I1210 22:50:37.814488 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 22:50:37 crc kubenswrapper[4791]: I1210 22:50:37.814499 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 22:50:37 crc kubenswrapper[4791]: I1210 22:50:37.814513 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 22:50:37 crc kubenswrapper[4791]: I1210 22:50:37.814523 4791 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T22:50:37Z","lastTransitionTime":"2025-12-10T22:50:37Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 22:50:37 crc kubenswrapper[4791]: I1210 22:50:37.895751 4791 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-machine-config-operator/kube-rbac-proxy-crio-crc"] Dec 10 22:50:37 crc kubenswrapper[4791]: I1210 22:50:37.916763 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 22:50:37 crc kubenswrapper[4791]: I1210 22:50:37.916816 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 22:50:37 crc kubenswrapper[4791]: I1210 22:50:37.916830 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 22:50:37 crc kubenswrapper[4791]: I1210 22:50:37.916846 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 22:50:37 crc kubenswrapper[4791]: I1210 22:50:37.916857 4791 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T22:50:37Z","lastTransitionTime":"2025-12-10T22:50:37Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 22:50:38 crc kubenswrapper[4791]: I1210 22:50:38.019476 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 22:50:38 crc kubenswrapper[4791]: I1210 22:50:38.019573 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 22:50:38 crc kubenswrapper[4791]: I1210 22:50:38.019593 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 22:50:38 crc kubenswrapper[4791]: I1210 22:50:38.019614 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 22:50:38 crc kubenswrapper[4791]: I1210 22:50:38.019626 4791 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T22:50:38Z","lastTransitionTime":"2025-12-10T22:50:38Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 22:50:38 crc kubenswrapper[4791]: I1210 22:50:38.122237 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 22:50:38 crc kubenswrapper[4791]: I1210 22:50:38.122273 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 22:50:38 crc kubenswrapper[4791]: I1210 22:50:38.122282 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 22:50:38 crc kubenswrapper[4791]: I1210 22:50:38.122298 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 22:50:38 crc kubenswrapper[4791]: I1210 22:50:38.122307 4791 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T22:50:38Z","lastTransitionTime":"2025-12-10T22:50:38Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 22:50:38 crc kubenswrapper[4791]: I1210 22:50:38.224896 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 22:50:38 crc kubenswrapper[4791]: I1210 22:50:38.224935 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 22:50:38 crc kubenswrapper[4791]: I1210 22:50:38.224944 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 22:50:38 crc kubenswrapper[4791]: I1210 22:50:38.224958 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 22:50:38 crc kubenswrapper[4791]: I1210 22:50:38.224968 4791 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T22:50:38Z","lastTransitionTime":"2025-12-10T22:50:38Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 22:50:38 crc kubenswrapper[4791]: I1210 22:50:38.328725 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 22:50:38 crc kubenswrapper[4791]: I1210 22:50:38.328770 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 22:50:38 crc kubenswrapper[4791]: I1210 22:50:38.328782 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 22:50:38 crc kubenswrapper[4791]: I1210 22:50:38.328798 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 22:50:38 crc kubenswrapper[4791]: I1210 22:50:38.328810 4791 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T22:50:38Z","lastTransitionTime":"2025-12-10T22:50:38Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 22:50:38 crc kubenswrapper[4791]: I1210 22:50:38.431827 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 22:50:38 crc kubenswrapper[4791]: I1210 22:50:38.431888 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 22:50:38 crc kubenswrapper[4791]: I1210 22:50:38.431905 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 22:50:38 crc kubenswrapper[4791]: I1210 22:50:38.431930 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 22:50:38 crc kubenswrapper[4791]: I1210 22:50:38.431947 4791 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T22:50:38Z","lastTransitionTime":"2025-12-10T22:50:38Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 22:50:38 crc kubenswrapper[4791]: I1210 22:50:38.462247 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 22:50:38 crc kubenswrapper[4791]: I1210 22:50:38.462275 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 22:50:38 crc kubenswrapper[4791]: I1210 22:50:38.462307 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 22:50:38 crc kubenswrapper[4791]: I1210 22:50:38.462321 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 22:50:38 crc kubenswrapper[4791]: I1210 22:50:38.462331 4791 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T22:50:38Z","lastTransitionTime":"2025-12-10T22:50:38Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 22:50:38 crc kubenswrapper[4791]: E1210 22:50:38.477434 4791 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-12-10T22:50:38Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-10T22:50:38Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-10T22:50:38Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-10T22:50:38Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-10T22:50:38Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-10T22:50:38Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-10T22:50:38Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-10T22:50:38Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"4fab3af6-f657-4dfc-8ef2-12b7c978c94f\\\",\\\"systemUUID\\\":\\\"442c0b68-8dce-47e0-bd68-d5210d7e0493\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T22:50:38Z is after 2025-08-24T17:21:41Z" Dec 10 22:50:38 crc kubenswrapper[4791]: I1210 22:50:38.482854 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 22:50:38 crc kubenswrapper[4791]: I1210 22:50:38.482888 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 22:50:38 crc kubenswrapper[4791]: I1210 22:50:38.482897 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 22:50:38 crc kubenswrapper[4791]: I1210 22:50:38.482912 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 22:50:38 crc kubenswrapper[4791]: I1210 22:50:38.482922 4791 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T22:50:38Z","lastTransitionTime":"2025-12-10T22:50:38Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 22:50:38 crc kubenswrapper[4791]: E1210 22:50:38.497184 4791 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-12-10T22:50:38Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-10T22:50:38Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-10T22:50:38Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-10T22:50:38Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-10T22:50:38Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-10T22:50:38Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-10T22:50:38Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-10T22:50:38Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"4fab3af6-f657-4dfc-8ef2-12b7c978c94f\\\",\\\"systemUUID\\\":\\\"442c0b68-8dce-47e0-bd68-d5210d7e0493\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T22:50:38Z is after 2025-08-24T17:21:41Z" Dec 10 22:50:38 crc kubenswrapper[4791]: I1210 22:50:38.502844 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 22:50:38 crc kubenswrapper[4791]: I1210 22:50:38.502875 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 22:50:38 crc kubenswrapper[4791]: I1210 22:50:38.502888 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 22:50:38 crc kubenswrapper[4791]: I1210 22:50:38.502905 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 22:50:38 crc kubenswrapper[4791]: I1210 22:50:38.502916 4791 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T22:50:38Z","lastTransitionTime":"2025-12-10T22:50:38Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 22:50:38 crc kubenswrapper[4791]: E1210 22:50:38.516538 4791 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-12-10T22:50:38Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-10T22:50:38Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-10T22:50:38Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-10T22:50:38Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-10T22:50:38Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-10T22:50:38Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-10T22:50:38Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-10T22:50:38Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"4fab3af6-f657-4dfc-8ef2-12b7c978c94f\\\",\\\"systemUUID\\\":\\\"442c0b68-8dce-47e0-bd68-d5210d7e0493\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T22:50:38Z is after 2025-08-24T17:21:41Z" Dec 10 22:50:38 crc kubenswrapper[4791]: I1210 22:50:38.520593 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 22:50:38 crc kubenswrapper[4791]: I1210 22:50:38.520624 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 22:50:38 crc kubenswrapper[4791]: I1210 22:50:38.520635 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 22:50:38 crc kubenswrapper[4791]: I1210 22:50:38.520671 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 22:50:38 crc kubenswrapper[4791]: I1210 22:50:38.520682 4791 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T22:50:38Z","lastTransitionTime":"2025-12-10T22:50:38Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 22:50:38 crc kubenswrapper[4791]: E1210 22:50:38.534704 4791 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-12-10T22:50:38Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-10T22:50:38Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-10T22:50:38Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-10T22:50:38Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-10T22:50:38Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-10T22:50:38Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-10T22:50:38Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-10T22:50:38Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"4fab3af6-f657-4dfc-8ef2-12b7c978c94f\\\",\\\"systemUUID\\\":\\\"442c0b68-8dce-47e0-bd68-d5210d7e0493\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T22:50:38Z is after 2025-08-24T17:21:41Z" Dec 10 22:50:38 crc kubenswrapper[4791]: I1210 22:50:38.538865 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 22:50:38 crc kubenswrapper[4791]: I1210 22:50:38.538895 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 22:50:38 crc kubenswrapper[4791]: I1210 22:50:38.538905 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 22:50:38 crc kubenswrapper[4791]: I1210 22:50:38.538936 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 22:50:38 crc kubenswrapper[4791]: I1210 22:50:38.538945 4791 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T22:50:38Z","lastTransitionTime":"2025-12-10T22:50:38Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 22:50:38 crc kubenswrapper[4791]: E1210 22:50:38.552970 4791 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-12-10T22:50:38Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-10T22:50:38Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-10T22:50:38Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-10T22:50:38Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-10T22:50:38Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-10T22:50:38Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-10T22:50:38Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-10T22:50:38Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"4fab3af6-f657-4dfc-8ef2-12b7c978c94f\\\",\\\"systemUUID\\\":\\\"442c0b68-8dce-47e0-bd68-d5210d7e0493\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T22:50:38Z is after 2025-08-24T17:21:41Z" Dec 10 22:50:38 crc kubenswrapper[4791]: E1210 22:50:38.553149 4791 kubelet_node_status.go:572] "Unable to update node status" err="update node status exceeds retry count" Dec 10 22:50:38 crc kubenswrapper[4791]: I1210 22:50:38.554865 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 22:50:38 crc kubenswrapper[4791]: I1210 22:50:38.554921 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 22:50:38 crc kubenswrapper[4791]: I1210 22:50:38.554941 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 22:50:38 crc kubenswrapper[4791]: I1210 22:50:38.554966 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 22:50:38 crc kubenswrapper[4791]: I1210 22:50:38.554986 4791 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T22:50:38Z","lastTransitionTime":"2025-12-10T22:50:38Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 22:50:38 crc kubenswrapper[4791]: I1210 22:50:38.657290 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 22:50:38 crc kubenswrapper[4791]: I1210 22:50:38.657319 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 22:50:38 crc kubenswrapper[4791]: I1210 22:50:38.657328 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 22:50:38 crc kubenswrapper[4791]: I1210 22:50:38.657362 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 22:50:38 crc kubenswrapper[4791]: I1210 22:50:38.657375 4791 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T22:50:38Z","lastTransitionTime":"2025-12-10T22:50:38Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 22:50:38 crc kubenswrapper[4791]: I1210 22:50:38.759861 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 22:50:38 crc kubenswrapper[4791]: I1210 22:50:38.759908 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 22:50:38 crc kubenswrapper[4791]: I1210 22:50:38.759917 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 22:50:38 crc kubenswrapper[4791]: I1210 22:50:38.759929 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 22:50:38 crc kubenswrapper[4791]: I1210 22:50:38.759937 4791 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T22:50:38Z","lastTransitionTime":"2025-12-10T22:50:38Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 22:50:38 crc kubenswrapper[4791]: I1210 22:50:38.862724 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 22:50:38 crc kubenswrapper[4791]: I1210 22:50:38.862794 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 22:50:38 crc kubenswrapper[4791]: I1210 22:50:38.862810 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 22:50:38 crc kubenswrapper[4791]: I1210 22:50:38.862834 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 22:50:38 crc kubenswrapper[4791]: I1210 22:50:38.862851 4791 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T22:50:38Z","lastTransitionTime":"2025-12-10T22:50:38Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 22:50:38 crc kubenswrapper[4791]: I1210 22:50:38.884467 4791 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 10 22:50:38 crc kubenswrapper[4791]: I1210 22:50:38.884488 4791 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-b6kf6" Dec 10 22:50:38 crc kubenswrapper[4791]: I1210 22:50:38.884533 4791 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 10 22:50:38 crc kubenswrapper[4791]: I1210 22:50:38.884492 4791 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 10 22:50:38 crc kubenswrapper[4791]: E1210 22:50:38.884611 4791 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 10 22:50:38 crc kubenswrapper[4791]: E1210 22:50:38.884774 4791 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-b6kf6" podUID="b4527e71-bb58-4a39-a99c-fe614aa4c5f9" Dec 10 22:50:38 crc kubenswrapper[4791]: E1210 22:50:38.884927 4791 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 10 22:50:38 crc kubenswrapper[4791]: E1210 22:50:38.885043 4791 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 10 22:50:38 crc kubenswrapper[4791]: I1210 22:50:38.966507 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 22:50:38 crc kubenswrapper[4791]: I1210 22:50:38.966550 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 22:50:38 crc kubenswrapper[4791]: I1210 22:50:38.966559 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 22:50:38 crc kubenswrapper[4791]: I1210 22:50:38.966576 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 22:50:38 crc kubenswrapper[4791]: I1210 22:50:38.966586 4791 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T22:50:38Z","lastTransitionTime":"2025-12-10T22:50:38Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 22:50:39 crc kubenswrapper[4791]: I1210 22:50:39.069189 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 22:50:39 crc kubenswrapper[4791]: I1210 22:50:39.069228 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 22:50:39 crc kubenswrapper[4791]: I1210 22:50:39.069239 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 22:50:39 crc kubenswrapper[4791]: I1210 22:50:39.069254 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 22:50:39 crc kubenswrapper[4791]: I1210 22:50:39.069264 4791 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T22:50:39Z","lastTransitionTime":"2025-12-10T22:50:39Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 22:50:39 crc kubenswrapper[4791]: I1210 22:50:39.172487 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 22:50:39 crc kubenswrapper[4791]: I1210 22:50:39.172550 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 22:50:39 crc kubenswrapper[4791]: I1210 22:50:39.172560 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 22:50:39 crc kubenswrapper[4791]: I1210 22:50:39.172578 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 22:50:39 crc kubenswrapper[4791]: I1210 22:50:39.172588 4791 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T22:50:39Z","lastTransitionTime":"2025-12-10T22:50:39Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 22:50:39 crc kubenswrapper[4791]: I1210 22:50:39.274208 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 22:50:39 crc kubenswrapper[4791]: I1210 22:50:39.274249 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 22:50:39 crc kubenswrapper[4791]: I1210 22:50:39.274259 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 22:50:39 crc kubenswrapper[4791]: I1210 22:50:39.274274 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 22:50:39 crc kubenswrapper[4791]: I1210 22:50:39.274284 4791 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T22:50:39Z","lastTransitionTime":"2025-12-10T22:50:39Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 22:50:39 crc kubenswrapper[4791]: I1210 22:50:39.376423 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 22:50:39 crc kubenswrapper[4791]: I1210 22:50:39.376463 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 22:50:39 crc kubenswrapper[4791]: I1210 22:50:39.376474 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 22:50:39 crc kubenswrapper[4791]: I1210 22:50:39.376491 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 22:50:39 crc kubenswrapper[4791]: I1210 22:50:39.376502 4791 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T22:50:39Z","lastTransitionTime":"2025-12-10T22:50:39Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 22:50:39 crc kubenswrapper[4791]: I1210 22:50:39.479377 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 22:50:39 crc kubenswrapper[4791]: I1210 22:50:39.479649 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 22:50:39 crc kubenswrapper[4791]: I1210 22:50:39.479729 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 22:50:39 crc kubenswrapper[4791]: I1210 22:50:39.479802 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 22:50:39 crc kubenswrapper[4791]: I1210 22:50:39.479876 4791 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T22:50:39Z","lastTransitionTime":"2025-12-10T22:50:39Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 22:50:39 crc kubenswrapper[4791]: I1210 22:50:39.511260 4791 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-multus_multus-4nwkq_672aa28c-8169-49ed-87b8-21187d13a80c/kube-multus/0.log" Dec 10 22:50:39 crc kubenswrapper[4791]: I1210 22:50:39.511323 4791 generic.go:334] "Generic (PLEG): container finished" podID="672aa28c-8169-49ed-87b8-21187d13a80c" containerID="a2f2a62315529e148a10a33ccb241e9417fbde0fe42afb61cb7ee2438c74636f" exitCode=1 Dec 10 22:50:39 crc kubenswrapper[4791]: I1210 22:50:39.511381 4791 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-4nwkq" event={"ID":"672aa28c-8169-49ed-87b8-21187d13a80c","Type":"ContainerDied","Data":"a2f2a62315529e148a10a33ccb241e9417fbde0fe42afb61cb7ee2438c74636f"} Dec 10 22:50:39 crc kubenswrapper[4791]: I1210 22:50:39.511805 4791 scope.go:117] "RemoveContainer" containerID="a2f2a62315529e148a10a33ccb241e9417fbde0fe42afb61cb7ee2438c74636f" Dec 10 22:50:39 crc kubenswrapper[4791]: I1210 22:50:39.531698 4791 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-tchxk" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"c0ab74e0-42a9-4f42-ac80-16030cf4ffff\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T22:49:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T22:49:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T22:49:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T22:49:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://04e0fd1b53256b7a938e8ccab93e2a4eb73dd1a59e0b9799e09c55e511baf7fe\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T22:49:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zn2qt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-10T22:49:50Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-tchxk\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T22:50:39Z is after 2025-08-24T17:21:41Z" Dec 10 22:50:39 crc kubenswrapper[4791]: I1210 22:50:39.546026 4791 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-rw584" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ff9beb90-69b0-4732-bf37-0b81f58ecc98\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T22:49:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T22:49:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T22:49:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T22:49:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ddb9df501dc6e8e5561fbc2aefd431037a3f28f73bf8e82a6598d261a4224807\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T22:49:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fksqn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ce67e54b696bb53ffbbe48c6ce71476b635e0868993e9c427b22f904891433ee\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ce67e54b696bb53ffbbe48c6ce71476b635e0868993e9c427b22f904891433ee\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-10T22:49:52Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-10T22:49:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fksqn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://818d89bfffb8a110670152c8480684665ea8db578e9768f524fbbe6a92f27c22\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://818d89bfffb8a110670152c8480684665ea8db578e9768f524fbbe6a92f27c22\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-10T22:49:53Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-10T22:49:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fksqn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://9aa9c06e894bb2a2ed3738defaeff68a0a0458ead48177f36b90770944c211a0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://9aa9c06e894bb2a2ed3738defaeff68a0a0458ead48177f36b90770944c211a0\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-10T22:49:54Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-10T22:49:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fksqn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://dfd6d03336dcea7529419a30e8d70d84e2290a0fa5b0e46d09ad341c190638d6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://dfd6d03336dcea7529419a30e8d70d84e2290a0fa5b0e46d09ad341c190638d6\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-10T22:49:55Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-10T22:49:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fksqn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b86738401ece6fbc3083782720c66a515850ed78d4054f34fcabab7ff525b85f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b86738401ece6fbc3083782720c66a515850ed78d4054f34fcabab7ff525b85f\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-10T22:49:56Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-10T22:49:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fksqn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://18b326f572a899ddd10197a745bf9264d4ddd975690ad1cfb0ec576ccd5cf102\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://18b326f572a899ddd10197a745bf9264d4ddd975690ad1cfb0ec576ccd5cf102\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-10T22:49:57Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-10T22:49:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fksqn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-10T22:49:51Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-rw584\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T22:50:39Z is after 2025-08-24T17:21:41Z" Dec 10 22:50:39 crc kubenswrapper[4791]: I1210 22:50:39.562522 4791 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-10T22:49:52Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-10T22:49:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-10T22:49:52Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://20651eb0c16e2c71222f56da1ff755cb0e7b084886acbbb24f0ccaf57041ed85\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T22:49:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T22:50:39Z is after 2025-08-24T17:21:41Z" Dec 10 22:50:39 crc kubenswrapper[4791]: I1210 22:50:39.576201 4791 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-10T22:49:50Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-10T22:49:50Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T22:50:39Z is after 2025-08-24T17:21:41Z" Dec 10 22:50:39 crc kubenswrapper[4791]: I1210 22:50:39.582811 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 22:50:39 crc kubenswrapper[4791]: I1210 22:50:39.582866 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 22:50:39 crc kubenswrapper[4791]: I1210 22:50:39.582883 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 22:50:39 crc kubenswrapper[4791]: I1210 22:50:39.582902 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 22:50:39 crc kubenswrapper[4791]: I1210 22:50:39.582919 4791 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T22:50:39Z","lastTransitionTime":"2025-12-10T22:50:39Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 22:50:39 crc kubenswrapper[4791]: I1210 22:50:39.588473 4791 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-10T22:49:50Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-10T22:49:50Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T22:50:39Z is after 2025-08-24T17:21:41Z" Dec 10 22:50:39 crc kubenswrapper[4791]: I1210 22:50:39.604175 4791 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-10T22:49:50Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-10T22:49:50Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T22:50:39Z is after 2025-08-24T17:21:41Z" Dec 10 22:50:39 crc kubenswrapper[4791]: I1210 22:50:39.614979 4791 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-10T22:49:54Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-10T22:49:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-10T22:49:54Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://62bfaf5d7a6d1b452cdaf88086447d0c8cb7306e94434e46752be5475124d56a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T22:49:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T22:50:39Z is after 2025-08-24T17:21:41Z" Dec 10 22:50:39 crc kubenswrapper[4791]: I1210 22:50:39.627790 4791 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-b6kf6" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b4527e71-bb58-4a39-a99c-fe614aa4c5f9\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T22:50:04Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T22:50:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T22:50:04Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T22:50:04Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ngpr4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ngpr4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-10T22:50:04Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-b6kf6\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T22:50:39Z is after 2025-08-24T17:21:41Z" Dec 10 22:50:39 crc kubenswrapper[4791]: I1210 22:50:39.643007 4791 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f8349fea-c49b-49a1-b5ee-032c5dffe021\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T22:49:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T22:49:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T22:49:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T22:49:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T22:49:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://acafe12186df99d1059a31260de0eafce35ff7771bcb8f5c6753b72665d6f60b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T22:49:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ea4c346507862a8c12756a3061d6e25e788a8d79f4621b62f81f4322185c82a8\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T22:49:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://2e6e324ec6f4f7bb69fdfab4ea3ac1aa55e480dd8e2a8984b0fe5b4c9f66636e\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T22:49:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://1b2e11ff2bf8789fa1d0d23ae648b3e6ef653106e5bc4bbb5dd6447f7cd6f05d\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d6aa7f8c5d908d90008f2df49860ac8a3480136da567e3d85f86135d779204f9\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-12-10T22:49:44Z\\\",\\\"message\\\":\\\"W1210 22:49:33.845733 1 cmd.go:257] Using insecure, self-signed certificates\\\\nI1210 22:49:33.846108 1 crypto.go:601] Generating new CA for check-endpoints-signer@1765406973 cert, and key in /tmp/serving-cert-2585701456/serving-signer.crt, /tmp/serving-cert-2585701456/serving-signer.key\\\\nI1210 22:49:34.529748 1 observer_polling.go:159] Starting file observer\\\\nW1210 22:49:34.532225 1 builder.go:272] unable to get owner reference (falling back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": dial tcp [::1]:6443: connect: connection refused\\\\nI1210 22:49:34.532599 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1210 22:49:34.533977 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-2585701456/tls.crt::/tmp/serving-cert-2585701456/tls.key\\\\\\\"\\\\nF1210 22:49:44.774710 1 cmd.go:182] error initializing delegating authentication: unable to load configmap based request-header-client-ca-file: Get \\\\\\\"https://localhost:6443/api/v1/namespaces/kube-system/configmaps/extension-apiserver-authentication\\\\\\\": net/http: TLS handshake timeout\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-10T22:49:33Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T22:49:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://0c4d036ab269f62561fe7ef1586012cb3a16e3583a08b9488a0616ec00c8e0ae\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T22:49:32Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c5856d6b26c19fda13b5c9ac2892a7e2f9c6f79feb8173ecfd9b33b673327d67\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c5856d6b26c19fda13b5c9ac2892a7e2f9c6f79feb8173ecfd9b33b673327d67\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-10T22:49:27Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-10T22:49:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-10T22:49:25Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T22:50:39Z is after 2025-08-24T17:21:41Z" Dec 10 22:50:39 crc kubenswrapper[4791]: I1210 22:50:39.654713 4791 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"02a949d8-38e2-4be3-989f-40a015407140\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T22:49:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T22:49:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T22:50:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T22:50:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T22:49:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://cd4158bc59b30f0f1716bf43c1ae7c9916a5b7d7164d074987fc1f19ba49b778\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T22:49:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://656065fd3cbdc6a25dab5d06c7baa00af16d8f62fc5da687efe9c6cc91945de0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T22:49:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://90ed7fff816b1772ba9ecd95587a59cbed7c30ac302fff7e7c5de934c912bcf4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T22:49:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9b7a5502803eaad9f76f5ffa91b29fbb2657402065a03a4d6433c096aac7264f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"wait-for-host-port\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://9b7a5502803eaad9f76f5ffa91b29fbb2657402065a03a4d6433c096aac7264f\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-10T22:49:27Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-10T22:49:27Z\\\"}}}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-10T22:49:25Z\\\"}}\" for pod \"openshift-kube-scheduler\"/\"openshift-kube-scheduler-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T22:50:39Z is after 2025-08-24T17:21:41Z" Dec 10 22:50:39 crc kubenswrapper[4791]: I1210 22:50:39.677944 4791 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"90250e13-6c63-443a-a5e5-6e4e341a2289\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T22:49:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T22:49:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T22:49:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T22:49:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T22:49:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ba54f95b1e00843b854db9a10fe6897592159ccacf30ed0f76002093a13eedd3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T22:49:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5aa8ffa518f3a68b60b275f8df4851b2564aa4984b61efdc9470b73072133e43\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T22:49:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://16f989503edc78d9451638e2cf6cf4b19a2cf2b8809b10b237541cb829303091\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T22:49:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://82556824d63ce6e46df14c45689ff1f81ebf729abe902061815c040b323898e9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T22:49:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://61db531940fa6f2effdae5a005b4c91bd2698bd5d1bff9588c171a65ad550eca\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T22:49:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0f65c864d885319ce2a08d61c30be24b3ffb7d23ea30c82ed7faf039cb2c006d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://0f65c864d885319ce2a08d61c30be24b3ffb7d23ea30c82ed7faf039cb2c006d\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-10T22:49:27Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-10T22:49:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e31a76d1d5b2812ba83e24e38268e25ec8bed98fc3da9adcd06bda5822b6ace1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e31a76d1d5b2812ba83e24e38268e25ec8bed98fc3da9adcd06bda5822b6ace1\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-10T22:49:30Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-10T22:49:29Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://b5b5eceae562f3986b543d41de1e977efb85a91d84c98e09159ca6d092eda691\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b5b5eceae562f3986b543d41de1e977efb85a91d84c98e09159ca6d092eda691\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-10T22:49:31Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-10T22:49:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-10T22:49:25Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T22:50:39Z is after 2025-08-24T17:21:41Z" Dec 10 22:50:39 crc kubenswrapper[4791]: I1210 22:50:39.685276 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 22:50:39 crc kubenswrapper[4791]: I1210 22:50:39.685312 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 22:50:39 crc kubenswrapper[4791]: I1210 22:50:39.685323 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 22:50:39 crc kubenswrapper[4791]: I1210 22:50:39.685358 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 22:50:39 crc kubenswrapper[4791]: I1210 22:50:39.685369 4791 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T22:50:39Z","lastTransitionTime":"2025-12-10T22:50:39Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 22:50:39 crc kubenswrapper[4791]: I1210 22:50:39.689664 4791 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-5rb5l" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ba35653c-6e06-4cee-a4d6-137764090d18\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T22:49:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T22:49:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T22:49:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T22:49:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://dbb5c875730a98ad2daf316cfafc08fcdce21ac16148154780ac9193ed405dab\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T22:49:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fvvrs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://185eb2ff44221a514878283ec873773a6c9a9073dc64e296f982e60dee1338e4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T22:49:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fvvrs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-10T22:49:51Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-5rb5l\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T22:50:39Z is after 2025-08-24T17:21:41Z" Dec 10 22:50:39 crc kubenswrapper[4791]: I1210 22:50:39.699033 4791 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-s6rk8" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5c9d0978-a6f9-49c5-b377-6619b4e3ff52\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T22:50:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T22:50:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T22:50:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T22:50:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8e55baaf460c794ab196e237d51d12de059aa038617dc2d5366d49c529f5d5c7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T22:50:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7vkgm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://51d1b9b086c60bae9223cb51d5a998d02790a0d91b745aab10392c23e617eb9f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T22:50:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7vkgm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-10T22:50:03Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-s6rk8\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T22:50:39Z is after 2025-08-24T17:21:41Z" Dec 10 22:50:39 crc kubenswrapper[4791]: I1210 22:50:39.709163 4791 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-vczjq" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b9f00f25-3fa1-4b63-9710-fd352224b01b\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T22:49:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T22:49:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T22:49:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T22:49:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4698cf4c6b55b41e628448d37c2eb847c2777d0efbd93afcbf367d5e78623b7e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T22:49:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lrrpv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-10T22:49:50Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-vczjq\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T22:50:39Z is after 2025-08-24T17:21:41Z" Dec 10 22:50:39 crc kubenswrapper[4791]: I1210 22:50:39.722123 4791 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-4nwkq" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"672aa28c-8169-49ed-87b8-21187d13a80c\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T22:49:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T22:49:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T22:50:39Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T22:50:39Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a2f2a62315529e148a10a33ccb241e9417fbde0fe42afb61cb7ee2438c74636f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a2f2a62315529e148a10a33ccb241e9417fbde0fe42afb61cb7ee2438c74636f\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-12-10T22:50:38Z\\\",\\\"message\\\":\\\"2025-12-10T22:49:52+00:00 [cnibincopy] Successfully copied files in /usr/src/multus-cni/rhel9/bin/ to /host/opt/cni/bin/upgrade_e48d9423-4bca-43c4-9eb9-b55a19a5ffd6\\\\n2025-12-10T22:49:52+00:00 [cnibincopy] Successfully moved files in /host/opt/cni/bin/upgrade_e48d9423-4bca-43c4-9eb9-b55a19a5ffd6 to /host/opt/cni/bin/\\\\n2025-12-10T22:49:52Z [verbose] multus-daemon started\\\\n2025-12-10T22:49:52Z [verbose] Readiness Indicator file check\\\\n2025-12-10T22:50:37Z [error] have you checked that your default network is ready? still waiting for readinessindicatorfile @ /host/run/multus/cni/net.d/10-ovn-kubernetes.conf. pollimmediate error: timed out waiting for the condition\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-10T22:49:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xl62j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-10T22:49:51Z\\\"}}\" for pod \"openshift-multus\"/\"multus-4nwkq\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T22:50:39Z is after 2025-08-24T17:21:41Z" Dec 10 22:50:39 crc kubenswrapper[4791]: I1210 22:50:39.749360 4791 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-zhq64" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3cd47739-0fa9-4321-aff1-220f8721a0b3\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T22:49:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T22:49:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T22:49:52Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T22:49:52Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://68795a29944e56870ffe94fa732e9dcf0bfb344c0d85d1e590f58a876129c61f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T22:49:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wvpw8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://bc721b530c08b433756e978ec0423cc18aabf6771b61eb96f903b4d57ea335ff\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T22:49:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wvpw8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://152a7618be8f38fadf4e564b3948980503cedcdb466c37344ccc20f4a3173a5f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T22:49:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wvpw8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3d18f9cb7207efb1e32de75a30d33e024f656a08f0503db95e5488bc5280971f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T22:49:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wvpw8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://937137ae9d7480cdc7c59bc4a1df0a4c10e72b3b63693af285278789953ed218\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T22:49:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wvpw8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://1a52d19b7671e8b143017c8a981a6f8d3ee3b84650e4b223b814a531ac76f364\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T22:49:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wvpw8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b6d739a4314c4fb536810b8623f691d6aaf1090b9c83f17b484d2b22f7d3ad01\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b6d739a4314c4fb536810b8623f691d6aaf1090b9c83f17b484d2b22f7d3ad01\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-12-10T22:50:21Z\\\",\\\"message\\\":\\\"llocateLoadBalancerNodePorts:nil,LoadBalancerClass:nil,InternalTrafficPolicy:*Cluster,TrafficDistribution:nil,},Status:ServiceStatus{LoadBalancer:LoadBalancerStatus{Ingress:[]LoadBalancerIngress{},},Conditions:[]Condition{},},}\\\\nI1210 22:50:21.614710 6424 metrics.go:553] Stopping metrics server at address \\\\\\\"127.0.0.1:29103\\\\\\\"\\\\nI1210 22:50:21.614731 6424 lb_config.go:1031] Cluster endpoints for openshift-kube-scheduler/scheduler for network=default are: map[]\\\\nI1210 22:50:21.614619 6424 services_controller.go:360] Finished syncing service ovn-kubernetes-control-plane on namespace openshift-ovn-kubernetes for network=default : 7.02µs\\\\nI1210 22:50:21.614757 6424 services_controller.go:356] Processing sync for service openshift-route-controller-manager/route-controller-manager for network=default\\\\nI1210 22:50:21.614615 6424 services_controller.go:444] Built service openshift-machine-config-operator/machine-config-daemon LB per-node configs for network=default: []services.lbConfig(nil)\\\\nF1210 22:50:21.614772 6424 ovnkube.go:137] failed to run ovnkube: [failed to start network controller: failed to start default network controller: unable to create admin network policy controller, err: could not add Event Handler for anpInformer during admin network policy controller initialization, handler {0x1fcc6e0 0x1fcc3c0 0x1fcc360} was not added to\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-10T22:50:20Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":2,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 20s restarting failed container=ovnkube-controller pod=ovnkube-node-zhq64_openshift-ovn-kubernetes(3cd47739-0fa9-4321-aff1-220f8721a0b3)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wvpw8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://dd3f80e5713221f96ed50da7f2e7fd5b9fabfcd5c1656da8d8a86d818a5fdc3f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T22:49:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wvpw8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3afbd1ffcd2c4b40324aca6e52d6fc822db131c3f35a9af3a6f005c2cffdb7bf\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3afbd1ffcd2c4b40324aca6e52d6fc822db131c3f35a9af3a6f005c2cffdb7bf\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-10T22:49:53Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-10T22:49:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wvpw8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-10T22:49:52Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-zhq64\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T22:50:39Z is after 2025-08-24T17:21:41Z" Dec 10 22:50:39 crc kubenswrapper[4791]: I1210 22:50:39.762716 4791 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"698bbe9a-fbe5-436d-a835-5e80df77ba24\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T22:49:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T22:49:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T22:49:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T22:49:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T22:49:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://224d73713312376a58ec6cc2bc8fdb6ffa44c4749ae0e7691d5fe27b9afca09f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T22:49:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://372eeaffd40992c6f06efc4df6dfcb44b31de64ee8256c21508cdd7c74ecb96e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T22:49:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://76b6013b062cac5b0b4d12a94f38ce7b43f9c7ea216063455d88a956ac04d26e\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T22:49:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://78e950e122e761d950da1a3c60c06476e53920e8795f6c61db22c01e22d5136e\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T22:49:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-10T22:49:25Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T22:50:39Z is after 2025-08-24T17:21:41Z" Dec 10 22:50:39 crc kubenswrapper[4791]: I1210 22:50:39.794662 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 22:50:39 crc kubenswrapper[4791]: I1210 22:50:39.794707 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 22:50:39 crc kubenswrapper[4791]: I1210 22:50:39.794720 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 22:50:39 crc kubenswrapper[4791]: I1210 22:50:39.794741 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 22:50:39 crc kubenswrapper[4791]: I1210 22:50:39.794753 4791 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T22:50:39Z","lastTransitionTime":"2025-12-10T22:50:39Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 22:50:39 crc kubenswrapper[4791]: I1210 22:50:39.836826 4791 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7a9a3709-32fb-48b4-97f4-e1a19282a085\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T22:49:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T22:49:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T22:49:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T22:49:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T22:49:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://511bd056768a15ea0f650a2b1a6e87e3c71dfc75696222a54f7376227ad480c6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-crio\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T22:49:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kube\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"var-lib-kubelet\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8c489105d0dfb302732d4f5bd2fc9d000967a071802735b778d49bdbc428b264\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://8c489105d0dfb302732d4f5bd2fc9d000967a071802735b778d49bdbc428b264\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-10T22:49:27Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-10T22:49:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var\\\",\\\"name\\\":\\\"var-lib-kubelet\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-10T22:49:25Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"kube-rbac-proxy-crio-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T22:50:39Z is after 2025-08-24T17:21:41Z" Dec 10 22:50:39 crc kubenswrapper[4791]: I1210 22:50:39.850797 4791 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-10T22:49:52Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-10T22:49:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-10T22:49:52Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3734b110802828632d44263ed0a43cabc4b710833f0be120d72db0940fe783e2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T22:49:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://794803ac9be3c52509e6dcb10854d8864e1a9e406762e57c7971b07c4da1220e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T22:49:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T22:50:39Z is after 2025-08-24T17:21:41Z" Dec 10 22:50:39 crc kubenswrapper[4791]: I1210 22:50:39.897458 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 22:50:39 crc kubenswrapper[4791]: I1210 22:50:39.897521 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 22:50:39 crc kubenswrapper[4791]: I1210 22:50:39.897536 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 22:50:39 crc kubenswrapper[4791]: I1210 22:50:39.897552 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 22:50:39 crc kubenswrapper[4791]: I1210 22:50:39.897564 4791 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T22:50:39Z","lastTransitionTime":"2025-12-10T22:50:39Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 22:50:40 crc kubenswrapper[4791]: I1210 22:50:40.000926 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 22:50:40 crc kubenswrapper[4791]: I1210 22:50:40.000964 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 22:50:40 crc kubenswrapper[4791]: I1210 22:50:40.000975 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 22:50:40 crc kubenswrapper[4791]: I1210 22:50:40.000988 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 22:50:40 crc kubenswrapper[4791]: I1210 22:50:40.000997 4791 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T22:50:40Z","lastTransitionTime":"2025-12-10T22:50:40Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 22:50:40 crc kubenswrapper[4791]: I1210 22:50:40.103818 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 22:50:40 crc kubenswrapper[4791]: I1210 22:50:40.103876 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 22:50:40 crc kubenswrapper[4791]: I1210 22:50:40.103893 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 22:50:40 crc kubenswrapper[4791]: I1210 22:50:40.103918 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 22:50:40 crc kubenswrapper[4791]: I1210 22:50:40.103934 4791 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T22:50:40Z","lastTransitionTime":"2025-12-10T22:50:40Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 22:50:40 crc kubenswrapper[4791]: I1210 22:50:40.207199 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 22:50:40 crc kubenswrapper[4791]: I1210 22:50:40.207380 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 22:50:40 crc kubenswrapper[4791]: I1210 22:50:40.207416 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 22:50:40 crc kubenswrapper[4791]: I1210 22:50:40.207440 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 22:50:40 crc kubenswrapper[4791]: I1210 22:50:40.207456 4791 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T22:50:40Z","lastTransitionTime":"2025-12-10T22:50:40Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 22:50:40 crc kubenswrapper[4791]: I1210 22:50:40.309536 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 22:50:40 crc kubenswrapper[4791]: I1210 22:50:40.309568 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 22:50:40 crc kubenswrapper[4791]: I1210 22:50:40.309594 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 22:50:40 crc kubenswrapper[4791]: I1210 22:50:40.309608 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 22:50:40 crc kubenswrapper[4791]: I1210 22:50:40.309616 4791 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T22:50:40Z","lastTransitionTime":"2025-12-10T22:50:40Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 22:50:40 crc kubenswrapper[4791]: I1210 22:50:40.412756 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 22:50:40 crc kubenswrapper[4791]: I1210 22:50:40.412849 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 22:50:40 crc kubenswrapper[4791]: I1210 22:50:40.412914 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 22:50:40 crc kubenswrapper[4791]: I1210 22:50:40.412946 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 22:50:40 crc kubenswrapper[4791]: I1210 22:50:40.412967 4791 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T22:50:40Z","lastTransitionTime":"2025-12-10T22:50:40Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 22:50:40 crc kubenswrapper[4791]: I1210 22:50:40.514930 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 22:50:40 crc kubenswrapper[4791]: I1210 22:50:40.514961 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 22:50:40 crc kubenswrapper[4791]: I1210 22:50:40.514989 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 22:50:40 crc kubenswrapper[4791]: I1210 22:50:40.515026 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 22:50:40 crc kubenswrapper[4791]: I1210 22:50:40.515036 4791 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T22:50:40Z","lastTransitionTime":"2025-12-10T22:50:40Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 22:50:40 crc kubenswrapper[4791]: I1210 22:50:40.516062 4791 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-multus_multus-4nwkq_672aa28c-8169-49ed-87b8-21187d13a80c/kube-multus/0.log" Dec 10 22:50:40 crc kubenswrapper[4791]: I1210 22:50:40.516134 4791 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-4nwkq" event={"ID":"672aa28c-8169-49ed-87b8-21187d13a80c","Type":"ContainerStarted","Data":"f337f881ffd71f4a2ee19e92bb0b92c76cbb9b8352511f741d0c87e7da4cdd18"} Dec 10 22:50:40 crc kubenswrapper[4791]: I1210 22:50:40.530585 4791 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-5rb5l" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ba35653c-6e06-4cee-a4d6-137764090d18\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T22:49:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T22:49:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T22:49:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T22:49:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://dbb5c875730a98ad2daf316cfafc08fcdce21ac16148154780ac9193ed405dab\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T22:49:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fvvrs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://185eb2ff44221a514878283ec873773a6c9a9073dc64e296f982e60dee1338e4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T22:49:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fvvrs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-10T22:49:51Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-5rb5l\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T22:50:40Z is after 2025-08-24T17:21:41Z" Dec 10 22:50:40 crc kubenswrapper[4791]: I1210 22:50:40.545226 4791 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-s6rk8" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5c9d0978-a6f9-49c5-b377-6619b4e3ff52\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T22:50:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T22:50:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T22:50:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T22:50:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8e55baaf460c794ab196e237d51d12de059aa038617dc2d5366d49c529f5d5c7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T22:50:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7vkgm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://51d1b9b086c60bae9223cb51d5a998d02790a0d91b745aab10392c23e617eb9f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T22:50:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7vkgm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-10T22:50:03Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-s6rk8\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T22:50:40Z is after 2025-08-24T17:21:41Z" Dec 10 22:50:40 crc kubenswrapper[4791]: I1210 22:50:40.556740 4791 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-10T22:49:52Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-10T22:49:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-10T22:49:52Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3734b110802828632d44263ed0a43cabc4b710833f0be120d72db0940fe783e2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T22:49:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://794803ac9be3c52509e6dcb10854d8864e1a9e406762e57c7971b07c4da1220e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T22:49:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T22:50:40Z is after 2025-08-24T17:21:41Z" Dec 10 22:50:40 crc kubenswrapper[4791]: I1210 22:50:40.570299 4791 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-vczjq" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b9f00f25-3fa1-4b63-9710-fd352224b01b\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T22:49:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T22:49:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T22:49:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T22:49:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4698cf4c6b55b41e628448d37c2eb847c2777d0efbd93afcbf367d5e78623b7e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T22:49:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lrrpv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-10T22:49:50Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-vczjq\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T22:50:40Z is after 2025-08-24T17:21:41Z" Dec 10 22:50:40 crc kubenswrapper[4791]: I1210 22:50:40.583985 4791 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-4nwkq" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"672aa28c-8169-49ed-87b8-21187d13a80c\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T22:49:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T22:49:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T22:50:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T22:50:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f337f881ffd71f4a2ee19e92bb0b92c76cbb9b8352511f741d0c87e7da4cdd18\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a2f2a62315529e148a10a33ccb241e9417fbde0fe42afb61cb7ee2438c74636f\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-12-10T22:50:38Z\\\",\\\"message\\\":\\\"2025-12-10T22:49:52+00:00 [cnibincopy] Successfully copied files in /usr/src/multus-cni/rhel9/bin/ to /host/opt/cni/bin/upgrade_e48d9423-4bca-43c4-9eb9-b55a19a5ffd6\\\\n2025-12-10T22:49:52+00:00 [cnibincopy] Successfully moved files in /host/opt/cni/bin/upgrade_e48d9423-4bca-43c4-9eb9-b55a19a5ffd6 to /host/opt/cni/bin/\\\\n2025-12-10T22:49:52Z [verbose] multus-daemon started\\\\n2025-12-10T22:49:52Z [verbose] Readiness Indicator file check\\\\n2025-12-10T22:50:37Z [error] have you checked that your default network is ready? still waiting for readinessindicatorfile @ /host/run/multus/cni/net.d/10-ovn-kubernetes.conf. pollimmediate error: timed out waiting for the condition\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-10T22:49:52Z\\\"}},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T22:50:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xl62j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-10T22:49:51Z\\\"}}\" for pod \"openshift-multus\"/\"multus-4nwkq\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T22:50:40Z is after 2025-08-24T17:21:41Z" Dec 10 22:50:40 crc kubenswrapper[4791]: I1210 22:50:40.604680 4791 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-zhq64" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3cd47739-0fa9-4321-aff1-220f8721a0b3\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T22:49:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T22:49:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T22:49:52Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T22:49:52Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://68795a29944e56870ffe94fa732e9dcf0bfb344c0d85d1e590f58a876129c61f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T22:49:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wvpw8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://bc721b530c08b433756e978ec0423cc18aabf6771b61eb96f903b4d57ea335ff\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T22:49:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wvpw8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://152a7618be8f38fadf4e564b3948980503cedcdb466c37344ccc20f4a3173a5f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T22:49:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wvpw8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3d18f9cb7207efb1e32de75a30d33e024f656a08f0503db95e5488bc5280971f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T22:49:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wvpw8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://937137ae9d7480cdc7c59bc4a1df0a4c10e72b3b63693af285278789953ed218\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T22:49:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wvpw8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://1a52d19b7671e8b143017c8a981a6f8d3ee3b84650e4b223b814a531ac76f364\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T22:49:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wvpw8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b6d739a4314c4fb536810b8623f691d6aaf1090b9c83f17b484d2b22f7d3ad01\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b6d739a4314c4fb536810b8623f691d6aaf1090b9c83f17b484d2b22f7d3ad01\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-12-10T22:50:21Z\\\",\\\"message\\\":\\\"llocateLoadBalancerNodePorts:nil,LoadBalancerClass:nil,InternalTrafficPolicy:*Cluster,TrafficDistribution:nil,},Status:ServiceStatus{LoadBalancer:LoadBalancerStatus{Ingress:[]LoadBalancerIngress{},},Conditions:[]Condition{},},}\\\\nI1210 22:50:21.614710 6424 metrics.go:553] Stopping metrics server at address \\\\\\\"127.0.0.1:29103\\\\\\\"\\\\nI1210 22:50:21.614731 6424 lb_config.go:1031] Cluster endpoints for openshift-kube-scheduler/scheduler for network=default are: map[]\\\\nI1210 22:50:21.614619 6424 services_controller.go:360] Finished syncing service ovn-kubernetes-control-plane on namespace openshift-ovn-kubernetes for network=default : 7.02µs\\\\nI1210 22:50:21.614757 6424 services_controller.go:356] Processing sync for service openshift-route-controller-manager/route-controller-manager for network=default\\\\nI1210 22:50:21.614615 6424 services_controller.go:444] Built service openshift-machine-config-operator/machine-config-daemon LB per-node configs for network=default: []services.lbConfig(nil)\\\\nF1210 22:50:21.614772 6424 ovnkube.go:137] failed to run ovnkube: [failed to start network controller: failed to start default network controller: unable to create admin network policy controller, err: could not add Event Handler for anpInformer during admin network policy controller initialization, handler {0x1fcc6e0 0x1fcc3c0 0x1fcc360} was not added to\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-10T22:50:20Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":2,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 20s restarting failed container=ovnkube-controller pod=ovnkube-node-zhq64_openshift-ovn-kubernetes(3cd47739-0fa9-4321-aff1-220f8721a0b3)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wvpw8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://dd3f80e5713221f96ed50da7f2e7fd5b9fabfcd5c1656da8d8a86d818a5fdc3f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T22:49:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wvpw8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3afbd1ffcd2c4b40324aca6e52d6fc822db131c3f35a9af3a6f005c2cffdb7bf\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3afbd1ffcd2c4b40324aca6e52d6fc822db131c3f35a9af3a6f005c2cffdb7bf\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-10T22:49:53Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-10T22:49:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wvpw8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-10T22:49:52Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-zhq64\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T22:50:40Z is after 2025-08-24T17:21:41Z" Dec 10 22:50:40 crc kubenswrapper[4791]: I1210 22:50:40.617316 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 22:50:40 crc kubenswrapper[4791]: I1210 22:50:40.617457 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 22:50:40 crc kubenswrapper[4791]: I1210 22:50:40.617482 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 22:50:40 crc kubenswrapper[4791]: I1210 22:50:40.617509 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 22:50:40 crc kubenswrapper[4791]: I1210 22:50:40.617531 4791 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T22:50:40Z","lastTransitionTime":"2025-12-10T22:50:40Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 22:50:40 crc kubenswrapper[4791]: I1210 22:50:40.622130 4791 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"698bbe9a-fbe5-436d-a835-5e80df77ba24\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T22:49:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T22:49:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T22:49:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T22:49:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T22:49:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://224d73713312376a58ec6cc2bc8fdb6ffa44c4749ae0e7691d5fe27b9afca09f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T22:49:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://372eeaffd40992c6f06efc4df6dfcb44b31de64ee8256c21508cdd7c74ecb96e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T22:49:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://76b6013b062cac5b0b4d12a94f38ce7b43f9c7ea216063455d88a956ac04d26e\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T22:49:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://78e950e122e761d950da1a3c60c06476e53920e8795f6c61db22c01e22d5136e\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T22:49:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-10T22:49:25Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T22:50:40Z is after 2025-08-24T17:21:41Z" Dec 10 22:50:40 crc kubenswrapper[4791]: I1210 22:50:40.634538 4791 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7a9a3709-32fb-48b4-97f4-e1a19282a085\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T22:49:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T22:49:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T22:49:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T22:49:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T22:49:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://511bd056768a15ea0f650a2b1a6e87e3c71dfc75696222a54f7376227ad480c6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-crio\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T22:49:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kube\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"var-lib-kubelet\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8c489105d0dfb302732d4f5bd2fc9d000967a071802735b778d49bdbc428b264\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://8c489105d0dfb302732d4f5bd2fc9d000967a071802735b778d49bdbc428b264\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-10T22:49:27Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-10T22:49:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var\\\",\\\"name\\\":\\\"var-lib-kubelet\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-10T22:49:25Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"kube-rbac-proxy-crio-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T22:50:40Z is after 2025-08-24T17:21:41Z" Dec 10 22:50:40 crc kubenswrapper[4791]: I1210 22:50:40.652056 4791 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-10T22:49:50Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-10T22:49:50Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T22:50:40Z is after 2025-08-24T17:21:41Z" Dec 10 22:50:40 crc kubenswrapper[4791]: I1210 22:50:40.664531 4791 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-tchxk" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"c0ab74e0-42a9-4f42-ac80-16030cf4ffff\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T22:49:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T22:49:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T22:49:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T22:49:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://04e0fd1b53256b7a938e8ccab93e2a4eb73dd1a59e0b9799e09c55e511baf7fe\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T22:49:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zn2qt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-10T22:49:50Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-tchxk\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T22:50:40Z is after 2025-08-24T17:21:41Z" Dec 10 22:50:40 crc kubenswrapper[4791]: I1210 22:50:40.682066 4791 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-rw584" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ff9beb90-69b0-4732-bf37-0b81f58ecc98\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T22:49:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T22:49:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T22:49:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T22:49:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ddb9df501dc6e8e5561fbc2aefd431037a3f28f73bf8e82a6598d261a4224807\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T22:49:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fksqn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ce67e54b696bb53ffbbe48c6ce71476b635e0868993e9c427b22f904891433ee\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ce67e54b696bb53ffbbe48c6ce71476b635e0868993e9c427b22f904891433ee\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-10T22:49:52Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-10T22:49:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fksqn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://818d89bfffb8a110670152c8480684665ea8db578e9768f524fbbe6a92f27c22\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://818d89bfffb8a110670152c8480684665ea8db578e9768f524fbbe6a92f27c22\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-10T22:49:53Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-10T22:49:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fksqn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://9aa9c06e894bb2a2ed3738defaeff68a0a0458ead48177f36b90770944c211a0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://9aa9c06e894bb2a2ed3738defaeff68a0a0458ead48177f36b90770944c211a0\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-10T22:49:54Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-10T22:49:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fksqn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://dfd6d03336dcea7529419a30e8d70d84e2290a0fa5b0e46d09ad341c190638d6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://dfd6d03336dcea7529419a30e8d70d84e2290a0fa5b0e46d09ad341c190638d6\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-10T22:49:55Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-10T22:49:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fksqn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b86738401ece6fbc3083782720c66a515850ed78d4054f34fcabab7ff525b85f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b86738401ece6fbc3083782720c66a515850ed78d4054f34fcabab7ff525b85f\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-10T22:49:56Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-10T22:49:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fksqn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://18b326f572a899ddd10197a745bf9264d4ddd975690ad1cfb0ec576ccd5cf102\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://18b326f572a899ddd10197a745bf9264d4ddd975690ad1cfb0ec576ccd5cf102\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-10T22:49:57Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-10T22:49:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fksqn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-10T22:49:51Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-rw584\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T22:50:40Z is after 2025-08-24T17:21:41Z" Dec 10 22:50:40 crc kubenswrapper[4791]: I1210 22:50:40.698193 4791 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-10T22:49:52Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-10T22:49:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-10T22:49:52Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://20651eb0c16e2c71222f56da1ff755cb0e7b084886acbbb24f0ccaf57041ed85\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T22:49:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T22:50:40Z is after 2025-08-24T17:21:41Z" Dec 10 22:50:40 crc kubenswrapper[4791]: I1210 22:50:40.712442 4791 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-10T22:49:50Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-10T22:49:50Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T22:50:40Z is after 2025-08-24T17:21:41Z" Dec 10 22:50:40 crc kubenswrapper[4791]: I1210 22:50:40.720248 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 22:50:40 crc kubenswrapper[4791]: I1210 22:50:40.720375 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 22:50:40 crc kubenswrapper[4791]: I1210 22:50:40.720390 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 22:50:40 crc kubenswrapper[4791]: I1210 22:50:40.720417 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 22:50:40 crc kubenswrapper[4791]: I1210 22:50:40.720432 4791 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T22:50:40Z","lastTransitionTime":"2025-12-10T22:50:40Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 22:50:40 crc kubenswrapper[4791]: I1210 22:50:40.737574 4791 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"90250e13-6c63-443a-a5e5-6e4e341a2289\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T22:49:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T22:49:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T22:49:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T22:49:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T22:49:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ba54f95b1e00843b854db9a10fe6897592159ccacf30ed0f76002093a13eedd3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T22:49:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5aa8ffa518f3a68b60b275f8df4851b2564aa4984b61efdc9470b73072133e43\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T22:49:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://16f989503edc78d9451638e2cf6cf4b19a2cf2b8809b10b237541cb829303091\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T22:49:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://82556824d63ce6e46df14c45689ff1f81ebf729abe902061815c040b323898e9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T22:49:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://61db531940fa6f2effdae5a005b4c91bd2698bd5d1bff9588c171a65ad550eca\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T22:49:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0f65c864d885319ce2a08d61c30be24b3ffb7d23ea30c82ed7faf039cb2c006d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://0f65c864d885319ce2a08d61c30be24b3ffb7d23ea30c82ed7faf039cb2c006d\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-10T22:49:27Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-10T22:49:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e31a76d1d5b2812ba83e24e38268e25ec8bed98fc3da9adcd06bda5822b6ace1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e31a76d1d5b2812ba83e24e38268e25ec8bed98fc3da9adcd06bda5822b6ace1\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-10T22:49:30Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-10T22:49:29Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://b5b5eceae562f3986b543d41de1e977efb85a91d84c98e09159ca6d092eda691\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b5b5eceae562f3986b543d41de1e977efb85a91d84c98e09159ca6d092eda691\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-10T22:49:31Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-10T22:49:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-10T22:49:25Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T22:50:40Z is after 2025-08-24T17:21:41Z" Dec 10 22:50:40 crc kubenswrapper[4791]: I1210 22:50:40.752765 4791 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-10T22:49:50Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-10T22:49:50Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T22:50:40Z is after 2025-08-24T17:21:41Z" Dec 10 22:50:40 crc kubenswrapper[4791]: I1210 22:50:40.767631 4791 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-10T22:49:54Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-10T22:49:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-10T22:49:54Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://62bfaf5d7a6d1b452cdaf88086447d0c8cb7306e94434e46752be5475124d56a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T22:49:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T22:50:40Z is after 2025-08-24T17:21:41Z" Dec 10 22:50:40 crc kubenswrapper[4791]: I1210 22:50:40.779844 4791 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-b6kf6" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b4527e71-bb58-4a39-a99c-fe614aa4c5f9\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T22:50:04Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T22:50:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T22:50:04Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T22:50:04Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ngpr4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ngpr4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-10T22:50:04Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-b6kf6\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T22:50:40Z is after 2025-08-24T17:21:41Z" Dec 10 22:50:40 crc kubenswrapper[4791]: I1210 22:50:40.792723 4791 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f8349fea-c49b-49a1-b5ee-032c5dffe021\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T22:49:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T22:49:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T22:49:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T22:49:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T22:49:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://acafe12186df99d1059a31260de0eafce35ff7771bcb8f5c6753b72665d6f60b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T22:49:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ea4c346507862a8c12756a3061d6e25e788a8d79f4621b62f81f4322185c82a8\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T22:49:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://2e6e324ec6f4f7bb69fdfab4ea3ac1aa55e480dd8e2a8984b0fe5b4c9f66636e\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T22:49:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://1b2e11ff2bf8789fa1d0d23ae648b3e6ef653106e5bc4bbb5dd6447f7cd6f05d\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d6aa7f8c5d908d90008f2df49860ac8a3480136da567e3d85f86135d779204f9\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-12-10T22:49:44Z\\\",\\\"message\\\":\\\"W1210 22:49:33.845733 1 cmd.go:257] Using insecure, self-signed certificates\\\\nI1210 22:49:33.846108 1 crypto.go:601] Generating new CA for check-endpoints-signer@1765406973 cert, and key in /tmp/serving-cert-2585701456/serving-signer.crt, /tmp/serving-cert-2585701456/serving-signer.key\\\\nI1210 22:49:34.529748 1 observer_polling.go:159] Starting file observer\\\\nW1210 22:49:34.532225 1 builder.go:272] unable to get owner reference (falling back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": dial tcp [::1]:6443: connect: connection refused\\\\nI1210 22:49:34.532599 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1210 22:49:34.533977 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-2585701456/tls.crt::/tmp/serving-cert-2585701456/tls.key\\\\\\\"\\\\nF1210 22:49:44.774710 1 cmd.go:182] error initializing delegating authentication: unable to load configmap based request-header-client-ca-file: Get \\\\\\\"https://localhost:6443/api/v1/namespaces/kube-system/configmaps/extension-apiserver-authentication\\\\\\\": net/http: TLS handshake timeout\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-10T22:49:33Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T22:49:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://0c4d036ab269f62561fe7ef1586012cb3a16e3583a08b9488a0616ec00c8e0ae\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T22:49:32Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c5856d6b26c19fda13b5c9ac2892a7e2f9c6f79feb8173ecfd9b33b673327d67\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c5856d6b26c19fda13b5c9ac2892a7e2f9c6f79feb8173ecfd9b33b673327d67\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-10T22:49:27Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-10T22:49:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-10T22:49:25Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T22:50:40Z is after 2025-08-24T17:21:41Z" Dec 10 22:50:40 crc kubenswrapper[4791]: I1210 22:50:40.803715 4791 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"02a949d8-38e2-4be3-989f-40a015407140\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T22:49:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T22:49:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T22:50:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T22:50:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T22:49:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://cd4158bc59b30f0f1716bf43c1ae7c9916a5b7d7164d074987fc1f19ba49b778\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T22:49:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://656065fd3cbdc6a25dab5d06c7baa00af16d8f62fc5da687efe9c6cc91945de0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T22:49:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://90ed7fff816b1772ba9ecd95587a59cbed7c30ac302fff7e7c5de934c912bcf4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T22:49:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9b7a5502803eaad9f76f5ffa91b29fbb2657402065a03a4d6433c096aac7264f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"wait-for-host-port\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://9b7a5502803eaad9f76f5ffa91b29fbb2657402065a03a4d6433c096aac7264f\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-10T22:49:27Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-10T22:49:27Z\\\"}}}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-10T22:49:25Z\\\"}}\" for pod \"openshift-kube-scheduler\"/\"openshift-kube-scheduler-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T22:50:40Z is after 2025-08-24T17:21:41Z" Dec 10 22:50:40 crc kubenswrapper[4791]: I1210 22:50:40.822629 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 22:50:40 crc kubenswrapper[4791]: I1210 22:50:40.822688 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 22:50:40 crc kubenswrapper[4791]: I1210 22:50:40.822705 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 22:50:40 crc kubenswrapper[4791]: I1210 22:50:40.822730 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 22:50:40 crc kubenswrapper[4791]: I1210 22:50:40.822747 4791 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T22:50:40Z","lastTransitionTime":"2025-12-10T22:50:40Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 22:50:40 crc kubenswrapper[4791]: I1210 22:50:40.884292 4791 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-b6kf6" Dec 10 22:50:40 crc kubenswrapper[4791]: I1210 22:50:40.884465 4791 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 10 22:50:40 crc kubenswrapper[4791]: E1210 22:50:40.884657 4791 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-b6kf6" podUID="b4527e71-bb58-4a39-a99c-fe614aa4c5f9" Dec 10 22:50:40 crc kubenswrapper[4791]: I1210 22:50:40.884693 4791 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 10 22:50:40 crc kubenswrapper[4791]: I1210 22:50:40.884720 4791 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 10 22:50:40 crc kubenswrapper[4791]: E1210 22:50:40.884799 4791 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 10 22:50:40 crc kubenswrapper[4791]: E1210 22:50:40.884912 4791 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 10 22:50:40 crc kubenswrapper[4791]: E1210 22:50:40.885018 4791 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 10 22:50:40 crc kubenswrapper[4791]: I1210 22:50:40.925438 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 22:50:40 crc kubenswrapper[4791]: I1210 22:50:40.925477 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 22:50:40 crc kubenswrapper[4791]: I1210 22:50:40.925487 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 22:50:40 crc kubenswrapper[4791]: I1210 22:50:40.925502 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 22:50:40 crc kubenswrapper[4791]: I1210 22:50:40.925514 4791 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T22:50:40Z","lastTransitionTime":"2025-12-10T22:50:40Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 22:50:41 crc kubenswrapper[4791]: I1210 22:50:41.027736 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 22:50:41 crc kubenswrapper[4791]: I1210 22:50:41.027814 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 22:50:41 crc kubenswrapper[4791]: I1210 22:50:41.027836 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 22:50:41 crc kubenswrapper[4791]: I1210 22:50:41.027860 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 22:50:41 crc kubenswrapper[4791]: I1210 22:50:41.027878 4791 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T22:50:41Z","lastTransitionTime":"2025-12-10T22:50:41Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 22:50:41 crc kubenswrapper[4791]: I1210 22:50:41.131427 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 22:50:41 crc kubenswrapper[4791]: I1210 22:50:41.131645 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 22:50:41 crc kubenswrapper[4791]: I1210 22:50:41.131672 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 22:50:41 crc kubenswrapper[4791]: I1210 22:50:41.131701 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 22:50:41 crc kubenswrapper[4791]: I1210 22:50:41.131718 4791 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T22:50:41Z","lastTransitionTime":"2025-12-10T22:50:41Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 22:50:41 crc kubenswrapper[4791]: I1210 22:50:41.235206 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 22:50:41 crc kubenswrapper[4791]: I1210 22:50:41.235322 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 22:50:41 crc kubenswrapper[4791]: I1210 22:50:41.235382 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 22:50:41 crc kubenswrapper[4791]: I1210 22:50:41.235408 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 22:50:41 crc kubenswrapper[4791]: I1210 22:50:41.235427 4791 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T22:50:41Z","lastTransitionTime":"2025-12-10T22:50:41Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 22:50:41 crc kubenswrapper[4791]: I1210 22:50:41.338248 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 22:50:41 crc kubenswrapper[4791]: I1210 22:50:41.338293 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 22:50:41 crc kubenswrapper[4791]: I1210 22:50:41.338310 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 22:50:41 crc kubenswrapper[4791]: I1210 22:50:41.338328 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 22:50:41 crc kubenswrapper[4791]: I1210 22:50:41.338355 4791 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T22:50:41Z","lastTransitionTime":"2025-12-10T22:50:41Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 22:50:41 crc kubenswrapper[4791]: I1210 22:50:41.440242 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 22:50:41 crc kubenswrapper[4791]: I1210 22:50:41.440291 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 22:50:41 crc kubenswrapper[4791]: I1210 22:50:41.440302 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 22:50:41 crc kubenswrapper[4791]: I1210 22:50:41.440319 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 22:50:41 crc kubenswrapper[4791]: I1210 22:50:41.440330 4791 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T22:50:41Z","lastTransitionTime":"2025-12-10T22:50:41Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 22:50:41 crc kubenswrapper[4791]: I1210 22:50:41.543805 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 22:50:41 crc kubenswrapper[4791]: I1210 22:50:41.543870 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 22:50:41 crc kubenswrapper[4791]: I1210 22:50:41.543887 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 22:50:41 crc kubenswrapper[4791]: I1210 22:50:41.543911 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 22:50:41 crc kubenswrapper[4791]: I1210 22:50:41.543928 4791 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T22:50:41Z","lastTransitionTime":"2025-12-10T22:50:41Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 22:50:41 crc kubenswrapper[4791]: I1210 22:50:41.647533 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 22:50:41 crc kubenswrapper[4791]: I1210 22:50:41.647623 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 22:50:41 crc kubenswrapper[4791]: I1210 22:50:41.647642 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 22:50:41 crc kubenswrapper[4791]: I1210 22:50:41.647671 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 22:50:41 crc kubenswrapper[4791]: I1210 22:50:41.647692 4791 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T22:50:41Z","lastTransitionTime":"2025-12-10T22:50:41Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 22:50:41 crc kubenswrapper[4791]: I1210 22:50:41.750582 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 22:50:41 crc kubenswrapper[4791]: I1210 22:50:41.750663 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 22:50:41 crc kubenswrapper[4791]: I1210 22:50:41.750678 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 22:50:41 crc kubenswrapper[4791]: I1210 22:50:41.750701 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 22:50:41 crc kubenswrapper[4791]: I1210 22:50:41.750716 4791 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T22:50:41Z","lastTransitionTime":"2025-12-10T22:50:41Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 22:50:41 crc kubenswrapper[4791]: I1210 22:50:41.854810 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 22:50:41 crc kubenswrapper[4791]: I1210 22:50:41.854878 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 22:50:41 crc kubenswrapper[4791]: I1210 22:50:41.854888 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 22:50:41 crc kubenswrapper[4791]: I1210 22:50:41.854912 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 22:50:41 crc kubenswrapper[4791]: I1210 22:50:41.854925 4791 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T22:50:41Z","lastTransitionTime":"2025-12-10T22:50:41Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 22:50:41 crc kubenswrapper[4791]: I1210 22:50:41.957642 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 22:50:41 crc kubenswrapper[4791]: I1210 22:50:41.957693 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 22:50:41 crc kubenswrapper[4791]: I1210 22:50:41.957705 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 22:50:41 crc kubenswrapper[4791]: I1210 22:50:41.957723 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 22:50:41 crc kubenswrapper[4791]: I1210 22:50:41.957734 4791 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T22:50:41Z","lastTransitionTime":"2025-12-10T22:50:41Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 22:50:42 crc kubenswrapper[4791]: I1210 22:50:42.060665 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 22:50:42 crc kubenswrapper[4791]: I1210 22:50:42.060747 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 22:50:42 crc kubenswrapper[4791]: I1210 22:50:42.060768 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 22:50:42 crc kubenswrapper[4791]: I1210 22:50:42.060799 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 22:50:42 crc kubenswrapper[4791]: I1210 22:50:42.060821 4791 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T22:50:42Z","lastTransitionTime":"2025-12-10T22:50:42Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 22:50:42 crc kubenswrapper[4791]: I1210 22:50:42.163505 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 22:50:42 crc kubenswrapper[4791]: I1210 22:50:42.163614 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 22:50:42 crc kubenswrapper[4791]: I1210 22:50:42.163634 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 22:50:42 crc kubenswrapper[4791]: I1210 22:50:42.163661 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 22:50:42 crc kubenswrapper[4791]: I1210 22:50:42.163678 4791 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T22:50:42Z","lastTransitionTime":"2025-12-10T22:50:42Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 22:50:42 crc kubenswrapper[4791]: I1210 22:50:42.266831 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 22:50:42 crc kubenswrapper[4791]: I1210 22:50:42.266886 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 22:50:42 crc kubenswrapper[4791]: I1210 22:50:42.266901 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 22:50:42 crc kubenswrapper[4791]: I1210 22:50:42.266925 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 22:50:42 crc kubenswrapper[4791]: I1210 22:50:42.266943 4791 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T22:50:42Z","lastTransitionTime":"2025-12-10T22:50:42Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 22:50:42 crc kubenswrapper[4791]: I1210 22:50:42.370427 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 22:50:42 crc kubenswrapper[4791]: I1210 22:50:42.370486 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 22:50:42 crc kubenswrapper[4791]: I1210 22:50:42.370499 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 22:50:42 crc kubenswrapper[4791]: I1210 22:50:42.370521 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 22:50:42 crc kubenswrapper[4791]: I1210 22:50:42.370534 4791 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T22:50:42Z","lastTransitionTime":"2025-12-10T22:50:42Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 22:50:42 crc kubenswrapper[4791]: I1210 22:50:42.474224 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 22:50:42 crc kubenswrapper[4791]: I1210 22:50:42.474274 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 22:50:42 crc kubenswrapper[4791]: I1210 22:50:42.474285 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 22:50:42 crc kubenswrapper[4791]: I1210 22:50:42.474304 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 22:50:42 crc kubenswrapper[4791]: I1210 22:50:42.474317 4791 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T22:50:42Z","lastTransitionTime":"2025-12-10T22:50:42Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 22:50:42 crc kubenswrapper[4791]: I1210 22:50:42.577710 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 22:50:42 crc kubenswrapper[4791]: I1210 22:50:42.577746 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 22:50:42 crc kubenswrapper[4791]: I1210 22:50:42.577755 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 22:50:42 crc kubenswrapper[4791]: I1210 22:50:42.577769 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 22:50:42 crc kubenswrapper[4791]: I1210 22:50:42.577778 4791 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T22:50:42Z","lastTransitionTime":"2025-12-10T22:50:42Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 22:50:42 crc kubenswrapper[4791]: I1210 22:50:42.680746 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 22:50:42 crc kubenswrapper[4791]: I1210 22:50:42.680798 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 22:50:42 crc kubenswrapper[4791]: I1210 22:50:42.680811 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 22:50:42 crc kubenswrapper[4791]: I1210 22:50:42.680829 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 22:50:42 crc kubenswrapper[4791]: I1210 22:50:42.680888 4791 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T22:50:42Z","lastTransitionTime":"2025-12-10T22:50:42Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 22:50:42 crc kubenswrapper[4791]: I1210 22:50:42.783975 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 22:50:42 crc kubenswrapper[4791]: I1210 22:50:42.784047 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 22:50:42 crc kubenswrapper[4791]: I1210 22:50:42.784073 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 22:50:42 crc kubenswrapper[4791]: I1210 22:50:42.784101 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 22:50:42 crc kubenswrapper[4791]: I1210 22:50:42.784119 4791 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T22:50:42Z","lastTransitionTime":"2025-12-10T22:50:42Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 22:50:42 crc kubenswrapper[4791]: I1210 22:50:42.885557 4791 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 10 22:50:42 crc kubenswrapper[4791]: I1210 22:50:42.885589 4791 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-b6kf6" Dec 10 22:50:42 crc kubenswrapper[4791]: I1210 22:50:42.885733 4791 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 10 22:50:42 crc kubenswrapper[4791]: E1210 22:50:42.885780 4791 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 10 22:50:42 crc kubenswrapper[4791]: E1210 22:50:42.885857 4791 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-b6kf6" podUID="b4527e71-bb58-4a39-a99c-fe614aa4c5f9" Dec 10 22:50:42 crc kubenswrapper[4791]: E1210 22:50:42.885951 4791 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 10 22:50:42 crc kubenswrapper[4791]: I1210 22:50:42.886525 4791 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 10 22:50:42 crc kubenswrapper[4791]: E1210 22:50:42.886709 4791 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 10 22:50:42 crc kubenswrapper[4791]: I1210 22:50:42.887951 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 22:50:42 crc kubenswrapper[4791]: I1210 22:50:42.887979 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 22:50:42 crc kubenswrapper[4791]: I1210 22:50:42.887989 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 22:50:42 crc kubenswrapper[4791]: I1210 22:50:42.888003 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 22:50:42 crc kubenswrapper[4791]: I1210 22:50:42.888017 4791 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T22:50:42Z","lastTransitionTime":"2025-12-10T22:50:42Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 22:50:42 crc kubenswrapper[4791]: I1210 22:50:42.991279 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 22:50:42 crc kubenswrapper[4791]: I1210 22:50:42.991380 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 22:50:42 crc kubenswrapper[4791]: I1210 22:50:42.991400 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 22:50:42 crc kubenswrapper[4791]: I1210 22:50:42.991424 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 22:50:42 crc kubenswrapper[4791]: I1210 22:50:42.991442 4791 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T22:50:42Z","lastTransitionTime":"2025-12-10T22:50:42Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 22:50:43 crc kubenswrapper[4791]: I1210 22:50:43.094651 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 22:50:43 crc kubenswrapper[4791]: I1210 22:50:43.094711 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 22:50:43 crc kubenswrapper[4791]: I1210 22:50:43.094734 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 22:50:43 crc kubenswrapper[4791]: I1210 22:50:43.094763 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 22:50:43 crc kubenswrapper[4791]: I1210 22:50:43.094786 4791 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T22:50:43Z","lastTransitionTime":"2025-12-10T22:50:43Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 22:50:43 crc kubenswrapper[4791]: I1210 22:50:43.197541 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 22:50:43 crc kubenswrapper[4791]: I1210 22:50:43.197599 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 22:50:43 crc kubenswrapper[4791]: I1210 22:50:43.197617 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 22:50:43 crc kubenswrapper[4791]: I1210 22:50:43.197639 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 22:50:43 crc kubenswrapper[4791]: I1210 22:50:43.197655 4791 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T22:50:43Z","lastTransitionTime":"2025-12-10T22:50:43Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 22:50:43 crc kubenswrapper[4791]: I1210 22:50:43.300601 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 22:50:43 crc kubenswrapper[4791]: I1210 22:50:43.300665 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 22:50:43 crc kubenswrapper[4791]: I1210 22:50:43.300682 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 22:50:43 crc kubenswrapper[4791]: I1210 22:50:43.300707 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 22:50:43 crc kubenswrapper[4791]: I1210 22:50:43.300726 4791 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T22:50:43Z","lastTransitionTime":"2025-12-10T22:50:43Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 22:50:43 crc kubenswrapper[4791]: I1210 22:50:43.403674 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 22:50:43 crc kubenswrapper[4791]: I1210 22:50:43.403740 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 22:50:43 crc kubenswrapper[4791]: I1210 22:50:43.403752 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 22:50:43 crc kubenswrapper[4791]: I1210 22:50:43.403773 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 22:50:43 crc kubenswrapper[4791]: I1210 22:50:43.404079 4791 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T22:50:43Z","lastTransitionTime":"2025-12-10T22:50:43Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 22:50:43 crc kubenswrapper[4791]: I1210 22:50:43.506959 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 22:50:43 crc kubenswrapper[4791]: I1210 22:50:43.507014 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 22:50:43 crc kubenswrapper[4791]: I1210 22:50:43.507023 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 22:50:43 crc kubenswrapper[4791]: I1210 22:50:43.507044 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 22:50:43 crc kubenswrapper[4791]: I1210 22:50:43.507054 4791 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T22:50:43Z","lastTransitionTime":"2025-12-10T22:50:43Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 22:50:43 crc kubenswrapper[4791]: I1210 22:50:43.609973 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 22:50:43 crc kubenswrapper[4791]: I1210 22:50:43.610026 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 22:50:43 crc kubenswrapper[4791]: I1210 22:50:43.610042 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 22:50:43 crc kubenswrapper[4791]: I1210 22:50:43.610069 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 22:50:43 crc kubenswrapper[4791]: I1210 22:50:43.610090 4791 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T22:50:43Z","lastTransitionTime":"2025-12-10T22:50:43Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 22:50:43 crc kubenswrapper[4791]: I1210 22:50:43.713222 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 22:50:43 crc kubenswrapper[4791]: I1210 22:50:43.713283 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 22:50:43 crc kubenswrapper[4791]: I1210 22:50:43.713300 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 22:50:43 crc kubenswrapper[4791]: I1210 22:50:43.713327 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 22:50:43 crc kubenswrapper[4791]: I1210 22:50:43.713402 4791 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T22:50:43Z","lastTransitionTime":"2025-12-10T22:50:43Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 22:50:43 crc kubenswrapper[4791]: I1210 22:50:43.816074 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 22:50:43 crc kubenswrapper[4791]: I1210 22:50:43.816164 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 22:50:43 crc kubenswrapper[4791]: I1210 22:50:43.816185 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 22:50:43 crc kubenswrapper[4791]: I1210 22:50:43.816212 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 22:50:43 crc kubenswrapper[4791]: I1210 22:50:43.816233 4791 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T22:50:43Z","lastTransitionTime":"2025-12-10T22:50:43Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 22:50:43 crc kubenswrapper[4791]: I1210 22:50:43.919479 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 22:50:43 crc kubenswrapper[4791]: I1210 22:50:43.919552 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 22:50:43 crc kubenswrapper[4791]: I1210 22:50:43.919575 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 22:50:43 crc kubenswrapper[4791]: I1210 22:50:43.919607 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 22:50:43 crc kubenswrapper[4791]: I1210 22:50:43.919637 4791 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T22:50:43Z","lastTransitionTime":"2025-12-10T22:50:43Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 22:50:44 crc kubenswrapper[4791]: I1210 22:50:44.022429 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 22:50:44 crc kubenswrapper[4791]: I1210 22:50:44.022469 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 22:50:44 crc kubenswrapper[4791]: I1210 22:50:44.022478 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 22:50:44 crc kubenswrapper[4791]: I1210 22:50:44.022491 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 22:50:44 crc kubenswrapper[4791]: I1210 22:50:44.022499 4791 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T22:50:44Z","lastTransitionTime":"2025-12-10T22:50:44Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 22:50:44 crc kubenswrapper[4791]: I1210 22:50:44.125530 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 22:50:44 crc kubenswrapper[4791]: I1210 22:50:44.125603 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 22:50:44 crc kubenswrapper[4791]: I1210 22:50:44.125629 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 22:50:44 crc kubenswrapper[4791]: I1210 22:50:44.125656 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 22:50:44 crc kubenswrapper[4791]: I1210 22:50:44.125678 4791 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T22:50:44Z","lastTransitionTime":"2025-12-10T22:50:44Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 22:50:44 crc kubenswrapper[4791]: I1210 22:50:44.228294 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 22:50:44 crc kubenswrapper[4791]: I1210 22:50:44.228403 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 22:50:44 crc kubenswrapper[4791]: I1210 22:50:44.228414 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 22:50:44 crc kubenswrapper[4791]: I1210 22:50:44.228430 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 22:50:44 crc kubenswrapper[4791]: I1210 22:50:44.228442 4791 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T22:50:44Z","lastTransitionTime":"2025-12-10T22:50:44Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 22:50:44 crc kubenswrapper[4791]: I1210 22:50:44.331287 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 22:50:44 crc kubenswrapper[4791]: I1210 22:50:44.331326 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 22:50:44 crc kubenswrapper[4791]: I1210 22:50:44.331453 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 22:50:44 crc kubenswrapper[4791]: I1210 22:50:44.331479 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 22:50:44 crc kubenswrapper[4791]: I1210 22:50:44.331491 4791 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T22:50:44Z","lastTransitionTime":"2025-12-10T22:50:44Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 22:50:44 crc kubenswrapper[4791]: I1210 22:50:44.433891 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 22:50:44 crc kubenswrapper[4791]: I1210 22:50:44.433929 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 22:50:44 crc kubenswrapper[4791]: I1210 22:50:44.433939 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 22:50:44 crc kubenswrapper[4791]: I1210 22:50:44.433952 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 22:50:44 crc kubenswrapper[4791]: I1210 22:50:44.433961 4791 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T22:50:44Z","lastTransitionTime":"2025-12-10T22:50:44Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 22:50:44 crc kubenswrapper[4791]: I1210 22:50:44.537024 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 22:50:44 crc kubenswrapper[4791]: I1210 22:50:44.537078 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 22:50:44 crc kubenswrapper[4791]: I1210 22:50:44.537090 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 22:50:44 crc kubenswrapper[4791]: I1210 22:50:44.537113 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 22:50:44 crc kubenswrapper[4791]: I1210 22:50:44.537131 4791 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T22:50:44Z","lastTransitionTime":"2025-12-10T22:50:44Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 22:50:44 crc kubenswrapper[4791]: I1210 22:50:44.639449 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 22:50:44 crc kubenswrapper[4791]: I1210 22:50:44.639487 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 22:50:44 crc kubenswrapper[4791]: I1210 22:50:44.639500 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 22:50:44 crc kubenswrapper[4791]: I1210 22:50:44.639516 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 22:50:44 crc kubenswrapper[4791]: I1210 22:50:44.639528 4791 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T22:50:44Z","lastTransitionTime":"2025-12-10T22:50:44Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 22:50:44 crc kubenswrapper[4791]: I1210 22:50:44.741961 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 22:50:44 crc kubenswrapper[4791]: I1210 22:50:44.741993 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 22:50:44 crc kubenswrapper[4791]: I1210 22:50:44.742023 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 22:50:44 crc kubenswrapper[4791]: I1210 22:50:44.742040 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 22:50:44 crc kubenswrapper[4791]: I1210 22:50:44.742052 4791 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T22:50:44Z","lastTransitionTime":"2025-12-10T22:50:44Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 22:50:44 crc kubenswrapper[4791]: I1210 22:50:44.844739 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 22:50:44 crc kubenswrapper[4791]: I1210 22:50:44.844782 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 22:50:44 crc kubenswrapper[4791]: I1210 22:50:44.844795 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 22:50:44 crc kubenswrapper[4791]: I1210 22:50:44.844810 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 22:50:44 crc kubenswrapper[4791]: I1210 22:50:44.844823 4791 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T22:50:44Z","lastTransitionTime":"2025-12-10T22:50:44Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 22:50:44 crc kubenswrapper[4791]: I1210 22:50:44.884458 4791 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-b6kf6" Dec 10 22:50:44 crc kubenswrapper[4791]: I1210 22:50:44.884487 4791 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 10 22:50:44 crc kubenswrapper[4791]: I1210 22:50:44.884533 4791 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 10 22:50:44 crc kubenswrapper[4791]: I1210 22:50:44.884634 4791 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 10 22:50:44 crc kubenswrapper[4791]: E1210 22:50:44.884815 4791 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-b6kf6" podUID="b4527e71-bb58-4a39-a99c-fe614aa4c5f9" Dec 10 22:50:44 crc kubenswrapper[4791]: E1210 22:50:44.884920 4791 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 10 22:50:44 crc kubenswrapper[4791]: E1210 22:50:44.885031 4791 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 10 22:50:44 crc kubenswrapper[4791]: E1210 22:50:44.885101 4791 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 10 22:50:44 crc kubenswrapper[4791]: I1210 22:50:44.946887 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 22:50:44 crc kubenswrapper[4791]: I1210 22:50:44.946916 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 22:50:44 crc kubenswrapper[4791]: I1210 22:50:44.946927 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 22:50:44 crc kubenswrapper[4791]: I1210 22:50:44.946940 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 22:50:44 crc kubenswrapper[4791]: I1210 22:50:44.946949 4791 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T22:50:44Z","lastTransitionTime":"2025-12-10T22:50:44Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 22:50:45 crc kubenswrapper[4791]: I1210 22:50:45.049487 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 22:50:45 crc kubenswrapper[4791]: I1210 22:50:45.049524 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 22:50:45 crc kubenswrapper[4791]: I1210 22:50:45.049536 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 22:50:45 crc kubenswrapper[4791]: I1210 22:50:45.049554 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 22:50:45 crc kubenswrapper[4791]: I1210 22:50:45.049566 4791 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T22:50:45Z","lastTransitionTime":"2025-12-10T22:50:45Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 22:50:45 crc kubenswrapper[4791]: I1210 22:50:45.151509 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 22:50:45 crc kubenswrapper[4791]: I1210 22:50:45.151555 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 22:50:45 crc kubenswrapper[4791]: I1210 22:50:45.151565 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 22:50:45 crc kubenswrapper[4791]: I1210 22:50:45.151579 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 22:50:45 crc kubenswrapper[4791]: I1210 22:50:45.151588 4791 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T22:50:45Z","lastTransitionTime":"2025-12-10T22:50:45Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 22:50:45 crc kubenswrapper[4791]: I1210 22:50:45.255040 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 22:50:45 crc kubenswrapper[4791]: I1210 22:50:45.255108 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 22:50:45 crc kubenswrapper[4791]: I1210 22:50:45.255125 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 22:50:45 crc kubenswrapper[4791]: I1210 22:50:45.255148 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 22:50:45 crc kubenswrapper[4791]: I1210 22:50:45.255165 4791 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T22:50:45Z","lastTransitionTime":"2025-12-10T22:50:45Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 22:50:45 crc kubenswrapper[4791]: I1210 22:50:45.357973 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 22:50:45 crc kubenswrapper[4791]: I1210 22:50:45.358012 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 22:50:45 crc kubenswrapper[4791]: I1210 22:50:45.358021 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 22:50:45 crc kubenswrapper[4791]: I1210 22:50:45.358061 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 22:50:45 crc kubenswrapper[4791]: I1210 22:50:45.358071 4791 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T22:50:45Z","lastTransitionTime":"2025-12-10T22:50:45Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 22:50:45 crc kubenswrapper[4791]: I1210 22:50:45.459985 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 22:50:45 crc kubenswrapper[4791]: I1210 22:50:45.460046 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 22:50:45 crc kubenswrapper[4791]: I1210 22:50:45.460059 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 22:50:45 crc kubenswrapper[4791]: I1210 22:50:45.460075 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 22:50:45 crc kubenswrapper[4791]: I1210 22:50:45.460085 4791 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T22:50:45Z","lastTransitionTime":"2025-12-10T22:50:45Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 22:50:45 crc kubenswrapper[4791]: I1210 22:50:45.562920 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 22:50:45 crc kubenswrapper[4791]: I1210 22:50:45.562969 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 22:50:45 crc kubenswrapper[4791]: I1210 22:50:45.562979 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 22:50:45 crc kubenswrapper[4791]: I1210 22:50:45.562995 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 22:50:45 crc kubenswrapper[4791]: I1210 22:50:45.563008 4791 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T22:50:45Z","lastTransitionTime":"2025-12-10T22:50:45Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 22:50:45 crc kubenswrapper[4791]: I1210 22:50:45.665171 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 22:50:45 crc kubenswrapper[4791]: I1210 22:50:45.665820 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 22:50:45 crc kubenswrapper[4791]: I1210 22:50:45.666015 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 22:50:45 crc kubenswrapper[4791]: I1210 22:50:45.666207 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 22:50:45 crc kubenswrapper[4791]: I1210 22:50:45.666636 4791 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T22:50:45Z","lastTransitionTime":"2025-12-10T22:50:45Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 22:50:45 crc kubenswrapper[4791]: I1210 22:50:45.769283 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 22:50:45 crc kubenswrapper[4791]: I1210 22:50:45.769530 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 22:50:45 crc kubenswrapper[4791]: I1210 22:50:45.769600 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 22:50:45 crc kubenswrapper[4791]: I1210 22:50:45.769693 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 22:50:45 crc kubenswrapper[4791]: I1210 22:50:45.769751 4791 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T22:50:45Z","lastTransitionTime":"2025-12-10T22:50:45Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 22:50:45 crc kubenswrapper[4791]: I1210 22:50:45.872286 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 22:50:45 crc kubenswrapper[4791]: I1210 22:50:45.872633 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 22:50:45 crc kubenswrapper[4791]: I1210 22:50:45.872726 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 22:50:45 crc kubenswrapper[4791]: I1210 22:50:45.872810 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 22:50:45 crc kubenswrapper[4791]: I1210 22:50:45.872890 4791 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T22:50:45Z","lastTransitionTime":"2025-12-10T22:50:45Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 22:50:45 crc kubenswrapper[4791]: I1210 22:50:45.904105 4791 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"90250e13-6c63-443a-a5e5-6e4e341a2289\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T22:49:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T22:49:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T22:49:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T22:49:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T22:49:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ba54f95b1e00843b854db9a10fe6897592159ccacf30ed0f76002093a13eedd3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T22:49:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5aa8ffa518f3a68b60b275f8df4851b2564aa4984b61efdc9470b73072133e43\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T22:49:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://16f989503edc78d9451638e2cf6cf4b19a2cf2b8809b10b237541cb829303091\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T22:49:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://82556824d63ce6e46df14c45689ff1f81ebf729abe902061815c040b323898e9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T22:49:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://61db531940fa6f2effdae5a005b4c91bd2698bd5d1bff9588c171a65ad550eca\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T22:49:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0f65c864d885319ce2a08d61c30be24b3ffb7d23ea30c82ed7faf039cb2c006d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://0f65c864d885319ce2a08d61c30be24b3ffb7d23ea30c82ed7faf039cb2c006d\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-10T22:49:27Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-10T22:49:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e31a76d1d5b2812ba83e24e38268e25ec8bed98fc3da9adcd06bda5822b6ace1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e31a76d1d5b2812ba83e24e38268e25ec8bed98fc3da9adcd06bda5822b6ace1\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-10T22:49:30Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-10T22:49:29Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://b5b5eceae562f3986b543d41de1e977efb85a91d84c98e09159ca6d092eda691\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b5b5eceae562f3986b543d41de1e977efb85a91d84c98e09159ca6d092eda691\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-10T22:49:31Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-10T22:49:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-10T22:49:25Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T22:50:45Z is after 2025-08-24T17:21:41Z" Dec 10 22:50:45 crc kubenswrapper[4791]: I1210 22:50:45.916955 4791 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-10T22:49:50Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-10T22:49:50Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T22:50:45Z is after 2025-08-24T17:21:41Z" Dec 10 22:50:45 crc kubenswrapper[4791]: I1210 22:50:45.930732 4791 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-10T22:49:54Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-10T22:49:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-10T22:49:54Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://62bfaf5d7a6d1b452cdaf88086447d0c8cb7306e94434e46752be5475124d56a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T22:49:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T22:50:45Z is after 2025-08-24T17:21:41Z" Dec 10 22:50:45 crc kubenswrapper[4791]: I1210 22:50:45.943314 4791 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-b6kf6" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b4527e71-bb58-4a39-a99c-fe614aa4c5f9\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T22:50:04Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T22:50:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T22:50:04Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T22:50:04Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ngpr4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ngpr4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-10T22:50:04Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-b6kf6\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T22:50:45Z is after 2025-08-24T17:21:41Z" Dec 10 22:50:45 crc kubenswrapper[4791]: I1210 22:50:45.961183 4791 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f8349fea-c49b-49a1-b5ee-032c5dffe021\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T22:49:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T22:49:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T22:49:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T22:49:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T22:49:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://acafe12186df99d1059a31260de0eafce35ff7771bcb8f5c6753b72665d6f60b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T22:49:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ea4c346507862a8c12756a3061d6e25e788a8d79f4621b62f81f4322185c82a8\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T22:49:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://2e6e324ec6f4f7bb69fdfab4ea3ac1aa55e480dd8e2a8984b0fe5b4c9f66636e\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T22:49:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://1b2e11ff2bf8789fa1d0d23ae648b3e6ef653106e5bc4bbb5dd6447f7cd6f05d\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d6aa7f8c5d908d90008f2df49860ac8a3480136da567e3d85f86135d779204f9\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-12-10T22:49:44Z\\\",\\\"message\\\":\\\"W1210 22:49:33.845733 1 cmd.go:257] Using insecure, self-signed certificates\\\\nI1210 22:49:33.846108 1 crypto.go:601] Generating new CA for check-endpoints-signer@1765406973 cert, and key in /tmp/serving-cert-2585701456/serving-signer.crt, /tmp/serving-cert-2585701456/serving-signer.key\\\\nI1210 22:49:34.529748 1 observer_polling.go:159] Starting file observer\\\\nW1210 22:49:34.532225 1 builder.go:272] unable to get owner reference (falling back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": dial tcp [::1]:6443: connect: connection refused\\\\nI1210 22:49:34.532599 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1210 22:49:34.533977 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-2585701456/tls.crt::/tmp/serving-cert-2585701456/tls.key\\\\\\\"\\\\nF1210 22:49:44.774710 1 cmd.go:182] error initializing delegating authentication: unable to load configmap based request-header-client-ca-file: Get \\\\\\\"https://localhost:6443/api/v1/namespaces/kube-system/configmaps/extension-apiserver-authentication\\\\\\\": net/http: TLS handshake timeout\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-10T22:49:33Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T22:49:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://0c4d036ab269f62561fe7ef1586012cb3a16e3583a08b9488a0616ec00c8e0ae\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T22:49:32Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c5856d6b26c19fda13b5c9ac2892a7e2f9c6f79feb8173ecfd9b33b673327d67\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c5856d6b26c19fda13b5c9ac2892a7e2f9c6f79feb8173ecfd9b33b673327d67\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-10T22:49:27Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-10T22:49:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-10T22:49:25Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T22:50:45Z is after 2025-08-24T17:21:41Z" Dec 10 22:50:45 crc kubenswrapper[4791]: I1210 22:50:45.974105 4791 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"02a949d8-38e2-4be3-989f-40a015407140\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T22:49:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T22:49:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T22:50:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T22:50:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T22:49:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://cd4158bc59b30f0f1716bf43c1ae7c9916a5b7d7164d074987fc1f19ba49b778\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T22:49:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://656065fd3cbdc6a25dab5d06c7baa00af16d8f62fc5da687efe9c6cc91945de0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T22:49:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://90ed7fff816b1772ba9ecd95587a59cbed7c30ac302fff7e7c5de934c912bcf4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T22:49:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9b7a5502803eaad9f76f5ffa91b29fbb2657402065a03a4d6433c096aac7264f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"wait-for-host-port\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://9b7a5502803eaad9f76f5ffa91b29fbb2657402065a03a4d6433c096aac7264f\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-10T22:49:27Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-10T22:49:27Z\\\"}}}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-10T22:49:25Z\\\"}}\" for pod \"openshift-kube-scheduler\"/\"openshift-kube-scheduler-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T22:50:45Z is after 2025-08-24T17:21:41Z" Dec 10 22:50:45 crc kubenswrapper[4791]: I1210 22:50:45.974827 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 22:50:45 crc kubenswrapper[4791]: I1210 22:50:45.974873 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 22:50:45 crc kubenswrapper[4791]: I1210 22:50:45.974891 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 22:50:45 crc kubenswrapper[4791]: I1210 22:50:45.974913 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 22:50:45 crc kubenswrapper[4791]: I1210 22:50:45.974929 4791 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T22:50:45Z","lastTransitionTime":"2025-12-10T22:50:45Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 22:50:45 crc kubenswrapper[4791]: I1210 22:50:45.988014 4791 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-5rb5l" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ba35653c-6e06-4cee-a4d6-137764090d18\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T22:49:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T22:49:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T22:49:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T22:49:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://dbb5c875730a98ad2daf316cfafc08fcdce21ac16148154780ac9193ed405dab\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T22:49:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fvvrs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://185eb2ff44221a514878283ec873773a6c9a9073dc64e296f982e60dee1338e4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T22:49:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fvvrs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-10T22:49:51Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-5rb5l\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T22:50:45Z is after 2025-08-24T17:21:41Z" Dec 10 22:50:46 crc kubenswrapper[4791]: I1210 22:50:46.001292 4791 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-s6rk8" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5c9d0978-a6f9-49c5-b377-6619b4e3ff52\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T22:50:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T22:50:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T22:50:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T22:50:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8e55baaf460c794ab196e237d51d12de059aa038617dc2d5366d49c529f5d5c7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T22:50:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7vkgm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://51d1b9b086c60bae9223cb51d5a998d02790a0d91b745aab10392c23e617eb9f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T22:50:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7vkgm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-10T22:50:03Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-s6rk8\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T22:50:45Z is after 2025-08-24T17:21:41Z" Dec 10 22:50:46 crc kubenswrapper[4791]: I1210 22:50:46.022115 4791 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-10T22:49:52Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-10T22:49:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-10T22:49:52Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3734b110802828632d44263ed0a43cabc4b710833f0be120d72db0940fe783e2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T22:49:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://794803ac9be3c52509e6dcb10854d8864e1a9e406762e57c7971b07c4da1220e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T22:49:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T22:50:46Z is after 2025-08-24T17:21:41Z" Dec 10 22:50:46 crc kubenswrapper[4791]: I1210 22:50:46.035002 4791 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-vczjq" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b9f00f25-3fa1-4b63-9710-fd352224b01b\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T22:49:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T22:49:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T22:49:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T22:49:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4698cf4c6b55b41e628448d37c2eb847c2777d0efbd93afcbf367d5e78623b7e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T22:49:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lrrpv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-10T22:49:50Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-vczjq\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T22:50:46Z is after 2025-08-24T17:21:41Z" Dec 10 22:50:46 crc kubenswrapper[4791]: I1210 22:50:46.049711 4791 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-4nwkq" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"672aa28c-8169-49ed-87b8-21187d13a80c\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T22:49:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T22:49:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T22:50:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T22:50:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f337f881ffd71f4a2ee19e92bb0b92c76cbb9b8352511f741d0c87e7da4cdd18\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a2f2a62315529e148a10a33ccb241e9417fbde0fe42afb61cb7ee2438c74636f\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-12-10T22:50:38Z\\\",\\\"message\\\":\\\"2025-12-10T22:49:52+00:00 [cnibincopy] Successfully copied files in /usr/src/multus-cni/rhel9/bin/ to /host/opt/cni/bin/upgrade_e48d9423-4bca-43c4-9eb9-b55a19a5ffd6\\\\n2025-12-10T22:49:52+00:00 [cnibincopy] Successfully moved files in /host/opt/cni/bin/upgrade_e48d9423-4bca-43c4-9eb9-b55a19a5ffd6 to /host/opt/cni/bin/\\\\n2025-12-10T22:49:52Z [verbose] multus-daemon started\\\\n2025-12-10T22:49:52Z [verbose] Readiness Indicator file check\\\\n2025-12-10T22:50:37Z [error] have you checked that your default network is ready? still waiting for readinessindicatorfile @ /host/run/multus/cni/net.d/10-ovn-kubernetes.conf. pollimmediate error: timed out waiting for the condition\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-10T22:49:52Z\\\"}},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T22:50:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xl62j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-10T22:49:51Z\\\"}}\" for pod \"openshift-multus\"/\"multus-4nwkq\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T22:50:46Z is after 2025-08-24T17:21:41Z" Dec 10 22:50:46 crc kubenswrapper[4791]: I1210 22:50:46.077126 4791 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-zhq64" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3cd47739-0fa9-4321-aff1-220f8721a0b3\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T22:49:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T22:49:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T22:49:52Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T22:49:52Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://68795a29944e56870ffe94fa732e9dcf0bfb344c0d85d1e590f58a876129c61f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T22:49:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wvpw8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://bc721b530c08b433756e978ec0423cc18aabf6771b61eb96f903b4d57ea335ff\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T22:49:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wvpw8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://152a7618be8f38fadf4e564b3948980503cedcdb466c37344ccc20f4a3173a5f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T22:49:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wvpw8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3d18f9cb7207efb1e32de75a30d33e024f656a08f0503db95e5488bc5280971f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T22:49:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wvpw8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://937137ae9d7480cdc7c59bc4a1df0a4c10e72b3b63693af285278789953ed218\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T22:49:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wvpw8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://1a52d19b7671e8b143017c8a981a6f8d3ee3b84650e4b223b814a531ac76f364\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T22:49:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wvpw8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b6d739a4314c4fb536810b8623f691d6aaf1090b9c83f17b484d2b22f7d3ad01\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b6d739a4314c4fb536810b8623f691d6aaf1090b9c83f17b484d2b22f7d3ad01\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-12-10T22:50:21Z\\\",\\\"message\\\":\\\"llocateLoadBalancerNodePorts:nil,LoadBalancerClass:nil,InternalTrafficPolicy:*Cluster,TrafficDistribution:nil,},Status:ServiceStatus{LoadBalancer:LoadBalancerStatus{Ingress:[]LoadBalancerIngress{},},Conditions:[]Condition{},},}\\\\nI1210 22:50:21.614710 6424 metrics.go:553] Stopping metrics server at address \\\\\\\"127.0.0.1:29103\\\\\\\"\\\\nI1210 22:50:21.614731 6424 lb_config.go:1031] Cluster endpoints for openshift-kube-scheduler/scheduler for network=default are: map[]\\\\nI1210 22:50:21.614619 6424 services_controller.go:360] Finished syncing service ovn-kubernetes-control-plane on namespace openshift-ovn-kubernetes for network=default : 7.02µs\\\\nI1210 22:50:21.614757 6424 services_controller.go:356] Processing sync for service openshift-route-controller-manager/route-controller-manager for network=default\\\\nI1210 22:50:21.614615 6424 services_controller.go:444] Built service openshift-machine-config-operator/machine-config-daemon LB per-node configs for network=default: []services.lbConfig(nil)\\\\nF1210 22:50:21.614772 6424 ovnkube.go:137] failed to run ovnkube: [failed to start network controller: failed to start default network controller: unable to create admin network policy controller, err: could not add Event Handler for anpInformer during admin network policy controller initialization, handler {0x1fcc6e0 0x1fcc3c0 0x1fcc360} was not added to\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-10T22:50:20Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":2,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 20s restarting failed container=ovnkube-controller pod=ovnkube-node-zhq64_openshift-ovn-kubernetes(3cd47739-0fa9-4321-aff1-220f8721a0b3)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wvpw8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://dd3f80e5713221f96ed50da7f2e7fd5b9fabfcd5c1656da8d8a86d818a5fdc3f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T22:49:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wvpw8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3afbd1ffcd2c4b40324aca6e52d6fc822db131c3f35a9af3a6f005c2cffdb7bf\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3afbd1ffcd2c4b40324aca6e52d6fc822db131c3f35a9af3a6f005c2cffdb7bf\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-10T22:49:53Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-10T22:49:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wvpw8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-10T22:49:52Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-zhq64\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T22:50:46Z is after 2025-08-24T17:21:41Z" Dec 10 22:50:46 crc kubenswrapper[4791]: I1210 22:50:46.077727 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 22:50:46 crc kubenswrapper[4791]: I1210 22:50:46.077769 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 22:50:46 crc kubenswrapper[4791]: I1210 22:50:46.077780 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 22:50:46 crc kubenswrapper[4791]: I1210 22:50:46.077795 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 22:50:46 crc kubenswrapper[4791]: I1210 22:50:46.077806 4791 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T22:50:46Z","lastTransitionTime":"2025-12-10T22:50:46Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 22:50:46 crc kubenswrapper[4791]: I1210 22:50:46.090386 4791 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"698bbe9a-fbe5-436d-a835-5e80df77ba24\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T22:49:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T22:49:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T22:49:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T22:49:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T22:49:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://224d73713312376a58ec6cc2bc8fdb6ffa44c4749ae0e7691d5fe27b9afca09f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T22:49:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://372eeaffd40992c6f06efc4df6dfcb44b31de64ee8256c21508cdd7c74ecb96e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T22:49:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://76b6013b062cac5b0b4d12a94f38ce7b43f9c7ea216063455d88a956ac04d26e\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T22:49:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://78e950e122e761d950da1a3c60c06476e53920e8795f6c61db22c01e22d5136e\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T22:49:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-10T22:49:25Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T22:50:46Z is after 2025-08-24T17:21:41Z" Dec 10 22:50:46 crc kubenswrapper[4791]: I1210 22:50:46.100380 4791 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7a9a3709-32fb-48b4-97f4-e1a19282a085\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T22:49:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T22:49:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T22:49:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T22:49:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T22:49:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://511bd056768a15ea0f650a2b1a6e87e3c71dfc75696222a54f7376227ad480c6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-crio\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T22:49:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kube\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"var-lib-kubelet\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8c489105d0dfb302732d4f5bd2fc9d000967a071802735b778d49bdbc428b264\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://8c489105d0dfb302732d4f5bd2fc9d000967a071802735b778d49bdbc428b264\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-10T22:49:27Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-10T22:49:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var\\\",\\\"name\\\":\\\"var-lib-kubelet\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-10T22:49:25Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"kube-rbac-proxy-crio-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T22:50:46Z is after 2025-08-24T17:21:41Z" Dec 10 22:50:46 crc kubenswrapper[4791]: I1210 22:50:46.113616 4791 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-10T22:49:50Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-10T22:49:50Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T22:50:46Z is after 2025-08-24T17:21:41Z" Dec 10 22:50:46 crc kubenswrapper[4791]: I1210 22:50:46.126820 4791 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-tchxk" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"c0ab74e0-42a9-4f42-ac80-16030cf4ffff\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T22:49:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T22:49:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T22:49:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T22:49:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://04e0fd1b53256b7a938e8ccab93e2a4eb73dd1a59e0b9799e09c55e511baf7fe\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T22:49:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zn2qt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-10T22:49:50Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-tchxk\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T22:50:46Z is after 2025-08-24T17:21:41Z" Dec 10 22:50:46 crc kubenswrapper[4791]: I1210 22:50:46.143538 4791 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-rw584" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ff9beb90-69b0-4732-bf37-0b81f58ecc98\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T22:49:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T22:49:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T22:49:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T22:49:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ddb9df501dc6e8e5561fbc2aefd431037a3f28f73bf8e82a6598d261a4224807\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T22:49:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fksqn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ce67e54b696bb53ffbbe48c6ce71476b635e0868993e9c427b22f904891433ee\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ce67e54b696bb53ffbbe48c6ce71476b635e0868993e9c427b22f904891433ee\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-10T22:49:52Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-10T22:49:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fksqn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://818d89bfffb8a110670152c8480684665ea8db578e9768f524fbbe6a92f27c22\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://818d89bfffb8a110670152c8480684665ea8db578e9768f524fbbe6a92f27c22\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-10T22:49:53Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-10T22:49:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fksqn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://9aa9c06e894bb2a2ed3738defaeff68a0a0458ead48177f36b90770944c211a0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://9aa9c06e894bb2a2ed3738defaeff68a0a0458ead48177f36b90770944c211a0\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-10T22:49:54Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-10T22:49:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fksqn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://dfd6d03336dcea7529419a30e8d70d84e2290a0fa5b0e46d09ad341c190638d6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://dfd6d03336dcea7529419a30e8d70d84e2290a0fa5b0e46d09ad341c190638d6\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-10T22:49:55Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-10T22:49:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fksqn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b86738401ece6fbc3083782720c66a515850ed78d4054f34fcabab7ff525b85f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b86738401ece6fbc3083782720c66a515850ed78d4054f34fcabab7ff525b85f\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-10T22:49:56Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-10T22:49:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fksqn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://18b326f572a899ddd10197a745bf9264d4ddd975690ad1cfb0ec576ccd5cf102\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://18b326f572a899ddd10197a745bf9264d4ddd975690ad1cfb0ec576ccd5cf102\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-10T22:49:57Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-10T22:49:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fksqn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-10T22:49:51Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-rw584\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T22:50:46Z is after 2025-08-24T17:21:41Z" Dec 10 22:50:46 crc kubenswrapper[4791]: I1210 22:50:46.159039 4791 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-10T22:49:52Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-10T22:49:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-10T22:49:52Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://20651eb0c16e2c71222f56da1ff755cb0e7b084886acbbb24f0ccaf57041ed85\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T22:49:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T22:50:46Z is after 2025-08-24T17:21:41Z" Dec 10 22:50:46 crc kubenswrapper[4791]: I1210 22:50:46.171404 4791 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-10T22:49:50Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-10T22:49:50Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T22:50:46Z is after 2025-08-24T17:21:41Z" Dec 10 22:50:46 crc kubenswrapper[4791]: I1210 22:50:46.180158 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 22:50:46 crc kubenswrapper[4791]: I1210 22:50:46.180206 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 22:50:46 crc kubenswrapper[4791]: I1210 22:50:46.180218 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 22:50:46 crc kubenswrapper[4791]: I1210 22:50:46.180237 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 22:50:46 crc kubenswrapper[4791]: I1210 22:50:46.180249 4791 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T22:50:46Z","lastTransitionTime":"2025-12-10T22:50:46Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 22:50:46 crc kubenswrapper[4791]: I1210 22:50:46.282801 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 22:50:46 crc kubenswrapper[4791]: I1210 22:50:46.282841 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 22:50:46 crc kubenswrapper[4791]: I1210 22:50:46.282852 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 22:50:46 crc kubenswrapper[4791]: I1210 22:50:46.282871 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 22:50:46 crc kubenswrapper[4791]: I1210 22:50:46.282882 4791 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T22:50:46Z","lastTransitionTime":"2025-12-10T22:50:46Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 22:50:46 crc kubenswrapper[4791]: I1210 22:50:46.385363 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 22:50:46 crc kubenswrapper[4791]: I1210 22:50:46.385419 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 22:50:46 crc kubenswrapper[4791]: I1210 22:50:46.385438 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 22:50:46 crc kubenswrapper[4791]: I1210 22:50:46.385495 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 22:50:46 crc kubenswrapper[4791]: I1210 22:50:46.385514 4791 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T22:50:46Z","lastTransitionTime":"2025-12-10T22:50:46Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 22:50:46 crc kubenswrapper[4791]: I1210 22:50:46.487590 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 22:50:46 crc kubenswrapper[4791]: I1210 22:50:46.487851 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 22:50:46 crc kubenswrapper[4791]: I1210 22:50:46.487938 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 22:50:46 crc kubenswrapper[4791]: I1210 22:50:46.488011 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 22:50:46 crc kubenswrapper[4791]: I1210 22:50:46.488085 4791 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T22:50:46Z","lastTransitionTime":"2025-12-10T22:50:46Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 22:50:46 crc kubenswrapper[4791]: I1210 22:50:46.590793 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 22:50:46 crc kubenswrapper[4791]: I1210 22:50:46.590866 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 22:50:46 crc kubenswrapper[4791]: I1210 22:50:46.590877 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 22:50:46 crc kubenswrapper[4791]: I1210 22:50:46.590896 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 22:50:46 crc kubenswrapper[4791]: I1210 22:50:46.590910 4791 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T22:50:46Z","lastTransitionTime":"2025-12-10T22:50:46Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 22:50:46 crc kubenswrapper[4791]: I1210 22:50:46.693173 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 22:50:46 crc kubenswrapper[4791]: I1210 22:50:46.693248 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 22:50:46 crc kubenswrapper[4791]: I1210 22:50:46.693271 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 22:50:46 crc kubenswrapper[4791]: I1210 22:50:46.693300 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 22:50:46 crc kubenswrapper[4791]: I1210 22:50:46.693324 4791 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T22:50:46Z","lastTransitionTime":"2025-12-10T22:50:46Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 22:50:46 crc kubenswrapper[4791]: I1210 22:50:46.795860 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 22:50:46 crc kubenswrapper[4791]: I1210 22:50:46.795898 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 22:50:46 crc kubenswrapper[4791]: I1210 22:50:46.795908 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 22:50:46 crc kubenswrapper[4791]: I1210 22:50:46.795922 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 22:50:46 crc kubenswrapper[4791]: I1210 22:50:46.795931 4791 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T22:50:46Z","lastTransitionTime":"2025-12-10T22:50:46Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 22:50:46 crc kubenswrapper[4791]: I1210 22:50:46.884304 4791 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-b6kf6" Dec 10 22:50:46 crc kubenswrapper[4791]: I1210 22:50:46.884380 4791 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 10 22:50:46 crc kubenswrapper[4791]: I1210 22:50:46.884610 4791 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 10 22:50:46 crc kubenswrapper[4791]: I1210 22:50:46.884640 4791 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 10 22:50:46 crc kubenswrapper[4791]: E1210 22:50:46.884709 4791 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 10 22:50:46 crc kubenswrapper[4791]: I1210 22:50:46.884989 4791 scope.go:117] "RemoveContainer" containerID="b6d739a4314c4fb536810b8623f691d6aaf1090b9c83f17b484d2b22f7d3ad01" Dec 10 22:50:46 crc kubenswrapper[4791]: E1210 22:50:46.885049 4791 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 10 22:50:46 crc kubenswrapper[4791]: E1210 22:50:46.885145 4791 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 10 22:50:46 crc kubenswrapper[4791]: E1210 22:50:46.885220 4791 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-b6kf6" podUID="b4527e71-bb58-4a39-a99c-fe614aa4c5f9" Dec 10 22:50:46 crc kubenswrapper[4791]: I1210 22:50:46.898016 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 22:50:46 crc kubenswrapper[4791]: I1210 22:50:46.898064 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 22:50:46 crc kubenswrapper[4791]: I1210 22:50:46.898077 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 22:50:46 crc kubenswrapper[4791]: I1210 22:50:46.898104 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 22:50:46 crc kubenswrapper[4791]: I1210 22:50:46.898120 4791 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T22:50:46Z","lastTransitionTime":"2025-12-10T22:50:46Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 22:50:47 crc kubenswrapper[4791]: I1210 22:50:46.999852 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 22:50:47 crc kubenswrapper[4791]: I1210 22:50:47.000124 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 22:50:47 crc kubenswrapper[4791]: I1210 22:50:47.000134 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 22:50:47 crc kubenswrapper[4791]: I1210 22:50:47.000148 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 22:50:47 crc kubenswrapper[4791]: I1210 22:50:47.000159 4791 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T22:50:47Z","lastTransitionTime":"2025-12-10T22:50:47Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 22:50:47 crc kubenswrapper[4791]: I1210 22:50:47.102446 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 22:50:47 crc kubenswrapper[4791]: I1210 22:50:47.102487 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 22:50:47 crc kubenswrapper[4791]: I1210 22:50:47.102497 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 22:50:47 crc kubenswrapper[4791]: I1210 22:50:47.102512 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 22:50:47 crc kubenswrapper[4791]: I1210 22:50:47.102520 4791 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T22:50:47Z","lastTransitionTime":"2025-12-10T22:50:47Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 22:50:47 crc kubenswrapper[4791]: I1210 22:50:47.204850 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 22:50:47 crc kubenswrapper[4791]: I1210 22:50:47.204878 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 22:50:47 crc kubenswrapper[4791]: I1210 22:50:47.204886 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 22:50:47 crc kubenswrapper[4791]: I1210 22:50:47.204902 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 22:50:47 crc kubenswrapper[4791]: I1210 22:50:47.204912 4791 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T22:50:47Z","lastTransitionTime":"2025-12-10T22:50:47Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 22:50:47 crc kubenswrapper[4791]: I1210 22:50:47.308106 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 22:50:47 crc kubenswrapper[4791]: I1210 22:50:47.308185 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 22:50:47 crc kubenswrapper[4791]: I1210 22:50:47.308205 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 22:50:47 crc kubenswrapper[4791]: I1210 22:50:47.308235 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 22:50:47 crc kubenswrapper[4791]: I1210 22:50:47.308255 4791 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T22:50:47Z","lastTransitionTime":"2025-12-10T22:50:47Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 22:50:47 crc kubenswrapper[4791]: I1210 22:50:47.410904 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 22:50:47 crc kubenswrapper[4791]: I1210 22:50:47.410945 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 22:50:47 crc kubenswrapper[4791]: I1210 22:50:47.410955 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 22:50:47 crc kubenswrapper[4791]: I1210 22:50:47.410970 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 22:50:47 crc kubenswrapper[4791]: I1210 22:50:47.410983 4791 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T22:50:47Z","lastTransitionTime":"2025-12-10T22:50:47Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 22:50:47 crc kubenswrapper[4791]: I1210 22:50:47.513051 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 22:50:47 crc kubenswrapper[4791]: I1210 22:50:47.513106 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 22:50:47 crc kubenswrapper[4791]: I1210 22:50:47.513117 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 22:50:47 crc kubenswrapper[4791]: I1210 22:50:47.513135 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 22:50:47 crc kubenswrapper[4791]: I1210 22:50:47.513147 4791 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T22:50:47Z","lastTransitionTime":"2025-12-10T22:50:47Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 22:50:47 crc kubenswrapper[4791]: I1210 22:50:47.543152 4791 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-zhq64_3cd47739-0fa9-4321-aff1-220f8721a0b3/ovnkube-controller/2.log" Dec 10 22:50:47 crc kubenswrapper[4791]: I1210 22:50:47.545440 4791 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-zhq64" event={"ID":"3cd47739-0fa9-4321-aff1-220f8721a0b3","Type":"ContainerStarted","Data":"5f299a2b60de0ef5fb17a1e463d8552d6c4d11d166ae32e490104b3d90116d25"} Dec 10 22:50:47 crc kubenswrapper[4791]: I1210 22:50:47.545831 4791 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-ovn-kubernetes/ovnkube-node-zhq64" Dec 10 22:50:47 crc kubenswrapper[4791]: I1210 22:50:47.556957 4791 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-10T22:49:50Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-10T22:49:50Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T22:50:47Z is after 2025-08-24T17:21:41Z" Dec 10 22:50:47 crc kubenswrapper[4791]: I1210 22:50:47.566609 4791 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-10T22:49:54Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-10T22:49:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-10T22:49:54Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://62bfaf5d7a6d1b452cdaf88086447d0c8cb7306e94434e46752be5475124d56a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T22:49:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T22:50:47Z is after 2025-08-24T17:21:41Z" Dec 10 22:50:47 crc kubenswrapper[4791]: I1210 22:50:47.576524 4791 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-b6kf6" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b4527e71-bb58-4a39-a99c-fe614aa4c5f9\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T22:50:04Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T22:50:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T22:50:04Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T22:50:04Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ngpr4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ngpr4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-10T22:50:04Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-b6kf6\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T22:50:47Z is after 2025-08-24T17:21:41Z" Dec 10 22:50:47 crc kubenswrapper[4791]: I1210 22:50:47.591156 4791 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f8349fea-c49b-49a1-b5ee-032c5dffe021\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T22:49:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T22:49:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T22:49:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T22:49:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T22:49:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://acafe12186df99d1059a31260de0eafce35ff7771bcb8f5c6753b72665d6f60b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T22:49:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ea4c346507862a8c12756a3061d6e25e788a8d79f4621b62f81f4322185c82a8\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T22:49:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://2e6e324ec6f4f7bb69fdfab4ea3ac1aa55e480dd8e2a8984b0fe5b4c9f66636e\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T22:49:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://1b2e11ff2bf8789fa1d0d23ae648b3e6ef653106e5bc4bbb5dd6447f7cd6f05d\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d6aa7f8c5d908d90008f2df49860ac8a3480136da567e3d85f86135d779204f9\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-12-10T22:49:44Z\\\",\\\"message\\\":\\\"W1210 22:49:33.845733 1 cmd.go:257] Using insecure, self-signed certificates\\\\nI1210 22:49:33.846108 1 crypto.go:601] Generating new CA for check-endpoints-signer@1765406973 cert, and key in /tmp/serving-cert-2585701456/serving-signer.crt, /tmp/serving-cert-2585701456/serving-signer.key\\\\nI1210 22:49:34.529748 1 observer_polling.go:159] Starting file observer\\\\nW1210 22:49:34.532225 1 builder.go:272] unable to get owner reference (falling back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": dial tcp [::1]:6443: connect: connection refused\\\\nI1210 22:49:34.532599 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1210 22:49:34.533977 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-2585701456/tls.crt::/tmp/serving-cert-2585701456/tls.key\\\\\\\"\\\\nF1210 22:49:44.774710 1 cmd.go:182] error initializing delegating authentication: unable to load configmap based request-header-client-ca-file: Get \\\\\\\"https://localhost:6443/api/v1/namespaces/kube-system/configmaps/extension-apiserver-authentication\\\\\\\": net/http: TLS handshake timeout\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-10T22:49:33Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T22:49:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://0c4d036ab269f62561fe7ef1586012cb3a16e3583a08b9488a0616ec00c8e0ae\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T22:49:32Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c5856d6b26c19fda13b5c9ac2892a7e2f9c6f79feb8173ecfd9b33b673327d67\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c5856d6b26c19fda13b5c9ac2892a7e2f9c6f79feb8173ecfd9b33b673327d67\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-10T22:49:27Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-10T22:49:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-10T22:49:25Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T22:50:47Z is after 2025-08-24T17:21:41Z" Dec 10 22:50:47 crc kubenswrapper[4791]: I1210 22:50:47.601674 4791 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"02a949d8-38e2-4be3-989f-40a015407140\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T22:49:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T22:49:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T22:50:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T22:50:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T22:49:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://cd4158bc59b30f0f1716bf43c1ae7c9916a5b7d7164d074987fc1f19ba49b778\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T22:49:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://656065fd3cbdc6a25dab5d06c7baa00af16d8f62fc5da687efe9c6cc91945de0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T22:49:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://90ed7fff816b1772ba9ecd95587a59cbed7c30ac302fff7e7c5de934c912bcf4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T22:49:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9b7a5502803eaad9f76f5ffa91b29fbb2657402065a03a4d6433c096aac7264f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"wait-for-host-port\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://9b7a5502803eaad9f76f5ffa91b29fbb2657402065a03a4d6433c096aac7264f\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-10T22:49:27Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-10T22:49:27Z\\\"}}}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-10T22:49:25Z\\\"}}\" for pod \"openshift-kube-scheduler\"/\"openshift-kube-scheduler-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T22:50:47Z is after 2025-08-24T17:21:41Z" Dec 10 22:50:47 crc kubenswrapper[4791]: I1210 22:50:47.615810 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 22:50:47 crc kubenswrapper[4791]: I1210 22:50:47.615843 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 22:50:47 crc kubenswrapper[4791]: I1210 22:50:47.615851 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 22:50:47 crc kubenswrapper[4791]: I1210 22:50:47.615886 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 22:50:47 crc kubenswrapper[4791]: I1210 22:50:47.615895 4791 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T22:50:47Z","lastTransitionTime":"2025-12-10T22:50:47Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 22:50:47 crc kubenswrapper[4791]: I1210 22:50:47.617986 4791 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"90250e13-6c63-443a-a5e5-6e4e341a2289\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T22:49:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T22:49:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T22:49:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T22:49:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T22:49:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ba54f95b1e00843b854db9a10fe6897592159ccacf30ed0f76002093a13eedd3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T22:49:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5aa8ffa518f3a68b60b275f8df4851b2564aa4984b61efdc9470b73072133e43\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T22:49:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://16f989503edc78d9451638e2cf6cf4b19a2cf2b8809b10b237541cb829303091\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T22:49:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://82556824d63ce6e46df14c45689ff1f81ebf729abe902061815c040b323898e9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T22:49:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://61db531940fa6f2effdae5a005b4c91bd2698bd5d1bff9588c171a65ad550eca\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T22:49:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0f65c864d885319ce2a08d61c30be24b3ffb7d23ea30c82ed7faf039cb2c006d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://0f65c864d885319ce2a08d61c30be24b3ffb7d23ea30c82ed7faf039cb2c006d\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-10T22:49:27Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-10T22:49:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e31a76d1d5b2812ba83e24e38268e25ec8bed98fc3da9adcd06bda5822b6ace1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e31a76d1d5b2812ba83e24e38268e25ec8bed98fc3da9adcd06bda5822b6ace1\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-10T22:49:30Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-10T22:49:29Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://b5b5eceae562f3986b543d41de1e977efb85a91d84c98e09159ca6d092eda691\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b5b5eceae562f3986b543d41de1e977efb85a91d84c98e09159ca6d092eda691\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-10T22:49:31Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-10T22:49:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-10T22:49:25Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T22:50:47Z is after 2025-08-24T17:21:41Z" Dec 10 22:50:47 crc kubenswrapper[4791]: I1210 22:50:47.628367 4791 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-5rb5l" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ba35653c-6e06-4cee-a4d6-137764090d18\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T22:49:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T22:49:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T22:49:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T22:49:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://dbb5c875730a98ad2daf316cfafc08fcdce21ac16148154780ac9193ed405dab\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T22:49:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fvvrs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://185eb2ff44221a514878283ec873773a6c9a9073dc64e296f982e60dee1338e4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T22:49:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fvvrs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-10T22:49:51Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-5rb5l\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T22:50:47Z is after 2025-08-24T17:21:41Z" Dec 10 22:50:47 crc kubenswrapper[4791]: I1210 22:50:47.637444 4791 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-s6rk8" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5c9d0978-a6f9-49c5-b377-6619b4e3ff52\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T22:50:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T22:50:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T22:50:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T22:50:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8e55baaf460c794ab196e237d51d12de059aa038617dc2d5366d49c529f5d5c7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T22:50:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7vkgm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://51d1b9b086c60bae9223cb51d5a998d02790a0d91b745aab10392c23e617eb9f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T22:50:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7vkgm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-10T22:50:03Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-s6rk8\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T22:50:47Z is after 2025-08-24T17:21:41Z" Dec 10 22:50:47 crc kubenswrapper[4791]: I1210 22:50:47.647242 4791 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-vczjq" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b9f00f25-3fa1-4b63-9710-fd352224b01b\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T22:49:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T22:49:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T22:49:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T22:49:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4698cf4c6b55b41e628448d37c2eb847c2777d0efbd93afcbf367d5e78623b7e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T22:49:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lrrpv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-10T22:49:50Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-vczjq\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T22:50:47Z is after 2025-08-24T17:21:41Z" Dec 10 22:50:47 crc kubenswrapper[4791]: I1210 22:50:47.658210 4791 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-4nwkq" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"672aa28c-8169-49ed-87b8-21187d13a80c\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T22:49:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T22:49:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T22:50:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T22:50:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f337f881ffd71f4a2ee19e92bb0b92c76cbb9b8352511f741d0c87e7da4cdd18\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a2f2a62315529e148a10a33ccb241e9417fbde0fe42afb61cb7ee2438c74636f\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-12-10T22:50:38Z\\\",\\\"message\\\":\\\"2025-12-10T22:49:52+00:00 [cnibincopy] Successfully copied files in /usr/src/multus-cni/rhel9/bin/ to /host/opt/cni/bin/upgrade_e48d9423-4bca-43c4-9eb9-b55a19a5ffd6\\\\n2025-12-10T22:49:52+00:00 [cnibincopy] Successfully moved files in /host/opt/cni/bin/upgrade_e48d9423-4bca-43c4-9eb9-b55a19a5ffd6 to /host/opt/cni/bin/\\\\n2025-12-10T22:49:52Z [verbose] multus-daemon started\\\\n2025-12-10T22:49:52Z [verbose] Readiness Indicator file check\\\\n2025-12-10T22:50:37Z [error] have you checked that your default network is ready? still waiting for readinessindicatorfile @ /host/run/multus/cni/net.d/10-ovn-kubernetes.conf. pollimmediate error: timed out waiting for the condition\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-10T22:49:52Z\\\"}},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T22:50:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xl62j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-10T22:49:51Z\\\"}}\" for pod \"openshift-multus\"/\"multus-4nwkq\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T22:50:47Z is after 2025-08-24T17:21:41Z" Dec 10 22:50:47 crc kubenswrapper[4791]: I1210 22:50:47.676688 4791 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-zhq64" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3cd47739-0fa9-4321-aff1-220f8721a0b3\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T22:49:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T22:49:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T22:49:52Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T22:49:52Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://68795a29944e56870ffe94fa732e9dcf0bfb344c0d85d1e590f58a876129c61f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T22:49:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wvpw8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://bc721b530c08b433756e978ec0423cc18aabf6771b61eb96f903b4d57ea335ff\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T22:49:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wvpw8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://152a7618be8f38fadf4e564b3948980503cedcdb466c37344ccc20f4a3173a5f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T22:49:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wvpw8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3d18f9cb7207efb1e32de75a30d33e024f656a08f0503db95e5488bc5280971f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T22:49:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wvpw8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://937137ae9d7480cdc7c59bc4a1df0a4c10e72b3b63693af285278789953ed218\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T22:49:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wvpw8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://1a52d19b7671e8b143017c8a981a6f8d3ee3b84650e4b223b814a531ac76f364\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T22:49:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wvpw8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5f299a2b60de0ef5fb17a1e463d8552d6c4d11d166ae32e490104b3d90116d25\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b6d739a4314c4fb536810b8623f691d6aaf1090b9c83f17b484d2b22f7d3ad01\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-12-10T22:50:21Z\\\",\\\"message\\\":\\\"llocateLoadBalancerNodePorts:nil,LoadBalancerClass:nil,InternalTrafficPolicy:*Cluster,TrafficDistribution:nil,},Status:ServiceStatus{LoadBalancer:LoadBalancerStatus{Ingress:[]LoadBalancerIngress{},},Conditions:[]Condition{},},}\\\\nI1210 22:50:21.614710 6424 metrics.go:553] Stopping metrics server at address \\\\\\\"127.0.0.1:29103\\\\\\\"\\\\nI1210 22:50:21.614731 6424 lb_config.go:1031] Cluster endpoints for openshift-kube-scheduler/scheduler for network=default are: map[]\\\\nI1210 22:50:21.614619 6424 services_controller.go:360] Finished syncing service ovn-kubernetes-control-plane on namespace openshift-ovn-kubernetes for network=default : 7.02µs\\\\nI1210 22:50:21.614757 6424 services_controller.go:356] Processing sync for service openshift-route-controller-manager/route-controller-manager for network=default\\\\nI1210 22:50:21.614615 6424 services_controller.go:444] Built service openshift-machine-config-operator/machine-config-daemon LB per-node configs for network=default: []services.lbConfig(nil)\\\\nF1210 22:50:21.614772 6424 ovnkube.go:137] failed to run ovnkube: [failed to start network controller: failed to start default network controller: unable to create admin network policy controller, err: could not add Event Handler for anpInformer during admin network policy controller initialization, handler {0x1fcc6e0 0x1fcc3c0 0x1fcc360} was not added to\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-10T22:50:20Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T22:50:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wvpw8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://dd3f80e5713221f96ed50da7f2e7fd5b9fabfcd5c1656da8d8a86d818a5fdc3f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T22:49:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wvpw8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3afbd1ffcd2c4b40324aca6e52d6fc822db131c3f35a9af3a6f005c2cffdb7bf\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3afbd1ffcd2c4b40324aca6e52d6fc822db131c3f35a9af3a6f005c2cffdb7bf\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-10T22:49:53Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-10T22:49:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wvpw8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-10T22:49:52Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-zhq64\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T22:50:47Z is after 2025-08-24T17:21:41Z" Dec 10 22:50:47 crc kubenswrapper[4791]: I1210 22:50:47.690138 4791 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"698bbe9a-fbe5-436d-a835-5e80df77ba24\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T22:49:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T22:49:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T22:49:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T22:49:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T22:49:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://224d73713312376a58ec6cc2bc8fdb6ffa44c4749ae0e7691d5fe27b9afca09f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T22:49:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://372eeaffd40992c6f06efc4df6dfcb44b31de64ee8256c21508cdd7c74ecb96e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T22:49:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://76b6013b062cac5b0b4d12a94f38ce7b43f9c7ea216063455d88a956ac04d26e\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T22:49:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://78e950e122e761d950da1a3c60c06476e53920e8795f6c61db22c01e22d5136e\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T22:49:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-10T22:49:25Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T22:50:47Z is after 2025-08-24T17:21:41Z" Dec 10 22:50:47 crc kubenswrapper[4791]: I1210 22:50:47.702359 4791 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7a9a3709-32fb-48b4-97f4-e1a19282a085\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T22:49:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T22:49:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T22:49:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T22:49:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T22:49:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://511bd056768a15ea0f650a2b1a6e87e3c71dfc75696222a54f7376227ad480c6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-crio\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T22:49:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kube\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"var-lib-kubelet\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8c489105d0dfb302732d4f5bd2fc9d000967a071802735b778d49bdbc428b264\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://8c489105d0dfb302732d4f5bd2fc9d000967a071802735b778d49bdbc428b264\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-10T22:49:27Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-10T22:49:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var\\\",\\\"name\\\":\\\"var-lib-kubelet\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-10T22:49:25Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"kube-rbac-proxy-crio-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T22:50:47Z is after 2025-08-24T17:21:41Z" Dec 10 22:50:47 crc kubenswrapper[4791]: I1210 22:50:47.712826 4791 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-10T22:49:52Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-10T22:49:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-10T22:49:52Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3734b110802828632d44263ed0a43cabc4b710833f0be120d72db0940fe783e2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T22:49:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://794803ac9be3c52509e6dcb10854d8864e1a9e406762e57c7971b07c4da1220e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T22:49:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T22:50:47Z is after 2025-08-24T17:21:41Z" Dec 10 22:50:47 crc kubenswrapper[4791]: I1210 22:50:47.718785 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 22:50:47 crc kubenswrapper[4791]: I1210 22:50:47.718822 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 22:50:47 crc kubenswrapper[4791]: I1210 22:50:47.718834 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 22:50:47 crc kubenswrapper[4791]: I1210 22:50:47.718849 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 22:50:47 crc kubenswrapper[4791]: I1210 22:50:47.718860 4791 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T22:50:47Z","lastTransitionTime":"2025-12-10T22:50:47Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 22:50:47 crc kubenswrapper[4791]: I1210 22:50:47.727260 4791 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-tchxk" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"c0ab74e0-42a9-4f42-ac80-16030cf4ffff\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T22:49:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T22:49:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T22:49:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T22:49:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://04e0fd1b53256b7a938e8ccab93e2a4eb73dd1a59e0b9799e09c55e511baf7fe\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T22:49:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zn2qt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-10T22:49:50Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-tchxk\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T22:50:47Z is after 2025-08-24T17:21:41Z" Dec 10 22:50:47 crc kubenswrapper[4791]: I1210 22:50:47.741119 4791 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-rw584" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ff9beb90-69b0-4732-bf37-0b81f58ecc98\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T22:49:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T22:49:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T22:49:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T22:49:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ddb9df501dc6e8e5561fbc2aefd431037a3f28f73bf8e82a6598d261a4224807\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T22:49:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fksqn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ce67e54b696bb53ffbbe48c6ce71476b635e0868993e9c427b22f904891433ee\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ce67e54b696bb53ffbbe48c6ce71476b635e0868993e9c427b22f904891433ee\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-10T22:49:52Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-10T22:49:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fksqn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://818d89bfffb8a110670152c8480684665ea8db578e9768f524fbbe6a92f27c22\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://818d89bfffb8a110670152c8480684665ea8db578e9768f524fbbe6a92f27c22\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-10T22:49:53Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-10T22:49:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fksqn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://9aa9c06e894bb2a2ed3738defaeff68a0a0458ead48177f36b90770944c211a0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://9aa9c06e894bb2a2ed3738defaeff68a0a0458ead48177f36b90770944c211a0\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-10T22:49:54Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-10T22:49:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fksqn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://dfd6d03336dcea7529419a30e8d70d84e2290a0fa5b0e46d09ad341c190638d6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://dfd6d03336dcea7529419a30e8d70d84e2290a0fa5b0e46d09ad341c190638d6\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-10T22:49:55Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-10T22:49:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fksqn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b86738401ece6fbc3083782720c66a515850ed78d4054f34fcabab7ff525b85f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b86738401ece6fbc3083782720c66a515850ed78d4054f34fcabab7ff525b85f\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-10T22:49:56Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-10T22:49:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fksqn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://18b326f572a899ddd10197a745bf9264d4ddd975690ad1cfb0ec576ccd5cf102\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://18b326f572a899ddd10197a745bf9264d4ddd975690ad1cfb0ec576ccd5cf102\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-10T22:49:57Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-10T22:49:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fksqn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-10T22:49:51Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-rw584\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T22:50:47Z is after 2025-08-24T17:21:41Z" Dec 10 22:50:47 crc kubenswrapper[4791]: I1210 22:50:47.760663 4791 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-10T22:49:52Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-10T22:49:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-10T22:49:52Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://20651eb0c16e2c71222f56da1ff755cb0e7b084886acbbb24f0ccaf57041ed85\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T22:49:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T22:50:47Z is after 2025-08-24T17:21:41Z" Dec 10 22:50:47 crc kubenswrapper[4791]: I1210 22:50:47.782609 4791 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-10T22:49:50Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-10T22:49:50Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T22:50:47Z is after 2025-08-24T17:21:41Z" Dec 10 22:50:47 crc kubenswrapper[4791]: I1210 22:50:47.815533 4791 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-10T22:49:50Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-10T22:49:50Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T22:50:47Z is after 2025-08-24T17:21:41Z" Dec 10 22:50:47 crc kubenswrapper[4791]: I1210 22:50:47.821009 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 22:50:47 crc kubenswrapper[4791]: I1210 22:50:47.821041 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 22:50:47 crc kubenswrapper[4791]: I1210 22:50:47.821051 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 22:50:47 crc kubenswrapper[4791]: I1210 22:50:47.821063 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 22:50:47 crc kubenswrapper[4791]: I1210 22:50:47.821074 4791 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T22:50:47Z","lastTransitionTime":"2025-12-10T22:50:47Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 22:50:47 crc kubenswrapper[4791]: I1210 22:50:47.923773 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 22:50:47 crc kubenswrapper[4791]: I1210 22:50:47.923834 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 22:50:47 crc kubenswrapper[4791]: I1210 22:50:47.923848 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 22:50:47 crc kubenswrapper[4791]: I1210 22:50:47.923873 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 22:50:47 crc kubenswrapper[4791]: I1210 22:50:47.923888 4791 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T22:50:47Z","lastTransitionTime":"2025-12-10T22:50:47Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 22:50:48 crc kubenswrapper[4791]: I1210 22:50:48.026269 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 22:50:48 crc kubenswrapper[4791]: I1210 22:50:48.026316 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 22:50:48 crc kubenswrapper[4791]: I1210 22:50:48.026328 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 22:50:48 crc kubenswrapper[4791]: I1210 22:50:48.026370 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 22:50:48 crc kubenswrapper[4791]: I1210 22:50:48.026384 4791 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T22:50:48Z","lastTransitionTime":"2025-12-10T22:50:48Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 22:50:48 crc kubenswrapper[4791]: I1210 22:50:48.128784 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 22:50:48 crc kubenswrapper[4791]: I1210 22:50:48.128823 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 22:50:48 crc kubenswrapper[4791]: I1210 22:50:48.128833 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 22:50:48 crc kubenswrapper[4791]: I1210 22:50:48.128866 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 22:50:48 crc kubenswrapper[4791]: I1210 22:50:48.128876 4791 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T22:50:48Z","lastTransitionTime":"2025-12-10T22:50:48Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 22:50:48 crc kubenswrapper[4791]: I1210 22:50:48.231671 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 22:50:48 crc kubenswrapper[4791]: I1210 22:50:48.231734 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 22:50:48 crc kubenswrapper[4791]: I1210 22:50:48.231751 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 22:50:48 crc kubenswrapper[4791]: I1210 22:50:48.231770 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 22:50:48 crc kubenswrapper[4791]: I1210 22:50:48.231783 4791 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T22:50:48Z","lastTransitionTime":"2025-12-10T22:50:48Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 22:50:48 crc kubenswrapper[4791]: I1210 22:50:48.334297 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 22:50:48 crc kubenswrapper[4791]: I1210 22:50:48.334368 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 22:50:48 crc kubenswrapper[4791]: I1210 22:50:48.334381 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 22:50:48 crc kubenswrapper[4791]: I1210 22:50:48.334402 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 22:50:48 crc kubenswrapper[4791]: I1210 22:50:48.334414 4791 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T22:50:48Z","lastTransitionTime":"2025-12-10T22:50:48Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 22:50:48 crc kubenswrapper[4791]: I1210 22:50:48.436446 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 22:50:48 crc kubenswrapper[4791]: I1210 22:50:48.436491 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 22:50:48 crc kubenswrapper[4791]: I1210 22:50:48.436509 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 22:50:48 crc kubenswrapper[4791]: I1210 22:50:48.436525 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 22:50:48 crc kubenswrapper[4791]: I1210 22:50:48.436535 4791 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T22:50:48Z","lastTransitionTime":"2025-12-10T22:50:48Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 22:50:48 crc kubenswrapper[4791]: I1210 22:50:48.538821 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 22:50:48 crc kubenswrapper[4791]: I1210 22:50:48.538875 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 22:50:48 crc kubenswrapper[4791]: I1210 22:50:48.538902 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 22:50:48 crc kubenswrapper[4791]: I1210 22:50:48.538928 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 22:50:48 crc kubenswrapper[4791]: I1210 22:50:48.538944 4791 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T22:50:48Z","lastTransitionTime":"2025-12-10T22:50:48Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 22:50:48 crc kubenswrapper[4791]: I1210 22:50:48.550141 4791 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-zhq64_3cd47739-0fa9-4321-aff1-220f8721a0b3/ovnkube-controller/3.log" Dec 10 22:50:48 crc kubenswrapper[4791]: I1210 22:50:48.550808 4791 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-zhq64_3cd47739-0fa9-4321-aff1-220f8721a0b3/ovnkube-controller/2.log" Dec 10 22:50:48 crc kubenswrapper[4791]: I1210 22:50:48.553777 4791 generic.go:334] "Generic (PLEG): container finished" podID="3cd47739-0fa9-4321-aff1-220f8721a0b3" containerID="5f299a2b60de0ef5fb17a1e463d8552d6c4d11d166ae32e490104b3d90116d25" exitCode=1 Dec 10 22:50:48 crc kubenswrapper[4791]: I1210 22:50:48.553813 4791 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-zhq64" event={"ID":"3cd47739-0fa9-4321-aff1-220f8721a0b3","Type":"ContainerDied","Data":"5f299a2b60de0ef5fb17a1e463d8552d6c4d11d166ae32e490104b3d90116d25"} Dec 10 22:50:48 crc kubenswrapper[4791]: I1210 22:50:48.553873 4791 scope.go:117] "RemoveContainer" containerID="b6d739a4314c4fb536810b8623f691d6aaf1090b9c83f17b484d2b22f7d3ad01" Dec 10 22:50:48 crc kubenswrapper[4791]: I1210 22:50:48.556014 4791 scope.go:117] "RemoveContainer" containerID="5f299a2b60de0ef5fb17a1e463d8552d6c4d11d166ae32e490104b3d90116d25" Dec 10 22:50:48 crc kubenswrapper[4791]: E1210 22:50:48.556608 4791 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"ovnkube-controller\" with CrashLoopBackOff: \"back-off 40s restarting failed container=ovnkube-controller pod=ovnkube-node-zhq64_openshift-ovn-kubernetes(3cd47739-0fa9-4321-aff1-220f8721a0b3)\"" pod="openshift-ovn-kubernetes/ovnkube-node-zhq64" podUID="3cd47739-0fa9-4321-aff1-220f8721a0b3" Dec 10 22:50:48 crc kubenswrapper[4791]: I1210 22:50:48.573438 4791 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-10T22:49:54Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-10T22:49:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-10T22:49:54Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://62bfaf5d7a6d1b452cdaf88086447d0c8cb7306e94434e46752be5475124d56a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T22:49:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T22:50:48Z is after 2025-08-24T17:21:41Z" Dec 10 22:50:48 crc kubenswrapper[4791]: I1210 22:50:48.583755 4791 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-b6kf6" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b4527e71-bb58-4a39-a99c-fe614aa4c5f9\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T22:50:04Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T22:50:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T22:50:04Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T22:50:04Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ngpr4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ngpr4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-10T22:50:04Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-b6kf6\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T22:50:48Z is after 2025-08-24T17:21:41Z" Dec 10 22:50:48 crc kubenswrapper[4791]: I1210 22:50:48.597111 4791 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f8349fea-c49b-49a1-b5ee-032c5dffe021\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T22:49:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T22:49:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T22:49:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T22:49:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T22:49:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://acafe12186df99d1059a31260de0eafce35ff7771bcb8f5c6753b72665d6f60b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T22:49:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ea4c346507862a8c12756a3061d6e25e788a8d79f4621b62f81f4322185c82a8\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T22:49:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://2e6e324ec6f4f7bb69fdfab4ea3ac1aa55e480dd8e2a8984b0fe5b4c9f66636e\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T22:49:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://1b2e11ff2bf8789fa1d0d23ae648b3e6ef653106e5bc4bbb5dd6447f7cd6f05d\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d6aa7f8c5d908d90008f2df49860ac8a3480136da567e3d85f86135d779204f9\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-12-10T22:49:44Z\\\",\\\"message\\\":\\\"W1210 22:49:33.845733 1 cmd.go:257] Using insecure, self-signed certificates\\\\nI1210 22:49:33.846108 1 crypto.go:601] Generating new CA for check-endpoints-signer@1765406973 cert, and key in /tmp/serving-cert-2585701456/serving-signer.crt, /tmp/serving-cert-2585701456/serving-signer.key\\\\nI1210 22:49:34.529748 1 observer_polling.go:159] Starting file observer\\\\nW1210 22:49:34.532225 1 builder.go:272] unable to get owner reference (falling back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": dial tcp [::1]:6443: connect: connection refused\\\\nI1210 22:49:34.532599 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1210 22:49:34.533977 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-2585701456/tls.crt::/tmp/serving-cert-2585701456/tls.key\\\\\\\"\\\\nF1210 22:49:44.774710 1 cmd.go:182] error initializing delegating authentication: unable to load configmap based request-header-client-ca-file: Get \\\\\\\"https://localhost:6443/api/v1/namespaces/kube-system/configmaps/extension-apiserver-authentication\\\\\\\": net/http: TLS handshake timeout\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-10T22:49:33Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T22:49:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://0c4d036ab269f62561fe7ef1586012cb3a16e3583a08b9488a0616ec00c8e0ae\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T22:49:32Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c5856d6b26c19fda13b5c9ac2892a7e2f9c6f79feb8173ecfd9b33b673327d67\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c5856d6b26c19fda13b5c9ac2892a7e2f9c6f79feb8173ecfd9b33b673327d67\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-10T22:49:27Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-10T22:49:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-10T22:49:25Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T22:50:48Z is after 2025-08-24T17:21:41Z" Dec 10 22:50:48 crc kubenswrapper[4791]: I1210 22:50:48.610305 4791 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"02a949d8-38e2-4be3-989f-40a015407140\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T22:49:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T22:49:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T22:50:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T22:50:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T22:49:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://cd4158bc59b30f0f1716bf43c1ae7c9916a5b7d7164d074987fc1f19ba49b778\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T22:49:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://656065fd3cbdc6a25dab5d06c7baa00af16d8f62fc5da687efe9c6cc91945de0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T22:49:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://90ed7fff816b1772ba9ecd95587a59cbed7c30ac302fff7e7c5de934c912bcf4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T22:49:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9b7a5502803eaad9f76f5ffa91b29fbb2657402065a03a4d6433c096aac7264f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"wait-for-host-port\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://9b7a5502803eaad9f76f5ffa91b29fbb2657402065a03a4d6433c096aac7264f\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-10T22:49:27Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-10T22:49:27Z\\\"}}}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-10T22:49:25Z\\\"}}\" for pod \"openshift-kube-scheduler\"/\"openshift-kube-scheduler-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T22:50:48Z is after 2025-08-24T17:21:41Z" Dec 10 22:50:48 crc kubenswrapper[4791]: I1210 22:50:48.636785 4791 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"90250e13-6c63-443a-a5e5-6e4e341a2289\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T22:49:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T22:49:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T22:49:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T22:49:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T22:49:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ba54f95b1e00843b854db9a10fe6897592159ccacf30ed0f76002093a13eedd3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T22:49:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5aa8ffa518f3a68b60b275f8df4851b2564aa4984b61efdc9470b73072133e43\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T22:49:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://16f989503edc78d9451638e2cf6cf4b19a2cf2b8809b10b237541cb829303091\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T22:49:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://82556824d63ce6e46df14c45689ff1f81ebf729abe902061815c040b323898e9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T22:49:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://61db531940fa6f2effdae5a005b4c91bd2698bd5d1bff9588c171a65ad550eca\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T22:49:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0f65c864d885319ce2a08d61c30be24b3ffb7d23ea30c82ed7faf039cb2c006d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://0f65c864d885319ce2a08d61c30be24b3ffb7d23ea30c82ed7faf039cb2c006d\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-10T22:49:27Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-10T22:49:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e31a76d1d5b2812ba83e24e38268e25ec8bed98fc3da9adcd06bda5822b6ace1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e31a76d1d5b2812ba83e24e38268e25ec8bed98fc3da9adcd06bda5822b6ace1\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-10T22:49:30Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-10T22:49:29Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://b5b5eceae562f3986b543d41de1e977efb85a91d84c98e09159ca6d092eda691\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b5b5eceae562f3986b543d41de1e977efb85a91d84c98e09159ca6d092eda691\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-10T22:49:31Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-10T22:49:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-10T22:49:25Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T22:50:48Z is after 2025-08-24T17:21:41Z" Dec 10 22:50:48 crc kubenswrapper[4791]: I1210 22:50:48.641215 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 22:50:48 crc kubenswrapper[4791]: I1210 22:50:48.641238 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 22:50:48 crc kubenswrapper[4791]: I1210 22:50:48.641247 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 22:50:48 crc kubenswrapper[4791]: I1210 22:50:48.641260 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 22:50:48 crc kubenswrapper[4791]: I1210 22:50:48.641270 4791 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T22:50:48Z","lastTransitionTime":"2025-12-10T22:50:48Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 22:50:48 crc kubenswrapper[4791]: I1210 22:50:48.649517 4791 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-10T22:49:50Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-10T22:49:50Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T22:50:48Z is after 2025-08-24T17:21:41Z" Dec 10 22:50:48 crc kubenswrapper[4791]: I1210 22:50:48.660429 4791 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-5rb5l" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ba35653c-6e06-4cee-a4d6-137764090d18\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T22:49:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T22:49:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T22:49:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T22:49:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://dbb5c875730a98ad2daf316cfafc08fcdce21ac16148154780ac9193ed405dab\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T22:49:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fvvrs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://185eb2ff44221a514878283ec873773a6c9a9073dc64e296f982e60dee1338e4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T22:49:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fvvrs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-10T22:49:51Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-5rb5l\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T22:50:48Z is after 2025-08-24T17:21:41Z" Dec 10 22:50:48 crc kubenswrapper[4791]: I1210 22:50:48.670912 4791 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-s6rk8" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5c9d0978-a6f9-49c5-b377-6619b4e3ff52\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T22:50:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T22:50:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T22:50:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T22:50:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8e55baaf460c794ab196e237d51d12de059aa038617dc2d5366d49c529f5d5c7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T22:50:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7vkgm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://51d1b9b086c60bae9223cb51d5a998d02790a0d91b745aab10392c23e617eb9f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T22:50:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7vkgm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-10T22:50:03Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-s6rk8\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T22:50:48Z is after 2025-08-24T17:21:41Z" Dec 10 22:50:48 crc kubenswrapper[4791]: I1210 22:50:48.684449 4791 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-4nwkq" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"672aa28c-8169-49ed-87b8-21187d13a80c\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T22:49:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T22:49:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T22:50:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T22:50:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f337f881ffd71f4a2ee19e92bb0b92c76cbb9b8352511f741d0c87e7da4cdd18\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a2f2a62315529e148a10a33ccb241e9417fbde0fe42afb61cb7ee2438c74636f\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-12-10T22:50:38Z\\\",\\\"message\\\":\\\"2025-12-10T22:49:52+00:00 [cnibincopy] Successfully copied files in /usr/src/multus-cni/rhel9/bin/ to /host/opt/cni/bin/upgrade_e48d9423-4bca-43c4-9eb9-b55a19a5ffd6\\\\n2025-12-10T22:49:52+00:00 [cnibincopy] Successfully moved files in /host/opt/cni/bin/upgrade_e48d9423-4bca-43c4-9eb9-b55a19a5ffd6 to /host/opt/cni/bin/\\\\n2025-12-10T22:49:52Z [verbose] multus-daemon started\\\\n2025-12-10T22:49:52Z [verbose] Readiness Indicator file check\\\\n2025-12-10T22:50:37Z [error] have you checked that your default network is ready? still waiting for readinessindicatorfile @ /host/run/multus/cni/net.d/10-ovn-kubernetes.conf. pollimmediate error: timed out waiting for the condition\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-10T22:49:52Z\\\"}},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T22:50:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xl62j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-10T22:49:51Z\\\"}}\" for pod \"openshift-multus\"/\"multus-4nwkq\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T22:50:48Z is after 2025-08-24T17:21:41Z" Dec 10 22:50:48 crc kubenswrapper[4791]: I1210 22:50:48.711665 4791 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-zhq64" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3cd47739-0fa9-4321-aff1-220f8721a0b3\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T22:49:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T22:49:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T22:49:52Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T22:49:52Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://68795a29944e56870ffe94fa732e9dcf0bfb344c0d85d1e590f58a876129c61f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T22:49:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wvpw8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://bc721b530c08b433756e978ec0423cc18aabf6771b61eb96f903b4d57ea335ff\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T22:49:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wvpw8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://152a7618be8f38fadf4e564b3948980503cedcdb466c37344ccc20f4a3173a5f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T22:49:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wvpw8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3d18f9cb7207efb1e32de75a30d33e024f656a08f0503db95e5488bc5280971f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T22:49:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wvpw8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://937137ae9d7480cdc7c59bc4a1df0a4c10e72b3b63693af285278789953ed218\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T22:49:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wvpw8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://1a52d19b7671e8b143017c8a981a6f8d3ee3b84650e4b223b814a531ac76f364\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T22:49:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wvpw8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5f299a2b60de0ef5fb17a1e463d8552d6c4d11d166ae32e490104b3d90116d25\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b6d739a4314c4fb536810b8623f691d6aaf1090b9c83f17b484d2b22f7d3ad01\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-12-10T22:50:21Z\\\",\\\"message\\\":\\\"llocateLoadBalancerNodePorts:nil,LoadBalancerClass:nil,InternalTrafficPolicy:*Cluster,TrafficDistribution:nil,},Status:ServiceStatus{LoadBalancer:LoadBalancerStatus{Ingress:[]LoadBalancerIngress{},},Conditions:[]Condition{},},}\\\\nI1210 22:50:21.614710 6424 metrics.go:553] Stopping metrics server at address \\\\\\\"127.0.0.1:29103\\\\\\\"\\\\nI1210 22:50:21.614731 6424 lb_config.go:1031] Cluster endpoints for openshift-kube-scheduler/scheduler for network=default are: map[]\\\\nI1210 22:50:21.614619 6424 services_controller.go:360] Finished syncing service ovn-kubernetes-control-plane on namespace openshift-ovn-kubernetes for network=default : 7.02µs\\\\nI1210 22:50:21.614757 6424 services_controller.go:356] Processing sync for service openshift-route-controller-manager/route-controller-manager for network=default\\\\nI1210 22:50:21.614615 6424 services_controller.go:444] Built service openshift-machine-config-operator/machine-config-daemon LB per-node configs for network=default: []services.lbConfig(nil)\\\\nF1210 22:50:21.614772 6424 ovnkube.go:137] failed to run ovnkube: [failed to start network controller: failed to start default network controller: unable to create admin network policy controller, err: could not add Event Handler for anpInformer during admin network policy controller initialization, handler {0x1fcc6e0 0x1fcc3c0 0x1fcc360} was not added to\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-10T22:50:20Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5f299a2b60de0ef5fb17a1e463d8552d6c4d11d166ae32e490104b3d90116d25\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-12-10T22:50:48Z\\\",\\\"message\\\":\\\"11] Stopping reflector *v1.EgressIP (0s) from github.com/openshift/ovn-kubernetes/go-controller/pkg/crd/egressip/v1/apis/informers/externalversions/factory.go:140\\\\nI1210 22:50:47.868727 6852 reflector.go:311] Stopping reflector *v1.Pod (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI1210 22:50:47.868812 6852 reflector.go:311] Stopping reflector *v1.NetworkPolicy (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI1210 22:50:47.869024 6852 reflector.go:311] Stopping reflector *v1.EndpointSlice (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI1210 22:50:47.869223 6852 reflector.go:311] Stopping reflector *v1alpha1.BaselineAdminNetworkPolicy (0s) from sigs.k8s.io/network-policy-api/pkg/client/informers/externalversions/factory.go:141\\\\nI1210 22:50:47.869558 6852 handler.go:190] Sending *v1.Namespace event handler 1 for removal\\\\nI1210 22:50:47.869577 6852 handler.go:190] Sending *v1.Namespace event handler 5 for removal\\\\nI1210 22:50:47.869595 6852 handler.go:208] Removed *v1.Namespace event handler 5\\\\nI1210 22:50:47.869604 6852 handler.go:190] Sending *v1.NetworkPolicy event handler 4 for removal\\\\nI1210 22:50:47.869608 6852 handler.go:208] Removed *v1.Namespace event handler 1\\\\nI1210 22:50:47.869637 6852 factory.go:656] Stopping watch factory\\\\nI1210 22:50:47.869656 6852 ovnkube.go:599] Stopped ovnkube\\\\nI1210 22:50:\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-10T22:50:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wvpw8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://dd3f80e5713221f96ed50da7f2e7fd5b9fabfcd5c1656da8d8a86d818a5fdc3f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T22:49:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wvpw8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3afbd1ffcd2c4b40324aca6e52d6fc822db131c3f35a9af3a6f005c2cffdb7bf\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3afbd1ffcd2c4b40324aca6e52d6fc822db131c3f35a9af3a6f005c2cffdb7bf\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-10T22:49:53Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-10T22:49:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wvpw8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-10T22:49:52Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-zhq64\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T22:50:48Z is after 2025-08-24T17:21:41Z" Dec 10 22:50:48 crc kubenswrapper[4791]: I1210 22:50:48.724836 4791 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"698bbe9a-fbe5-436d-a835-5e80df77ba24\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T22:49:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T22:49:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T22:49:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T22:49:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T22:49:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://224d73713312376a58ec6cc2bc8fdb6ffa44c4749ae0e7691d5fe27b9afca09f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T22:49:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://372eeaffd40992c6f06efc4df6dfcb44b31de64ee8256c21508cdd7c74ecb96e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T22:49:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://76b6013b062cac5b0b4d12a94f38ce7b43f9c7ea216063455d88a956ac04d26e\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T22:49:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://78e950e122e761d950da1a3c60c06476e53920e8795f6c61db22c01e22d5136e\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T22:49:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-10T22:49:25Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T22:50:48Z is after 2025-08-24T17:21:41Z" Dec 10 22:50:48 crc kubenswrapper[4791]: I1210 22:50:48.735025 4791 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7a9a3709-32fb-48b4-97f4-e1a19282a085\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T22:49:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T22:49:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T22:49:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T22:49:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T22:49:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://511bd056768a15ea0f650a2b1a6e87e3c71dfc75696222a54f7376227ad480c6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-crio\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T22:49:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kube\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"var-lib-kubelet\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8c489105d0dfb302732d4f5bd2fc9d000967a071802735b778d49bdbc428b264\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://8c489105d0dfb302732d4f5bd2fc9d000967a071802735b778d49bdbc428b264\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-10T22:49:27Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-10T22:49:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var\\\",\\\"name\\\":\\\"var-lib-kubelet\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-10T22:49:25Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"kube-rbac-proxy-crio-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T22:50:48Z is after 2025-08-24T17:21:41Z" Dec 10 22:50:48 crc kubenswrapper[4791]: I1210 22:50:48.743698 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 22:50:48 crc kubenswrapper[4791]: I1210 22:50:48.743735 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 22:50:48 crc kubenswrapper[4791]: I1210 22:50:48.743745 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 22:50:48 crc kubenswrapper[4791]: I1210 22:50:48.743759 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 22:50:48 crc kubenswrapper[4791]: I1210 22:50:48.743771 4791 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T22:50:48Z","lastTransitionTime":"2025-12-10T22:50:48Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 22:50:48 crc kubenswrapper[4791]: I1210 22:50:48.747164 4791 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-10T22:49:52Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-10T22:49:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-10T22:49:52Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3734b110802828632d44263ed0a43cabc4b710833f0be120d72db0940fe783e2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T22:49:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://794803ac9be3c52509e6dcb10854d8864e1a9e406762e57c7971b07c4da1220e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T22:49:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T22:50:48Z is after 2025-08-24T17:21:41Z" Dec 10 22:50:48 crc kubenswrapper[4791]: I1210 22:50:48.759611 4791 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-vczjq" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b9f00f25-3fa1-4b63-9710-fd352224b01b\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T22:49:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T22:49:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T22:49:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T22:49:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4698cf4c6b55b41e628448d37c2eb847c2777d0efbd93afcbf367d5e78623b7e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T22:49:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lrrpv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-10T22:49:50Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-vczjq\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T22:50:48Z is after 2025-08-24T17:21:41Z" Dec 10 22:50:48 crc kubenswrapper[4791]: I1210 22:50:48.774636 4791 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-rw584" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ff9beb90-69b0-4732-bf37-0b81f58ecc98\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T22:49:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T22:49:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T22:49:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T22:49:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ddb9df501dc6e8e5561fbc2aefd431037a3f28f73bf8e82a6598d261a4224807\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T22:49:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fksqn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ce67e54b696bb53ffbbe48c6ce71476b635e0868993e9c427b22f904891433ee\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ce67e54b696bb53ffbbe48c6ce71476b635e0868993e9c427b22f904891433ee\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-10T22:49:52Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-10T22:49:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fksqn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://818d89bfffb8a110670152c8480684665ea8db578e9768f524fbbe6a92f27c22\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://818d89bfffb8a110670152c8480684665ea8db578e9768f524fbbe6a92f27c22\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-10T22:49:53Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-10T22:49:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fksqn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://9aa9c06e894bb2a2ed3738defaeff68a0a0458ead48177f36b90770944c211a0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://9aa9c06e894bb2a2ed3738defaeff68a0a0458ead48177f36b90770944c211a0\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-10T22:49:54Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-10T22:49:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fksqn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://dfd6d03336dcea7529419a30e8d70d84e2290a0fa5b0e46d09ad341c190638d6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://dfd6d03336dcea7529419a30e8d70d84e2290a0fa5b0e46d09ad341c190638d6\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-10T22:49:55Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-10T22:49:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fksqn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b86738401ece6fbc3083782720c66a515850ed78d4054f34fcabab7ff525b85f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b86738401ece6fbc3083782720c66a515850ed78d4054f34fcabab7ff525b85f\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-10T22:49:56Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-10T22:49:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fksqn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://18b326f572a899ddd10197a745bf9264d4ddd975690ad1cfb0ec576ccd5cf102\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://18b326f572a899ddd10197a745bf9264d4ddd975690ad1cfb0ec576ccd5cf102\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-10T22:49:57Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-10T22:49:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fksqn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-10T22:49:51Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-rw584\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T22:50:48Z is after 2025-08-24T17:21:41Z" Dec 10 22:50:48 crc kubenswrapper[4791]: I1210 22:50:48.788957 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 22:50:48 crc kubenswrapper[4791]: I1210 22:50:48.789031 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 22:50:48 crc kubenswrapper[4791]: I1210 22:50:48.789048 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 22:50:48 crc kubenswrapper[4791]: I1210 22:50:48.789066 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 22:50:48 crc kubenswrapper[4791]: I1210 22:50:48.789078 4791 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T22:50:48Z","lastTransitionTime":"2025-12-10T22:50:48Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 22:50:48 crc kubenswrapper[4791]: I1210 22:50:48.789581 4791 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-10T22:49:52Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-10T22:49:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-10T22:49:52Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://20651eb0c16e2c71222f56da1ff755cb0e7b084886acbbb24f0ccaf57041ed85\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T22:49:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T22:50:48Z is after 2025-08-24T17:21:41Z" Dec 10 22:50:48 crc kubenswrapper[4791]: I1210 22:50:48.803059 4791 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-10T22:49:50Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-10T22:49:50Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T22:50:48Z is after 2025-08-24T17:21:41Z" Dec 10 22:50:48 crc kubenswrapper[4791]: E1210 22:50:48.805186 4791 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-12-10T22:50:48Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-10T22:50:48Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-10T22:50:48Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-10T22:50:48Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-10T22:50:48Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-10T22:50:48Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-10T22:50:48Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-10T22:50:48Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"4fab3af6-f657-4dfc-8ef2-12b7c978c94f\\\",\\\"systemUUID\\\":\\\"442c0b68-8dce-47e0-bd68-d5210d7e0493\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T22:50:48Z is after 2025-08-24T17:21:41Z" Dec 10 22:50:48 crc kubenswrapper[4791]: I1210 22:50:48.808684 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 22:50:48 crc kubenswrapper[4791]: I1210 22:50:48.808769 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 22:50:48 crc kubenswrapper[4791]: I1210 22:50:48.808780 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 22:50:48 crc kubenswrapper[4791]: I1210 22:50:48.808796 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 22:50:48 crc kubenswrapper[4791]: I1210 22:50:48.808808 4791 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T22:50:48Z","lastTransitionTime":"2025-12-10T22:50:48Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 22:50:48 crc kubenswrapper[4791]: I1210 22:50:48.816382 4791 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-10T22:49:50Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-10T22:49:50Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T22:50:48Z is after 2025-08-24T17:21:41Z" Dec 10 22:50:48 crc kubenswrapper[4791]: E1210 22:50:48.821982 4791 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-12-10T22:50:48Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-10T22:50:48Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-10T22:50:48Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-10T22:50:48Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-10T22:50:48Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-10T22:50:48Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-10T22:50:48Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-10T22:50:48Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"4fab3af6-f657-4dfc-8ef2-12b7c978c94f\\\",\\\"systemUUID\\\":\\\"442c0b68-8dce-47e0-bd68-d5210d7e0493\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T22:50:48Z is after 2025-08-24T17:21:41Z" Dec 10 22:50:48 crc kubenswrapper[4791]: I1210 22:50:48.824723 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 22:50:48 crc kubenswrapper[4791]: I1210 22:50:48.824768 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 22:50:48 crc kubenswrapper[4791]: I1210 22:50:48.824781 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 22:50:48 crc kubenswrapper[4791]: I1210 22:50:48.824806 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 22:50:48 crc kubenswrapper[4791]: I1210 22:50:48.824822 4791 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T22:50:48Z","lastTransitionTime":"2025-12-10T22:50:48Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 22:50:48 crc kubenswrapper[4791]: I1210 22:50:48.826945 4791 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-tchxk" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"c0ab74e0-42a9-4f42-ac80-16030cf4ffff\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T22:49:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T22:49:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T22:49:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T22:49:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://04e0fd1b53256b7a938e8ccab93e2a4eb73dd1a59e0b9799e09c55e511baf7fe\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T22:49:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zn2qt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-10T22:49:50Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-tchxk\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T22:50:48Z is after 2025-08-24T17:21:41Z" Dec 10 22:50:48 crc kubenswrapper[4791]: E1210 22:50:48.836878 4791 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-12-10T22:50:48Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-10T22:50:48Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-10T22:50:48Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-10T22:50:48Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-10T22:50:48Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-10T22:50:48Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-10T22:50:48Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-10T22:50:48Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"4fab3af6-f657-4dfc-8ef2-12b7c978c94f\\\",\\\"systemUUID\\\":\\\"442c0b68-8dce-47e0-bd68-d5210d7e0493\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T22:50:48Z is after 2025-08-24T17:21:41Z" Dec 10 22:50:48 crc kubenswrapper[4791]: I1210 22:50:48.840903 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 22:50:48 crc kubenswrapper[4791]: I1210 22:50:48.840938 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 22:50:48 crc kubenswrapper[4791]: I1210 22:50:48.840950 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 22:50:48 crc kubenswrapper[4791]: I1210 22:50:48.840968 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 22:50:48 crc kubenswrapper[4791]: I1210 22:50:48.840979 4791 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T22:50:48Z","lastTransitionTime":"2025-12-10T22:50:48Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 22:50:48 crc kubenswrapper[4791]: E1210 22:50:48.852549 4791 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-12-10T22:50:48Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-10T22:50:48Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-10T22:50:48Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-10T22:50:48Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-10T22:50:48Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-10T22:50:48Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-10T22:50:48Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-10T22:50:48Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"4fab3af6-f657-4dfc-8ef2-12b7c978c94f\\\",\\\"systemUUID\\\":\\\"442c0b68-8dce-47e0-bd68-d5210d7e0493\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T22:50:48Z is after 2025-08-24T17:21:41Z" Dec 10 22:50:48 crc kubenswrapper[4791]: I1210 22:50:48.856648 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 22:50:48 crc kubenswrapper[4791]: I1210 22:50:48.856687 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 22:50:48 crc kubenswrapper[4791]: I1210 22:50:48.856698 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 22:50:48 crc kubenswrapper[4791]: I1210 22:50:48.856717 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 22:50:48 crc kubenswrapper[4791]: I1210 22:50:48.856728 4791 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T22:50:48Z","lastTransitionTime":"2025-12-10T22:50:48Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 22:50:48 crc kubenswrapper[4791]: E1210 22:50:48.868519 4791 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-12-10T22:50:48Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-10T22:50:48Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-10T22:50:48Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-10T22:50:48Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-10T22:50:48Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-10T22:50:48Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-10T22:50:48Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-10T22:50:48Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"4fab3af6-f657-4dfc-8ef2-12b7c978c94f\\\",\\\"systemUUID\\\":\\\"442c0b68-8dce-47e0-bd68-d5210d7e0493\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T22:50:48Z is after 2025-08-24T17:21:41Z" Dec 10 22:50:48 crc kubenswrapper[4791]: E1210 22:50:48.868626 4791 kubelet_node_status.go:572] "Unable to update node status" err="update node status exceeds retry count" Dec 10 22:50:48 crc kubenswrapper[4791]: I1210 22:50:48.870218 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 22:50:48 crc kubenswrapper[4791]: I1210 22:50:48.870252 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 22:50:48 crc kubenswrapper[4791]: I1210 22:50:48.870267 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 22:50:48 crc kubenswrapper[4791]: I1210 22:50:48.870282 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 22:50:48 crc kubenswrapper[4791]: I1210 22:50:48.870293 4791 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T22:50:48Z","lastTransitionTime":"2025-12-10T22:50:48Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 22:50:48 crc kubenswrapper[4791]: I1210 22:50:48.884715 4791 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-b6kf6" Dec 10 22:50:48 crc kubenswrapper[4791]: I1210 22:50:48.884811 4791 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 10 22:50:48 crc kubenswrapper[4791]: E1210 22:50:48.884936 4791 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-b6kf6" podUID="b4527e71-bb58-4a39-a99c-fe614aa4c5f9" Dec 10 22:50:48 crc kubenswrapper[4791]: I1210 22:50:48.884967 4791 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 10 22:50:48 crc kubenswrapper[4791]: I1210 22:50:48.885024 4791 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 10 22:50:48 crc kubenswrapper[4791]: E1210 22:50:48.885165 4791 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 10 22:50:48 crc kubenswrapper[4791]: E1210 22:50:48.885182 4791 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 10 22:50:48 crc kubenswrapper[4791]: E1210 22:50:48.885328 4791 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 10 22:50:48 crc kubenswrapper[4791]: I1210 22:50:48.973257 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 22:50:48 crc kubenswrapper[4791]: I1210 22:50:48.973385 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 22:50:48 crc kubenswrapper[4791]: I1210 22:50:48.973421 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 22:50:48 crc kubenswrapper[4791]: I1210 22:50:48.973457 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 22:50:48 crc kubenswrapper[4791]: I1210 22:50:48.973482 4791 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T22:50:48Z","lastTransitionTime":"2025-12-10T22:50:48Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 22:50:49 crc kubenswrapper[4791]: I1210 22:50:49.076762 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 22:50:49 crc kubenswrapper[4791]: I1210 22:50:49.076819 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 22:50:49 crc kubenswrapper[4791]: I1210 22:50:49.076836 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 22:50:49 crc kubenswrapper[4791]: I1210 22:50:49.076859 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 22:50:49 crc kubenswrapper[4791]: I1210 22:50:49.076878 4791 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T22:50:49Z","lastTransitionTime":"2025-12-10T22:50:49Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 22:50:49 crc kubenswrapper[4791]: I1210 22:50:49.178956 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 22:50:49 crc kubenswrapper[4791]: I1210 22:50:49.178996 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 22:50:49 crc kubenswrapper[4791]: I1210 22:50:49.179040 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 22:50:49 crc kubenswrapper[4791]: I1210 22:50:49.179055 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 22:50:49 crc kubenswrapper[4791]: I1210 22:50:49.179064 4791 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T22:50:49Z","lastTransitionTime":"2025-12-10T22:50:49Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 22:50:49 crc kubenswrapper[4791]: I1210 22:50:49.281699 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 22:50:49 crc kubenswrapper[4791]: I1210 22:50:49.281765 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 22:50:49 crc kubenswrapper[4791]: I1210 22:50:49.281776 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 22:50:49 crc kubenswrapper[4791]: I1210 22:50:49.281817 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 22:50:49 crc kubenswrapper[4791]: I1210 22:50:49.281833 4791 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T22:50:49Z","lastTransitionTime":"2025-12-10T22:50:49Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 22:50:49 crc kubenswrapper[4791]: I1210 22:50:49.384843 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 22:50:49 crc kubenswrapper[4791]: I1210 22:50:49.384882 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 22:50:49 crc kubenswrapper[4791]: I1210 22:50:49.384892 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 22:50:49 crc kubenswrapper[4791]: I1210 22:50:49.384909 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 22:50:49 crc kubenswrapper[4791]: I1210 22:50:49.384920 4791 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T22:50:49Z","lastTransitionTime":"2025-12-10T22:50:49Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 22:50:49 crc kubenswrapper[4791]: I1210 22:50:49.487361 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 22:50:49 crc kubenswrapper[4791]: I1210 22:50:49.487411 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 22:50:49 crc kubenswrapper[4791]: I1210 22:50:49.487425 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 22:50:49 crc kubenswrapper[4791]: I1210 22:50:49.487443 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 22:50:49 crc kubenswrapper[4791]: I1210 22:50:49.487458 4791 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T22:50:49Z","lastTransitionTime":"2025-12-10T22:50:49Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 22:50:49 crc kubenswrapper[4791]: I1210 22:50:49.559137 4791 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-zhq64_3cd47739-0fa9-4321-aff1-220f8721a0b3/ovnkube-controller/3.log" Dec 10 22:50:49 crc kubenswrapper[4791]: I1210 22:50:49.563521 4791 scope.go:117] "RemoveContainer" containerID="5f299a2b60de0ef5fb17a1e463d8552d6c4d11d166ae32e490104b3d90116d25" Dec 10 22:50:49 crc kubenswrapper[4791]: E1210 22:50:49.563825 4791 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"ovnkube-controller\" with CrashLoopBackOff: \"back-off 40s restarting failed container=ovnkube-controller pod=ovnkube-node-zhq64_openshift-ovn-kubernetes(3cd47739-0fa9-4321-aff1-220f8721a0b3)\"" pod="openshift-ovn-kubernetes/ovnkube-node-zhq64" podUID="3cd47739-0fa9-4321-aff1-220f8721a0b3" Dec 10 22:50:49 crc kubenswrapper[4791]: I1210 22:50:49.588394 4791 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-zhq64" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3cd47739-0fa9-4321-aff1-220f8721a0b3\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T22:49:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T22:49:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T22:49:52Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T22:49:52Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://68795a29944e56870ffe94fa732e9dcf0bfb344c0d85d1e590f58a876129c61f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T22:49:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wvpw8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://bc721b530c08b433756e978ec0423cc18aabf6771b61eb96f903b4d57ea335ff\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T22:49:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wvpw8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://152a7618be8f38fadf4e564b3948980503cedcdb466c37344ccc20f4a3173a5f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T22:49:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wvpw8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3d18f9cb7207efb1e32de75a30d33e024f656a08f0503db95e5488bc5280971f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T22:49:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wvpw8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://937137ae9d7480cdc7c59bc4a1df0a4c10e72b3b63693af285278789953ed218\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T22:49:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wvpw8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://1a52d19b7671e8b143017c8a981a6f8d3ee3b84650e4b223b814a531ac76f364\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T22:49:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wvpw8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5f299a2b60de0ef5fb17a1e463d8552d6c4d11d166ae32e490104b3d90116d25\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5f299a2b60de0ef5fb17a1e463d8552d6c4d11d166ae32e490104b3d90116d25\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-12-10T22:50:48Z\\\",\\\"message\\\":\\\"11] Stopping reflector *v1.EgressIP (0s) from github.com/openshift/ovn-kubernetes/go-controller/pkg/crd/egressip/v1/apis/informers/externalversions/factory.go:140\\\\nI1210 22:50:47.868727 6852 reflector.go:311] Stopping reflector *v1.Pod (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI1210 22:50:47.868812 6852 reflector.go:311] Stopping reflector *v1.NetworkPolicy (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI1210 22:50:47.869024 6852 reflector.go:311] Stopping reflector *v1.EndpointSlice (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI1210 22:50:47.869223 6852 reflector.go:311] Stopping reflector *v1alpha1.BaselineAdminNetworkPolicy (0s) from sigs.k8s.io/network-policy-api/pkg/client/informers/externalversions/factory.go:141\\\\nI1210 22:50:47.869558 6852 handler.go:190] Sending *v1.Namespace event handler 1 for removal\\\\nI1210 22:50:47.869577 6852 handler.go:190] Sending *v1.Namespace event handler 5 for removal\\\\nI1210 22:50:47.869595 6852 handler.go:208] Removed *v1.Namespace event handler 5\\\\nI1210 22:50:47.869604 6852 handler.go:190] Sending *v1.NetworkPolicy event handler 4 for removal\\\\nI1210 22:50:47.869608 6852 handler.go:208] Removed *v1.Namespace event handler 1\\\\nI1210 22:50:47.869637 6852 factory.go:656] Stopping watch factory\\\\nI1210 22:50:47.869656 6852 ovnkube.go:599] Stopped ovnkube\\\\nI1210 22:50:\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-10T22:50:47Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 40s restarting failed container=ovnkube-controller pod=ovnkube-node-zhq64_openshift-ovn-kubernetes(3cd47739-0fa9-4321-aff1-220f8721a0b3)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wvpw8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://dd3f80e5713221f96ed50da7f2e7fd5b9fabfcd5c1656da8d8a86d818a5fdc3f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T22:49:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wvpw8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3afbd1ffcd2c4b40324aca6e52d6fc822db131c3f35a9af3a6f005c2cffdb7bf\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3afbd1ffcd2c4b40324aca6e52d6fc822db131c3f35a9af3a6f005c2cffdb7bf\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-10T22:49:53Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-10T22:49:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wvpw8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-10T22:49:52Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-zhq64\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T22:50:49Z is after 2025-08-24T17:21:41Z" Dec 10 22:50:49 crc kubenswrapper[4791]: I1210 22:50:49.590016 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 22:50:49 crc kubenswrapper[4791]: I1210 22:50:49.590055 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 22:50:49 crc kubenswrapper[4791]: I1210 22:50:49.590067 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 22:50:49 crc kubenswrapper[4791]: I1210 22:50:49.590086 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 22:50:49 crc kubenswrapper[4791]: I1210 22:50:49.590099 4791 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T22:50:49Z","lastTransitionTime":"2025-12-10T22:50:49Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 22:50:49 crc kubenswrapper[4791]: I1210 22:50:49.607459 4791 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"698bbe9a-fbe5-436d-a835-5e80df77ba24\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T22:49:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T22:49:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T22:49:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T22:49:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T22:49:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://224d73713312376a58ec6cc2bc8fdb6ffa44c4749ae0e7691d5fe27b9afca09f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T22:49:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://372eeaffd40992c6f06efc4df6dfcb44b31de64ee8256c21508cdd7c74ecb96e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T22:49:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://76b6013b062cac5b0b4d12a94f38ce7b43f9c7ea216063455d88a956ac04d26e\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T22:49:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://78e950e122e761d950da1a3c60c06476e53920e8795f6c61db22c01e22d5136e\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T22:49:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-10T22:49:25Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T22:50:49Z is after 2025-08-24T17:21:41Z" Dec 10 22:50:49 crc kubenswrapper[4791]: I1210 22:50:49.617147 4791 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7a9a3709-32fb-48b4-97f4-e1a19282a085\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T22:49:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T22:49:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T22:49:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T22:49:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T22:49:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://511bd056768a15ea0f650a2b1a6e87e3c71dfc75696222a54f7376227ad480c6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-crio\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T22:49:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kube\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"var-lib-kubelet\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8c489105d0dfb302732d4f5bd2fc9d000967a071802735b778d49bdbc428b264\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://8c489105d0dfb302732d4f5bd2fc9d000967a071802735b778d49bdbc428b264\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-10T22:49:27Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-10T22:49:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var\\\",\\\"name\\\":\\\"var-lib-kubelet\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-10T22:49:25Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"kube-rbac-proxy-crio-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T22:50:49Z is after 2025-08-24T17:21:41Z" Dec 10 22:50:49 crc kubenswrapper[4791]: I1210 22:50:49.631934 4791 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-10T22:49:52Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-10T22:49:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-10T22:49:52Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3734b110802828632d44263ed0a43cabc4b710833f0be120d72db0940fe783e2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T22:49:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://794803ac9be3c52509e6dcb10854d8864e1a9e406762e57c7971b07c4da1220e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T22:49:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T22:50:49Z is after 2025-08-24T17:21:41Z" Dec 10 22:50:49 crc kubenswrapper[4791]: I1210 22:50:49.640774 4791 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-vczjq" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b9f00f25-3fa1-4b63-9710-fd352224b01b\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T22:49:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T22:49:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T22:49:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T22:49:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4698cf4c6b55b41e628448d37c2eb847c2777d0efbd93afcbf367d5e78623b7e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T22:49:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lrrpv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-10T22:49:50Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-vczjq\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T22:50:49Z is after 2025-08-24T17:21:41Z" Dec 10 22:50:49 crc kubenswrapper[4791]: I1210 22:50:49.653272 4791 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-4nwkq" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"672aa28c-8169-49ed-87b8-21187d13a80c\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T22:49:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T22:49:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T22:50:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T22:50:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f337f881ffd71f4a2ee19e92bb0b92c76cbb9b8352511f741d0c87e7da4cdd18\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a2f2a62315529e148a10a33ccb241e9417fbde0fe42afb61cb7ee2438c74636f\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-12-10T22:50:38Z\\\",\\\"message\\\":\\\"2025-12-10T22:49:52+00:00 [cnibincopy] Successfully copied files in /usr/src/multus-cni/rhel9/bin/ to /host/opt/cni/bin/upgrade_e48d9423-4bca-43c4-9eb9-b55a19a5ffd6\\\\n2025-12-10T22:49:52+00:00 [cnibincopy] Successfully moved files in /host/opt/cni/bin/upgrade_e48d9423-4bca-43c4-9eb9-b55a19a5ffd6 to /host/opt/cni/bin/\\\\n2025-12-10T22:49:52Z [verbose] multus-daemon started\\\\n2025-12-10T22:49:52Z [verbose] Readiness Indicator file check\\\\n2025-12-10T22:50:37Z [error] have you checked that your default network is ready? still waiting for readinessindicatorfile @ /host/run/multus/cni/net.d/10-ovn-kubernetes.conf. pollimmediate error: timed out waiting for the condition\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-10T22:49:52Z\\\"}},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T22:50:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xl62j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-10T22:49:51Z\\\"}}\" for pod \"openshift-multus\"/\"multus-4nwkq\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T22:50:49Z is after 2025-08-24T17:21:41Z" Dec 10 22:50:49 crc kubenswrapper[4791]: I1210 22:50:49.664887 4791 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-10T22:49:52Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-10T22:49:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-10T22:49:52Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://20651eb0c16e2c71222f56da1ff755cb0e7b084886acbbb24f0ccaf57041ed85\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T22:49:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T22:50:49Z is after 2025-08-24T17:21:41Z" Dec 10 22:50:49 crc kubenswrapper[4791]: I1210 22:50:49.676616 4791 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-10T22:49:50Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-10T22:49:50Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T22:50:49Z is after 2025-08-24T17:21:41Z" Dec 10 22:50:49 crc kubenswrapper[4791]: I1210 22:50:49.688679 4791 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-10T22:49:50Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-10T22:49:50Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T22:50:49Z is after 2025-08-24T17:21:41Z" Dec 10 22:50:49 crc kubenswrapper[4791]: I1210 22:50:49.692289 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 22:50:49 crc kubenswrapper[4791]: I1210 22:50:49.692364 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 22:50:49 crc kubenswrapper[4791]: I1210 22:50:49.692375 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 22:50:49 crc kubenswrapper[4791]: I1210 22:50:49.692389 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 22:50:49 crc kubenswrapper[4791]: I1210 22:50:49.692399 4791 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T22:50:49Z","lastTransitionTime":"2025-12-10T22:50:49Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 22:50:49 crc kubenswrapper[4791]: I1210 22:50:49.698933 4791 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-tchxk" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"c0ab74e0-42a9-4f42-ac80-16030cf4ffff\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T22:49:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T22:49:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T22:49:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T22:49:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://04e0fd1b53256b7a938e8ccab93e2a4eb73dd1a59e0b9799e09c55e511baf7fe\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T22:49:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zn2qt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-10T22:49:50Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-tchxk\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T22:50:49Z is after 2025-08-24T17:21:41Z" Dec 10 22:50:49 crc kubenswrapper[4791]: I1210 22:50:49.710683 4791 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-rw584" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ff9beb90-69b0-4732-bf37-0b81f58ecc98\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T22:49:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T22:49:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T22:49:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T22:49:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ddb9df501dc6e8e5561fbc2aefd431037a3f28f73bf8e82a6598d261a4224807\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T22:49:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fksqn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ce67e54b696bb53ffbbe48c6ce71476b635e0868993e9c427b22f904891433ee\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ce67e54b696bb53ffbbe48c6ce71476b635e0868993e9c427b22f904891433ee\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-10T22:49:52Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-10T22:49:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fksqn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://818d89bfffb8a110670152c8480684665ea8db578e9768f524fbbe6a92f27c22\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://818d89bfffb8a110670152c8480684665ea8db578e9768f524fbbe6a92f27c22\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-10T22:49:53Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-10T22:49:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fksqn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://9aa9c06e894bb2a2ed3738defaeff68a0a0458ead48177f36b90770944c211a0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://9aa9c06e894bb2a2ed3738defaeff68a0a0458ead48177f36b90770944c211a0\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-10T22:49:54Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-10T22:49:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fksqn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://dfd6d03336dcea7529419a30e8d70d84e2290a0fa5b0e46d09ad341c190638d6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://dfd6d03336dcea7529419a30e8d70d84e2290a0fa5b0e46d09ad341c190638d6\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-10T22:49:55Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-10T22:49:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fksqn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b86738401ece6fbc3083782720c66a515850ed78d4054f34fcabab7ff525b85f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b86738401ece6fbc3083782720c66a515850ed78d4054f34fcabab7ff525b85f\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-10T22:49:56Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-10T22:49:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fksqn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://18b326f572a899ddd10197a745bf9264d4ddd975690ad1cfb0ec576ccd5cf102\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://18b326f572a899ddd10197a745bf9264d4ddd975690ad1cfb0ec576ccd5cf102\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-10T22:49:57Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-10T22:49:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fksqn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-10T22:49:51Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-rw584\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T22:50:49Z is after 2025-08-24T17:21:41Z" Dec 10 22:50:49 crc kubenswrapper[4791]: I1210 22:50:49.720201 4791 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-b6kf6" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b4527e71-bb58-4a39-a99c-fe614aa4c5f9\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T22:50:04Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T22:50:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T22:50:04Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T22:50:04Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ngpr4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ngpr4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-10T22:50:04Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-b6kf6\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T22:50:49Z is after 2025-08-24T17:21:41Z" Dec 10 22:50:49 crc kubenswrapper[4791]: I1210 22:50:49.730599 4791 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f8349fea-c49b-49a1-b5ee-032c5dffe021\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T22:49:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T22:49:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T22:49:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T22:49:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T22:49:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://acafe12186df99d1059a31260de0eafce35ff7771bcb8f5c6753b72665d6f60b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T22:49:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ea4c346507862a8c12756a3061d6e25e788a8d79f4621b62f81f4322185c82a8\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T22:49:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://2e6e324ec6f4f7bb69fdfab4ea3ac1aa55e480dd8e2a8984b0fe5b4c9f66636e\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T22:49:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://1b2e11ff2bf8789fa1d0d23ae648b3e6ef653106e5bc4bbb5dd6447f7cd6f05d\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d6aa7f8c5d908d90008f2df49860ac8a3480136da567e3d85f86135d779204f9\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-12-10T22:49:44Z\\\",\\\"message\\\":\\\"W1210 22:49:33.845733 1 cmd.go:257] Using insecure, self-signed certificates\\\\nI1210 22:49:33.846108 1 crypto.go:601] Generating new CA for check-endpoints-signer@1765406973 cert, and key in /tmp/serving-cert-2585701456/serving-signer.crt, /tmp/serving-cert-2585701456/serving-signer.key\\\\nI1210 22:49:34.529748 1 observer_polling.go:159] Starting file observer\\\\nW1210 22:49:34.532225 1 builder.go:272] unable to get owner reference (falling back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": dial tcp [::1]:6443: connect: connection refused\\\\nI1210 22:49:34.532599 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1210 22:49:34.533977 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-2585701456/tls.crt::/tmp/serving-cert-2585701456/tls.key\\\\\\\"\\\\nF1210 22:49:44.774710 1 cmd.go:182] error initializing delegating authentication: unable to load configmap based request-header-client-ca-file: Get \\\\\\\"https://localhost:6443/api/v1/namespaces/kube-system/configmaps/extension-apiserver-authentication\\\\\\\": net/http: TLS handshake timeout\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-10T22:49:33Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T22:49:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://0c4d036ab269f62561fe7ef1586012cb3a16e3583a08b9488a0616ec00c8e0ae\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T22:49:32Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c5856d6b26c19fda13b5c9ac2892a7e2f9c6f79feb8173ecfd9b33b673327d67\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c5856d6b26c19fda13b5c9ac2892a7e2f9c6f79feb8173ecfd9b33b673327d67\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-10T22:49:27Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-10T22:49:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-10T22:49:25Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T22:50:49Z is after 2025-08-24T17:21:41Z" Dec 10 22:50:49 crc kubenswrapper[4791]: I1210 22:50:49.740091 4791 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"02a949d8-38e2-4be3-989f-40a015407140\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T22:49:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T22:49:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T22:50:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T22:50:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T22:49:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://cd4158bc59b30f0f1716bf43c1ae7c9916a5b7d7164d074987fc1f19ba49b778\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T22:49:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://656065fd3cbdc6a25dab5d06c7baa00af16d8f62fc5da687efe9c6cc91945de0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T22:49:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://90ed7fff816b1772ba9ecd95587a59cbed7c30ac302fff7e7c5de934c912bcf4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T22:49:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9b7a5502803eaad9f76f5ffa91b29fbb2657402065a03a4d6433c096aac7264f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"wait-for-host-port\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://9b7a5502803eaad9f76f5ffa91b29fbb2657402065a03a4d6433c096aac7264f\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-10T22:49:27Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-10T22:49:27Z\\\"}}}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-10T22:49:25Z\\\"}}\" for pod \"openshift-kube-scheduler\"/\"openshift-kube-scheduler-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T22:50:49Z is after 2025-08-24T17:21:41Z" Dec 10 22:50:49 crc kubenswrapper[4791]: I1210 22:50:49.758069 4791 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"90250e13-6c63-443a-a5e5-6e4e341a2289\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T22:49:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T22:49:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T22:49:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T22:49:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T22:49:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ba54f95b1e00843b854db9a10fe6897592159ccacf30ed0f76002093a13eedd3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T22:49:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5aa8ffa518f3a68b60b275f8df4851b2564aa4984b61efdc9470b73072133e43\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T22:49:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://16f989503edc78d9451638e2cf6cf4b19a2cf2b8809b10b237541cb829303091\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T22:49:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://82556824d63ce6e46df14c45689ff1f81ebf729abe902061815c040b323898e9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T22:49:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://61db531940fa6f2effdae5a005b4c91bd2698bd5d1bff9588c171a65ad550eca\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T22:49:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0f65c864d885319ce2a08d61c30be24b3ffb7d23ea30c82ed7faf039cb2c006d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://0f65c864d885319ce2a08d61c30be24b3ffb7d23ea30c82ed7faf039cb2c006d\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-10T22:49:27Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-10T22:49:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e31a76d1d5b2812ba83e24e38268e25ec8bed98fc3da9adcd06bda5822b6ace1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e31a76d1d5b2812ba83e24e38268e25ec8bed98fc3da9adcd06bda5822b6ace1\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-10T22:49:30Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-10T22:49:29Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://b5b5eceae562f3986b543d41de1e977efb85a91d84c98e09159ca6d092eda691\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b5b5eceae562f3986b543d41de1e977efb85a91d84c98e09159ca6d092eda691\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-10T22:49:31Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-10T22:49:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-10T22:49:25Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T22:50:49Z is after 2025-08-24T17:21:41Z" Dec 10 22:50:49 crc kubenswrapper[4791]: I1210 22:50:49.770297 4791 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-10T22:49:50Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-10T22:49:50Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T22:50:49Z is after 2025-08-24T17:21:41Z" Dec 10 22:50:49 crc kubenswrapper[4791]: I1210 22:50:49.779889 4791 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-10T22:49:54Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-10T22:49:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-10T22:49:54Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://62bfaf5d7a6d1b452cdaf88086447d0c8cb7306e94434e46752be5475124d56a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T22:49:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T22:50:49Z is after 2025-08-24T17:21:41Z" Dec 10 22:50:49 crc kubenswrapper[4791]: I1210 22:50:49.788779 4791 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-5rb5l" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ba35653c-6e06-4cee-a4d6-137764090d18\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T22:49:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T22:49:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T22:49:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T22:49:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://dbb5c875730a98ad2daf316cfafc08fcdce21ac16148154780ac9193ed405dab\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T22:49:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fvvrs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://185eb2ff44221a514878283ec873773a6c9a9073dc64e296f982e60dee1338e4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T22:49:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fvvrs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-10T22:49:51Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-5rb5l\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T22:50:49Z is after 2025-08-24T17:21:41Z" Dec 10 22:50:49 crc kubenswrapper[4791]: I1210 22:50:49.794455 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 22:50:49 crc kubenswrapper[4791]: I1210 22:50:49.794494 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 22:50:49 crc kubenswrapper[4791]: I1210 22:50:49.794507 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 22:50:49 crc kubenswrapper[4791]: I1210 22:50:49.794526 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 22:50:49 crc kubenswrapper[4791]: I1210 22:50:49.794540 4791 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T22:50:49Z","lastTransitionTime":"2025-12-10T22:50:49Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 22:50:49 crc kubenswrapper[4791]: I1210 22:50:49.799741 4791 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-s6rk8" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5c9d0978-a6f9-49c5-b377-6619b4e3ff52\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T22:50:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T22:50:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T22:50:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T22:50:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8e55baaf460c794ab196e237d51d12de059aa038617dc2d5366d49c529f5d5c7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T22:50:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7vkgm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://51d1b9b086c60bae9223cb51d5a998d02790a0d91b745aab10392c23e617eb9f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T22:50:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7vkgm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-10T22:50:03Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-s6rk8\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T22:50:49Z is after 2025-08-24T17:21:41Z" Dec 10 22:50:49 crc kubenswrapper[4791]: I1210 22:50:49.896730 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 22:50:49 crc kubenswrapper[4791]: I1210 22:50:49.896768 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 22:50:49 crc kubenswrapper[4791]: I1210 22:50:49.896778 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 22:50:49 crc kubenswrapper[4791]: I1210 22:50:49.896790 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 22:50:49 crc kubenswrapper[4791]: I1210 22:50:49.896800 4791 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T22:50:49Z","lastTransitionTime":"2025-12-10T22:50:49Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 22:50:50 crc kubenswrapper[4791]: I1210 22:50:50.009736 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 22:50:50 crc kubenswrapper[4791]: I1210 22:50:50.009804 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 22:50:50 crc kubenswrapper[4791]: I1210 22:50:50.009822 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 22:50:50 crc kubenswrapper[4791]: I1210 22:50:50.010214 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 22:50:50 crc kubenswrapper[4791]: I1210 22:50:50.010266 4791 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T22:50:50Z","lastTransitionTime":"2025-12-10T22:50:50Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 22:50:50 crc kubenswrapper[4791]: I1210 22:50:50.113064 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 22:50:50 crc kubenswrapper[4791]: I1210 22:50:50.113104 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 22:50:50 crc kubenswrapper[4791]: I1210 22:50:50.113113 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 22:50:50 crc kubenswrapper[4791]: I1210 22:50:50.113126 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 22:50:50 crc kubenswrapper[4791]: I1210 22:50:50.113135 4791 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T22:50:50Z","lastTransitionTime":"2025-12-10T22:50:50Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 22:50:50 crc kubenswrapper[4791]: I1210 22:50:50.216540 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 22:50:50 crc kubenswrapper[4791]: I1210 22:50:50.216617 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 22:50:50 crc kubenswrapper[4791]: I1210 22:50:50.216649 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 22:50:50 crc kubenswrapper[4791]: I1210 22:50:50.216680 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 22:50:50 crc kubenswrapper[4791]: I1210 22:50:50.216701 4791 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T22:50:50Z","lastTransitionTime":"2025-12-10T22:50:50Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 22:50:50 crc kubenswrapper[4791]: I1210 22:50:50.319488 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 22:50:50 crc kubenswrapper[4791]: I1210 22:50:50.319532 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 22:50:50 crc kubenswrapper[4791]: I1210 22:50:50.319544 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 22:50:50 crc kubenswrapper[4791]: I1210 22:50:50.319559 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 22:50:50 crc kubenswrapper[4791]: I1210 22:50:50.319571 4791 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T22:50:50Z","lastTransitionTime":"2025-12-10T22:50:50Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 22:50:50 crc kubenswrapper[4791]: I1210 22:50:50.421945 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 22:50:50 crc kubenswrapper[4791]: I1210 22:50:50.422047 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 22:50:50 crc kubenswrapper[4791]: I1210 22:50:50.422063 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 22:50:50 crc kubenswrapper[4791]: I1210 22:50:50.422081 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 22:50:50 crc kubenswrapper[4791]: I1210 22:50:50.422092 4791 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T22:50:50Z","lastTransitionTime":"2025-12-10T22:50:50Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 22:50:50 crc kubenswrapper[4791]: I1210 22:50:50.524743 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 22:50:50 crc kubenswrapper[4791]: I1210 22:50:50.524821 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 22:50:50 crc kubenswrapper[4791]: I1210 22:50:50.524838 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 22:50:50 crc kubenswrapper[4791]: I1210 22:50:50.524857 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 22:50:50 crc kubenswrapper[4791]: I1210 22:50:50.524868 4791 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T22:50:50Z","lastTransitionTime":"2025-12-10T22:50:50Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 22:50:50 crc kubenswrapper[4791]: I1210 22:50:50.626916 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 22:50:50 crc kubenswrapper[4791]: I1210 22:50:50.626978 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 22:50:50 crc kubenswrapper[4791]: I1210 22:50:50.626991 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 22:50:50 crc kubenswrapper[4791]: I1210 22:50:50.627012 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 22:50:50 crc kubenswrapper[4791]: I1210 22:50:50.627042 4791 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T22:50:50Z","lastTransitionTime":"2025-12-10T22:50:50Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 22:50:50 crc kubenswrapper[4791]: I1210 22:50:50.729620 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 22:50:50 crc kubenswrapper[4791]: I1210 22:50:50.729666 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 22:50:50 crc kubenswrapper[4791]: I1210 22:50:50.729675 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 22:50:50 crc kubenswrapper[4791]: I1210 22:50:50.729689 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 22:50:50 crc kubenswrapper[4791]: I1210 22:50:50.729698 4791 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T22:50:50Z","lastTransitionTime":"2025-12-10T22:50:50Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 22:50:50 crc kubenswrapper[4791]: I1210 22:50:50.832198 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 22:50:50 crc kubenswrapper[4791]: I1210 22:50:50.832237 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 22:50:50 crc kubenswrapper[4791]: I1210 22:50:50.832247 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 22:50:50 crc kubenswrapper[4791]: I1210 22:50:50.832265 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 22:50:50 crc kubenswrapper[4791]: I1210 22:50:50.832277 4791 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T22:50:50Z","lastTransitionTime":"2025-12-10T22:50:50Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 22:50:50 crc kubenswrapper[4791]: I1210 22:50:50.884243 4791 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 10 22:50:50 crc kubenswrapper[4791]: I1210 22:50:50.884309 4791 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 10 22:50:50 crc kubenswrapper[4791]: I1210 22:50:50.884264 4791 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 10 22:50:50 crc kubenswrapper[4791]: I1210 22:50:50.884242 4791 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-b6kf6" Dec 10 22:50:50 crc kubenswrapper[4791]: E1210 22:50:50.884406 4791 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 10 22:50:50 crc kubenswrapper[4791]: E1210 22:50:50.884462 4791 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 10 22:50:50 crc kubenswrapper[4791]: E1210 22:50:50.884555 4791 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-b6kf6" podUID="b4527e71-bb58-4a39-a99c-fe614aa4c5f9" Dec 10 22:50:50 crc kubenswrapper[4791]: E1210 22:50:50.884650 4791 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 10 22:50:50 crc kubenswrapper[4791]: I1210 22:50:50.934993 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 22:50:50 crc kubenswrapper[4791]: I1210 22:50:50.935030 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 22:50:50 crc kubenswrapper[4791]: I1210 22:50:50.935041 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 22:50:50 crc kubenswrapper[4791]: I1210 22:50:50.935057 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 22:50:50 crc kubenswrapper[4791]: I1210 22:50:50.935069 4791 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T22:50:50Z","lastTransitionTime":"2025-12-10T22:50:50Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 22:50:51 crc kubenswrapper[4791]: I1210 22:50:51.037866 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 22:50:51 crc kubenswrapper[4791]: I1210 22:50:51.037939 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 22:50:51 crc kubenswrapper[4791]: I1210 22:50:51.037962 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 22:50:51 crc kubenswrapper[4791]: I1210 22:50:51.037993 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 22:50:51 crc kubenswrapper[4791]: I1210 22:50:51.038015 4791 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T22:50:51Z","lastTransitionTime":"2025-12-10T22:50:51Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 22:50:51 crc kubenswrapper[4791]: I1210 22:50:51.140288 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 22:50:51 crc kubenswrapper[4791]: I1210 22:50:51.140452 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 22:50:51 crc kubenswrapper[4791]: I1210 22:50:51.140480 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 22:50:51 crc kubenswrapper[4791]: I1210 22:50:51.140503 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 22:50:51 crc kubenswrapper[4791]: I1210 22:50:51.140520 4791 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T22:50:51Z","lastTransitionTime":"2025-12-10T22:50:51Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 22:50:51 crc kubenswrapper[4791]: I1210 22:50:51.243038 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 22:50:51 crc kubenswrapper[4791]: I1210 22:50:51.243079 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 22:50:51 crc kubenswrapper[4791]: I1210 22:50:51.243090 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 22:50:51 crc kubenswrapper[4791]: I1210 22:50:51.243110 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 22:50:51 crc kubenswrapper[4791]: I1210 22:50:51.243122 4791 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T22:50:51Z","lastTransitionTime":"2025-12-10T22:50:51Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 22:50:51 crc kubenswrapper[4791]: I1210 22:50:51.346052 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 22:50:51 crc kubenswrapper[4791]: I1210 22:50:51.346106 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 22:50:51 crc kubenswrapper[4791]: I1210 22:50:51.346118 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 22:50:51 crc kubenswrapper[4791]: I1210 22:50:51.346141 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 22:50:51 crc kubenswrapper[4791]: I1210 22:50:51.346150 4791 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T22:50:51Z","lastTransitionTime":"2025-12-10T22:50:51Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 22:50:51 crc kubenswrapper[4791]: I1210 22:50:51.449431 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 22:50:51 crc kubenswrapper[4791]: I1210 22:50:51.449515 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 22:50:51 crc kubenswrapper[4791]: I1210 22:50:51.449540 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 22:50:51 crc kubenswrapper[4791]: I1210 22:50:51.449579 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 22:50:51 crc kubenswrapper[4791]: I1210 22:50:51.449606 4791 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T22:50:51Z","lastTransitionTime":"2025-12-10T22:50:51Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 22:50:51 crc kubenswrapper[4791]: I1210 22:50:51.551824 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 22:50:51 crc kubenswrapper[4791]: I1210 22:50:51.551861 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 22:50:51 crc kubenswrapper[4791]: I1210 22:50:51.551869 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 22:50:51 crc kubenswrapper[4791]: I1210 22:50:51.551882 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 22:50:51 crc kubenswrapper[4791]: I1210 22:50:51.551890 4791 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T22:50:51Z","lastTransitionTime":"2025-12-10T22:50:51Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 22:50:51 crc kubenswrapper[4791]: I1210 22:50:51.653950 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 22:50:51 crc kubenswrapper[4791]: I1210 22:50:51.654011 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 22:50:51 crc kubenswrapper[4791]: I1210 22:50:51.654023 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 22:50:51 crc kubenswrapper[4791]: I1210 22:50:51.654039 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 22:50:51 crc kubenswrapper[4791]: I1210 22:50:51.654048 4791 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T22:50:51Z","lastTransitionTime":"2025-12-10T22:50:51Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 22:50:51 crc kubenswrapper[4791]: I1210 22:50:51.756612 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 22:50:51 crc kubenswrapper[4791]: I1210 22:50:51.756658 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 22:50:51 crc kubenswrapper[4791]: I1210 22:50:51.756674 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 22:50:51 crc kubenswrapper[4791]: I1210 22:50:51.756695 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 22:50:51 crc kubenswrapper[4791]: I1210 22:50:51.756710 4791 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T22:50:51Z","lastTransitionTime":"2025-12-10T22:50:51Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 22:50:51 crc kubenswrapper[4791]: I1210 22:50:51.859150 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 22:50:51 crc kubenswrapper[4791]: I1210 22:50:51.859210 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 22:50:51 crc kubenswrapper[4791]: I1210 22:50:51.859222 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 22:50:51 crc kubenswrapper[4791]: I1210 22:50:51.859239 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 22:50:51 crc kubenswrapper[4791]: I1210 22:50:51.859253 4791 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T22:50:51Z","lastTransitionTime":"2025-12-10T22:50:51Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 22:50:51 crc kubenswrapper[4791]: I1210 22:50:51.965261 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 22:50:51 crc kubenswrapper[4791]: I1210 22:50:51.965317 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 22:50:51 crc kubenswrapper[4791]: I1210 22:50:51.965333 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 22:50:51 crc kubenswrapper[4791]: I1210 22:50:51.965396 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 22:50:51 crc kubenswrapper[4791]: I1210 22:50:51.965415 4791 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T22:50:51Z","lastTransitionTime":"2025-12-10T22:50:51Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 22:50:52 crc kubenswrapper[4791]: I1210 22:50:52.067425 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 22:50:52 crc kubenswrapper[4791]: I1210 22:50:52.067501 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 22:50:52 crc kubenswrapper[4791]: I1210 22:50:52.067526 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 22:50:52 crc kubenswrapper[4791]: I1210 22:50:52.067556 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 22:50:52 crc kubenswrapper[4791]: I1210 22:50:52.067584 4791 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T22:50:52Z","lastTransitionTime":"2025-12-10T22:50:52Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 22:50:52 crc kubenswrapper[4791]: I1210 22:50:52.170142 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 22:50:52 crc kubenswrapper[4791]: I1210 22:50:52.170181 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 22:50:52 crc kubenswrapper[4791]: I1210 22:50:52.170192 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 22:50:52 crc kubenswrapper[4791]: I1210 22:50:52.170207 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 22:50:52 crc kubenswrapper[4791]: I1210 22:50:52.170217 4791 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T22:50:52Z","lastTransitionTime":"2025-12-10T22:50:52Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 22:50:52 crc kubenswrapper[4791]: I1210 22:50:52.280264 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 22:50:52 crc kubenswrapper[4791]: I1210 22:50:52.280317 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 22:50:52 crc kubenswrapper[4791]: I1210 22:50:52.280331 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 22:50:52 crc kubenswrapper[4791]: I1210 22:50:52.280368 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 22:50:52 crc kubenswrapper[4791]: I1210 22:50:52.280401 4791 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T22:50:52Z","lastTransitionTime":"2025-12-10T22:50:52Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 22:50:52 crc kubenswrapper[4791]: I1210 22:50:52.382577 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 22:50:52 crc kubenswrapper[4791]: I1210 22:50:52.382648 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 22:50:52 crc kubenswrapper[4791]: I1210 22:50:52.382672 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 22:50:52 crc kubenswrapper[4791]: I1210 22:50:52.382704 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 22:50:52 crc kubenswrapper[4791]: I1210 22:50:52.382727 4791 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T22:50:52Z","lastTransitionTime":"2025-12-10T22:50:52Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 22:50:52 crc kubenswrapper[4791]: I1210 22:50:52.484687 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 22:50:52 crc kubenswrapper[4791]: I1210 22:50:52.484766 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 22:50:52 crc kubenswrapper[4791]: I1210 22:50:52.484781 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 22:50:52 crc kubenswrapper[4791]: I1210 22:50:52.484800 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 22:50:52 crc kubenswrapper[4791]: I1210 22:50:52.484824 4791 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T22:50:52Z","lastTransitionTime":"2025-12-10T22:50:52Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 22:50:52 crc kubenswrapper[4791]: I1210 22:50:52.586606 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 22:50:52 crc kubenswrapper[4791]: I1210 22:50:52.586647 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 22:50:52 crc kubenswrapper[4791]: I1210 22:50:52.586658 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 22:50:52 crc kubenswrapper[4791]: I1210 22:50:52.586674 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 22:50:52 crc kubenswrapper[4791]: I1210 22:50:52.586685 4791 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T22:50:52Z","lastTransitionTime":"2025-12-10T22:50:52Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 22:50:52 crc kubenswrapper[4791]: I1210 22:50:52.689125 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 22:50:52 crc kubenswrapper[4791]: I1210 22:50:52.689167 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 22:50:52 crc kubenswrapper[4791]: I1210 22:50:52.689179 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 22:50:52 crc kubenswrapper[4791]: I1210 22:50:52.689196 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 22:50:52 crc kubenswrapper[4791]: I1210 22:50:52.689208 4791 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T22:50:52Z","lastTransitionTime":"2025-12-10T22:50:52Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 22:50:52 crc kubenswrapper[4791]: I1210 22:50:52.791770 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 22:50:52 crc kubenswrapper[4791]: I1210 22:50:52.791831 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 22:50:52 crc kubenswrapper[4791]: I1210 22:50:52.791849 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 22:50:52 crc kubenswrapper[4791]: I1210 22:50:52.791875 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 22:50:52 crc kubenswrapper[4791]: I1210 22:50:52.791895 4791 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T22:50:52Z","lastTransitionTime":"2025-12-10T22:50:52Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 22:50:52 crc kubenswrapper[4791]: I1210 22:50:52.894485 4791 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 10 22:50:52 crc kubenswrapper[4791]: I1210 22:50:52.894515 4791 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 10 22:50:52 crc kubenswrapper[4791]: I1210 22:50:52.894557 4791 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-b6kf6" Dec 10 22:50:52 crc kubenswrapper[4791]: I1210 22:50:52.894581 4791 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 10 22:50:52 crc kubenswrapper[4791]: E1210 22:50:52.894676 4791 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 10 22:50:52 crc kubenswrapper[4791]: I1210 22:50:52.894705 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 22:50:52 crc kubenswrapper[4791]: I1210 22:50:52.894762 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 22:50:52 crc kubenswrapper[4791]: I1210 22:50:52.894787 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 22:50:52 crc kubenswrapper[4791]: I1210 22:50:52.894817 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 22:50:52 crc kubenswrapper[4791]: E1210 22:50:52.894773 4791 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 10 22:50:52 crc kubenswrapper[4791]: I1210 22:50:52.894840 4791 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T22:50:52Z","lastTransitionTime":"2025-12-10T22:50:52Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 22:50:52 crc kubenswrapper[4791]: E1210 22:50:52.894912 4791 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-b6kf6" podUID="b4527e71-bb58-4a39-a99c-fe614aa4c5f9" Dec 10 22:50:52 crc kubenswrapper[4791]: E1210 22:50:52.894855 4791 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 10 22:50:52 crc kubenswrapper[4791]: I1210 22:50:52.997684 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 22:50:52 crc kubenswrapper[4791]: I1210 22:50:52.997741 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 22:50:52 crc kubenswrapper[4791]: I1210 22:50:52.997761 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 22:50:52 crc kubenswrapper[4791]: I1210 22:50:52.997781 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 22:50:52 crc kubenswrapper[4791]: I1210 22:50:52.997814 4791 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T22:50:52Z","lastTransitionTime":"2025-12-10T22:50:52Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 22:50:53 crc kubenswrapper[4791]: I1210 22:50:53.099836 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 22:50:53 crc kubenswrapper[4791]: I1210 22:50:53.099873 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 22:50:53 crc kubenswrapper[4791]: I1210 22:50:53.099884 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 22:50:53 crc kubenswrapper[4791]: I1210 22:50:53.099901 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 22:50:53 crc kubenswrapper[4791]: I1210 22:50:53.099912 4791 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T22:50:53Z","lastTransitionTime":"2025-12-10T22:50:53Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 22:50:53 crc kubenswrapper[4791]: I1210 22:50:53.202060 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 22:50:53 crc kubenswrapper[4791]: I1210 22:50:53.202108 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 22:50:53 crc kubenswrapper[4791]: I1210 22:50:53.202120 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 22:50:53 crc kubenswrapper[4791]: I1210 22:50:53.202142 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 22:50:53 crc kubenswrapper[4791]: I1210 22:50:53.202158 4791 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T22:50:53Z","lastTransitionTime":"2025-12-10T22:50:53Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 22:50:53 crc kubenswrapper[4791]: I1210 22:50:53.305632 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 22:50:53 crc kubenswrapper[4791]: I1210 22:50:53.305695 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 22:50:53 crc kubenswrapper[4791]: I1210 22:50:53.305709 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 22:50:53 crc kubenswrapper[4791]: I1210 22:50:53.305731 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 22:50:53 crc kubenswrapper[4791]: I1210 22:50:53.305745 4791 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T22:50:53Z","lastTransitionTime":"2025-12-10T22:50:53Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 22:50:53 crc kubenswrapper[4791]: I1210 22:50:53.408678 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 22:50:53 crc kubenswrapper[4791]: I1210 22:50:53.408748 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 22:50:53 crc kubenswrapper[4791]: I1210 22:50:53.408766 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 22:50:53 crc kubenswrapper[4791]: I1210 22:50:53.408792 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 22:50:53 crc kubenswrapper[4791]: I1210 22:50:53.408812 4791 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T22:50:53Z","lastTransitionTime":"2025-12-10T22:50:53Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 22:50:53 crc kubenswrapper[4791]: I1210 22:50:53.510656 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 22:50:53 crc kubenswrapper[4791]: I1210 22:50:53.510714 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 22:50:53 crc kubenswrapper[4791]: I1210 22:50:53.510726 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 22:50:53 crc kubenswrapper[4791]: I1210 22:50:53.510746 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 22:50:53 crc kubenswrapper[4791]: I1210 22:50:53.510761 4791 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T22:50:53Z","lastTransitionTime":"2025-12-10T22:50:53Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 22:50:53 crc kubenswrapper[4791]: I1210 22:50:53.613132 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 22:50:53 crc kubenswrapper[4791]: I1210 22:50:53.613164 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 22:50:53 crc kubenswrapper[4791]: I1210 22:50:53.613172 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 22:50:53 crc kubenswrapper[4791]: I1210 22:50:53.613186 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 22:50:53 crc kubenswrapper[4791]: I1210 22:50:53.613194 4791 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T22:50:53Z","lastTransitionTime":"2025-12-10T22:50:53Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 22:50:53 crc kubenswrapper[4791]: I1210 22:50:53.715177 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 22:50:53 crc kubenswrapper[4791]: I1210 22:50:53.715227 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 22:50:53 crc kubenswrapper[4791]: I1210 22:50:53.715245 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 22:50:53 crc kubenswrapper[4791]: I1210 22:50:53.715266 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 22:50:53 crc kubenswrapper[4791]: I1210 22:50:53.715282 4791 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T22:50:53Z","lastTransitionTime":"2025-12-10T22:50:53Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 22:50:53 crc kubenswrapper[4791]: I1210 22:50:53.818273 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 22:50:53 crc kubenswrapper[4791]: I1210 22:50:53.818361 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 22:50:53 crc kubenswrapper[4791]: I1210 22:50:53.818388 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 22:50:53 crc kubenswrapper[4791]: I1210 22:50:53.818410 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 22:50:53 crc kubenswrapper[4791]: I1210 22:50:53.818422 4791 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T22:50:53Z","lastTransitionTime":"2025-12-10T22:50:53Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 22:50:53 crc kubenswrapper[4791]: I1210 22:50:53.920913 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 22:50:53 crc kubenswrapper[4791]: I1210 22:50:53.920988 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 22:50:53 crc kubenswrapper[4791]: I1210 22:50:53.921007 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 22:50:53 crc kubenswrapper[4791]: I1210 22:50:53.921035 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 22:50:53 crc kubenswrapper[4791]: I1210 22:50:53.921056 4791 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T22:50:53Z","lastTransitionTime":"2025-12-10T22:50:53Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 22:50:54 crc kubenswrapper[4791]: I1210 22:50:54.023829 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 22:50:54 crc kubenswrapper[4791]: I1210 22:50:54.023866 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 22:50:54 crc kubenswrapper[4791]: I1210 22:50:54.023876 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 22:50:54 crc kubenswrapper[4791]: I1210 22:50:54.023890 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 22:50:54 crc kubenswrapper[4791]: I1210 22:50:54.023899 4791 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T22:50:54Z","lastTransitionTime":"2025-12-10T22:50:54Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 22:50:54 crc kubenswrapper[4791]: I1210 22:50:54.127511 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 22:50:54 crc kubenswrapper[4791]: I1210 22:50:54.127582 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 22:50:54 crc kubenswrapper[4791]: I1210 22:50:54.127604 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 22:50:54 crc kubenswrapper[4791]: I1210 22:50:54.127631 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 22:50:54 crc kubenswrapper[4791]: I1210 22:50:54.127650 4791 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T22:50:54Z","lastTransitionTime":"2025-12-10T22:50:54Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 22:50:54 crc kubenswrapper[4791]: I1210 22:50:54.230796 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 22:50:54 crc kubenswrapper[4791]: I1210 22:50:54.231017 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 22:50:54 crc kubenswrapper[4791]: I1210 22:50:54.231051 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 22:50:54 crc kubenswrapper[4791]: I1210 22:50:54.231094 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 22:50:54 crc kubenswrapper[4791]: I1210 22:50:54.231116 4791 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T22:50:54Z","lastTransitionTime":"2025-12-10T22:50:54Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 22:50:54 crc kubenswrapper[4791]: I1210 22:50:54.333709 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 22:50:54 crc kubenswrapper[4791]: I1210 22:50:54.333798 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 22:50:54 crc kubenswrapper[4791]: I1210 22:50:54.333817 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 22:50:54 crc kubenswrapper[4791]: I1210 22:50:54.333841 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 22:50:54 crc kubenswrapper[4791]: I1210 22:50:54.334148 4791 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T22:50:54Z","lastTransitionTime":"2025-12-10T22:50:54Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 22:50:54 crc kubenswrapper[4791]: I1210 22:50:54.439991 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 22:50:54 crc kubenswrapper[4791]: I1210 22:50:54.440049 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 22:50:54 crc kubenswrapper[4791]: I1210 22:50:54.440065 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 22:50:54 crc kubenswrapper[4791]: I1210 22:50:54.440086 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 22:50:54 crc kubenswrapper[4791]: I1210 22:50:54.440102 4791 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T22:50:54Z","lastTransitionTime":"2025-12-10T22:50:54Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 22:50:54 crc kubenswrapper[4791]: I1210 22:50:54.542277 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 22:50:54 crc kubenswrapper[4791]: I1210 22:50:54.542385 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 22:50:54 crc kubenswrapper[4791]: I1210 22:50:54.542398 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 22:50:54 crc kubenswrapper[4791]: I1210 22:50:54.542423 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 22:50:54 crc kubenswrapper[4791]: I1210 22:50:54.542437 4791 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T22:50:54Z","lastTransitionTime":"2025-12-10T22:50:54Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 22:50:54 crc kubenswrapper[4791]: I1210 22:50:54.645547 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 22:50:54 crc kubenswrapper[4791]: I1210 22:50:54.645616 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 22:50:54 crc kubenswrapper[4791]: I1210 22:50:54.645631 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 22:50:54 crc kubenswrapper[4791]: I1210 22:50:54.645654 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 22:50:54 crc kubenswrapper[4791]: I1210 22:50:54.645670 4791 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T22:50:54Z","lastTransitionTime":"2025-12-10T22:50:54Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 22:50:54 crc kubenswrapper[4791]: I1210 22:50:54.748456 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 22:50:54 crc kubenswrapper[4791]: I1210 22:50:54.748510 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 22:50:54 crc kubenswrapper[4791]: I1210 22:50:54.748527 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 22:50:54 crc kubenswrapper[4791]: I1210 22:50:54.748550 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 22:50:54 crc kubenswrapper[4791]: I1210 22:50:54.748567 4791 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T22:50:54Z","lastTransitionTime":"2025-12-10T22:50:54Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 22:50:54 crc kubenswrapper[4791]: I1210 22:50:54.817466 4791 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 10 22:50:54 crc kubenswrapper[4791]: E1210 22:50:54.817595 4791 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-10 22:51:58.817575019 +0000 UTC m=+153.247192642 (durationBeforeRetry 1m4s). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 10 22:50:54 crc kubenswrapper[4791]: I1210 22:50:54.851851 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 22:50:54 crc kubenswrapper[4791]: I1210 22:50:54.851965 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 22:50:54 crc kubenswrapper[4791]: I1210 22:50:54.851984 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 22:50:54 crc kubenswrapper[4791]: I1210 22:50:54.852007 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 22:50:54 crc kubenswrapper[4791]: I1210 22:50:54.852022 4791 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T22:50:54Z","lastTransitionTime":"2025-12-10T22:50:54Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 22:50:54 crc kubenswrapper[4791]: I1210 22:50:54.884655 4791 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-b6kf6" Dec 10 22:50:54 crc kubenswrapper[4791]: I1210 22:50:54.884706 4791 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 10 22:50:54 crc kubenswrapper[4791]: I1210 22:50:54.884743 4791 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 10 22:50:54 crc kubenswrapper[4791]: I1210 22:50:54.884813 4791 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 10 22:50:54 crc kubenswrapper[4791]: E1210 22:50:54.884807 4791 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-b6kf6" podUID="b4527e71-bb58-4a39-a99c-fe614aa4c5f9" Dec 10 22:50:54 crc kubenswrapper[4791]: E1210 22:50:54.884901 4791 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 10 22:50:54 crc kubenswrapper[4791]: E1210 22:50:54.884986 4791 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 10 22:50:54 crc kubenswrapper[4791]: E1210 22:50:54.885126 4791 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 10 22:50:54 crc kubenswrapper[4791]: I1210 22:50:54.919183 4791 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cqllr\" (UniqueName: \"kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr\") pod \"network-check-target-xd92c\" (UID: \"3b6479f0-333b-4a96-9adf-2099afdc2447\") " pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 10 22:50:54 crc kubenswrapper[4791]: I1210 22:50:54.919242 4791 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-s2dwl\" (UniqueName: \"kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl\") pod \"network-check-source-55646444c4-trplf\" (UID: \"9d751cbb-f2e2-430d-9754-c882a5e924a5\") " pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 10 22:50:54 crc kubenswrapper[4791]: I1210 22:50:54.919276 4791 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"networking-console-plugin-cert\" (UniqueName: \"kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 10 22:50:54 crc kubenswrapper[4791]: I1210 22:50:54.919308 4791 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nginx-conf\" (UniqueName: \"kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 10 22:50:54 crc kubenswrapper[4791]: E1210 22:50:54.919436 4791 configmap.go:193] Couldn't get configMap openshift-network-console/networking-console-plugin: object "openshift-network-console"/"networking-console-plugin" not registered Dec 10 22:50:54 crc kubenswrapper[4791]: E1210 22:50:54.919490 4791 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2025-12-10 22:51:58.919473333 +0000 UTC m=+153.349090946 (durationBeforeRetry 1m4s). Error: MountVolume.SetUp failed for volume "nginx-conf" (UniqueName: "kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin" not registered Dec 10 22:50:54 crc kubenswrapper[4791]: E1210 22:50:54.919719 4791 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Dec 10 22:50:54 crc kubenswrapper[4791]: E1210 22:50:54.919750 4791 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Dec 10 22:50:54 crc kubenswrapper[4791]: E1210 22:50:54.919762 4791 projected.go:194] Error preparing data for projected volume kube-api-access-cqllr for pod openshift-network-diagnostics/network-check-target-xd92c: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Dec 10 22:50:54 crc kubenswrapper[4791]: E1210 22:50:54.919792 4791 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr podName:3b6479f0-333b-4a96-9adf-2099afdc2447 nodeName:}" failed. No retries permitted until 2025-12-10 22:51:58.919782561 +0000 UTC m=+153.349400184 (durationBeforeRetry 1m4s). Error: MountVolume.SetUp failed for volume "kube-api-access-cqllr" (UniqueName: "kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr") pod "network-check-target-xd92c" (UID: "3b6479f0-333b-4a96-9adf-2099afdc2447") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Dec 10 22:50:54 crc kubenswrapper[4791]: E1210 22:50:54.919935 4791 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Dec 10 22:50:54 crc kubenswrapper[4791]: E1210 22:50:54.919957 4791 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Dec 10 22:50:54 crc kubenswrapper[4791]: E1210 22:50:54.919969 4791 projected.go:194] Error preparing data for projected volume kube-api-access-s2dwl for pod openshift-network-diagnostics/network-check-source-55646444c4-trplf: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Dec 10 22:50:54 crc kubenswrapper[4791]: E1210 22:50:54.919996 4791 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl podName:9d751cbb-f2e2-430d-9754-c882a5e924a5 nodeName:}" failed. No retries permitted until 2025-12-10 22:51:58.919987256 +0000 UTC m=+153.349604869 (durationBeforeRetry 1m4s). Error: MountVolume.SetUp failed for volume "kube-api-access-s2dwl" (UniqueName: "kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl") pod "network-check-source-55646444c4-trplf" (UID: "9d751cbb-f2e2-430d-9754-c882a5e924a5") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Dec 10 22:50:54 crc kubenswrapper[4791]: E1210 22:50:54.920129 4791 secret.go:188] Couldn't get secret openshift-network-console/networking-console-plugin-cert: object "openshift-network-console"/"networking-console-plugin-cert" not registered Dec 10 22:50:54 crc kubenswrapper[4791]: E1210 22:50:54.920167 4791 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2025-12-10 22:51:58.920158671 +0000 UTC m=+153.349776284 (durationBeforeRetry 1m4s). Error: MountVolume.SetUp failed for volume "networking-console-plugin-cert" (UniqueName: "kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin-cert" not registered Dec 10 22:50:54 crc kubenswrapper[4791]: I1210 22:50:54.954804 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 22:50:54 crc kubenswrapper[4791]: I1210 22:50:54.954890 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 22:50:54 crc kubenswrapper[4791]: I1210 22:50:54.954917 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 22:50:54 crc kubenswrapper[4791]: I1210 22:50:54.954954 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 22:50:54 crc kubenswrapper[4791]: I1210 22:50:54.954980 4791 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T22:50:54Z","lastTransitionTime":"2025-12-10T22:50:54Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 22:50:55 crc kubenswrapper[4791]: I1210 22:50:55.057386 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 22:50:55 crc kubenswrapper[4791]: I1210 22:50:55.057434 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 22:50:55 crc kubenswrapper[4791]: I1210 22:50:55.057447 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 22:50:55 crc kubenswrapper[4791]: I1210 22:50:55.057469 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 22:50:55 crc kubenswrapper[4791]: I1210 22:50:55.057481 4791 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T22:50:55Z","lastTransitionTime":"2025-12-10T22:50:55Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 22:50:55 crc kubenswrapper[4791]: I1210 22:50:55.160008 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 22:50:55 crc kubenswrapper[4791]: I1210 22:50:55.160062 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 22:50:55 crc kubenswrapper[4791]: I1210 22:50:55.160079 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 22:50:55 crc kubenswrapper[4791]: I1210 22:50:55.160099 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 22:50:55 crc kubenswrapper[4791]: I1210 22:50:55.160113 4791 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T22:50:55Z","lastTransitionTime":"2025-12-10T22:50:55Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 22:50:55 crc kubenswrapper[4791]: I1210 22:50:55.262530 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 22:50:55 crc kubenswrapper[4791]: I1210 22:50:55.262574 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 22:50:55 crc kubenswrapper[4791]: I1210 22:50:55.262586 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 22:50:55 crc kubenswrapper[4791]: I1210 22:50:55.262602 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 22:50:55 crc kubenswrapper[4791]: I1210 22:50:55.262613 4791 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T22:50:55Z","lastTransitionTime":"2025-12-10T22:50:55Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 22:50:55 crc kubenswrapper[4791]: I1210 22:50:55.365621 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 22:50:55 crc kubenswrapper[4791]: I1210 22:50:55.365661 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 22:50:55 crc kubenswrapper[4791]: I1210 22:50:55.365674 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 22:50:55 crc kubenswrapper[4791]: I1210 22:50:55.365690 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 22:50:55 crc kubenswrapper[4791]: I1210 22:50:55.365701 4791 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T22:50:55Z","lastTransitionTime":"2025-12-10T22:50:55Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 22:50:55 crc kubenswrapper[4791]: I1210 22:50:55.468019 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 22:50:55 crc kubenswrapper[4791]: I1210 22:50:55.468086 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 22:50:55 crc kubenswrapper[4791]: I1210 22:50:55.468095 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 22:50:55 crc kubenswrapper[4791]: I1210 22:50:55.468108 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 22:50:55 crc kubenswrapper[4791]: I1210 22:50:55.468117 4791 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T22:50:55Z","lastTransitionTime":"2025-12-10T22:50:55Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 22:50:55 crc kubenswrapper[4791]: I1210 22:50:55.570690 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 22:50:55 crc kubenswrapper[4791]: I1210 22:50:55.570750 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 22:50:55 crc kubenswrapper[4791]: I1210 22:50:55.570765 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 22:50:55 crc kubenswrapper[4791]: I1210 22:50:55.570787 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 22:50:55 crc kubenswrapper[4791]: I1210 22:50:55.570802 4791 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T22:50:55Z","lastTransitionTime":"2025-12-10T22:50:55Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 22:50:55 crc kubenswrapper[4791]: I1210 22:50:55.673468 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 22:50:55 crc kubenswrapper[4791]: I1210 22:50:55.673547 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 22:50:55 crc kubenswrapper[4791]: I1210 22:50:55.673564 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 22:50:55 crc kubenswrapper[4791]: I1210 22:50:55.673584 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 22:50:55 crc kubenswrapper[4791]: I1210 22:50:55.673599 4791 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T22:50:55Z","lastTransitionTime":"2025-12-10T22:50:55Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 22:50:55 crc kubenswrapper[4791]: I1210 22:50:55.775723 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 22:50:55 crc kubenswrapper[4791]: I1210 22:50:55.775774 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 22:50:55 crc kubenswrapper[4791]: I1210 22:50:55.775787 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 22:50:55 crc kubenswrapper[4791]: I1210 22:50:55.775803 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 22:50:55 crc kubenswrapper[4791]: I1210 22:50:55.775814 4791 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T22:50:55Z","lastTransitionTime":"2025-12-10T22:50:55Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 22:50:55 crc kubenswrapper[4791]: I1210 22:50:55.878441 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 22:50:55 crc kubenswrapper[4791]: I1210 22:50:55.878736 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 22:50:55 crc kubenswrapper[4791]: I1210 22:50:55.878749 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 22:50:55 crc kubenswrapper[4791]: I1210 22:50:55.878767 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 22:50:55 crc kubenswrapper[4791]: I1210 22:50:55.878779 4791 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T22:50:55Z","lastTransitionTime":"2025-12-10T22:50:55Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 22:50:55 crc kubenswrapper[4791]: I1210 22:50:55.905092 4791 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"90250e13-6c63-443a-a5e5-6e4e341a2289\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T22:49:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T22:49:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T22:49:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T22:49:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T22:49:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ba54f95b1e00843b854db9a10fe6897592159ccacf30ed0f76002093a13eedd3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T22:49:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5aa8ffa518f3a68b60b275f8df4851b2564aa4984b61efdc9470b73072133e43\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T22:49:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://16f989503edc78d9451638e2cf6cf4b19a2cf2b8809b10b237541cb829303091\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T22:49:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://82556824d63ce6e46df14c45689ff1f81ebf729abe902061815c040b323898e9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T22:49:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://61db531940fa6f2effdae5a005b4c91bd2698bd5d1bff9588c171a65ad550eca\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T22:49:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0f65c864d885319ce2a08d61c30be24b3ffb7d23ea30c82ed7faf039cb2c006d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://0f65c864d885319ce2a08d61c30be24b3ffb7d23ea30c82ed7faf039cb2c006d\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-10T22:49:27Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-10T22:49:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e31a76d1d5b2812ba83e24e38268e25ec8bed98fc3da9adcd06bda5822b6ace1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e31a76d1d5b2812ba83e24e38268e25ec8bed98fc3da9adcd06bda5822b6ace1\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-10T22:49:30Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-10T22:49:29Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://b5b5eceae562f3986b543d41de1e977efb85a91d84c98e09159ca6d092eda691\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b5b5eceae562f3986b543d41de1e977efb85a91d84c98e09159ca6d092eda691\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-10T22:49:31Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-10T22:49:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-10T22:49:25Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T22:50:55Z is after 2025-08-24T17:21:41Z" Dec 10 22:50:55 crc kubenswrapper[4791]: I1210 22:50:55.919510 4791 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-10T22:49:50Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-10T22:49:50Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T22:50:55Z is after 2025-08-24T17:21:41Z" Dec 10 22:50:55 crc kubenswrapper[4791]: I1210 22:50:55.932651 4791 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-10T22:49:54Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-10T22:49:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-10T22:49:54Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://62bfaf5d7a6d1b452cdaf88086447d0c8cb7306e94434e46752be5475124d56a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T22:49:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T22:50:55Z is after 2025-08-24T17:21:41Z" Dec 10 22:50:55 crc kubenswrapper[4791]: I1210 22:50:55.944519 4791 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-b6kf6" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b4527e71-bb58-4a39-a99c-fe614aa4c5f9\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T22:50:04Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T22:50:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T22:50:04Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T22:50:04Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ngpr4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ngpr4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-10T22:50:04Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-b6kf6\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T22:50:55Z is after 2025-08-24T17:21:41Z" Dec 10 22:50:55 crc kubenswrapper[4791]: I1210 22:50:55.960154 4791 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f8349fea-c49b-49a1-b5ee-032c5dffe021\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T22:49:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T22:49:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T22:49:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T22:49:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T22:49:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://acafe12186df99d1059a31260de0eafce35ff7771bcb8f5c6753b72665d6f60b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T22:49:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ea4c346507862a8c12756a3061d6e25e788a8d79f4621b62f81f4322185c82a8\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T22:49:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://2e6e324ec6f4f7bb69fdfab4ea3ac1aa55e480dd8e2a8984b0fe5b4c9f66636e\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T22:49:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://1b2e11ff2bf8789fa1d0d23ae648b3e6ef653106e5bc4bbb5dd6447f7cd6f05d\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d6aa7f8c5d908d90008f2df49860ac8a3480136da567e3d85f86135d779204f9\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-12-10T22:49:44Z\\\",\\\"message\\\":\\\"W1210 22:49:33.845733 1 cmd.go:257] Using insecure, self-signed certificates\\\\nI1210 22:49:33.846108 1 crypto.go:601] Generating new CA for check-endpoints-signer@1765406973 cert, and key in /tmp/serving-cert-2585701456/serving-signer.crt, /tmp/serving-cert-2585701456/serving-signer.key\\\\nI1210 22:49:34.529748 1 observer_polling.go:159] Starting file observer\\\\nW1210 22:49:34.532225 1 builder.go:272] unable to get owner reference (falling back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": dial tcp [::1]:6443: connect: connection refused\\\\nI1210 22:49:34.532599 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1210 22:49:34.533977 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-2585701456/tls.crt::/tmp/serving-cert-2585701456/tls.key\\\\\\\"\\\\nF1210 22:49:44.774710 1 cmd.go:182] error initializing delegating authentication: unable to load configmap based request-header-client-ca-file: Get \\\\\\\"https://localhost:6443/api/v1/namespaces/kube-system/configmaps/extension-apiserver-authentication\\\\\\\": net/http: TLS handshake timeout\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-10T22:49:33Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T22:49:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://0c4d036ab269f62561fe7ef1586012cb3a16e3583a08b9488a0616ec00c8e0ae\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T22:49:32Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c5856d6b26c19fda13b5c9ac2892a7e2f9c6f79feb8173ecfd9b33b673327d67\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c5856d6b26c19fda13b5c9ac2892a7e2f9c6f79feb8173ecfd9b33b673327d67\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-10T22:49:27Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-10T22:49:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-10T22:49:25Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T22:50:55Z is after 2025-08-24T17:21:41Z" Dec 10 22:50:55 crc kubenswrapper[4791]: I1210 22:50:55.980623 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 22:50:55 crc kubenswrapper[4791]: I1210 22:50:55.980664 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 22:50:55 crc kubenswrapper[4791]: I1210 22:50:55.980678 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 22:50:55 crc kubenswrapper[4791]: I1210 22:50:55.980700 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 22:50:55 crc kubenswrapper[4791]: I1210 22:50:55.980716 4791 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T22:50:55Z","lastTransitionTime":"2025-12-10T22:50:55Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 22:50:55 crc kubenswrapper[4791]: I1210 22:50:55.981104 4791 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"02a949d8-38e2-4be3-989f-40a015407140\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T22:49:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T22:49:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T22:50:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T22:50:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T22:49:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://cd4158bc59b30f0f1716bf43c1ae7c9916a5b7d7164d074987fc1f19ba49b778\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T22:49:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://656065fd3cbdc6a25dab5d06c7baa00af16d8f62fc5da687efe9c6cc91945de0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T22:49:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://90ed7fff816b1772ba9ecd95587a59cbed7c30ac302fff7e7c5de934c912bcf4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T22:49:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9b7a5502803eaad9f76f5ffa91b29fbb2657402065a03a4d6433c096aac7264f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"wait-for-host-port\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://9b7a5502803eaad9f76f5ffa91b29fbb2657402065a03a4d6433c096aac7264f\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-10T22:49:27Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-10T22:49:27Z\\\"}}}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-10T22:49:25Z\\\"}}\" for pod \"openshift-kube-scheduler\"/\"openshift-kube-scheduler-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T22:50:55Z is after 2025-08-24T17:21:41Z" Dec 10 22:50:55 crc kubenswrapper[4791]: I1210 22:50:55.992878 4791 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-5rb5l" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ba35653c-6e06-4cee-a4d6-137764090d18\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T22:49:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T22:49:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T22:49:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T22:49:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://dbb5c875730a98ad2daf316cfafc08fcdce21ac16148154780ac9193ed405dab\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T22:49:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fvvrs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://185eb2ff44221a514878283ec873773a6c9a9073dc64e296f982e60dee1338e4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T22:49:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fvvrs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-10T22:49:51Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-5rb5l\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T22:50:55Z is after 2025-08-24T17:21:41Z" Dec 10 22:50:56 crc kubenswrapper[4791]: I1210 22:50:56.006649 4791 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-s6rk8" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5c9d0978-a6f9-49c5-b377-6619b4e3ff52\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T22:50:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T22:50:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T22:50:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T22:50:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8e55baaf460c794ab196e237d51d12de059aa038617dc2d5366d49c529f5d5c7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T22:50:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7vkgm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://51d1b9b086c60bae9223cb51d5a998d02790a0d91b745aab10392c23e617eb9f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T22:50:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7vkgm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-10T22:50:03Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-s6rk8\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T22:50:56Z is after 2025-08-24T17:21:41Z" Dec 10 22:50:56 crc kubenswrapper[4791]: I1210 22:50:56.019620 4791 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-10T22:49:52Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-10T22:49:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-10T22:49:52Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3734b110802828632d44263ed0a43cabc4b710833f0be120d72db0940fe783e2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T22:49:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://794803ac9be3c52509e6dcb10854d8864e1a9e406762e57c7971b07c4da1220e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T22:49:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T22:50:56Z is after 2025-08-24T17:21:41Z" Dec 10 22:50:56 crc kubenswrapper[4791]: I1210 22:50:56.028808 4791 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-vczjq" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b9f00f25-3fa1-4b63-9710-fd352224b01b\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T22:49:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T22:49:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T22:49:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T22:49:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4698cf4c6b55b41e628448d37c2eb847c2777d0efbd93afcbf367d5e78623b7e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T22:49:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lrrpv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-10T22:49:50Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-vczjq\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T22:50:56Z is after 2025-08-24T17:21:41Z" Dec 10 22:50:56 crc kubenswrapper[4791]: I1210 22:50:56.040695 4791 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-4nwkq" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"672aa28c-8169-49ed-87b8-21187d13a80c\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T22:49:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T22:49:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T22:50:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T22:50:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f337f881ffd71f4a2ee19e92bb0b92c76cbb9b8352511f741d0c87e7da4cdd18\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a2f2a62315529e148a10a33ccb241e9417fbde0fe42afb61cb7ee2438c74636f\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-12-10T22:50:38Z\\\",\\\"message\\\":\\\"2025-12-10T22:49:52+00:00 [cnibincopy] Successfully copied files in /usr/src/multus-cni/rhel9/bin/ to /host/opt/cni/bin/upgrade_e48d9423-4bca-43c4-9eb9-b55a19a5ffd6\\\\n2025-12-10T22:49:52+00:00 [cnibincopy] Successfully moved files in /host/opt/cni/bin/upgrade_e48d9423-4bca-43c4-9eb9-b55a19a5ffd6 to /host/opt/cni/bin/\\\\n2025-12-10T22:49:52Z [verbose] multus-daemon started\\\\n2025-12-10T22:49:52Z [verbose] Readiness Indicator file check\\\\n2025-12-10T22:50:37Z [error] have you checked that your default network is ready? still waiting for readinessindicatorfile @ /host/run/multus/cni/net.d/10-ovn-kubernetes.conf. pollimmediate error: timed out waiting for the condition\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-10T22:49:52Z\\\"}},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T22:50:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xl62j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-10T22:49:51Z\\\"}}\" for pod \"openshift-multus\"/\"multus-4nwkq\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T22:50:56Z is after 2025-08-24T17:21:41Z" Dec 10 22:50:56 crc kubenswrapper[4791]: I1210 22:50:56.055702 4791 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-zhq64" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3cd47739-0fa9-4321-aff1-220f8721a0b3\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T22:49:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T22:49:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T22:49:52Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T22:49:52Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://68795a29944e56870ffe94fa732e9dcf0bfb344c0d85d1e590f58a876129c61f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T22:49:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wvpw8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://bc721b530c08b433756e978ec0423cc18aabf6771b61eb96f903b4d57ea335ff\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T22:49:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wvpw8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://152a7618be8f38fadf4e564b3948980503cedcdb466c37344ccc20f4a3173a5f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T22:49:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wvpw8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3d18f9cb7207efb1e32de75a30d33e024f656a08f0503db95e5488bc5280971f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T22:49:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wvpw8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://937137ae9d7480cdc7c59bc4a1df0a4c10e72b3b63693af285278789953ed218\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T22:49:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wvpw8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://1a52d19b7671e8b143017c8a981a6f8d3ee3b84650e4b223b814a531ac76f364\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T22:49:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wvpw8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5f299a2b60de0ef5fb17a1e463d8552d6c4d11d166ae32e490104b3d90116d25\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5f299a2b60de0ef5fb17a1e463d8552d6c4d11d166ae32e490104b3d90116d25\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-12-10T22:50:48Z\\\",\\\"message\\\":\\\"11] Stopping reflector *v1.EgressIP (0s) from github.com/openshift/ovn-kubernetes/go-controller/pkg/crd/egressip/v1/apis/informers/externalversions/factory.go:140\\\\nI1210 22:50:47.868727 6852 reflector.go:311] Stopping reflector *v1.Pod (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI1210 22:50:47.868812 6852 reflector.go:311] Stopping reflector *v1.NetworkPolicy (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI1210 22:50:47.869024 6852 reflector.go:311] Stopping reflector *v1.EndpointSlice (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI1210 22:50:47.869223 6852 reflector.go:311] Stopping reflector *v1alpha1.BaselineAdminNetworkPolicy (0s) from sigs.k8s.io/network-policy-api/pkg/client/informers/externalversions/factory.go:141\\\\nI1210 22:50:47.869558 6852 handler.go:190] Sending *v1.Namespace event handler 1 for removal\\\\nI1210 22:50:47.869577 6852 handler.go:190] Sending *v1.Namespace event handler 5 for removal\\\\nI1210 22:50:47.869595 6852 handler.go:208] Removed *v1.Namespace event handler 5\\\\nI1210 22:50:47.869604 6852 handler.go:190] Sending *v1.NetworkPolicy event handler 4 for removal\\\\nI1210 22:50:47.869608 6852 handler.go:208] Removed *v1.Namespace event handler 1\\\\nI1210 22:50:47.869637 6852 factory.go:656] Stopping watch factory\\\\nI1210 22:50:47.869656 6852 ovnkube.go:599] Stopped ovnkube\\\\nI1210 22:50:\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-10T22:50:47Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 40s restarting failed container=ovnkube-controller pod=ovnkube-node-zhq64_openshift-ovn-kubernetes(3cd47739-0fa9-4321-aff1-220f8721a0b3)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wvpw8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://dd3f80e5713221f96ed50da7f2e7fd5b9fabfcd5c1656da8d8a86d818a5fdc3f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T22:49:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wvpw8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3afbd1ffcd2c4b40324aca6e52d6fc822db131c3f35a9af3a6f005c2cffdb7bf\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3afbd1ffcd2c4b40324aca6e52d6fc822db131c3f35a9af3a6f005c2cffdb7bf\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-10T22:49:53Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-10T22:49:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wvpw8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-10T22:49:52Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-zhq64\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T22:50:56Z is after 2025-08-24T17:21:41Z" Dec 10 22:50:56 crc kubenswrapper[4791]: I1210 22:50:56.069280 4791 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"698bbe9a-fbe5-436d-a835-5e80df77ba24\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T22:49:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T22:49:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T22:49:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T22:49:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T22:49:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://224d73713312376a58ec6cc2bc8fdb6ffa44c4749ae0e7691d5fe27b9afca09f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T22:49:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://372eeaffd40992c6f06efc4df6dfcb44b31de64ee8256c21508cdd7c74ecb96e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T22:49:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://76b6013b062cac5b0b4d12a94f38ce7b43f9c7ea216063455d88a956ac04d26e\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T22:49:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://78e950e122e761d950da1a3c60c06476e53920e8795f6c61db22c01e22d5136e\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T22:49:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-10T22:49:25Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T22:50:56Z is after 2025-08-24T17:21:41Z" Dec 10 22:50:56 crc kubenswrapper[4791]: I1210 22:50:56.079113 4791 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7a9a3709-32fb-48b4-97f4-e1a19282a085\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T22:49:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T22:49:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T22:49:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T22:49:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T22:49:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://511bd056768a15ea0f650a2b1a6e87e3c71dfc75696222a54f7376227ad480c6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-crio\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T22:49:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kube\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"var-lib-kubelet\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8c489105d0dfb302732d4f5bd2fc9d000967a071802735b778d49bdbc428b264\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://8c489105d0dfb302732d4f5bd2fc9d000967a071802735b778d49bdbc428b264\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-10T22:49:27Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-10T22:49:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var\\\",\\\"name\\\":\\\"var-lib-kubelet\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-10T22:49:25Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"kube-rbac-proxy-crio-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T22:50:56Z is after 2025-08-24T17:21:41Z" Dec 10 22:50:56 crc kubenswrapper[4791]: I1210 22:50:56.082490 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 22:50:56 crc kubenswrapper[4791]: I1210 22:50:56.082518 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 22:50:56 crc kubenswrapper[4791]: I1210 22:50:56.082525 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 22:50:56 crc kubenswrapper[4791]: I1210 22:50:56.082539 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 22:50:56 crc kubenswrapper[4791]: I1210 22:50:56.082550 4791 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T22:50:56Z","lastTransitionTime":"2025-12-10T22:50:56Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 22:50:56 crc kubenswrapper[4791]: I1210 22:50:56.094752 4791 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-10T22:49:50Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-10T22:49:50Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T22:50:56Z is after 2025-08-24T17:21:41Z" Dec 10 22:50:56 crc kubenswrapper[4791]: I1210 22:50:56.104034 4791 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-tchxk" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"c0ab74e0-42a9-4f42-ac80-16030cf4ffff\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T22:49:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T22:49:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T22:49:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T22:49:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://04e0fd1b53256b7a938e8ccab93e2a4eb73dd1a59e0b9799e09c55e511baf7fe\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T22:49:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zn2qt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-10T22:49:50Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-tchxk\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T22:50:56Z is after 2025-08-24T17:21:41Z" Dec 10 22:50:56 crc kubenswrapper[4791]: I1210 22:50:56.117199 4791 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-rw584" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ff9beb90-69b0-4732-bf37-0b81f58ecc98\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T22:49:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T22:49:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T22:49:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T22:49:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ddb9df501dc6e8e5561fbc2aefd431037a3f28f73bf8e82a6598d261a4224807\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T22:49:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fksqn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ce67e54b696bb53ffbbe48c6ce71476b635e0868993e9c427b22f904891433ee\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ce67e54b696bb53ffbbe48c6ce71476b635e0868993e9c427b22f904891433ee\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-10T22:49:52Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-10T22:49:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fksqn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://818d89bfffb8a110670152c8480684665ea8db578e9768f524fbbe6a92f27c22\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://818d89bfffb8a110670152c8480684665ea8db578e9768f524fbbe6a92f27c22\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-10T22:49:53Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-10T22:49:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fksqn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://9aa9c06e894bb2a2ed3738defaeff68a0a0458ead48177f36b90770944c211a0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://9aa9c06e894bb2a2ed3738defaeff68a0a0458ead48177f36b90770944c211a0\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-10T22:49:54Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-10T22:49:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fksqn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://dfd6d03336dcea7529419a30e8d70d84e2290a0fa5b0e46d09ad341c190638d6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://dfd6d03336dcea7529419a30e8d70d84e2290a0fa5b0e46d09ad341c190638d6\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-10T22:49:55Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-10T22:49:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fksqn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b86738401ece6fbc3083782720c66a515850ed78d4054f34fcabab7ff525b85f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b86738401ece6fbc3083782720c66a515850ed78d4054f34fcabab7ff525b85f\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-10T22:49:56Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-10T22:49:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fksqn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://18b326f572a899ddd10197a745bf9264d4ddd975690ad1cfb0ec576ccd5cf102\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://18b326f572a899ddd10197a745bf9264d4ddd975690ad1cfb0ec576ccd5cf102\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-10T22:49:57Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-10T22:49:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fksqn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-10T22:49:51Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-rw584\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T22:50:56Z is after 2025-08-24T17:21:41Z" Dec 10 22:50:56 crc kubenswrapper[4791]: I1210 22:50:56.132559 4791 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-10T22:49:52Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-10T22:49:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-10T22:49:52Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://20651eb0c16e2c71222f56da1ff755cb0e7b084886acbbb24f0ccaf57041ed85\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T22:49:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T22:50:56Z is after 2025-08-24T17:21:41Z" Dec 10 22:50:56 crc kubenswrapper[4791]: I1210 22:50:56.143139 4791 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-10T22:49:50Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-10T22:49:50Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T22:50:56Z is after 2025-08-24T17:21:41Z" Dec 10 22:50:56 crc kubenswrapper[4791]: I1210 22:50:56.185321 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 22:50:56 crc kubenswrapper[4791]: I1210 22:50:56.185377 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 22:50:56 crc kubenswrapper[4791]: I1210 22:50:56.185390 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 22:50:56 crc kubenswrapper[4791]: I1210 22:50:56.185407 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 22:50:56 crc kubenswrapper[4791]: I1210 22:50:56.185480 4791 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T22:50:56Z","lastTransitionTime":"2025-12-10T22:50:56Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 22:50:56 crc kubenswrapper[4791]: I1210 22:50:56.287418 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 22:50:56 crc kubenswrapper[4791]: I1210 22:50:56.287476 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 22:50:56 crc kubenswrapper[4791]: I1210 22:50:56.287486 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 22:50:56 crc kubenswrapper[4791]: I1210 22:50:56.287499 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 22:50:56 crc kubenswrapper[4791]: I1210 22:50:56.287507 4791 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T22:50:56Z","lastTransitionTime":"2025-12-10T22:50:56Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 22:50:56 crc kubenswrapper[4791]: I1210 22:50:56.390654 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 22:50:56 crc kubenswrapper[4791]: I1210 22:50:56.390730 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 22:50:56 crc kubenswrapper[4791]: I1210 22:50:56.390753 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 22:50:56 crc kubenswrapper[4791]: I1210 22:50:56.390780 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 22:50:56 crc kubenswrapper[4791]: I1210 22:50:56.390801 4791 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T22:50:56Z","lastTransitionTime":"2025-12-10T22:50:56Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 22:50:56 crc kubenswrapper[4791]: I1210 22:50:56.493286 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 22:50:56 crc kubenswrapper[4791]: I1210 22:50:56.493332 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 22:50:56 crc kubenswrapper[4791]: I1210 22:50:56.493371 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 22:50:56 crc kubenswrapper[4791]: I1210 22:50:56.493388 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 22:50:56 crc kubenswrapper[4791]: I1210 22:50:56.493402 4791 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T22:50:56Z","lastTransitionTime":"2025-12-10T22:50:56Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 22:50:56 crc kubenswrapper[4791]: I1210 22:50:56.595010 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 22:50:56 crc kubenswrapper[4791]: I1210 22:50:56.595055 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 22:50:56 crc kubenswrapper[4791]: I1210 22:50:56.595066 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 22:50:56 crc kubenswrapper[4791]: I1210 22:50:56.595082 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 22:50:56 crc kubenswrapper[4791]: I1210 22:50:56.595095 4791 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T22:50:56Z","lastTransitionTime":"2025-12-10T22:50:56Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 22:50:56 crc kubenswrapper[4791]: I1210 22:50:56.697136 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 22:50:56 crc kubenswrapper[4791]: I1210 22:50:56.697180 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 22:50:56 crc kubenswrapper[4791]: I1210 22:50:56.697196 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 22:50:56 crc kubenswrapper[4791]: I1210 22:50:56.697215 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 22:50:56 crc kubenswrapper[4791]: I1210 22:50:56.697375 4791 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T22:50:56Z","lastTransitionTime":"2025-12-10T22:50:56Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 22:50:56 crc kubenswrapper[4791]: I1210 22:50:56.799691 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 22:50:56 crc kubenswrapper[4791]: I1210 22:50:56.799734 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 22:50:56 crc kubenswrapper[4791]: I1210 22:50:56.799743 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 22:50:56 crc kubenswrapper[4791]: I1210 22:50:56.799757 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 22:50:56 crc kubenswrapper[4791]: I1210 22:50:56.799768 4791 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T22:50:56Z","lastTransitionTime":"2025-12-10T22:50:56Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 22:50:56 crc kubenswrapper[4791]: I1210 22:50:56.884642 4791 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-b6kf6" Dec 10 22:50:56 crc kubenswrapper[4791]: I1210 22:50:56.884682 4791 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 10 22:50:56 crc kubenswrapper[4791]: I1210 22:50:56.884768 4791 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 10 22:50:56 crc kubenswrapper[4791]: E1210 22:50:56.884783 4791 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-b6kf6" podUID="b4527e71-bb58-4a39-a99c-fe614aa4c5f9" Dec 10 22:50:56 crc kubenswrapper[4791]: I1210 22:50:56.884839 4791 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 10 22:50:56 crc kubenswrapper[4791]: E1210 22:50:56.884945 4791 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 10 22:50:56 crc kubenswrapper[4791]: E1210 22:50:56.885005 4791 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 10 22:50:56 crc kubenswrapper[4791]: E1210 22:50:56.885078 4791 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 10 22:50:56 crc kubenswrapper[4791]: I1210 22:50:56.902215 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 22:50:56 crc kubenswrapper[4791]: I1210 22:50:56.902243 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 22:50:56 crc kubenswrapper[4791]: I1210 22:50:56.902250 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 22:50:56 crc kubenswrapper[4791]: I1210 22:50:56.902264 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 22:50:56 crc kubenswrapper[4791]: I1210 22:50:56.902275 4791 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T22:50:56Z","lastTransitionTime":"2025-12-10T22:50:56Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 22:50:57 crc kubenswrapper[4791]: I1210 22:50:57.004517 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 22:50:57 crc kubenswrapper[4791]: I1210 22:50:57.004548 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 22:50:57 crc kubenswrapper[4791]: I1210 22:50:57.004557 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 22:50:57 crc kubenswrapper[4791]: I1210 22:50:57.004569 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 22:50:57 crc kubenswrapper[4791]: I1210 22:50:57.004577 4791 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T22:50:57Z","lastTransitionTime":"2025-12-10T22:50:57Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 22:50:57 crc kubenswrapper[4791]: I1210 22:50:57.108621 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 22:50:57 crc kubenswrapper[4791]: I1210 22:50:57.108683 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 22:50:57 crc kubenswrapper[4791]: I1210 22:50:57.108694 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 22:50:57 crc kubenswrapper[4791]: I1210 22:50:57.108714 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 22:50:57 crc kubenswrapper[4791]: I1210 22:50:57.108725 4791 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T22:50:57Z","lastTransitionTime":"2025-12-10T22:50:57Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 22:50:57 crc kubenswrapper[4791]: I1210 22:50:57.211522 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 22:50:57 crc kubenswrapper[4791]: I1210 22:50:57.211638 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 22:50:57 crc kubenswrapper[4791]: I1210 22:50:57.211663 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 22:50:57 crc kubenswrapper[4791]: I1210 22:50:57.211692 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 22:50:57 crc kubenswrapper[4791]: I1210 22:50:57.211714 4791 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T22:50:57Z","lastTransitionTime":"2025-12-10T22:50:57Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 22:50:57 crc kubenswrapper[4791]: I1210 22:50:57.315438 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 22:50:57 crc kubenswrapper[4791]: I1210 22:50:57.315506 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 22:50:57 crc kubenswrapper[4791]: I1210 22:50:57.315529 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 22:50:57 crc kubenswrapper[4791]: I1210 22:50:57.315606 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 22:50:57 crc kubenswrapper[4791]: I1210 22:50:57.315793 4791 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T22:50:57Z","lastTransitionTime":"2025-12-10T22:50:57Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 22:50:57 crc kubenswrapper[4791]: I1210 22:50:57.417985 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 22:50:57 crc kubenswrapper[4791]: I1210 22:50:57.418029 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 22:50:57 crc kubenswrapper[4791]: I1210 22:50:57.418040 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 22:50:57 crc kubenswrapper[4791]: I1210 22:50:57.418055 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 22:50:57 crc kubenswrapper[4791]: I1210 22:50:57.418066 4791 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T22:50:57Z","lastTransitionTime":"2025-12-10T22:50:57Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 22:50:57 crc kubenswrapper[4791]: I1210 22:50:57.520186 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 22:50:57 crc kubenswrapper[4791]: I1210 22:50:57.520229 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 22:50:57 crc kubenswrapper[4791]: I1210 22:50:57.520238 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 22:50:57 crc kubenswrapper[4791]: I1210 22:50:57.520253 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 22:50:57 crc kubenswrapper[4791]: I1210 22:50:57.520262 4791 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T22:50:57Z","lastTransitionTime":"2025-12-10T22:50:57Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 22:50:57 crc kubenswrapper[4791]: I1210 22:50:57.623554 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 22:50:57 crc kubenswrapper[4791]: I1210 22:50:57.623622 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 22:50:57 crc kubenswrapper[4791]: I1210 22:50:57.623649 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 22:50:57 crc kubenswrapper[4791]: I1210 22:50:57.623675 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 22:50:57 crc kubenswrapper[4791]: I1210 22:50:57.623690 4791 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T22:50:57Z","lastTransitionTime":"2025-12-10T22:50:57Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 22:50:57 crc kubenswrapper[4791]: I1210 22:50:57.725951 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 22:50:57 crc kubenswrapper[4791]: I1210 22:50:57.726043 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 22:50:57 crc kubenswrapper[4791]: I1210 22:50:57.726061 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 22:50:57 crc kubenswrapper[4791]: I1210 22:50:57.726086 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 22:50:57 crc kubenswrapper[4791]: I1210 22:50:57.726103 4791 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T22:50:57Z","lastTransitionTime":"2025-12-10T22:50:57Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 22:50:57 crc kubenswrapper[4791]: I1210 22:50:57.829197 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 22:50:57 crc kubenswrapper[4791]: I1210 22:50:57.829265 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 22:50:57 crc kubenswrapper[4791]: I1210 22:50:57.829282 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 22:50:57 crc kubenswrapper[4791]: I1210 22:50:57.829303 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 22:50:57 crc kubenswrapper[4791]: I1210 22:50:57.829320 4791 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T22:50:57Z","lastTransitionTime":"2025-12-10T22:50:57Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 22:50:57 crc kubenswrapper[4791]: I1210 22:50:57.931135 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 22:50:57 crc kubenswrapper[4791]: I1210 22:50:57.931206 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 22:50:57 crc kubenswrapper[4791]: I1210 22:50:57.931217 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 22:50:57 crc kubenswrapper[4791]: I1210 22:50:57.931569 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 22:50:57 crc kubenswrapper[4791]: I1210 22:50:57.931617 4791 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T22:50:57Z","lastTransitionTime":"2025-12-10T22:50:57Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 22:50:58 crc kubenswrapper[4791]: I1210 22:50:58.036411 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 22:50:58 crc kubenswrapper[4791]: I1210 22:50:58.036445 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 22:50:58 crc kubenswrapper[4791]: I1210 22:50:58.036456 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 22:50:58 crc kubenswrapper[4791]: I1210 22:50:58.036473 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 22:50:58 crc kubenswrapper[4791]: I1210 22:50:58.036484 4791 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T22:50:58Z","lastTransitionTime":"2025-12-10T22:50:58Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 22:50:58 crc kubenswrapper[4791]: I1210 22:50:58.138865 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 22:50:58 crc kubenswrapper[4791]: I1210 22:50:58.138906 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 22:50:58 crc kubenswrapper[4791]: I1210 22:50:58.138923 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 22:50:58 crc kubenswrapper[4791]: I1210 22:50:58.138943 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 22:50:58 crc kubenswrapper[4791]: I1210 22:50:58.138954 4791 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T22:50:58Z","lastTransitionTime":"2025-12-10T22:50:58Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 22:50:58 crc kubenswrapper[4791]: I1210 22:50:58.240798 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 22:50:58 crc kubenswrapper[4791]: I1210 22:50:58.240833 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 22:50:58 crc kubenswrapper[4791]: I1210 22:50:58.240844 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 22:50:58 crc kubenswrapper[4791]: I1210 22:50:58.240859 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 22:50:58 crc kubenswrapper[4791]: I1210 22:50:58.240869 4791 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T22:50:58Z","lastTransitionTime":"2025-12-10T22:50:58Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 22:50:58 crc kubenswrapper[4791]: I1210 22:50:58.343236 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 22:50:58 crc kubenswrapper[4791]: I1210 22:50:58.343267 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 22:50:58 crc kubenswrapper[4791]: I1210 22:50:58.343275 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 22:50:58 crc kubenswrapper[4791]: I1210 22:50:58.343287 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 22:50:58 crc kubenswrapper[4791]: I1210 22:50:58.343295 4791 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T22:50:58Z","lastTransitionTime":"2025-12-10T22:50:58Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 22:50:58 crc kubenswrapper[4791]: I1210 22:50:58.446305 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 22:50:58 crc kubenswrapper[4791]: I1210 22:50:58.446352 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 22:50:58 crc kubenswrapper[4791]: I1210 22:50:58.446365 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 22:50:58 crc kubenswrapper[4791]: I1210 22:50:58.446381 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 22:50:58 crc kubenswrapper[4791]: I1210 22:50:58.446412 4791 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T22:50:58Z","lastTransitionTime":"2025-12-10T22:50:58Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 22:50:58 crc kubenswrapper[4791]: I1210 22:50:58.549546 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 22:50:58 crc kubenswrapper[4791]: I1210 22:50:58.549607 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 22:50:58 crc kubenswrapper[4791]: I1210 22:50:58.549625 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 22:50:58 crc kubenswrapper[4791]: I1210 22:50:58.549661 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 22:50:58 crc kubenswrapper[4791]: I1210 22:50:58.549681 4791 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T22:50:58Z","lastTransitionTime":"2025-12-10T22:50:58Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 22:50:58 crc kubenswrapper[4791]: I1210 22:50:58.653267 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 22:50:58 crc kubenswrapper[4791]: I1210 22:50:58.653317 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 22:50:58 crc kubenswrapper[4791]: I1210 22:50:58.653328 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 22:50:58 crc kubenswrapper[4791]: I1210 22:50:58.653365 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 22:50:58 crc kubenswrapper[4791]: I1210 22:50:58.653375 4791 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T22:50:58Z","lastTransitionTime":"2025-12-10T22:50:58Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 22:50:58 crc kubenswrapper[4791]: I1210 22:50:58.756050 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 22:50:58 crc kubenswrapper[4791]: I1210 22:50:58.756103 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 22:50:58 crc kubenswrapper[4791]: I1210 22:50:58.756117 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 22:50:58 crc kubenswrapper[4791]: I1210 22:50:58.756139 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 22:50:58 crc kubenswrapper[4791]: I1210 22:50:58.756153 4791 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T22:50:58Z","lastTransitionTime":"2025-12-10T22:50:58Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 22:50:58 crc kubenswrapper[4791]: I1210 22:50:58.859375 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 22:50:58 crc kubenswrapper[4791]: I1210 22:50:58.859443 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 22:50:58 crc kubenswrapper[4791]: I1210 22:50:58.859460 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 22:50:58 crc kubenswrapper[4791]: I1210 22:50:58.859483 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 22:50:58 crc kubenswrapper[4791]: I1210 22:50:58.859502 4791 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T22:50:58Z","lastTransitionTime":"2025-12-10T22:50:58Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 22:50:58 crc kubenswrapper[4791]: I1210 22:50:58.884196 4791 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 10 22:50:58 crc kubenswrapper[4791]: I1210 22:50:58.884249 4791 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-b6kf6" Dec 10 22:50:58 crc kubenswrapper[4791]: I1210 22:50:58.884288 4791 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 10 22:50:58 crc kubenswrapper[4791]: E1210 22:50:58.884368 4791 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 10 22:50:58 crc kubenswrapper[4791]: I1210 22:50:58.884212 4791 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 10 22:50:58 crc kubenswrapper[4791]: E1210 22:50:58.884546 4791 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 10 22:50:58 crc kubenswrapper[4791]: E1210 22:50:58.884698 4791 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-b6kf6" podUID="b4527e71-bb58-4a39-a99c-fe614aa4c5f9" Dec 10 22:50:58 crc kubenswrapper[4791]: E1210 22:50:58.884748 4791 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 10 22:50:58 crc kubenswrapper[4791]: I1210 22:50:58.962673 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 22:50:58 crc kubenswrapper[4791]: I1210 22:50:58.962752 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 22:50:58 crc kubenswrapper[4791]: I1210 22:50:58.962773 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 22:50:58 crc kubenswrapper[4791]: I1210 22:50:58.962799 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 22:50:58 crc kubenswrapper[4791]: I1210 22:50:58.962818 4791 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T22:50:58Z","lastTransitionTime":"2025-12-10T22:50:58Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 22:50:59 crc kubenswrapper[4791]: I1210 22:50:59.066056 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 22:50:59 crc kubenswrapper[4791]: I1210 22:50:59.066126 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 22:50:59 crc kubenswrapper[4791]: I1210 22:50:59.066144 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 22:50:59 crc kubenswrapper[4791]: I1210 22:50:59.066170 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 22:50:59 crc kubenswrapper[4791]: I1210 22:50:59.066188 4791 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T22:50:59Z","lastTransitionTime":"2025-12-10T22:50:59Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 22:50:59 crc kubenswrapper[4791]: I1210 22:50:59.169780 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 22:50:59 crc kubenswrapper[4791]: I1210 22:50:59.169836 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 22:50:59 crc kubenswrapper[4791]: I1210 22:50:59.169854 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 22:50:59 crc kubenswrapper[4791]: I1210 22:50:59.169876 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 22:50:59 crc kubenswrapper[4791]: I1210 22:50:59.169893 4791 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T22:50:59Z","lastTransitionTime":"2025-12-10T22:50:59Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 22:50:59 crc kubenswrapper[4791]: I1210 22:50:59.184813 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 22:50:59 crc kubenswrapper[4791]: I1210 22:50:59.184872 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 22:50:59 crc kubenswrapper[4791]: I1210 22:50:59.184885 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 22:50:59 crc kubenswrapper[4791]: I1210 22:50:59.184906 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 22:50:59 crc kubenswrapper[4791]: I1210 22:50:59.184920 4791 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T22:50:59Z","lastTransitionTime":"2025-12-10T22:50:59Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 22:50:59 crc kubenswrapper[4791]: E1210 22:50:59.202556 4791 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-12-10T22:50:59Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-10T22:50:59Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-10T22:50:59Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-10T22:50:59Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-10T22:50:59Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-10T22:50:59Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-10T22:50:59Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-10T22:50:59Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"4fab3af6-f657-4dfc-8ef2-12b7c978c94f\\\",\\\"systemUUID\\\":\\\"442c0b68-8dce-47e0-bd68-d5210d7e0493\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T22:50:59Z is after 2025-08-24T17:21:41Z" Dec 10 22:50:59 crc kubenswrapper[4791]: I1210 22:50:59.208378 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 22:50:59 crc kubenswrapper[4791]: I1210 22:50:59.208419 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 22:50:59 crc kubenswrapper[4791]: I1210 22:50:59.208430 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 22:50:59 crc kubenswrapper[4791]: I1210 22:50:59.208449 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 22:50:59 crc kubenswrapper[4791]: I1210 22:50:59.208462 4791 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T22:50:59Z","lastTransitionTime":"2025-12-10T22:50:59Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 22:50:59 crc kubenswrapper[4791]: E1210 22:50:59.227745 4791 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-12-10T22:50:59Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-10T22:50:59Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-10T22:50:59Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-10T22:50:59Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-10T22:50:59Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-10T22:50:59Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-10T22:50:59Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-10T22:50:59Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"4fab3af6-f657-4dfc-8ef2-12b7c978c94f\\\",\\\"systemUUID\\\":\\\"442c0b68-8dce-47e0-bd68-d5210d7e0493\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T22:50:59Z is after 2025-08-24T17:21:41Z" Dec 10 22:50:59 crc kubenswrapper[4791]: I1210 22:50:59.232617 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 22:50:59 crc kubenswrapper[4791]: I1210 22:50:59.232672 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 22:50:59 crc kubenswrapper[4791]: I1210 22:50:59.232693 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 22:50:59 crc kubenswrapper[4791]: I1210 22:50:59.232717 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 22:50:59 crc kubenswrapper[4791]: I1210 22:50:59.232739 4791 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T22:50:59Z","lastTransitionTime":"2025-12-10T22:50:59Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 22:50:59 crc kubenswrapper[4791]: E1210 22:50:59.253614 4791 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-12-10T22:50:59Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-10T22:50:59Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-10T22:50:59Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-10T22:50:59Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-10T22:50:59Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-10T22:50:59Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-10T22:50:59Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-10T22:50:59Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"4fab3af6-f657-4dfc-8ef2-12b7c978c94f\\\",\\\"systemUUID\\\":\\\"442c0b68-8dce-47e0-bd68-d5210d7e0493\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T22:50:59Z is after 2025-08-24T17:21:41Z" Dec 10 22:50:59 crc kubenswrapper[4791]: I1210 22:50:59.258919 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 22:50:59 crc kubenswrapper[4791]: I1210 22:50:59.258994 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 22:50:59 crc kubenswrapper[4791]: I1210 22:50:59.259006 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 22:50:59 crc kubenswrapper[4791]: I1210 22:50:59.259022 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 22:50:59 crc kubenswrapper[4791]: I1210 22:50:59.259033 4791 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T22:50:59Z","lastTransitionTime":"2025-12-10T22:50:59Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 22:50:59 crc kubenswrapper[4791]: E1210 22:50:59.275703 4791 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-12-10T22:50:59Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-10T22:50:59Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-10T22:50:59Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-10T22:50:59Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-10T22:50:59Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-10T22:50:59Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-10T22:50:59Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-10T22:50:59Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"4fab3af6-f657-4dfc-8ef2-12b7c978c94f\\\",\\\"systemUUID\\\":\\\"442c0b68-8dce-47e0-bd68-d5210d7e0493\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T22:50:59Z is after 2025-08-24T17:21:41Z" Dec 10 22:50:59 crc kubenswrapper[4791]: I1210 22:50:59.280120 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 22:50:59 crc kubenswrapper[4791]: I1210 22:50:59.280159 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 22:50:59 crc kubenswrapper[4791]: I1210 22:50:59.280171 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 22:50:59 crc kubenswrapper[4791]: I1210 22:50:59.280186 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 22:50:59 crc kubenswrapper[4791]: I1210 22:50:59.280197 4791 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T22:50:59Z","lastTransitionTime":"2025-12-10T22:50:59Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 22:50:59 crc kubenswrapper[4791]: E1210 22:50:59.295450 4791 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-12-10T22:50:59Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-10T22:50:59Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-10T22:50:59Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-10T22:50:59Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-10T22:50:59Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-10T22:50:59Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-10T22:50:59Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-10T22:50:59Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"4fab3af6-f657-4dfc-8ef2-12b7c978c94f\\\",\\\"systemUUID\\\":\\\"442c0b68-8dce-47e0-bd68-d5210d7e0493\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T22:50:59Z is after 2025-08-24T17:21:41Z" Dec 10 22:50:59 crc kubenswrapper[4791]: E1210 22:50:59.295680 4791 kubelet_node_status.go:572] "Unable to update node status" err="update node status exceeds retry count" Dec 10 22:50:59 crc kubenswrapper[4791]: I1210 22:50:59.297205 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 22:50:59 crc kubenswrapper[4791]: I1210 22:50:59.297238 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 22:50:59 crc kubenswrapper[4791]: I1210 22:50:59.297248 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 22:50:59 crc kubenswrapper[4791]: I1210 22:50:59.297265 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 22:50:59 crc kubenswrapper[4791]: I1210 22:50:59.297278 4791 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T22:50:59Z","lastTransitionTime":"2025-12-10T22:50:59Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 22:50:59 crc kubenswrapper[4791]: I1210 22:50:59.400685 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 22:50:59 crc kubenswrapper[4791]: I1210 22:50:59.400744 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 22:50:59 crc kubenswrapper[4791]: I1210 22:50:59.400762 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 22:50:59 crc kubenswrapper[4791]: I1210 22:50:59.400785 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 22:50:59 crc kubenswrapper[4791]: I1210 22:50:59.400801 4791 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T22:50:59Z","lastTransitionTime":"2025-12-10T22:50:59Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 22:50:59 crc kubenswrapper[4791]: I1210 22:50:59.503498 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 22:50:59 crc kubenswrapper[4791]: I1210 22:50:59.503568 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 22:50:59 crc kubenswrapper[4791]: I1210 22:50:59.503590 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 22:50:59 crc kubenswrapper[4791]: I1210 22:50:59.504044 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 22:50:59 crc kubenswrapper[4791]: I1210 22:50:59.504090 4791 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T22:50:59Z","lastTransitionTime":"2025-12-10T22:50:59Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 22:50:59 crc kubenswrapper[4791]: I1210 22:50:59.606792 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 22:50:59 crc kubenswrapper[4791]: I1210 22:50:59.606839 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 22:50:59 crc kubenswrapper[4791]: I1210 22:50:59.606849 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 22:50:59 crc kubenswrapper[4791]: I1210 22:50:59.606864 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 22:50:59 crc kubenswrapper[4791]: I1210 22:50:59.606877 4791 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T22:50:59Z","lastTransitionTime":"2025-12-10T22:50:59Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 22:50:59 crc kubenswrapper[4791]: I1210 22:50:59.709883 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 22:50:59 crc kubenswrapper[4791]: I1210 22:50:59.709941 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 22:50:59 crc kubenswrapper[4791]: I1210 22:50:59.709962 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 22:50:59 crc kubenswrapper[4791]: I1210 22:50:59.710012 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 22:50:59 crc kubenswrapper[4791]: I1210 22:50:59.710032 4791 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T22:50:59Z","lastTransitionTime":"2025-12-10T22:50:59Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 22:50:59 crc kubenswrapper[4791]: I1210 22:50:59.813427 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 22:50:59 crc kubenswrapper[4791]: I1210 22:50:59.813504 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 22:50:59 crc kubenswrapper[4791]: I1210 22:50:59.813533 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 22:50:59 crc kubenswrapper[4791]: I1210 22:50:59.813567 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 22:50:59 crc kubenswrapper[4791]: I1210 22:50:59.813594 4791 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T22:50:59Z","lastTransitionTime":"2025-12-10T22:50:59Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 22:50:59 crc kubenswrapper[4791]: I1210 22:50:59.916418 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 22:50:59 crc kubenswrapper[4791]: I1210 22:50:59.916832 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 22:50:59 crc kubenswrapper[4791]: I1210 22:50:59.916978 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 22:50:59 crc kubenswrapper[4791]: I1210 22:50:59.917109 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 22:50:59 crc kubenswrapper[4791]: I1210 22:50:59.917253 4791 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T22:50:59Z","lastTransitionTime":"2025-12-10T22:50:59Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 22:51:00 crc kubenswrapper[4791]: I1210 22:51:00.020761 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 22:51:00 crc kubenswrapper[4791]: I1210 22:51:00.020821 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 22:51:00 crc kubenswrapper[4791]: I1210 22:51:00.020837 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 22:51:00 crc kubenswrapper[4791]: I1210 22:51:00.020860 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 22:51:00 crc kubenswrapper[4791]: I1210 22:51:00.020876 4791 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T22:51:00Z","lastTransitionTime":"2025-12-10T22:51:00Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 22:51:00 crc kubenswrapper[4791]: I1210 22:51:00.124483 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 22:51:00 crc kubenswrapper[4791]: I1210 22:51:00.124537 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 22:51:00 crc kubenswrapper[4791]: I1210 22:51:00.124552 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 22:51:00 crc kubenswrapper[4791]: I1210 22:51:00.124571 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 22:51:00 crc kubenswrapper[4791]: I1210 22:51:00.124586 4791 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T22:51:00Z","lastTransitionTime":"2025-12-10T22:51:00Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 22:51:00 crc kubenswrapper[4791]: I1210 22:51:00.228108 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 22:51:00 crc kubenswrapper[4791]: I1210 22:51:00.228212 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 22:51:00 crc kubenswrapper[4791]: I1210 22:51:00.228234 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 22:51:00 crc kubenswrapper[4791]: I1210 22:51:00.228292 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 22:51:00 crc kubenswrapper[4791]: I1210 22:51:00.228312 4791 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T22:51:00Z","lastTransitionTime":"2025-12-10T22:51:00Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 22:51:00 crc kubenswrapper[4791]: I1210 22:51:00.331940 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 22:51:00 crc kubenswrapper[4791]: I1210 22:51:00.332001 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 22:51:00 crc kubenswrapper[4791]: I1210 22:51:00.332020 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 22:51:00 crc kubenswrapper[4791]: I1210 22:51:00.332050 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 22:51:00 crc kubenswrapper[4791]: I1210 22:51:00.332072 4791 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T22:51:00Z","lastTransitionTime":"2025-12-10T22:51:00Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 22:51:00 crc kubenswrapper[4791]: I1210 22:51:00.435768 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 22:51:00 crc kubenswrapper[4791]: I1210 22:51:00.435823 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 22:51:00 crc kubenswrapper[4791]: I1210 22:51:00.435840 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 22:51:00 crc kubenswrapper[4791]: I1210 22:51:00.435863 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 22:51:00 crc kubenswrapper[4791]: I1210 22:51:00.435880 4791 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T22:51:00Z","lastTransitionTime":"2025-12-10T22:51:00Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 22:51:00 crc kubenswrapper[4791]: I1210 22:51:00.539103 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 22:51:00 crc kubenswrapper[4791]: I1210 22:51:00.539152 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 22:51:00 crc kubenswrapper[4791]: I1210 22:51:00.539163 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 22:51:00 crc kubenswrapper[4791]: I1210 22:51:00.539185 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 22:51:00 crc kubenswrapper[4791]: I1210 22:51:00.539201 4791 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T22:51:00Z","lastTransitionTime":"2025-12-10T22:51:00Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 22:51:00 crc kubenswrapper[4791]: I1210 22:51:00.641849 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 22:51:00 crc kubenswrapper[4791]: I1210 22:51:00.641906 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 22:51:00 crc kubenswrapper[4791]: I1210 22:51:00.641918 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 22:51:00 crc kubenswrapper[4791]: I1210 22:51:00.641936 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 22:51:00 crc kubenswrapper[4791]: I1210 22:51:00.641949 4791 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T22:51:00Z","lastTransitionTime":"2025-12-10T22:51:00Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 22:51:00 crc kubenswrapper[4791]: I1210 22:51:00.744890 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 22:51:00 crc kubenswrapper[4791]: I1210 22:51:00.744973 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 22:51:00 crc kubenswrapper[4791]: I1210 22:51:00.744991 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 22:51:00 crc kubenswrapper[4791]: I1210 22:51:00.745015 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 22:51:00 crc kubenswrapper[4791]: I1210 22:51:00.745032 4791 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T22:51:00Z","lastTransitionTime":"2025-12-10T22:51:00Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 22:51:00 crc kubenswrapper[4791]: I1210 22:51:00.847698 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 22:51:00 crc kubenswrapper[4791]: I1210 22:51:00.847729 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 22:51:00 crc kubenswrapper[4791]: I1210 22:51:00.847738 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 22:51:00 crc kubenswrapper[4791]: I1210 22:51:00.847751 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 22:51:00 crc kubenswrapper[4791]: I1210 22:51:00.847760 4791 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T22:51:00Z","lastTransitionTime":"2025-12-10T22:51:00Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 22:51:00 crc kubenswrapper[4791]: I1210 22:51:00.884900 4791 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-b6kf6" Dec 10 22:51:00 crc kubenswrapper[4791]: I1210 22:51:00.884930 4791 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 10 22:51:00 crc kubenswrapper[4791]: I1210 22:51:00.884944 4791 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 10 22:51:00 crc kubenswrapper[4791]: I1210 22:51:00.885086 4791 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 10 22:51:00 crc kubenswrapper[4791]: E1210 22:51:00.885217 4791 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-b6kf6" podUID="b4527e71-bb58-4a39-a99c-fe614aa4c5f9" Dec 10 22:51:00 crc kubenswrapper[4791]: E1210 22:51:00.885387 4791 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 10 22:51:00 crc kubenswrapper[4791]: E1210 22:51:00.885504 4791 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 10 22:51:00 crc kubenswrapper[4791]: E1210 22:51:00.885627 4791 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 10 22:51:00 crc kubenswrapper[4791]: I1210 22:51:00.950188 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 22:51:00 crc kubenswrapper[4791]: I1210 22:51:00.950218 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 22:51:00 crc kubenswrapper[4791]: I1210 22:51:00.950225 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 22:51:00 crc kubenswrapper[4791]: I1210 22:51:00.950237 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 22:51:00 crc kubenswrapper[4791]: I1210 22:51:00.950246 4791 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T22:51:00Z","lastTransitionTime":"2025-12-10T22:51:00Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 22:51:01 crc kubenswrapper[4791]: I1210 22:51:01.052899 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 22:51:01 crc kubenswrapper[4791]: I1210 22:51:01.052948 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 22:51:01 crc kubenswrapper[4791]: I1210 22:51:01.052961 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 22:51:01 crc kubenswrapper[4791]: I1210 22:51:01.052978 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 22:51:01 crc kubenswrapper[4791]: I1210 22:51:01.052991 4791 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T22:51:01Z","lastTransitionTime":"2025-12-10T22:51:01Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 22:51:01 crc kubenswrapper[4791]: I1210 22:51:01.156182 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 22:51:01 crc kubenswrapper[4791]: I1210 22:51:01.156273 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 22:51:01 crc kubenswrapper[4791]: I1210 22:51:01.156295 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 22:51:01 crc kubenswrapper[4791]: I1210 22:51:01.156321 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 22:51:01 crc kubenswrapper[4791]: I1210 22:51:01.156371 4791 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T22:51:01Z","lastTransitionTime":"2025-12-10T22:51:01Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 22:51:01 crc kubenswrapper[4791]: I1210 22:51:01.259214 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 22:51:01 crc kubenswrapper[4791]: I1210 22:51:01.259255 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 22:51:01 crc kubenswrapper[4791]: I1210 22:51:01.259264 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 22:51:01 crc kubenswrapper[4791]: I1210 22:51:01.259278 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 22:51:01 crc kubenswrapper[4791]: I1210 22:51:01.259287 4791 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T22:51:01Z","lastTransitionTime":"2025-12-10T22:51:01Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 22:51:01 crc kubenswrapper[4791]: I1210 22:51:01.361493 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 22:51:01 crc kubenswrapper[4791]: I1210 22:51:01.361547 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 22:51:01 crc kubenswrapper[4791]: I1210 22:51:01.361559 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 22:51:01 crc kubenswrapper[4791]: I1210 22:51:01.361575 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 22:51:01 crc kubenswrapper[4791]: I1210 22:51:01.361586 4791 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T22:51:01Z","lastTransitionTime":"2025-12-10T22:51:01Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 22:51:01 crc kubenswrapper[4791]: I1210 22:51:01.463461 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 22:51:01 crc kubenswrapper[4791]: I1210 22:51:01.463524 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 22:51:01 crc kubenswrapper[4791]: I1210 22:51:01.463534 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 22:51:01 crc kubenswrapper[4791]: I1210 22:51:01.463552 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 22:51:01 crc kubenswrapper[4791]: I1210 22:51:01.463563 4791 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T22:51:01Z","lastTransitionTime":"2025-12-10T22:51:01Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 22:51:01 crc kubenswrapper[4791]: I1210 22:51:01.565756 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 22:51:01 crc kubenswrapper[4791]: I1210 22:51:01.565807 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 22:51:01 crc kubenswrapper[4791]: I1210 22:51:01.565825 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 22:51:01 crc kubenswrapper[4791]: I1210 22:51:01.565845 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 22:51:01 crc kubenswrapper[4791]: I1210 22:51:01.565860 4791 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T22:51:01Z","lastTransitionTime":"2025-12-10T22:51:01Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 22:51:01 crc kubenswrapper[4791]: I1210 22:51:01.669411 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 22:51:01 crc kubenswrapper[4791]: I1210 22:51:01.669483 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 22:51:01 crc kubenswrapper[4791]: I1210 22:51:01.669506 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 22:51:01 crc kubenswrapper[4791]: I1210 22:51:01.669535 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 22:51:01 crc kubenswrapper[4791]: I1210 22:51:01.669558 4791 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T22:51:01Z","lastTransitionTime":"2025-12-10T22:51:01Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 22:51:01 crc kubenswrapper[4791]: I1210 22:51:01.772101 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 22:51:01 crc kubenswrapper[4791]: I1210 22:51:01.772146 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 22:51:01 crc kubenswrapper[4791]: I1210 22:51:01.772156 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 22:51:01 crc kubenswrapper[4791]: I1210 22:51:01.772170 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 22:51:01 crc kubenswrapper[4791]: I1210 22:51:01.772182 4791 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T22:51:01Z","lastTransitionTime":"2025-12-10T22:51:01Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 22:51:01 crc kubenswrapper[4791]: I1210 22:51:01.875413 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 22:51:01 crc kubenswrapper[4791]: I1210 22:51:01.875511 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 22:51:01 crc kubenswrapper[4791]: I1210 22:51:01.875551 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 22:51:01 crc kubenswrapper[4791]: I1210 22:51:01.875582 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 22:51:01 crc kubenswrapper[4791]: I1210 22:51:01.875605 4791 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T22:51:01Z","lastTransitionTime":"2025-12-10T22:51:01Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 22:51:01 crc kubenswrapper[4791]: I1210 22:51:01.978767 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 22:51:01 crc kubenswrapper[4791]: I1210 22:51:01.978858 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 22:51:01 crc kubenswrapper[4791]: I1210 22:51:01.978882 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 22:51:01 crc kubenswrapper[4791]: I1210 22:51:01.978913 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 22:51:01 crc kubenswrapper[4791]: I1210 22:51:01.978938 4791 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T22:51:01Z","lastTransitionTime":"2025-12-10T22:51:01Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 22:51:02 crc kubenswrapper[4791]: I1210 22:51:02.081757 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 22:51:02 crc kubenswrapper[4791]: I1210 22:51:02.081840 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 22:51:02 crc kubenswrapper[4791]: I1210 22:51:02.081851 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 22:51:02 crc kubenswrapper[4791]: I1210 22:51:02.081870 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 22:51:02 crc kubenswrapper[4791]: I1210 22:51:02.081881 4791 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T22:51:02Z","lastTransitionTime":"2025-12-10T22:51:02Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 22:51:02 crc kubenswrapper[4791]: I1210 22:51:02.184845 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 22:51:02 crc kubenswrapper[4791]: I1210 22:51:02.184917 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 22:51:02 crc kubenswrapper[4791]: I1210 22:51:02.184943 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 22:51:02 crc kubenswrapper[4791]: I1210 22:51:02.184972 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 22:51:02 crc kubenswrapper[4791]: I1210 22:51:02.184995 4791 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T22:51:02Z","lastTransitionTime":"2025-12-10T22:51:02Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 22:51:02 crc kubenswrapper[4791]: I1210 22:51:02.287228 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 22:51:02 crc kubenswrapper[4791]: I1210 22:51:02.287280 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 22:51:02 crc kubenswrapper[4791]: I1210 22:51:02.287301 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 22:51:02 crc kubenswrapper[4791]: I1210 22:51:02.287318 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 22:51:02 crc kubenswrapper[4791]: I1210 22:51:02.287329 4791 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T22:51:02Z","lastTransitionTime":"2025-12-10T22:51:02Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 22:51:02 crc kubenswrapper[4791]: I1210 22:51:02.390309 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 22:51:02 crc kubenswrapper[4791]: I1210 22:51:02.390380 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 22:51:02 crc kubenswrapper[4791]: I1210 22:51:02.390395 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 22:51:02 crc kubenswrapper[4791]: I1210 22:51:02.390411 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 22:51:02 crc kubenswrapper[4791]: I1210 22:51:02.390422 4791 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T22:51:02Z","lastTransitionTime":"2025-12-10T22:51:02Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 22:51:02 crc kubenswrapper[4791]: I1210 22:51:02.493489 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 22:51:02 crc kubenswrapper[4791]: I1210 22:51:02.493542 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 22:51:02 crc kubenswrapper[4791]: I1210 22:51:02.493554 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 22:51:02 crc kubenswrapper[4791]: I1210 22:51:02.493571 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 22:51:02 crc kubenswrapper[4791]: I1210 22:51:02.493582 4791 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T22:51:02Z","lastTransitionTime":"2025-12-10T22:51:02Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 22:51:02 crc kubenswrapper[4791]: I1210 22:51:02.597024 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 22:51:02 crc kubenswrapper[4791]: I1210 22:51:02.597069 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 22:51:02 crc kubenswrapper[4791]: I1210 22:51:02.597085 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 22:51:02 crc kubenswrapper[4791]: I1210 22:51:02.597106 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 22:51:02 crc kubenswrapper[4791]: I1210 22:51:02.597121 4791 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T22:51:02Z","lastTransitionTime":"2025-12-10T22:51:02Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 22:51:02 crc kubenswrapper[4791]: I1210 22:51:02.700635 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 22:51:02 crc kubenswrapper[4791]: I1210 22:51:02.700696 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 22:51:02 crc kubenswrapper[4791]: I1210 22:51:02.700729 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 22:51:02 crc kubenswrapper[4791]: I1210 22:51:02.700753 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 22:51:02 crc kubenswrapper[4791]: I1210 22:51:02.700775 4791 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T22:51:02Z","lastTransitionTime":"2025-12-10T22:51:02Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 22:51:02 crc kubenswrapper[4791]: I1210 22:51:02.802704 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 22:51:02 crc kubenswrapper[4791]: I1210 22:51:02.802751 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 22:51:02 crc kubenswrapper[4791]: I1210 22:51:02.802764 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 22:51:02 crc kubenswrapper[4791]: I1210 22:51:02.802780 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 22:51:02 crc kubenswrapper[4791]: I1210 22:51:02.802793 4791 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T22:51:02Z","lastTransitionTime":"2025-12-10T22:51:02Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 22:51:02 crc kubenswrapper[4791]: I1210 22:51:02.883766 4791 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 10 22:51:02 crc kubenswrapper[4791]: I1210 22:51:02.883766 4791 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 10 22:51:02 crc kubenswrapper[4791]: I1210 22:51:02.884387 4791 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-b6kf6" Dec 10 22:51:02 crc kubenswrapper[4791]: I1210 22:51:02.884424 4791 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 10 22:51:02 crc kubenswrapper[4791]: E1210 22:51:02.884548 4791 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 10 22:51:02 crc kubenswrapper[4791]: E1210 22:51:02.884823 4791 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 10 22:51:02 crc kubenswrapper[4791]: E1210 22:51:02.885019 4791 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 10 22:51:02 crc kubenswrapper[4791]: I1210 22:51:02.885124 4791 scope.go:117] "RemoveContainer" containerID="5f299a2b60de0ef5fb17a1e463d8552d6c4d11d166ae32e490104b3d90116d25" Dec 10 22:51:02 crc kubenswrapper[4791]: E1210 22:51:02.885178 4791 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-b6kf6" podUID="b4527e71-bb58-4a39-a99c-fe614aa4c5f9" Dec 10 22:51:02 crc kubenswrapper[4791]: E1210 22:51:02.885481 4791 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"ovnkube-controller\" with CrashLoopBackOff: \"back-off 40s restarting failed container=ovnkube-controller pod=ovnkube-node-zhq64_openshift-ovn-kubernetes(3cd47739-0fa9-4321-aff1-220f8721a0b3)\"" pod="openshift-ovn-kubernetes/ovnkube-node-zhq64" podUID="3cd47739-0fa9-4321-aff1-220f8721a0b3" Dec 10 22:51:02 crc kubenswrapper[4791]: I1210 22:51:02.904644 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 22:51:02 crc kubenswrapper[4791]: I1210 22:51:02.904719 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 22:51:02 crc kubenswrapper[4791]: I1210 22:51:02.904746 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 22:51:02 crc kubenswrapper[4791]: I1210 22:51:02.904776 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 22:51:02 crc kubenswrapper[4791]: I1210 22:51:02.904798 4791 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T22:51:02Z","lastTransitionTime":"2025-12-10T22:51:02Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 22:51:03 crc kubenswrapper[4791]: I1210 22:51:03.006545 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 22:51:03 crc kubenswrapper[4791]: I1210 22:51:03.006569 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 22:51:03 crc kubenswrapper[4791]: I1210 22:51:03.006577 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 22:51:03 crc kubenswrapper[4791]: I1210 22:51:03.006589 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 22:51:03 crc kubenswrapper[4791]: I1210 22:51:03.006597 4791 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T22:51:03Z","lastTransitionTime":"2025-12-10T22:51:03Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 22:51:03 crc kubenswrapper[4791]: I1210 22:51:03.108918 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 22:51:03 crc kubenswrapper[4791]: I1210 22:51:03.108986 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 22:51:03 crc kubenswrapper[4791]: I1210 22:51:03.109001 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 22:51:03 crc kubenswrapper[4791]: I1210 22:51:03.109022 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 22:51:03 crc kubenswrapper[4791]: I1210 22:51:03.109039 4791 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T22:51:03Z","lastTransitionTime":"2025-12-10T22:51:03Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 22:51:03 crc kubenswrapper[4791]: I1210 22:51:03.211149 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 22:51:03 crc kubenswrapper[4791]: I1210 22:51:03.211194 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 22:51:03 crc kubenswrapper[4791]: I1210 22:51:03.211207 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 22:51:03 crc kubenswrapper[4791]: I1210 22:51:03.211222 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 22:51:03 crc kubenswrapper[4791]: I1210 22:51:03.211234 4791 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T22:51:03Z","lastTransitionTime":"2025-12-10T22:51:03Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 22:51:03 crc kubenswrapper[4791]: I1210 22:51:03.314023 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 22:51:03 crc kubenswrapper[4791]: I1210 22:51:03.314063 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 22:51:03 crc kubenswrapper[4791]: I1210 22:51:03.314093 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 22:51:03 crc kubenswrapper[4791]: I1210 22:51:03.314107 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 22:51:03 crc kubenswrapper[4791]: I1210 22:51:03.314116 4791 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T22:51:03Z","lastTransitionTime":"2025-12-10T22:51:03Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 22:51:03 crc kubenswrapper[4791]: I1210 22:51:03.417028 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 22:51:03 crc kubenswrapper[4791]: I1210 22:51:03.417087 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 22:51:03 crc kubenswrapper[4791]: I1210 22:51:03.417096 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 22:51:03 crc kubenswrapper[4791]: I1210 22:51:03.417110 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 22:51:03 crc kubenswrapper[4791]: I1210 22:51:03.417119 4791 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T22:51:03Z","lastTransitionTime":"2025-12-10T22:51:03Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 22:51:03 crc kubenswrapper[4791]: I1210 22:51:03.519458 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 22:51:03 crc kubenswrapper[4791]: I1210 22:51:03.519503 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 22:51:03 crc kubenswrapper[4791]: I1210 22:51:03.519514 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 22:51:03 crc kubenswrapper[4791]: I1210 22:51:03.519530 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 22:51:03 crc kubenswrapper[4791]: I1210 22:51:03.519543 4791 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T22:51:03Z","lastTransitionTime":"2025-12-10T22:51:03Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 22:51:03 crc kubenswrapper[4791]: I1210 22:51:03.621106 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 22:51:03 crc kubenswrapper[4791]: I1210 22:51:03.621153 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 22:51:03 crc kubenswrapper[4791]: I1210 22:51:03.621165 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 22:51:03 crc kubenswrapper[4791]: I1210 22:51:03.621181 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 22:51:03 crc kubenswrapper[4791]: I1210 22:51:03.621193 4791 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T22:51:03Z","lastTransitionTime":"2025-12-10T22:51:03Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 22:51:03 crc kubenswrapper[4791]: I1210 22:51:03.723817 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 22:51:03 crc kubenswrapper[4791]: I1210 22:51:03.723881 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 22:51:03 crc kubenswrapper[4791]: I1210 22:51:03.723903 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 22:51:03 crc kubenswrapper[4791]: I1210 22:51:03.723932 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 22:51:03 crc kubenswrapper[4791]: I1210 22:51:03.723958 4791 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T22:51:03Z","lastTransitionTime":"2025-12-10T22:51:03Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 22:51:03 crc kubenswrapper[4791]: I1210 22:51:03.827012 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 22:51:03 crc kubenswrapper[4791]: I1210 22:51:03.827073 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 22:51:03 crc kubenswrapper[4791]: I1210 22:51:03.827089 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 22:51:03 crc kubenswrapper[4791]: I1210 22:51:03.827112 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 22:51:03 crc kubenswrapper[4791]: I1210 22:51:03.827129 4791 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T22:51:03Z","lastTransitionTime":"2025-12-10T22:51:03Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 22:51:03 crc kubenswrapper[4791]: I1210 22:51:03.929845 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 22:51:03 crc kubenswrapper[4791]: I1210 22:51:03.929921 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 22:51:03 crc kubenswrapper[4791]: I1210 22:51:03.929944 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 22:51:03 crc kubenswrapper[4791]: I1210 22:51:03.929972 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 22:51:03 crc kubenswrapper[4791]: I1210 22:51:03.929994 4791 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T22:51:03Z","lastTransitionTime":"2025-12-10T22:51:03Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 22:51:04 crc kubenswrapper[4791]: I1210 22:51:04.032982 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 22:51:04 crc kubenswrapper[4791]: I1210 22:51:04.033039 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 22:51:04 crc kubenswrapper[4791]: I1210 22:51:04.033053 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 22:51:04 crc kubenswrapper[4791]: I1210 22:51:04.033069 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 22:51:04 crc kubenswrapper[4791]: I1210 22:51:04.033080 4791 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T22:51:04Z","lastTransitionTime":"2025-12-10T22:51:04Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 22:51:04 crc kubenswrapper[4791]: I1210 22:51:04.135427 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 22:51:04 crc kubenswrapper[4791]: I1210 22:51:04.135499 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 22:51:04 crc kubenswrapper[4791]: I1210 22:51:04.135516 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 22:51:04 crc kubenswrapper[4791]: I1210 22:51:04.135539 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 22:51:04 crc kubenswrapper[4791]: I1210 22:51:04.135556 4791 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T22:51:04Z","lastTransitionTime":"2025-12-10T22:51:04Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 22:51:04 crc kubenswrapper[4791]: I1210 22:51:04.238530 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 22:51:04 crc kubenswrapper[4791]: I1210 22:51:04.238591 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 22:51:04 crc kubenswrapper[4791]: I1210 22:51:04.238610 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 22:51:04 crc kubenswrapper[4791]: I1210 22:51:04.238636 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 22:51:04 crc kubenswrapper[4791]: I1210 22:51:04.238655 4791 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T22:51:04Z","lastTransitionTime":"2025-12-10T22:51:04Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 22:51:04 crc kubenswrapper[4791]: I1210 22:51:04.341014 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 22:51:04 crc kubenswrapper[4791]: I1210 22:51:04.341069 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 22:51:04 crc kubenswrapper[4791]: I1210 22:51:04.341086 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 22:51:04 crc kubenswrapper[4791]: I1210 22:51:04.341108 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 22:51:04 crc kubenswrapper[4791]: I1210 22:51:04.341124 4791 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T22:51:04Z","lastTransitionTime":"2025-12-10T22:51:04Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 22:51:04 crc kubenswrapper[4791]: I1210 22:51:04.443793 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 22:51:04 crc kubenswrapper[4791]: I1210 22:51:04.443840 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 22:51:04 crc kubenswrapper[4791]: I1210 22:51:04.443852 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 22:51:04 crc kubenswrapper[4791]: I1210 22:51:04.443868 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 22:51:04 crc kubenswrapper[4791]: I1210 22:51:04.443881 4791 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T22:51:04Z","lastTransitionTime":"2025-12-10T22:51:04Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 22:51:04 crc kubenswrapper[4791]: I1210 22:51:04.546327 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 22:51:04 crc kubenswrapper[4791]: I1210 22:51:04.546407 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 22:51:04 crc kubenswrapper[4791]: I1210 22:51:04.546420 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 22:51:04 crc kubenswrapper[4791]: I1210 22:51:04.546437 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 22:51:04 crc kubenswrapper[4791]: I1210 22:51:04.546449 4791 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T22:51:04Z","lastTransitionTime":"2025-12-10T22:51:04Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 22:51:04 crc kubenswrapper[4791]: I1210 22:51:04.649015 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 22:51:04 crc kubenswrapper[4791]: I1210 22:51:04.649084 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 22:51:04 crc kubenswrapper[4791]: I1210 22:51:04.649097 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 22:51:04 crc kubenswrapper[4791]: I1210 22:51:04.649114 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 22:51:04 crc kubenswrapper[4791]: I1210 22:51:04.649136 4791 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T22:51:04Z","lastTransitionTime":"2025-12-10T22:51:04Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 22:51:04 crc kubenswrapper[4791]: I1210 22:51:04.752259 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 22:51:04 crc kubenswrapper[4791]: I1210 22:51:04.752412 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 22:51:04 crc kubenswrapper[4791]: I1210 22:51:04.752438 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 22:51:04 crc kubenswrapper[4791]: I1210 22:51:04.752471 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 22:51:04 crc kubenswrapper[4791]: I1210 22:51:04.752494 4791 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T22:51:04Z","lastTransitionTime":"2025-12-10T22:51:04Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 22:51:04 crc kubenswrapper[4791]: I1210 22:51:04.854984 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 22:51:04 crc kubenswrapper[4791]: I1210 22:51:04.855037 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 22:51:04 crc kubenswrapper[4791]: I1210 22:51:04.855051 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 22:51:04 crc kubenswrapper[4791]: I1210 22:51:04.855071 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 22:51:04 crc kubenswrapper[4791]: I1210 22:51:04.855085 4791 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T22:51:04Z","lastTransitionTime":"2025-12-10T22:51:04Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 22:51:04 crc kubenswrapper[4791]: I1210 22:51:04.884459 4791 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 10 22:51:04 crc kubenswrapper[4791]: I1210 22:51:04.884559 4791 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 10 22:51:04 crc kubenswrapper[4791]: I1210 22:51:04.884557 4791 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 10 22:51:04 crc kubenswrapper[4791]: I1210 22:51:04.884555 4791 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-b6kf6" Dec 10 22:51:04 crc kubenswrapper[4791]: E1210 22:51:04.884797 4791 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 10 22:51:04 crc kubenswrapper[4791]: E1210 22:51:04.884858 4791 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 10 22:51:04 crc kubenswrapper[4791]: E1210 22:51:04.884914 4791 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-b6kf6" podUID="b4527e71-bb58-4a39-a99c-fe614aa4c5f9" Dec 10 22:51:04 crc kubenswrapper[4791]: E1210 22:51:04.884980 4791 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 10 22:51:04 crc kubenswrapper[4791]: I1210 22:51:04.957890 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 22:51:04 crc kubenswrapper[4791]: I1210 22:51:04.957935 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 22:51:04 crc kubenswrapper[4791]: I1210 22:51:04.957947 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 22:51:04 crc kubenswrapper[4791]: I1210 22:51:04.957962 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 22:51:04 crc kubenswrapper[4791]: I1210 22:51:04.957972 4791 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T22:51:04Z","lastTransitionTime":"2025-12-10T22:51:04Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 22:51:05 crc kubenswrapper[4791]: I1210 22:51:05.060292 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 22:51:05 crc kubenswrapper[4791]: I1210 22:51:05.060403 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 22:51:05 crc kubenswrapper[4791]: I1210 22:51:05.060435 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 22:51:05 crc kubenswrapper[4791]: I1210 22:51:05.060464 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 22:51:05 crc kubenswrapper[4791]: I1210 22:51:05.060486 4791 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T22:51:05Z","lastTransitionTime":"2025-12-10T22:51:05Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 22:51:05 crc kubenswrapper[4791]: I1210 22:51:05.162748 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 22:51:05 crc kubenswrapper[4791]: I1210 22:51:05.162816 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 22:51:05 crc kubenswrapper[4791]: I1210 22:51:05.162837 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 22:51:05 crc kubenswrapper[4791]: I1210 22:51:05.162864 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 22:51:05 crc kubenswrapper[4791]: I1210 22:51:05.162900 4791 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T22:51:05Z","lastTransitionTime":"2025-12-10T22:51:05Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 22:51:05 crc kubenswrapper[4791]: I1210 22:51:05.265002 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 22:51:05 crc kubenswrapper[4791]: I1210 22:51:05.265067 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 22:51:05 crc kubenswrapper[4791]: I1210 22:51:05.265084 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 22:51:05 crc kubenswrapper[4791]: I1210 22:51:05.265107 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 22:51:05 crc kubenswrapper[4791]: I1210 22:51:05.265123 4791 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T22:51:05Z","lastTransitionTime":"2025-12-10T22:51:05Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 22:51:05 crc kubenswrapper[4791]: I1210 22:51:05.368373 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 22:51:05 crc kubenswrapper[4791]: I1210 22:51:05.368420 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 22:51:05 crc kubenswrapper[4791]: I1210 22:51:05.368433 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 22:51:05 crc kubenswrapper[4791]: I1210 22:51:05.368452 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 22:51:05 crc kubenswrapper[4791]: I1210 22:51:05.368464 4791 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T22:51:05Z","lastTransitionTime":"2025-12-10T22:51:05Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 22:51:05 crc kubenswrapper[4791]: I1210 22:51:05.470833 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 22:51:05 crc kubenswrapper[4791]: I1210 22:51:05.470874 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 22:51:05 crc kubenswrapper[4791]: I1210 22:51:05.470885 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 22:51:05 crc kubenswrapper[4791]: I1210 22:51:05.470901 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 22:51:05 crc kubenswrapper[4791]: I1210 22:51:05.470912 4791 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T22:51:05Z","lastTransitionTime":"2025-12-10T22:51:05Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 22:51:05 crc kubenswrapper[4791]: I1210 22:51:05.573520 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 22:51:05 crc kubenswrapper[4791]: I1210 22:51:05.573571 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 22:51:05 crc kubenswrapper[4791]: I1210 22:51:05.573583 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 22:51:05 crc kubenswrapper[4791]: I1210 22:51:05.573601 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 22:51:05 crc kubenswrapper[4791]: I1210 22:51:05.573613 4791 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T22:51:05Z","lastTransitionTime":"2025-12-10T22:51:05Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 22:51:05 crc kubenswrapper[4791]: I1210 22:51:05.675505 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 22:51:05 crc kubenswrapper[4791]: I1210 22:51:05.675583 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 22:51:05 crc kubenswrapper[4791]: I1210 22:51:05.675615 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 22:51:05 crc kubenswrapper[4791]: I1210 22:51:05.675633 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 22:51:05 crc kubenswrapper[4791]: I1210 22:51:05.675645 4791 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T22:51:05Z","lastTransitionTime":"2025-12-10T22:51:05Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 22:51:05 crc kubenswrapper[4791]: I1210 22:51:05.778034 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 22:51:05 crc kubenswrapper[4791]: I1210 22:51:05.778096 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 22:51:05 crc kubenswrapper[4791]: I1210 22:51:05.778119 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 22:51:05 crc kubenswrapper[4791]: I1210 22:51:05.778148 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 22:51:05 crc kubenswrapper[4791]: I1210 22:51:05.778170 4791 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T22:51:05Z","lastTransitionTime":"2025-12-10T22:51:05Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 22:51:05 crc kubenswrapper[4791]: I1210 22:51:05.881185 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 22:51:05 crc kubenswrapper[4791]: I1210 22:51:05.881232 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 22:51:05 crc kubenswrapper[4791]: I1210 22:51:05.881243 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 22:51:05 crc kubenswrapper[4791]: I1210 22:51:05.881263 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 22:51:05 crc kubenswrapper[4791]: I1210 22:51:05.881274 4791 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T22:51:05Z","lastTransitionTime":"2025-12-10T22:51:05Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 22:51:05 crc kubenswrapper[4791]: I1210 22:51:05.902134 4791 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-5rb5l" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ba35653c-6e06-4cee-a4d6-137764090d18\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T22:49:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T22:49:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T22:49:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T22:49:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://dbb5c875730a98ad2daf316cfafc08fcdce21ac16148154780ac9193ed405dab\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T22:49:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fvvrs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://185eb2ff44221a514878283ec873773a6c9a9073dc64e296f982e60dee1338e4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T22:49:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fvvrs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-10T22:49:51Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-5rb5l\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T22:51:05Z is after 2025-08-24T17:21:41Z" Dec 10 22:51:05 crc kubenswrapper[4791]: I1210 22:51:05.913331 4791 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-s6rk8" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5c9d0978-a6f9-49c5-b377-6619b4e3ff52\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T22:50:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T22:50:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T22:50:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T22:50:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8e55baaf460c794ab196e237d51d12de059aa038617dc2d5366d49c529f5d5c7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T22:50:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7vkgm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://51d1b9b086c60bae9223cb51d5a998d02790a0d91b745aab10392c23e617eb9f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T22:50:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7vkgm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-10T22:50:03Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-s6rk8\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T22:51:05Z is after 2025-08-24T17:21:41Z" Dec 10 22:51:05 crc kubenswrapper[4791]: I1210 22:51:05.927252 4791 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"698bbe9a-fbe5-436d-a835-5e80df77ba24\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T22:49:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T22:49:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T22:49:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T22:49:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T22:49:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://224d73713312376a58ec6cc2bc8fdb6ffa44c4749ae0e7691d5fe27b9afca09f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T22:49:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://372eeaffd40992c6f06efc4df6dfcb44b31de64ee8256c21508cdd7c74ecb96e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T22:49:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://76b6013b062cac5b0b4d12a94f38ce7b43f9c7ea216063455d88a956ac04d26e\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T22:49:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://78e950e122e761d950da1a3c60c06476e53920e8795f6c61db22c01e22d5136e\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T22:49:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-10T22:49:25Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T22:51:05Z is after 2025-08-24T17:21:41Z" Dec 10 22:51:05 crc kubenswrapper[4791]: I1210 22:51:05.937916 4791 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7a9a3709-32fb-48b4-97f4-e1a19282a085\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T22:49:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T22:49:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T22:49:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T22:49:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T22:49:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://511bd056768a15ea0f650a2b1a6e87e3c71dfc75696222a54f7376227ad480c6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-crio\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T22:49:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kube\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"var-lib-kubelet\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8c489105d0dfb302732d4f5bd2fc9d000967a071802735b778d49bdbc428b264\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://8c489105d0dfb302732d4f5bd2fc9d000967a071802735b778d49bdbc428b264\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-10T22:49:27Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-10T22:49:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var\\\",\\\"name\\\":\\\"var-lib-kubelet\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-10T22:49:25Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"kube-rbac-proxy-crio-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T22:51:05Z is after 2025-08-24T17:21:41Z" Dec 10 22:51:05 crc kubenswrapper[4791]: I1210 22:51:05.951045 4791 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-10T22:49:52Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-10T22:49:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-10T22:49:52Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3734b110802828632d44263ed0a43cabc4b710833f0be120d72db0940fe783e2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T22:49:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://794803ac9be3c52509e6dcb10854d8864e1a9e406762e57c7971b07c4da1220e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T22:49:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T22:51:05Z is after 2025-08-24T17:21:41Z" Dec 10 22:51:05 crc kubenswrapper[4791]: I1210 22:51:05.959988 4791 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-vczjq" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b9f00f25-3fa1-4b63-9710-fd352224b01b\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T22:49:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T22:49:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T22:49:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T22:49:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4698cf4c6b55b41e628448d37c2eb847c2777d0efbd93afcbf367d5e78623b7e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T22:49:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lrrpv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-10T22:49:50Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-vczjq\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T22:51:05Z is after 2025-08-24T17:21:41Z" Dec 10 22:51:05 crc kubenswrapper[4791]: I1210 22:51:05.978574 4791 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-4nwkq" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"672aa28c-8169-49ed-87b8-21187d13a80c\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T22:49:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T22:49:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T22:50:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T22:50:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f337f881ffd71f4a2ee19e92bb0b92c76cbb9b8352511f741d0c87e7da4cdd18\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a2f2a62315529e148a10a33ccb241e9417fbde0fe42afb61cb7ee2438c74636f\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-12-10T22:50:38Z\\\",\\\"message\\\":\\\"2025-12-10T22:49:52+00:00 [cnibincopy] Successfully copied files in /usr/src/multus-cni/rhel9/bin/ to /host/opt/cni/bin/upgrade_e48d9423-4bca-43c4-9eb9-b55a19a5ffd6\\\\n2025-12-10T22:49:52+00:00 [cnibincopy] Successfully moved files in /host/opt/cni/bin/upgrade_e48d9423-4bca-43c4-9eb9-b55a19a5ffd6 to /host/opt/cni/bin/\\\\n2025-12-10T22:49:52Z [verbose] multus-daemon started\\\\n2025-12-10T22:49:52Z [verbose] Readiness Indicator file check\\\\n2025-12-10T22:50:37Z [error] have you checked that your default network is ready? still waiting for readinessindicatorfile @ /host/run/multus/cni/net.d/10-ovn-kubernetes.conf. pollimmediate error: timed out waiting for the condition\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-10T22:49:52Z\\\"}},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T22:50:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xl62j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-10T22:49:51Z\\\"}}\" for pod \"openshift-multus\"/\"multus-4nwkq\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T22:51:05Z is after 2025-08-24T17:21:41Z" Dec 10 22:51:05 crc kubenswrapper[4791]: I1210 22:51:05.983494 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 22:51:05 crc kubenswrapper[4791]: I1210 22:51:05.983544 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 22:51:05 crc kubenswrapper[4791]: I1210 22:51:05.983553 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 22:51:05 crc kubenswrapper[4791]: I1210 22:51:05.983567 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 22:51:05 crc kubenswrapper[4791]: I1210 22:51:05.983577 4791 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T22:51:05Z","lastTransitionTime":"2025-12-10T22:51:05Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 22:51:05 crc kubenswrapper[4791]: I1210 22:51:05.995301 4791 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-zhq64" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3cd47739-0fa9-4321-aff1-220f8721a0b3\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T22:49:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T22:49:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T22:49:52Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T22:49:52Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://68795a29944e56870ffe94fa732e9dcf0bfb344c0d85d1e590f58a876129c61f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T22:49:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wvpw8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://bc721b530c08b433756e978ec0423cc18aabf6771b61eb96f903b4d57ea335ff\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T22:49:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wvpw8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://152a7618be8f38fadf4e564b3948980503cedcdb466c37344ccc20f4a3173a5f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T22:49:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wvpw8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3d18f9cb7207efb1e32de75a30d33e024f656a08f0503db95e5488bc5280971f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T22:49:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wvpw8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://937137ae9d7480cdc7c59bc4a1df0a4c10e72b3b63693af285278789953ed218\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T22:49:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wvpw8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://1a52d19b7671e8b143017c8a981a6f8d3ee3b84650e4b223b814a531ac76f364\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T22:49:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wvpw8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5f299a2b60de0ef5fb17a1e463d8552d6c4d11d166ae32e490104b3d90116d25\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5f299a2b60de0ef5fb17a1e463d8552d6c4d11d166ae32e490104b3d90116d25\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-12-10T22:50:48Z\\\",\\\"message\\\":\\\"11] Stopping reflector *v1.EgressIP (0s) from github.com/openshift/ovn-kubernetes/go-controller/pkg/crd/egressip/v1/apis/informers/externalversions/factory.go:140\\\\nI1210 22:50:47.868727 6852 reflector.go:311] Stopping reflector *v1.Pod (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI1210 22:50:47.868812 6852 reflector.go:311] Stopping reflector *v1.NetworkPolicy (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI1210 22:50:47.869024 6852 reflector.go:311] Stopping reflector *v1.EndpointSlice (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI1210 22:50:47.869223 6852 reflector.go:311] Stopping reflector *v1alpha1.BaselineAdminNetworkPolicy (0s) from sigs.k8s.io/network-policy-api/pkg/client/informers/externalversions/factory.go:141\\\\nI1210 22:50:47.869558 6852 handler.go:190] Sending *v1.Namespace event handler 1 for removal\\\\nI1210 22:50:47.869577 6852 handler.go:190] Sending *v1.Namespace event handler 5 for removal\\\\nI1210 22:50:47.869595 6852 handler.go:208] Removed *v1.Namespace event handler 5\\\\nI1210 22:50:47.869604 6852 handler.go:190] Sending *v1.NetworkPolicy event handler 4 for removal\\\\nI1210 22:50:47.869608 6852 handler.go:208] Removed *v1.Namespace event handler 1\\\\nI1210 22:50:47.869637 6852 factory.go:656] Stopping watch factory\\\\nI1210 22:50:47.869656 6852 ovnkube.go:599] Stopped ovnkube\\\\nI1210 22:50:\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-10T22:50:47Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 40s restarting failed container=ovnkube-controller pod=ovnkube-node-zhq64_openshift-ovn-kubernetes(3cd47739-0fa9-4321-aff1-220f8721a0b3)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wvpw8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://dd3f80e5713221f96ed50da7f2e7fd5b9fabfcd5c1656da8d8a86d818a5fdc3f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T22:49:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wvpw8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3afbd1ffcd2c4b40324aca6e52d6fc822db131c3f35a9af3a6f005c2cffdb7bf\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3afbd1ffcd2c4b40324aca6e52d6fc822db131c3f35a9af3a6f005c2cffdb7bf\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-10T22:49:53Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-10T22:49:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wvpw8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-10T22:49:52Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-zhq64\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T22:51:05Z is after 2025-08-24T17:21:41Z" Dec 10 22:51:06 crc kubenswrapper[4791]: I1210 22:51:06.006602 4791 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-10T22:49:52Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-10T22:49:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-10T22:49:52Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://20651eb0c16e2c71222f56da1ff755cb0e7b084886acbbb24f0ccaf57041ed85\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T22:49:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T22:51:06Z is after 2025-08-24T17:21:41Z" Dec 10 22:51:06 crc kubenswrapper[4791]: I1210 22:51:06.016059 4791 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-10T22:49:50Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-10T22:49:50Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T22:51:06Z is after 2025-08-24T17:21:41Z" Dec 10 22:51:06 crc kubenswrapper[4791]: I1210 22:51:06.027374 4791 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-10T22:49:50Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-10T22:49:50Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T22:51:06Z is after 2025-08-24T17:21:41Z" Dec 10 22:51:06 crc kubenswrapper[4791]: I1210 22:51:06.037393 4791 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-tchxk" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"c0ab74e0-42a9-4f42-ac80-16030cf4ffff\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T22:49:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T22:49:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T22:49:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T22:49:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://04e0fd1b53256b7a938e8ccab93e2a4eb73dd1a59e0b9799e09c55e511baf7fe\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T22:49:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zn2qt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-10T22:49:50Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-tchxk\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T22:51:06Z is after 2025-08-24T17:21:41Z" Dec 10 22:51:06 crc kubenswrapper[4791]: I1210 22:51:06.048624 4791 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-rw584" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ff9beb90-69b0-4732-bf37-0b81f58ecc98\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T22:49:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T22:49:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T22:49:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T22:49:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ddb9df501dc6e8e5561fbc2aefd431037a3f28f73bf8e82a6598d261a4224807\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T22:49:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fksqn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ce67e54b696bb53ffbbe48c6ce71476b635e0868993e9c427b22f904891433ee\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ce67e54b696bb53ffbbe48c6ce71476b635e0868993e9c427b22f904891433ee\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-10T22:49:52Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-10T22:49:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fksqn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://818d89bfffb8a110670152c8480684665ea8db578e9768f524fbbe6a92f27c22\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://818d89bfffb8a110670152c8480684665ea8db578e9768f524fbbe6a92f27c22\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-10T22:49:53Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-10T22:49:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fksqn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://9aa9c06e894bb2a2ed3738defaeff68a0a0458ead48177f36b90770944c211a0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://9aa9c06e894bb2a2ed3738defaeff68a0a0458ead48177f36b90770944c211a0\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-10T22:49:54Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-10T22:49:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fksqn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://dfd6d03336dcea7529419a30e8d70d84e2290a0fa5b0e46d09ad341c190638d6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://dfd6d03336dcea7529419a30e8d70d84e2290a0fa5b0e46d09ad341c190638d6\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-10T22:49:55Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-10T22:49:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fksqn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b86738401ece6fbc3083782720c66a515850ed78d4054f34fcabab7ff525b85f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b86738401ece6fbc3083782720c66a515850ed78d4054f34fcabab7ff525b85f\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-10T22:49:56Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-10T22:49:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fksqn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://18b326f572a899ddd10197a745bf9264d4ddd975690ad1cfb0ec576ccd5cf102\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://18b326f572a899ddd10197a745bf9264d4ddd975690ad1cfb0ec576ccd5cf102\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-10T22:49:57Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-10T22:49:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fksqn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-10T22:49:51Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-rw584\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T22:51:06Z is after 2025-08-24T17:21:41Z" Dec 10 22:51:06 crc kubenswrapper[4791]: I1210 22:51:06.060744 4791 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f8349fea-c49b-49a1-b5ee-032c5dffe021\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T22:49:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T22:49:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T22:49:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T22:49:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T22:49:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://acafe12186df99d1059a31260de0eafce35ff7771bcb8f5c6753b72665d6f60b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T22:49:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ea4c346507862a8c12756a3061d6e25e788a8d79f4621b62f81f4322185c82a8\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T22:49:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://2e6e324ec6f4f7bb69fdfab4ea3ac1aa55e480dd8e2a8984b0fe5b4c9f66636e\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T22:49:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://1b2e11ff2bf8789fa1d0d23ae648b3e6ef653106e5bc4bbb5dd6447f7cd6f05d\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d6aa7f8c5d908d90008f2df49860ac8a3480136da567e3d85f86135d779204f9\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-12-10T22:49:44Z\\\",\\\"message\\\":\\\"W1210 22:49:33.845733 1 cmd.go:257] Using insecure, self-signed certificates\\\\nI1210 22:49:33.846108 1 crypto.go:601] Generating new CA for check-endpoints-signer@1765406973 cert, and key in /tmp/serving-cert-2585701456/serving-signer.crt, /tmp/serving-cert-2585701456/serving-signer.key\\\\nI1210 22:49:34.529748 1 observer_polling.go:159] Starting file observer\\\\nW1210 22:49:34.532225 1 builder.go:272] unable to get owner reference (falling back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": dial tcp [::1]:6443: connect: connection refused\\\\nI1210 22:49:34.532599 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1210 22:49:34.533977 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-2585701456/tls.crt::/tmp/serving-cert-2585701456/tls.key\\\\\\\"\\\\nF1210 22:49:44.774710 1 cmd.go:182] error initializing delegating authentication: unable to load configmap based request-header-client-ca-file: Get \\\\\\\"https://localhost:6443/api/v1/namespaces/kube-system/configmaps/extension-apiserver-authentication\\\\\\\": net/http: TLS handshake timeout\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-10T22:49:33Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T22:49:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://0c4d036ab269f62561fe7ef1586012cb3a16e3583a08b9488a0616ec00c8e0ae\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T22:49:32Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c5856d6b26c19fda13b5c9ac2892a7e2f9c6f79feb8173ecfd9b33b673327d67\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c5856d6b26c19fda13b5c9ac2892a7e2f9c6f79feb8173ecfd9b33b673327d67\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-10T22:49:27Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-10T22:49:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-10T22:49:25Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T22:51:06Z is after 2025-08-24T17:21:41Z" Dec 10 22:51:06 crc kubenswrapper[4791]: I1210 22:51:06.071385 4791 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"02a949d8-38e2-4be3-989f-40a015407140\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T22:49:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T22:49:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T22:50:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T22:50:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T22:49:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://cd4158bc59b30f0f1716bf43c1ae7c9916a5b7d7164d074987fc1f19ba49b778\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T22:49:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://656065fd3cbdc6a25dab5d06c7baa00af16d8f62fc5da687efe9c6cc91945de0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T22:49:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://90ed7fff816b1772ba9ecd95587a59cbed7c30ac302fff7e7c5de934c912bcf4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T22:49:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9b7a5502803eaad9f76f5ffa91b29fbb2657402065a03a4d6433c096aac7264f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"wait-for-host-port\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://9b7a5502803eaad9f76f5ffa91b29fbb2657402065a03a4d6433c096aac7264f\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-10T22:49:27Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-10T22:49:27Z\\\"}}}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-10T22:49:25Z\\\"}}\" for pod \"openshift-kube-scheduler\"/\"openshift-kube-scheduler-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T22:51:06Z is after 2025-08-24T17:21:41Z" Dec 10 22:51:06 crc kubenswrapper[4791]: I1210 22:51:06.085459 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 22:51:06 crc kubenswrapper[4791]: I1210 22:51:06.085497 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 22:51:06 crc kubenswrapper[4791]: I1210 22:51:06.085504 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 22:51:06 crc kubenswrapper[4791]: I1210 22:51:06.085518 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 22:51:06 crc kubenswrapper[4791]: I1210 22:51:06.085529 4791 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T22:51:06Z","lastTransitionTime":"2025-12-10T22:51:06Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 22:51:06 crc kubenswrapper[4791]: I1210 22:51:06.092554 4791 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"90250e13-6c63-443a-a5e5-6e4e341a2289\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T22:49:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T22:49:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T22:49:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T22:49:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T22:49:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ba54f95b1e00843b854db9a10fe6897592159ccacf30ed0f76002093a13eedd3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T22:49:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5aa8ffa518f3a68b60b275f8df4851b2564aa4984b61efdc9470b73072133e43\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T22:49:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://16f989503edc78d9451638e2cf6cf4b19a2cf2b8809b10b237541cb829303091\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T22:49:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://82556824d63ce6e46df14c45689ff1f81ebf729abe902061815c040b323898e9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T22:49:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://61db531940fa6f2effdae5a005b4c91bd2698bd5d1bff9588c171a65ad550eca\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T22:49:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0f65c864d885319ce2a08d61c30be24b3ffb7d23ea30c82ed7faf039cb2c006d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://0f65c864d885319ce2a08d61c30be24b3ffb7d23ea30c82ed7faf039cb2c006d\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-10T22:49:27Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-10T22:49:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e31a76d1d5b2812ba83e24e38268e25ec8bed98fc3da9adcd06bda5822b6ace1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e31a76d1d5b2812ba83e24e38268e25ec8bed98fc3da9adcd06bda5822b6ace1\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-10T22:49:30Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-10T22:49:29Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://b5b5eceae562f3986b543d41de1e977efb85a91d84c98e09159ca6d092eda691\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b5b5eceae562f3986b543d41de1e977efb85a91d84c98e09159ca6d092eda691\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-10T22:49:31Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-10T22:49:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-10T22:49:25Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T22:51:06Z is after 2025-08-24T17:21:41Z" Dec 10 22:51:06 crc kubenswrapper[4791]: I1210 22:51:06.102490 4791 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-10T22:49:50Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-10T22:49:50Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T22:51:06Z is after 2025-08-24T17:21:41Z" Dec 10 22:51:06 crc kubenswrapper[4791]: I1210 22:51:06.111264 4791 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-10T22:49:54Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-10T22:49:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-10T22:49:54Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://62bfaf5d7a6d1b452cdaf88086447d0c8cb7306e94434e46752be5475124d56a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T22:49:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T22:51:06Z is after 2025-08-24T17:21:41Z" Dec 10 22:51:06 crc kubenswrapper[4791]: I1210 22:51:06.120748 4791 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-b6kf6" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b4527e71-bb58-4a39-a99c-fe614aa4c5f9\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T22:50:04Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T22:50:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T22:50:04Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T22:50:04Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ngpr4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ngpr4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-10T22:50:04Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-b6kf6\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T22:51:06Z is after 2025-08-24T17:21:41Z" Dec 10 22:51:06 crc kubenswrapper[4791]: I1210 22:51:06.187791 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 22:51:06 crc kubenswrapper[4791]: I1210 22:51:06.187849 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 22:51:06 crc kubenswrapper[4791]: I1210 22:51:06.187861 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 22:51:06 crc kubenswrapper[4791]: I1210 22:51:06.187880 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 22:51:06 crc kubenswrapper[4791]: I1210 22:51:06.187892 4791 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T22:51:06Z","lastTransitionTime":"2025-12-10T22:51:06Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 22:51:06 crc kubenswrapper[4791]: I1210 22:51:06.290037 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 22:51:06 crc kubenswrapper[4791]: I1210 22:51:06.290081 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 22:51:06 crc kubenswrapper[4791]: I1210 22:51:06.290095 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 22:51:06 crc kubenswrapper[4791]: I1210 22:51:06.290109 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 22:51:06 crc kubenswrapper[4791]: I1210 22:51:06.290119 4791 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T22:51:06Z","lastTransitionTime":"2025-12-10T22:51:06Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 22:51:06 crc kubenswrapper[4791]: I1210 22:51:06.392461 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 22:51:06 crc kubenswrapper[4791]: I1210 22:51:06.392501 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 22:51:06 crc kubenswrapper[4791]: I1210 22:51:06.392512 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 22:51:06 crc kubenswrapper[4791]: I1210 22:51:06.392528 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 22:51:06 crc kubenswrapper[4791]: I1210 22:51:06.392538 4791 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T22:51:06Z","lastTransitionTime":"2025-12-10T22:51:06Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 22:51:06 crc kubenswrapper[4791]: I1210 22:51:06.495470 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 22:51:06 crc kubenswrapper[4791]: I1210 22:51:06.495542 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 22:51:06 crc kubenswrapper[4791]: I1210 22:51:06.495555 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 22:51:06 crc kubenswrapper[4791]: I1210 22:51:06.495577 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 22:51:06 crc kubenswrapper[4791]: I1210 22:51:06.495597 4791 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T22:51:06Z","lastTransitionTime":"2025-12-10T22:51:06Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 22:51:06 crc kubenswrapper[4791]: I1210 22:51:06.597544 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 22:51:06 crc kubenswrapper[4791]: I1210 22:51:06.597846 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 22:51:06 crc kubenswrapper[4791]: I1210 22:51:06.597961 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 22:51:06 crc kubenswrapper[4791]: I1210 22:51:06.598070 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 22:51:06 crc kubenswrapper[4791]: I1210 22:51:06.598188 4791 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T22:51:06Z","lastTransitionTime":"2025-12-10T22:51:06Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 22:51:06 crc kubenswrapper[4791]: I1210 22:51:06.700598 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 22:51:06 crc kubenswrapper[4791]: I1210 22:51:06.700676 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 22:51:06 crc kubenswrapper[4791]: I1210 22:51:06.700700 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 22:51:06 crc kubenswrapper[4791]: I1210 22:51:06.700731 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 22:51:06 crc kubenswrapper[4791]: I1210 22:51:06.700754 4791 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T22:51:06Z","lastTransitionTime":"2025-12-10T22:51:06Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 22:51:06 crc kubenswrapper[4791]: I1210 22:51:06.803264 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 22:51:06 crc kubenswrapper[4791]: I1210 22:51:06.803299 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 22:51:06 crc kubenswrapper[4791]: I1210 22:51:06.803306 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 22:51:06 crc kubenswrapper[4791]: I1210 22:51:06.803359 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 22:51:06 crc kubenswrapper[4791]: I1210 22:51:06.803372 4791 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T22:51:06Z","lastTransitionTime":"2025-12-10T22:51:06Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 22:51:06 crc kubenswrapper[4791]: I1210 22:51:06.884152 4791 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 10 22:51:06 crc kubenswrapper[4791]: I1210 22:51:06.884695 4791 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 10 22:51:06 crc kubenswrapper[4791]: I1210 22:51:06.884904 4791 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-b6kf6" Dec 10 22:51:06 crc kubenswrapper[4791]: I1210 22:51:06.884948 4791 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 10 22:51:06 crc kubenswrapper[4791]: E1210 22:51:06.885051 4791 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 10 22:51:06 crc kubenswrapper[4791]: E1210 22:51:06.885187 4791 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 10 22:51:06 crc kubenswrapper[4791]: E1210 22:51:06.885406 4791 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-b6kf6" podUID="b4527e71-bb58-4a39-a99c-fe614aa4c5f9" Dec 10 22:51:06 crc kubenswrapper[4791]: E1210 22:51:06.885438 4791 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 10 22:51:06 crc kubenswrapper[4791]: I1210 22:51:06.906285 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 22:51:06 crc kubenswrapper[4791]: I1210 22:51:06.906376 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 22:51:06 crc kubenswrapper[4791]: I1210 22:51:06.906420 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 22:51:06 crc kubenswrapper[4791]: I1210 22:51:06.906446 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 22:51:06 crc kubenswrapper[4791]: I1210 22:51:06.906462 4791 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T22:51:06Z","lastTransitionTime":"2025-12-10T22:51:06Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 22:51:07 crc kubenswrapper[4791]: I1210 22:51:07.009144 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 22:51:07 crc kubenswrapper[4791]: I1210 22:51:07.009180 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 22:51:07 crc kubenswrapper[4791]: I1210 22:51:07.009188 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 22:51:07 crc kubenswrapper[4791]: I1210 22:51:07.009201 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 22:51:07 crc kubenswrapper[4791]: I1210 22:51:07.009210 4791 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T22:51:07Z","lastTransitionTime":"2025-12-10T22:51:07Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 22:51:07 crc kubenswrapper[4791]: I1210 22:51:07.111217 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 22:51:07 crc kubenswrapper[4791]: I1210 22:51:07.111271 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 22:51:07 crc kubenswrapper[4791]: I1210 22:51:07.111287 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 22:51:07 crc kubenswrapper[4791]: I1210 22:51:07.111309 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 22:51:07 crc kubenswrapper[4791]: I1210 22:51:07.111324 4791 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T22:51:07Z","lastTransitionTime":"2025-12-10T22:51:07Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 22:51:07 crc kubenswrapper[4791]: I1210 22:51:07.213660 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 22:51:07 crc kubenswrapper[4791]: I1210 22:51:07.213692 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 22:51:07 crc kubenswrapper[4791]: I1210 22:51:07.213701 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 22:51:07 crc kubenswrapper[4791]: I1210 22:51:07.213730 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 22:51:07 crc kubenswrapper[4791]: I1210 22:51:07.213739 4791 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T22:51:07Z","lastTransitionTime":"2025-12-10T22:51:07Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 22:51:07 crc kubenswrapper[4791]: I1210 22:51:07.316682 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 22:51:07 crc kubenswrapper[4791]: I1210 22:51:07.316744 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 22:51:07 crc kubenswrapper[4791]: I1210 22:51:07.316765 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 22:51:07 crc kubenswrapper[4791]: I1210 22:51:07.316787 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 22:51:07 crc kubenswrapper[4791]: I1210 22:51:07.316806 4791 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T22:51:07Z","lastTransitionTime":"2025-12-10T22:51:07Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 22:51:07 crc kubenswrapper[4791]: I1210 22:51:07.419076 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 22:51:07 crc kubenswrapper[4791]: I1210 22:51:07.419292 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 22:51:07 crc kubenswrapper[4791]: I1210 22:51:07.419301 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 22:51:07 crc kubenswrapper[4791]: I1210 22:51:07.419314 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 22:51:07 crc kubenswrapper[4791]: I1210 22:51:07.419324 4791 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T22:51:07Z","lastTransitionTime":"2025-12-10T22:51:07Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 22:51:07 crc kubenswrapper[4791]: I1210 22:51:07.521780 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 22:51:07 crc kubenswrapper[4791]: I1210 22:51:07.521814 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 22:51:07 crc kubenswrapper[4791]: I1210 22:51:07.521822 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 22:51:07 crc kubenswrapper[4791]: I1210 22:51:07.521834 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 22:51:07 crc kubenswrapper[4791]: I1210 22:51:07.521842 4791 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T22:51:07Z","lastTransitionTime":"2025-12-10T22:51:07Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 22:51:07 crc kubenswrapper[4791]: I1210 22:51:07.623480 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 22:51:07 crc kubenswrapper[4791]: I1210 22:51:07.623534 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 22:51:07 crc kubenswrapper[4791]: I1210 22:51:07.623545 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 22:51:07 crc kubenswrapper[4791]: I1210 22:51:07.623559 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 22:51:07 crc kubenswrapper[4791]: I1210 22:51:07.623569 4791 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T22:51:07Z","lastTransitionTime":"2025-12-10T22:51:07Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 22:51:07 crc kubenswrapper[4791]: I1210 22:51:07.725992 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 22:51:07 crc kubenswrapper[4791]: I1210 22:51:07.726026 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 22:51:07 crc kubenswrapper[4791]: I1210 22:51:07.726034 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 22:51:07 crc kubenswrapper[4791]: I1210 22:51:07.726048 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 22:51:07 crc kubenswrapper[4791]: I1210 22:51:07.726057 4791 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T22:51:07Z","lastTransitionTime":"2025-12-10T22:51:07Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 22:51:07 crc kubenswrapper[4791]: I1210 22:51:07.828810 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 22:51:07 crc kubenswrapper[4791]: I1210 22:51:07.828847 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 22:51:07 crc kubenswrapper[4791]: I1210 22:51:07.828857 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 22:51:07 crc kubenswrapper[4791]: I1210 22:51:07.828871 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 22:51:07 crc kubenswrapper[4791]: I1210 22:51:07.828880 4791 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T22:51:07Z","lastTransitionTime":"2025-12-10T22:51:07Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 22:51:07 crc kubenswrapper[4791]: I1210 22:51:07.931059 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 22:51:07 crc kubenswrapper[4791]: I1210 22:51:07.931100 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 22:51:07 crc kubenswrapper[4791]: I1210 22:51:07.931116 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 22:51:07 crc kubenswrapper[4791]: I1210 22:51:07.931133 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 22:51:07 crc kubenswrapper[4791]: I1210 22:51:07.931145 4791 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T22:51:07Z","lastTransitionTime":"2025-12-10T22:51:07Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 22:51:08 crc kubenswrapper[4791]: I1210 22:51:08.033802 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 22:51:08 crc kubenswrapper[4791]: I1210 22:51:08.033867 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 22:51:08 crc kubenswrapper[4791]: I1210 22:51:08.033876 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 22:51:08 crc kubenswrapper[4791]: I1210 22:51:08.033890 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 22:51:08 crc kubenswrapper[4791]: I1210 22:51:08.033899 4791 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T22:51:08Z","lastTransitionTime":"2025-12-10T22:51:08Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 22:51:08 crc kubenswrapper[4791]: I1210 22:51:08.136719 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 22:51:08 crc kubenswrapper[4791]: I1210 22:51:08.137169 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 22:51:08 crc kubenswrapper[4791]: I1210 22:51:08.137211 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 22:51:08 crc kubenswrapper[4791]: I1210 22:51:08.137246 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 22:51:08 crc kubenswrapper[4791]: I1210 22:51:08.137271 4791 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T22:51:08Z","lastTransitionTime":"2025-12-10T22:51:08Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 22:51:08 crc kubenswrapper[4791]: I1210 22:51:08.240981 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 22:51:08 crc kubenswrapper[4791]: I1210 22:51:08.241027 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 22:51:08 crc kubenswrapper[4791]: I1210 22:51:08.241046 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 22:51:08 crc kubenswrapper[4791]: I1210 22:51:08.241070 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 22:51:08 crc kubenswrapper[4791]: I1210 22:51:08.241088 4791 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T22:51:08Z","lastTransitionTime":"2025-12-10T22:51:08Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 22:51:08 crc kubenswrapper[4791]: I1210 22:51:08.343164 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 22:51:08 crc kubenswrapper[4791]: I1210 22:51:08.343547 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 22:51:08 crc kubenswrapper[4791]: I1210 22:51:08.343701 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 22:51:08 crc kubenswrapper[4791]: I1210 22:51:08.343880 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 22:51:08 crc kubenswrapper[4791]: I1210 22:51:08.344014 4791 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T22:51:08Z","lastTransitionTime":"2025-12-10T22:51:08Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 22:51:08 crc kubenswrapper[4791]: I1210 22:51:08.447331 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 22:51:08 crc kubenswrapper[4791]: I1210 22:51:08.447421 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 22:51:08 crc kubenswrapper[4791]: I1210 22:51:08.447441 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 22:51:08 crc kubenswrapper[4791]: I1210 22:51:08.447474 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 22:51:08 crc kubenswrapper[4791]: I1210 22:51:08.447493 4791 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T22:51:08Z","lastTransitionTime":"2025-12-10T22:51:08Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 22:51:08 crc kubenswrapper[4791]: I1210 22:51:08.550850 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 22:51:08 crc kubenswrapper[4791]: I1210 22:51:08.550916 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 22:51:08 crc kubenswrapper[4791]: I1210 22:51:08.550935 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 22:51:08 crc kubenswrapper[4791]: I1210 22:51:08.550961 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 22:51:08 crc kubenswrapper[4791]: I1210 22:51:08.550987 4791 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T22:51:08Z","lastTransitionTime":"2025-12-10T22:51:08Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 22:51:08 crc kubenswrapper[4791]: I1210 22:51:08.653669 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 22:51:08 crc kubenswrapper[4791]: I1210 22:51:08.653724 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 22:51:08 crc kubenswrapper[4791]: I1210 22:51:08.653736 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 22:51:08 crc kubenswrapper[4791]: I1210 22:51:08.653754 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 22:51:08 crc kubenswrapper[4791]: I1210 22:51:08.653769 4791 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T22:51:08Z","lastTransitionTime":"2025-12-10T22:51:08Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 22:51:08 crc kubenswrapper[4791]: I1210 22:51:08.756599 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 22:51:08 crc kubenswrapper[4791]: I1210 22:51:08.756679 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 22:51:08 crc kubenswrapper[4791]: I1210 22:51:08.756704 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 22:51:08 crc kubenswrapper[4791]: I1210 22:51:08.756735 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 22:51:08 crc kubenswrapper[4791]: I1210 22:51:08.756759 4791 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T22:51:08Z","lastTransitionTime":"2025-12-10T22:51:08Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 22:51:08 crc kubenswrapper[4791]: I1210 22:51:08.860331 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 22:51:08 crc kubenswrapper[4791]: I1210 22:51:08.860435 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 22:51:08 crc kubenswrapper[4791]: I1210 22:51:08.860453 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 22:51:08 crc kubenswrapper[4791]: I1210 22:51:08.860478 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 22:51:08 crc kubenswrapper[4791]: I1210 22:51:08.860497 4791 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T22:51:08Z","lastTransitionTime":"2025-12-10T22:51:08Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 22:51:08 crc kubenswrapper[4791]: I1210 22:51:08.863029 4791 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/b4527e71-bb58-4a39-a99c-fe614aa4c5f9-metrics-certs\") pod \"network-metrics-daemon-b6kf6\" (UID: \"b4527e71-bb58-4a39-a99c-fe614aa4c5f9\") " pod="openshift-multus/network-metrics-daemon-b6kf6" Dec 10 22:51:08 crc kubenswrapper[4791]: E1210 22:51:08.863250 4791 secret.go:188] Couldn't get secret openshift-multus/metrics-daemon-secret: object "openshift-multus"/"metrics-daemon-secret" not registered Dec 10 22:51:08 crc kubenswrapper[4791]: E1210 22:51:08.863372 4791 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/b4527e71-bb58-4a39-a99c-fe614aa4c5f9-metrics-certs podName:b4527e71-bb58-4a39-a99c-fe614aa4c5f9 nodeName:}" failed. No retries permitted until 2025-12-10 22:52:12.863320291 +0000 UTC m=+167.292937934 (durationBeforeRetry 1m4s). Error: MountVolume.SetUp failed for volume "metrics-certs" (UniqueName: "kubernetes.io/secret/b4527e71-bb58-4a39-a99c-fe614aa4c5f9-metrics-certs") pod "network-metrics-daemon-b6kf6" (UID: "b4527e71-bb58-4a39-a99c-fe614aa4c5f9") : object "openshift-multus"/"metrics-daemon-secret" not registered Dec 10 22:51:08 crc kubenswrapper[4791]: I1210 22:51:08.883974 4791 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-b6kf6" Dec 10 22:51:08 crc kubenswrapper[4791]: I1210 22:51:08.884067 4791 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 10 22:51:08 crc kubenswrapper[4791]: E1210 22:51:08.884170 4791 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-b6kf6" podUID="b4527e71-bb58-4a39-a99c-fe614aa4c5f9" Dec 10 22:51:08 crc kubenswrapper[4791]: I1210 22:51:08.884274 4791 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 10 22:51:08 crc kubenswrapper[4791]: I1210 22:51:08.884374 4791 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 10 22:51:08 crc kubenswrapper[4791]: E1210 22:51:08.884536 4791 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 10 22:51:08 crc kubenswrapper[4791]: E1210 22:51:08.884676 4791 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 10 22:51:08 crc kubenswrapper[4791]: E1210 22:51:08.884895 4791 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 10 22:51:08 crc kubenswrapper[4791]: I1210 22:51:08.963572 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 22:51:08 crc kubenswrapper[4791]: I1210 22:51:08.963632 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 22:51:08 crc kubenswrapper[4791]: I1210 22:51:08.963650 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 22:51:08 crc kubenswrapper[4791]: I1210 22:51:08.963674 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 22:51:08 crc kubenswrapper[4791]: I1210 22:51:08.963691 4791 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T22:51:08Z","lastTransitionTime":"2025-12-10T22:51:08Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 22:51:09 crc kubenswrapper[4791]: I1210 22:51:09.066728 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 22:51:09 crc kubenswrapper[4791]: I1210 22:51:09.066777 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 22:51:09 crc kubenswrapper[4791]: I1210 22:51:09.066788 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 22:51:09 crc kubenswrapper[4791]: I1210 22:51:09.066803 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 22:51:09 crc kubenswrapper[4791]: I1210 22:51:09.066815 4791 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T22:51:09Z","lastTransitionTime":"2025-12-10T22:51:09Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 22:51:09 crc kubenswrapper[4791]: I1210 22:51:09.170037 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 22:51:09 crc kubenswrapper[4791]: I1210 22:51:09.170156 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 22:51:09 crc kubenswrapper[4791]: I1210 22:51:09.170225 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 22:51:09 crc kubenswrapper[4791]: I1210 22:51:09.170277 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 22:51:09 crc kubenswrapper[4791]: I1210 22:51:09.170301 4791 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T22:51:09Z","lastTransitionTime":"2025-12-10T22:51:09Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 22:51:09 crc kubenswrapper[4791]: I1210 22:51:09.273402 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 22:51:09 crc kubenswrapper[4791]: I1210 22:51:09.273482 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 22:51:09 crc kubenswrapper[4791]: I1210 22:51:09.273500 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 22:51:09 crc kubenswrapper[4791]: I1210 22:51:09.273523 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 22:51:09 crc kubenswrapper[4791]: I1210 22:51:09.273543 4791 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T22:51:09Z","lastTransitionTime":"2025-12-10T22:51:09Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 22:51:09 crc kubenswrapper[4791]: I1210 22:51:09.304972 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 22:51:09 crc kubenswrapper[4791]: I1210 22:51:09.305086 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 22:51:09 crc kubenswrapper[4791]: I1210 22:51:09.305106 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 22:51:09 crc kubenswrapper[4791]: I1210 22:51:09.305148 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 22:51:09 crc kubenswrapper[4791]: I1210 22:51:09.305166 4791 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T22:51:09Z","lastTransitionTime":"2025-12-10T22:51:09Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 22:51:09 crc kubenswrapper[4791]: E1210 22:51:09.327710 4791 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-12-10T22:51:09Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-10T22:51:09Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-10T22:51:09Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-10T22:51:09Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-10T22:51:09Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-10T22:51:09Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-10T22:51:09Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-10T22:51:09Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"4fab3af6-f657-4dfc-8ef2-12b7c978c94f\\\",\\\"systemUUID\\\":\\\"442c0b68-8dce-47e0-bd68-d5210d7e0493\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T22:51:09Z is after 2025-08-24T17:21:41Z" Dec 10 22:51:09 crc kubenswrapper[4791]: I1210 22:51:09.333104 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 22:51:09 crc kubenswrapper[4791]: I1210 22:51:09.333188 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 22:51:09 crc kubenswrapper[4791]: I1210 22:51:09.333206 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 22:51:09 crc kubenswrapper[4791]: I1210 22:51:09.333636 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 22:51:09 crc kubenswrapper[4791]: I1210 22:51:09.333696 4791 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T22:51:09Z","lastTransitionTime":"2025-12-10T22:51:09Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 22:51:09 crc kubenswrapper[4791]: E1210 22:51:09.350522 4791 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-12-10T22:51:09Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-10T22:51:09Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-10T22:51:09Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-10T22:51:09Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-10T22:51:09Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-10T22:51:09Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-10T22:51:09Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-10T22:51:09Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"4fab3af6-f657-4dfc-8ef2-12b7c978c94f\\\",\\\"systemUUID\\\":\\\"442c0b68-8dce-47e0-bd68-d5210d7e0493\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T22:51:09Z is after 2025-08-24T17:21:41Z" Dec 10 22:51:09 crc kubenswrapper[4791]: I1210 22:51:09.355214 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 22:51:09 crc kubenswrapper[4791]: I1210 22:51:09.355261 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 22:51:09 crc kubenswrapper[4791]: I1210 22:51:09.355273 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 22:51:09 crc kubenswrapper[4791]: I1210 22:51:09.355291 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 22:51:09 crc kubenswrapper[4791]: I1210 22:51:09.355303 4791 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T22:51:09Z","lastTransitionTime":"2025-12-10T22:51:09Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 22:51:09 crc kubenswrapper[4791]: E1210 22:51:09.374077 4791 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-12-10T22:51:09Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-10T22:51:09Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-10T22:51:09Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-10T22:51:09Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-10T22:51:09Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-10T22:51:09Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-10T22:51:09Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-10T22:51:09Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"4fab3af6-f657-4dfc-8ef2-12b7c978c94f\\\",\\\"systemUUID\\\":\\\"442c0b68-8dce-47e0-bd68-d5210d7e0493\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T22:51:09Z is after 2025-08-24T17:21:41Z" Dec 10 22:51:09 crc kubenswrapper[4791]: I1210 22:51:09.378148 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 22:51:09 crc kubenswrapper[4791]: I1210 22:51:09.378203 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 22:51:09 crc kubenswrapper[4791]: I1210 22:51:09.378219 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 22:51:09 crc kubenswrapper[4791]: I1210 22:51:09.378240 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 22:51:09 crc kubenswrapper[4791]: I1210 22:51:09.378256 4791 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T22:51:09Z","lastTransitionTime":"2025-12-10T22:51:09Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 22:51:09 crc kubenswrapper[4791]: E1210 22:51:09.400151 4791 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-12-10T22:51:09Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-10T22:51:09Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-10T22:51:09Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-10T22:51:09Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-10T22:51:09Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-10T22:51:09Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-10T22:51:09Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-10T22:51:09Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"4fab3af6-f657-4dfc-8ef2-12b7c978c94f\\\",\\\"systemUUID\\\":\\\"442c0b68-8dce-47e0-bd68-d5210d7e0493\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T22:51:09Z is after 2025-08-24T17:21:41Z" Dec 10 22:51:09 crc kubenswrapper[4791]: I1210 22:51:09.404607 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 22:51:09 crc kubenswrapper[4791]: I1210 22:51:09.404658 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 22:51:09 crc kubenswrapper[4791]: I1210 22:51:09.404672 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 22:51:09 crc kubenswrapper[4791]: I1210 22:51:09.404725 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 22:51:09 crc kubenswrapper[4791]: I1210 22:51:09.404738 4791 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T22:51:09Z","lastTransitionTime":"2025-12-10T22:51:09Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 22:51:09 crc kubenswrapper[4791]: E1210 22:51:09.424220 4791 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-12-10T22:51:09Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-10T22:51:09Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-10T22:51:09Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-10T22:51:09Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-10T22:51:09Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-10T22:51:09Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-10T22:51:09Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-10T22:51:09Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"4fab3af6-f657-4dfc-8ef2-12b7c978c94f\\\",\\\"systemUUID\\\":\\\"442c0b68-8dce-47e0-bd68-d5210d7e0493\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T22:51:09Z is after 2025-08-24T17:21:41Z" Dec 10 22:51:09 crc kubenswrapper[4791]: E1210 22:51:09.424503 4791 kubelet_node_status.go:572] "Unable to update node status" err="update node status exceeds retry count" Dec 10 22:51:09 crc kubenswrapper[4791]: I1210 22:51:09.426332 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 22:51:09 crc kubenswrapper[4791]: I1210 22:51:09.426405 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 22:51:09 crc kubenswrapper[4791]: I1210 22:51:09.426416 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 22:51:09 crc kubenswrapper[4791]: I1210 22:51:09.426431 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 22:51:09 crc kubenswrapper[4791]: I1210 22:51:09.426443 4791 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T22:51:09Z","lastTransitionTime":"2025-12-10T22:51:09Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 22:51:09 crc kubenswrapper[4791]: I1210 22:51:09.529093 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 22:51:09 crc kubenswrapper[4791]: I1210 22:51:09.529125 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 22:51:09 crc kubenswrapper[4791]: I1210 22:51:09.529135 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 22:51:09 crc kubenswrapper[4791]: I1210 22:51:09.529149 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 22:51:09 crc kubenswrapper[4791]: I1210 22:51:09.529158 4791 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T22:51:09Z","lastTransitionTime":"2025-12-10T22:51:09Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 22:51:09 crc kubenswrapper[4791]: I1210 22:51:09.631534 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 22:51:09 crc kubenswrapper[4791]: I1210 22:51:09.631626 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 22:51:09 crc kubenswrapper[4791]: I1210 22:51:09.631653 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 22:51:09 crc kubenswrapper[4791]: I1210 22:51:09.631690 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 22:51:09 crc kubenswrapper[4791]: I1210 22:51:09.631715 4791 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T22:51:09Z","lastTransitionTime":"2025-12-10T22:51:09Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 22:51:09 crc kubenswrapper[4791]: I1210 22:51:09.734170 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 22:51:09 crc kubenswrapper[4791]: I1210 22:51:09.734211 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 22:51:09 crc kubenswrapper[4791]: I1210 22:51:09.734221 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 22:51:09 crc kubenswrapper[4791]: I1210 22:51:09.734238 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 22:51:09 crc kubenswrapper[4791]: I1210 22:51:09.734249 4791 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T22:51:09Z","lastTransitionTime":"2025-12-10T22:51:09Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 22:51:09 crc kubenswrapper[4791]: I1210 22:51:09.837265 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 22:51:09 crc kubenswrapper[4791]: I1210 22:51:09.837306 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 22:51:09 crc kubenswrapper[4791]: I1210 22:51:09.837316 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 22:51:09 crc kubenswrapper[4791]: I1210 22:51:09.837331 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 22:51:09 crc kubenswrapper[4791]: I1210 22:51:09.837366 4791 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T22:51:09Z","lastTransitionTime":"2025-12-10T22:51:09Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 22:51:09 crc kubenswrapper[4791]: I1210 22:51:09.940152 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 22:51:09 crc kubenswrapper[4791]: I1210 22:51:09.940427 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 22:51:09 crc kubenswrapper[4791]: I1210 22:51:09.940457 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 22:51:09 crc kubenswrapper[4791]: I1210 22:51:09.940483 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 22:51:09 crc kubenswrapper[4791]: I1210 22:51:09.940503 4791 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T22:51:09Z","lastTransitionTime":"2025-12-10T22:51:09Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 22:51:10 crc kubenswrapper[4791]: I1210 22:51:10.043597 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 22:51:10 crc kubenswrapper[4791]: I1210 22:51:10.043673 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 22:51:10 crc kubenswrapper[4791]: I1210 22:51:10.043688 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 22:51:10 crc kubenswrapper[4791]: I1210 22:51:10.043709 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 22:51:10 crc kubenswrapper[4791]: I1210 22:51:10.043722 4791 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T22:51:10Z","lastTransitionTime":"2025-12-10T22:51:10Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 22:51:10 crc kubenswrapper[4791]: I1210 22:51:10.147245 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 22:51:10 crc kubenswrapper[4791]: I1210 22:51:10.147313 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 22:51:10 crc kubenswrapper[4791]: I1210 22:51:10.147334 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 22:51:10 crc kubenswrapper[4791]: I1210 22:51:10.147390 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 22:51:10 crc kubenswrapper[4791]: I1210 22:51:10.147411 4791 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T22:51:10Z","lastTransitionTime":"2025-12-10T22:51:10Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 22:51:10 crc kubenswrapper[4791]: I1210 22:51:10.250976 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 22:51:10 crc kubenswrapper[4791]: I1210 22:51:10.251047 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 22:51:10 crc kubenswrapper[4791]: I1210 22:51:10.251064 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 22:51:10 crc kubenswrapper[4791]: I1210 22:51:10.251092 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 22:51:10 crc kubenswrapper[4791]: I1210 22:51:10.251110 4791 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T22:51:10Z","lastTransitionTime":"2025-12-10T22:51:10Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 22:51:10 crc kubenswrapper[4791]: I1210 22:51:10.354763 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 22:51:10 crc kubenswrapper[4791]: I1210 22:51:10.354826 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 22:51:10 crc kubenswrapper[4791]: I1210 22:51:10.354845 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 22:51:10 crc kubenswrapper[4791]: I1210 22:51:10.354871 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 22:51:10 crc kubenswrapper[4791]: I1210 22:51:10.354890 4791 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T22:51:10Z","lastTransitionTime":"2025-12-10T22:51:10Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 22:51:10 crc kubenswrapper[4791]: I1210 22:51:10.458737 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 22:51:10 crc kubenswrapper[4791]: I1210 22:51:10.458808 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 22:51:10 crc kubenswrapper[4791]: I1210 22:51:10.458827 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 22:51:10 crc kubenswrapper[4791]: I1210 22:51:10.458852 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 22:51:10 crc kubenswrapper[4791]: I1210 22:51:10.458870 4791 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T22:51:10Z","lastTransitionTime":"2025-12-10T22:51:10Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 22:51:10 crc kubenswrapper[4791]: I1210 22:51:10.561809 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 22:51:10 crc kubenswrapper[4791]: I1210 22:51:10.561896 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 22:51:10 crc kubenswrapper[4791]: I1210 22:51:10.561920 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 22:51:10 crc kubenswrapper[4791]: I1210 22:51:10.561954 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 22:51:10 crc kubenswrapper[4791]: I1210 22:51:10.561980 4791 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T22:51:10Z","lastTransitionTime":"2025-12-10T22:51:10Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 22:51:10 crc kubenswrapper[4791]: I1210 22:51:10.664574 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 22:51:10 crc kubenswrapper[4791]: I1210 22:51:10.664647 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 22:51:10 crc kubenswrapper[4791]: I1210 22:51:10.664668 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 22:51:10 crc kubenswrapper[4791]: I1210 22:51:10.664695 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 22:51:10 crc kubenswrapper[4791]: I1210 22:51:10.664717 4791 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T22:51:10Z","lastTransitionTime":"2025-12-10T22:51:10Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 22:51:10 crc kubenswrapper[4791]: I1210 22:51:10.767277 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 22:51:10 crc kubenswrapper[4791]: I1210 22:51:10.767326 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 22:51:10 crc kubenswrapper[4791]: I1210 22:51:10.767353 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 22:51:10 crc kubenswrapper[4791]: I1210 22:51:10.767370 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 22:51:10 crc kubenswrapper[4791]: I1210 22:51:10.767382 4791 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T22:51:10Z","lastTransitionTime":"2025-12-10T22:51:10Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 22:51:10 crc kubenswrapper[4791]: I1210 22:51:10.869329 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 22:51:10 crc kubenswrapper[4791]: I1210 22:51:10.869394 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 22:51:10 crc kubenswrapper[4791]: I1210 22:51:10.869403 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 22:51:10 crc kubenswrapper[4791]: I1210 22:51:10.869420 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 22:51:10 crc kubenswrapper[4791]: I1210 22:51:10.869428 4791 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T22:51:10Z","lastTransitionTime":"2025-12-10T22:51:10Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 22:51:10 crc kubenswrapper[4791]: I1210 22:51:10.883968 4791 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-b6kf6" Dec 10 22:51:10 crc kubenswrapper[4791]: I1210 22:51:10.884017 4791 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 10 22:51:10 crc kubenswrapper[4791]: I1210 22:51:10.884052 4791 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 10 22:51:10 crc kubenswrapper[4791]: I1210 22:51:10.883973 4791 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 10 22:51:10 crc kubenswrapper[4791]: E1210 22:51:10.884181 4791 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-b6kf6" podUID="b4527e71-bb58-4a39-a99c-fe614aa4c5f9" Dec 10 22:51:10 crc kubenswrapper[4791]: E1210 22:51:10.884254 4791 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 10 22:51:10 crc kubenswrapper[4791]: E1210 22:51:10.884416 4791 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 10 22:51:10 crc kubenswrapper[4791]: E1210 22:51:10.884574 4791 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 10 22:51:10 crc kubenswrapper[4791]: I1210 22:51:10.971645 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 22:51:10 crc kubenswrapper[4791]: I1210 22:51:10.971699 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 22:51:10 crc kubenswrapper[4791]: I1210 22:51:10.971713 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 22:51:10 crc kubenswrapper[4791]: I1210 22:51:10.971733 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 22:51:10 crc kubenswrapper[4791]: I1210 22:51:10.971748 4791 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T22:51:10Z","lastTransitionTime":"2025-12-10T22:51:10Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 22:51:11 crc kubenswrapper[4791]: I1210 22:51:11.074584 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 22:51:11 crc kubenswrapper[4791]: I1210 22:51:11.074633 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 22:51:11 crc kubenswrapper[4791]: I1210 22:51:11.074644 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 22:51:11 crc kubenswrapper[4791]: I1210 22:51:11.074662 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 22:51:11 crc kubenswrapper[4791]: I1210 22:51:11.074672 4791 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T22:51:11Z","lastTransitionTime":"2025-12-10T22:51:11Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 22:51:11 crc kubenswrapper[4791]: I1210 22:51:11.177259 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 22:51:11 crc kubenswrapper[4791]: I1210 22:51:11.177316 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 22:51:11 crc kubenswrapper[4791]: I1210 22:51:11.177330 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 22:51:11 crc kubenswrapper[4791]: I1210 22:51:11.177386 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 22:51:11 crc kubenswrapper[4791]: I1210 22:51:11.177399 4791 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T22:51:11Z","lastTransitionTime":"2025-12-10T22:51:11Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 22:51:11 crc kubenswrapper[4791]: I1210 22:51:11.279899 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 22:51:11 crc kubenswrapper[4791]: I1210 22:51:11.279935 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 22:51:11 crc kubenswrapper[4791]: I1210 22:51:11.279944 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 22:51:11 crc kubenswrapper[4791]: I1210 22:51:11.279959 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 22:51:11 crc kubenswrapper[4791]: I1210 22:51:11.279969 4791 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T22:51:11Z","lastTransitionTime":"2025-12-10T22:51:11Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 22:51:11 crc kubenswrapper[4791]: I1210 22:51:11.382858 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 22:51:11 crc kubenswrapper[4791]: I1210 22:51:11.382922 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 22:51:11 crc kubenswrapper[4791]: I1210 22:51:11.382940 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 22:51:11 crc kubenswrapper[4791]: I1210 22:51:11.382962 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 22:51:11 crc kubenswrapper[4791]: I1210 22:51:11.382980 4791 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T22:51:11Z","lastTransitionTime":"2025-12-10T22:51:11Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 22:51:11 crc kubenswrapper[4791]: I1210 22:51:11.486636 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 22:51:11 crc kubenswrapper[4791]: I1210 22:51:11.487035 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 22:51:11 crc kubenswrapper[4791]: I1210 22:51:11.487252 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 22:51:11 crc kubenswrapper[4791]: I1210 22:51:11.487534 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 22:51:11 crc kubenswrapper[4791]: I1210 22:51:11.487761 4791 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T22:51:11Z","lastTransitionTime":"2025-12-10T22:51:11Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 22:51:11 crc kubenswrapper[4791]: I1210 22:51:11.590403 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 22:51:11 crc kubenswrapper[4791]: I1210 22:51:11.590442 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 22:51:11 crc kubenswrapper[4791]: I1210 22:51:11.590452 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 22:51:11 crc kubenswrapper[4791]: I1210 22:51:11.590467 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 22:51:11 crc kubenswrapper[4791]: I1210 22:51:11.590476 4791 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T22:51:11Z","lastTransitionTime":"2025-12-10T22:51:11Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 22:51:11 crc kubenswrapper[4791]: I1210 22:51:11.692759 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 22:51:11 crc kubenswrapper[4791]: I1210 22:51:11.692795 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 22:51:11 crc kubenswrapper[4791]: I1210 22:51:11.692806 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 22:51:11 crc kubenswrapper[4791]: I1210 22:51:11.692821 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 22:51:11 crc kubenswrapper[4791]: I1210 22:51:11.692834 4791 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T22:51:11Z","lastTransitionTime":"2025-12-10T22:51:11Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 22:51:11 crc kubenswrapper[4791]: I1210 22:51:11.794608 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 22:51:11 crc kubenswrapper[4791]: I1210 22:51:11.794649 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 22:51:11 crc kubenswrapper[4791]: I1210 22:51:11.794662 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 22:51:11 crc kubenswrapper[4791]: I1210 22:51:11.794677 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 22:51:11 crc kubenswrapper[4791]: I1210 22:51:11.794687 4791 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T22:51:11Z","lastTransitionTime":"2025-12-10T22:51:11Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 22:51:11 crc kubenswrapper[4791]: I1210 22:51:11.896471 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 22:51:11 crc kubenswrapper[4791]: I1210 22:51:11.896543 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 22:51:11 crc kubenswrapper[4791]: I1210 22:51:11.896562 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 22:51:11 crc kubenswrapper[4791]: I1210 22:51:11.896589 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 22:51:11 crc kubenswrapper[4791]: I1210 22:51:11.896606 4791 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T22:51:11Z","lastTransitionTime":"2025-12-10T22:51:11Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 22:51:11 crc kubenswrapper[4791]: I1210 22:51:11.999507 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 22:51:11 crc kubenswrapper[4791]: I1210 22:51:11.999551 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 22:51:11 crc kubenswrapper[4791]: I1210 22:51:11.999562 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 22:51:11 crc kubenswrapper[4791]: I1210 22:51:11.999580 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 22:51:11 crc kubenswrapper[4791]: I1210 22:51:11.999592 4791 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T22:51:11Z","lastTransitionTime":"2025-12-10T22:51:11Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 22:51:12 crc kubenswrapper[4791]: I1210 22:51:12.103515 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 22:51:12 crc kubenswrapper[4791]: I1210 22:51:12.103579 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 22:51:12 crc kubenswrapper[4791]: I1210 22:51:12.103588 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 22:51:12 crc kubenswrapper[4791]: I1210 22:51:12.103621 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 22:51:12 crc kubenswrapper[4791]: I1210 22:51:12.103631 4791 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T22:51:12Z","lastTransitionTime":"2025-12-10T22:51:12Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 22:51:12 crc kubenswrapper[4791]: I1210 22:51:12.205917 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 22:51:12 crc kubenswrapper[4791]: I1210 22:51:12.205948 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 22:51:12 crc kubenswrapper[4791]: I1210 22:51:12.205957 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 22:51:12 crc kubenswrapper[4791]: I1210 22:51:12.205970 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 22:51:12 crc kubenswrapper[4791]: I1210 22:51:12.205983 4791 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T22:51:12Z","lastTransitionTime":"2025-12-10T22:51:12Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 22:51:12 crc kubenswrapper[4791]: I1210 22:51:12.308715 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 22:51:12 crc kubenswrapper[4791]: I1210 22:51:12.308785 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 22:51:12 crc kubenswrapper[4791]: I1210 22:51:12.308809 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 22:51:12 crc kubenswrapper[4791]: I1210 22:51:12.308844 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 22:51:12 crc kubenswrapper[4791]: I1210 22:51:12.308871 4791 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T22:51:12Z","lastTransitionTime":"2025-12-10T22:51:12Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 22:51:12 crc kubenswrapper[4791]: I1210 22:51:12.411999 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 22:51:12 crc kubenswrapper[4791]: I1210 22:51:12.412064 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 22:51:12 crc kubenswrapper[4791]: I1210 22:51:12.412088 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 22:51:12 crc kubenswrapper[4791]: I1210 22:51:12.412115 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 22:51:12 crc kubenswrapper[4791]: I1210 22:51:12.412137 4791 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T22:51:12Z","lastTransitionTime":"2025-12-10T22:51:12Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 22:51:12 crc kubenswrapper[4791]: I1210 22:51:12.514369 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 22:51:12 crc kubenswrapper[4791]: I1210 22:51:12.514419 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 22:51:12 crc kubenswrapper[4791]: I1210 22:51:12.514432 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 22:51:12 crc kubenswrapper[4791]: I1210 22:51:12.514451 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 22:51:12 crc kubenswrapper[4791]: I1210 22:51:12.514464 4791 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T22:51:12Z","lastTransitionTime":"2025-12-10T22:51:12Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 22:51:12 crc kubenswrapper[4791]: I1210 22:51:12.617599 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 22:51:12 crc kubenswrapper[4791]: I1210 22:51:12.617656 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 22:51:12 crc kubenswrapper[4791]: I1210 22:51:12.617668 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 22:51:12 crc kubenswrapper[4791]: I1210 22:51:12.617687 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 22:51:12 crc kubenswrapper[4791]: I1210 22:51:12.617698 4791 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T22:51:12Z","lastTransitionTime":"2025-12-10T22:51:12Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 22:51:12 crc kubenswrapper[4791]: I1210 22:51:12.721263 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 22:51:12 crc kubenswrapper[4791]: I1210 22:51:12.721307 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 22:51:12 crc kubenswrapper[4791]: I1210 22:51:12.721322 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 22:51:12 crc kubenswrapper[4791]: I1210 22:51:12.721358 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 22:51:12 crc kubenswrapper[4791]: I1210 22:51:12.721372 4791 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T22:51:12Z","lastTransitionTime":"2025-12-10T22:51:12Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 22:51:12 crc kubenswrapper[4791]: I1210 22:51:12.823998 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 22:51:12 crc kubenswrapper[4791]: I1210 22:51:12.824058 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 22:51:12 crc kubenswrapper[4791]: I1210 22:51:12.824083 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 22:51:12 crc kubenswrapper[4791]: I1210 22:51:12.824112 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 22:51:12 crc kubenswrapper[4791]: I1210 22:51:12.824138 4791 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T22:51:12Z","lastTransitionTime":"2025-12-10T22:51:12Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 22:51:12 crc kubenswrapper[4791]: I1210 22:51:12.883884 4791 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-b6kf6" Dec 10 22:51:12 crc kubenswrapper[4791]: I1210 22:51:12.883964 4791 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 10 22:51:12 crc kubenswrapper[4791]: I1210 22:51:12.883994 4791 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 10 22:51:12 crc kubenswrapper[4791]: I1210 22:51:12.884164 4791 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 10 22:51:12 crc kubenswrapper[4791]: E1210 22:51:12.884158 4791 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-b6kf6" podUID="b4527e71-bb58-4a39-a99c-fe614aa4c5f9" Dec 10 22:51:12 crc kubenswrapper[4791]: E1210 22:51:12.884314 4791 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 10 22:51:12 crc kubenswrapper[4791]: E1210 22:51:12.884446 4791 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 10 22:51:12 crc kubenswrapper[4791]: E1210 22:51:12.884896 4791 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 10 22:51:12 crc kubenswrapper[4791]: I1210 22:51:12.927018 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 22:51:12 crc kubenswrapper[4791]: I1210 22:51:12.927065 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 22:51:12 crc kubenswrapper[4791]: I1210 22:51:12.927078 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 22:51:12 crc kubenswrapper[4791]: I1210 22:51:12.927095 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 22:51:12 crc kubenswrapper[4791]: I1210 22:51:12.927105 4791 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T22:51:12Z","lastTransitionTime":"2025-12-10T22:51:12Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 22:51:13 crc kubenswrapper[4791]: I1210 22:51:13.030470 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 22:51:13 crc kubenswrapper[4791]: I1210 22:51:13.030543 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 22:51:13 crc kubenswrapper[4791]: I1210 22:51:13.030561 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 22:51:13 crc kubenswrapper[4791]: I1210 22:51:13.030584 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 22:51:13 crc kubenswrapper[4791]: I1210 22:51:13.030600 4791 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T22:51:13Z","lastTransitionTime":"2025-12-10T22:51:13Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 22:51:13 crc kubenswrapper[4791]: I1210 22:51:13.132804 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 22:51:13 crc kubenswrapper[4791]: I1210 22:51:13.132975 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 22:51:13 crc kubenswrapper[4791]: I1210 22:51:13.132999 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 22:51:13 crc kubenswrapper[4791]: I1210 22:51:13.133023 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 22:51:13 crc kubenswrapper[4791]: I1210 22:51:13.133043 4791 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T22:51:13Z","lastTransitionTime":"2025-12-10T22:51:13Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 22:51:13 crc kubenswrapper[4791]: I1210 22:51:13.235742 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 22:51:13 crc kubenswrapper[4791]: I1210 22:51:13.235809 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 22:51:13 crc kubenswrapper[4791]: I1210 22:51:13.235827 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 22:51:13 crc kubenswrapper[4791]: I1210 22:51:13.235854 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 22:51:13 crc kubenswrapper[4791]: I1210 22:51:13.235873 4791 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T22:51:13Z","lastTransitionTime":"2025-12-10T22:51:13Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 22:51:13 crc kubenswrapper[4791]: I1210 22:51:13.338213 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 22:51:13 crc kubenswrapper[4791]: I1210 22:51:13.338249 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 22:51:13 crc kubenswrapper[4791]: I1210 22:51:13.338256 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 22:51:13 crc kubenswrapper[4791]: I1210 22:51:13.338270 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 22:51:13 crc kubenswrapper[4791]: I1210 22:51:13.338279 4791 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T22:51:13Z","lastTransitionTime":"2025-12-10T22:51:13Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 22:51:13 crc kubenswrapper[4791]: I1210 22:51:13.440786 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 22:51:13 crc kubenswrapper[4791]: I1210 22:51:13.440841 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 22:51:13 crc kubenswrapper[4791]: I1210 22:51:13.440857 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 22:51:13 crc kubenswrapper[4791]: I1210 22:51:13.440882 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 22:51:13 crc kubenswrapper[4791]: I1210 22:51:13.440902 4791 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T22:51:13Z","lastTransitionTime":"2025-12-10T22:51:13Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 22:51:13 crc kubenswrapper[4791]: I1210 22:51:13.543231 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 22:51:13 crc kubenswrapper[4791]: I1210 22:51:13.543315 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 22:51:13 crc kubenswrapper[4791]: I1210 22:51:13.543334 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 22:51:13 crc kubenswrapper[4791]: I1210 22:51:13.543404 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 22:51:13 crc kubenswrapper[4791]: I1210 22:51:13.543421 4791 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T22:51:13Z","lastTransitionTime":"2025-12-10T22:51:13Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 22:51:13 crc kubenswrapper[4791]: I1210 22:51:13.646257 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 22:51:13 crc kubenswrapper[4791]: I1210 22:51:13.646335 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 22:51:13 crc kubenswrapper[4791]: I1210 22:51:13.646382 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 22:51:13 crc kubenswrapper[4791]: I1210 22:51:13.646406 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 22:51:13 crc kubenswrapper[4791]: I1210 22:51:13.646420 4791 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T22:51:13Z","lastTransitionTime":"2025-12-10T22:51:13Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 22:51:13 crc kubenswrapper[4791]: I1210 22:51:13.749653 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 22:51:13 crc kubenswrapper[4791]: I1210 22:51:13.749725 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 22:51:13 crc kubenswrapper[4791]: I1210 22:51:13.749735 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 22:51:13 crc kubenswrapper[4791]: I1210 22:51:13.749758 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 22:51:13 crc kubenswrapper[4791]: I1210 22:51:13.749770 4791 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T22:51:13Z","lastTransitionTime":"2025-12-10T22:51:13Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 22:51:13 crc kubenswrapper[4791]: I1210 22:51:13.852843 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 22:51:13 crc kubenswrapper[4791]: I1210 22:51:13.852887 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 22:51:13 crc kubenswrapper[4791]: I1210 22:51:13.852898 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 22:51:13 crc kubenswrapper[4791]: I1210 22:51:13.852916 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 22:51:13 crc kubenswrapper[4791]: I1210 22:51:13.852928 4791 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T22:51:13Z","lastTransitionTime":"2025-12-10T22:51:13Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 22:51:13 crc kubenswrapper[4791]: I1210 22:51:13.886402 4791 scope.go:117] "RemoveContainer" containerID="5f299a2b60de0ef5fb17a1e463d8552d6c4d11d166ae32e490104b3d90116d25" Dec 10 22:51:13 crc kubenswrapper[4791]: E1210 22:51:13.887293 4791 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"ovnkube-controller\" with CrashLoopBackOff: \"back-off 40s restarting failed container=ovnkube-controller pod=ovnkube-node-zhq64_openshift-ovn-kubernetes(3cd47739-0fa9-4321-aff1-220f8721a0b3)\"" pod="openshift-ovn-kubernetes/ovnkube-node-zhq64" podUID="3cd47739-0fa9-4321-aff1-220f8721a0b3" Dec 10 22:51:13 crc kubenswrapper[4791]: I1210 22:51:13.955430 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 22:51:13 crc kubenswrapper[4791]: I1210 22:51:13.955505 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 22:51:13 crc kubenswrapper[4791]: I1210 22:51:13.955520 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 22:51:13 crc kubenswrapper[4791]: I1210 22:51:13.955542 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 22:51:13 crc kubenswrapper[4791]: I1210 22:51:13.955557 4791 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T22:51:13Z","lastTransitionTime":"2025-12-10T22:51:13Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 22:51:14 crc kubenswrapper[4791]: I1210 22:51:14.058080 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 22:51:14 crc kubenswrapper[4791]: I1210 22:51:14.058560 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 22:51:14 crc kubenswrapper[4791]: I1210 22:51:14.058708 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 22:51:14 crc kubenswrapper[4791]: I1210 22:51:14.058847 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 22:51:14 crc kubenswrapper[4791]: I1210 22:51:14.058984 4791 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T22:51:14Z","lastTransitionTime":"2025-12-10T22:51:14Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 22:51:14 crc kubenswrapper[4791]: I1210 22:51:14.162629 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 22:51:14 crc kubenswrapper[4791]: I1210 22:51:14.163513 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 22:51:14 crc kubenswrapper[4791]: I1210 22:51:14.163576 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 22:51:14 crc kubenswrapper[4791]: I1210 22:51:14.163609 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 22:51:14 crc kubenswrapper[4791]: I1210 22:51:14.163629 4791 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T22:51:14Z","lastTransitionTime":"2025-12-10T22:51:14Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 22:51:14 crc kubenswrapper[4791]: I1210 22:51:14.266932 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 22:51:14 crc kubenswrapper[4791]: I1210 22:51:14.266992 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 22:51:14 crc kubenswrapper[4791]: I1210 22:51:14.267009 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 22:51:14 crc kubenswrapper[4791]: I1210 22:51:14.267032 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 22:51:14 crc kubenswrapper[4791]: I1210 22:51:14.267049 4791 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T22:51:14Z","lastTransitionTime":"2025-12-10T22:51:14Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 22:51:14 crc kubenswrapper[4791]: I1210 22:51:14.370011 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 22:51:14 crc kubenswrapper[4791]: I1210 22:51:14.370089 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 22:51:14 crc kubenswrapper[4791]: I1210 22:51:14.370106 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 22:51:14 crc kubenswrapper[4791]: I1210 22:51:14.370133 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 22:51:14 crc kubenswrapper[4791]: I1210 22:51:14.370155 4791 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T22:51:14Z","lastTransitionTime":"2025-12-10T22:51:14Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 22:51:14 crc kubenswrapper[4791]: I1210 22:51:14.472816 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 22:51:14 crc kubenswrapper[4791]: I1210 22:51:14.472902 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 22:51:14 crc kubenswrapper[4791]: I1210 22:51:14.472956 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 22:51:14 crc kubenswrapper[4791]: I1210 22:51:14.472982 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 22:51:14 crc kubenswrapper[4791]: I1210 22:51:14.472999 4791 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T22:51:14Z","lastTransitionTime":"2025-12-10T22:51:14Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 22:51:14 crc kubenswrapper[4791]: I1210 22:51:14.576190 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 22:51:14 crc kubenswrapper[4791]: I1210 22:51:14.576301 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 22:51:14 crc kubenswrapper[4791]: I1210 22:51:14.576326 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 22:51:14 crc kubenswrapper[4791]: I1210 22:51:14.576357 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 22:51:14 crc kubenswrapper[4791]: I1210 22:51:14.576369 4791 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T22:51:14Z","lastTransitionTime":"2025-12-10T22:51:14Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 22:51:14 crc kubenswrapper[4791]: I1210 22:51:14.678894 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 22:51:14 crc kubenswrapper[4791]: I1210 22:51:14.678954 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 22:51:14 crc kubenswrapper[4791]: I1210 22:51:14.678972 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 22:51:14 crc kubenswrapper[4791]: I1210 22:51:14.679000 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 22:51:14 crc kubenswrapper[4791]: I1210 22:51:14.679017 4791 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T22:51:14Z","lastTransitionTime":"2025-12-10T22:51:14Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 22:51:14 crc kubenswrapper[4791]: I1210 22:51:14.782746 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 22:51:14 crc kubenswrapper[4791]: I1210 22:51:14.782820 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 22:51:14 crc kubenswrapper[4791]: I1210 22:51:14.782839 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 22:51:14 crc kubenswrapper[4791]: I1210 22:51:14.782864 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 22:51:14 crc kubenswrapper[4791]: I1210 22:51:14.782884 4791 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T22:51:14Z","lastTransitionTime":"2025-12-10T22:51:14Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 22:51:14 crc kubenswrapper[4791]: I1210 22:51:14.884048 4791 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-b6kf6" Dec 10 22:51:14 crc kubenswrapper[4791]: I1210 22:51:14.884153 4791 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 10 22:51:14 crc kubenswrapper[4791]: I1210 22:51:14.884174 4791 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 10 22:51:14 crc kubenswrapper[4791]: I1210 22:51:14.884096 4791 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 10 22:51:14 crc kubenswrapper[4791]: E1210 22:51:14.884295 4791 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-b6kf6" podUID="b4527e71-bb58-4a39-a99c-fe614aa4c5f9" Dec 10 22:51:14 crc kubenswrapper[4791]: E1210 22:51:14.884424 4791 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 10 22:51:14 crc kubenswrapper[4791]: E1210 22:51:14.884710 4791 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 10 22:51:14 crc kubenswrapper[4791]: E1210 22:51:14.884778 4791 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 10 22:51:14 crc kubenswrapper[4791]: I1210 22:51:14.886441 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 22:51:14 crc kubenswrapper[4791]: I1210 22:51:14.886551 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 22:51:14 crc kubenswrapper[4791]: I1210 22:51:14.886576 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 22:51:14 crc kubenswrapper[4791]: I1210 22:51:14.886606 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 22:51:14 crc kubenswrapper[4791]: I1210 22:51:14.886628 4791 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T22:51:14Z","lastTransitionTime":"2025-12-10T22:51:14Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 22:51:14 crc kubenswrapper[4791]: I1210 22:51:14.989665 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 22:51:14 crc kubenswrapper[4791]: I1210 22:51:14.989747 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 22:51:14 crc kubenswrapper[4791]: I1210 22:51:14.989772 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 22:51:14 crc kubenswrapper[4791]: I1210 22:51:14.989803 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 22:51:14 crc kubenswrapper[4791]: I1210 22:51:14.989826 4791 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T22:51:14Z","lastTransitionTime":"2025-12-10T22:51:14Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 22:51:15 crc kubenswrapper[4791]: I1210 22:51:15.093017 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 22:51:15 crc kubenswrapper[4791]: I1210 22:51:15.093483 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 22:51:15 crc kubenswrapper[4791]: I1210 22:51:15.093679 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 22:51:15 crc kubenswrapper[4791]: I1210 22:51:15.093853 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 22:51:15 crc kubenswrapper[4791]: I1210 22:51:15.093999 4791 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T22:51:15Z","lastTransitionTime":"2025-12-10T22:51:15Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 22:51:15 crc kubenswrapper[4791]: I1210 22:51:15.197023 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 22:51:15 crc kubenswrapper[4791]: I1210 22:51:15.197073 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 22:51:15 crc kubenswrapper[4791]: I1210 22:51:15.197089 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 22:51:15 crc kubenswrapper[4791]: I1210 22:51:15.197106 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 22:51:15 crc kubenswrapper[4791]: I1210 22:51:15.197117 4791 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T22:51:15Z","lastTransitionTime":"2025-12-10T22:51:15Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 22:51:15 crc kubenswrapper[4791]: I1210 22:51:15.299512 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 22:51:15 crc kubenswrapper[4791]: I1210 22:51:15.299563 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 22:51:15 crc kubenswrapper[4791]: I1210 22:51:15.299575 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 22:51:15 crc kubenswrapper[4791]: I1210 22:51:15.299593 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 22:51:15 crc kubenswrapper[4791]: I1210 22:51:15.299607 4791 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T22:51:15Z","lastTransitionTime":"2025-12-10T22:51:15Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 22:51:15 crc kubenswrapper[4791]: I1210 22:51:15.402646 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 22:51:15 crc kubenswrapper[4791]: I1210 22:51:15.402711 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 22:51:15 crc kubenswrapper[4791]: I1210 22:51:15.402730 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 22:51:15 crc kubenswrapper[4791]: I1210 22:51:15.402755 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 22:51:15 crc kubenswrapper[4791]: I1210 22:51:15.402772 4791 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T22:51:15Z","lastTransitionTime":"2025-12-10T22:51:15Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 22:51:15 crc kubenswrapper[4791]: I1210 22:51:15.505175 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 22:51:15 crc kubenswrapper[4791]: I1210 22:51:15.505222 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 22:51:15 crc kubenswrapper[4791]: I1210 22:51:15.505237 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 22:51:15 crc kubenswrapper[4791]: I1210 22:51:15.505257 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 22:51:15 crc kubenswrapper[4791]: I1210 22:51:15.505271 4791 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T22:51:15Z","lastTransitionTime":"2025-12-10T22:51:15Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 22:51:15 crc kubenswrapper[4791]: I1210 22:51:15.608478 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 22:51:15 crc kubenswrapper[4791]: I1210 22:51:15.608545 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 22:51:15 crc kubenswrapper[4791]: I1210 22:51:15.608562 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 22:51:15 crc kubenswrapper[4791]: I1210 22:51:15.608590 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 22:51:15 crc kubenswrapper[4791]: I1210 22:51:15.608608 4791 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T22:51:15Z","lastTransitionTime":"2025-12-10T22:51:15Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 22:51:15 crc kubenswrapper[4791]: I1210 22:51:15.711475 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 22:51:15 crc kubenswrapper[4791]: I1210 22:51:15.711543 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 22:51:15 crc kubenswrapper[4791]: I1210 22:51:15.711568 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 22:51:15 crc kubenswrapper[4791]: I1210 22:51:15.711601 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 22:51:15 crc kubenswrapper[4791]: I1210 22:51:15.711625 4791 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T22:51:15Z","lastTransitionTime":"2025-12-10T22:51:15Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 22:51:15 crc kubenswrapper[4791]: I1210 22:51:15.813939 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 22:51:15 crc kubenswrapper[4791]: I1210 22:51:15.814031 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 22:51:15 crc kubenswrapper[4791]: I1210 22:51:15.814047 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 22:51:15 crc kubenswrapper[4791]: I1210 22:51:15.814069 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 22:51:15 crc kubenswrapper[4791]: I1210 22:51:15.814108 4791 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T22:51:15Z","lastTransitionTime":"2025-12-10T22:51:15Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 22:51:15 crc kubenswrapper[4791]: I1210 22:51:15.897674 4791 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-s6rk8" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5c9d0978-a6f9-49c5-b377-6619b4e3ff52\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T22:50:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T22:50:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T22:50:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T22:50:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8e55baaf460c794ab196e237d51d12de059aa038617dc2d5366d49c529f5d5c7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T22:50:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7vkgm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://51d1b9b086c60bae9223cb51d5a998d02790a0d91b745aab10392c23e617eb9f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T22:50:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7vkgm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-10T22:50:03Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-s6rk8\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T22:51:15Z is after 2025-08-24T17:21:41Z" Dec 10 22:51:15 crc kubenswrapper[4791]: I1210 22:51:15.915941 4791 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-5rb5l" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ba35653c-6e06-4cee-a4d6-137764090d18\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T22:49:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T22:49:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T22:49:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T22:49:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://dbb5c875730a98ad2daf316cfafc08fcdce21ac16148154780ac9193ed405dab\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T22:49:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fvvrs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://185eb2ff44221a514878283ec873773a6c9a9073dc64e296f982e60dee1338e4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T22:49:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fvvrs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-10T22:49:51Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-5rb5l\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T22:51:15Z is after 2025-08-24T17:21:41Z" Dec 10 22:51:15 crc kubenswrapper[4791]: I1210 22:51:15.917322 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 22:51:15 crc kubenswrapper[4791]: I1210 22:51:15.917418 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 22:51:15 crc kubenswrapper[4791]: I1210 22:51:15.917444 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 22:51:15 crc kubenswrapper[4791]: I1210 22:51:15.917474 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 22:51:15 crc kubenswrapper[4791]: I1210 22:51:15.917498 4791 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T22:51:15Z","lastTransitionTime":"2025-12-10T22:51:15Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 22:51:15 crc kubenswrapper[4791]: I1210 22:51:15.932756 4791 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7a9a3709-32fb-48b4-97f4-e1a19282a085\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T22:49:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T22:49:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T22:49:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T22:49:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T22:49:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://511bd056768a15ea0f650a2b1a6e87e3c71dfc75696222a54f7376227ad480c6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-crio\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T22:49:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kube\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"var-lib-kubelet\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8c489105d0dfb302732d4f5bd2fc9d000967a071802735b778d49bdbc428b264\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://8c489105d0dfb302732d4f5bd2fc9d000967a071802735b778d49bdbc428b264\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-10T22:49:27Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-10T22:49:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var\\\",\\\"name\\\":\\\"var-lib-kubelet\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-10T22:49:25Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"kube-rbac-proxy-crio-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T22:51:15Z is after 2025-08-24T17:21:41Z" Dec 10 22:51:15 crc kubenswrapper[4791]: I1210 22:51:15.952261 4791 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-10T22:49:52Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-10T22:49:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-10T22:49:52Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3734b110802828632d44263ed0a43cabc4b710833f0be120d72db0940fe783e2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T22:49:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://794803ac9be3c52509e6dcb10854d8864e1a9e406762e57c7971b07c4da1220e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T22:49:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T22:51:15Z is after 2025-08-24T17:21:41Z" Dec 10 22:51:15 crc kubenswrapper[4791]: I1210 22:51:15.961695 4791 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-vczjq" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b9f00f25-3fa1-4b63-9710-fd352224b01b\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T22:49:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T22:49:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T22:49:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T22:49:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4698cf4c6b55b41e628448d37c2eb847c2777d0efbd93afcbf367d5e78623b7e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T22:49:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lrrpv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-10T22:49:50Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-vczjq\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T22:51:15Z is after 2025-08-24T17:21:41Z" Dec 10 22:51:15 crc kubenswrapper[4791]: I1210 22:51:15.983952 4791 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-4nwkq" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"672aa28c-8169-49ed-87b8-21187d13a80c\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T22:49:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T22:49:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T22:50:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T22:50:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f337f881ffd71f4a2ee19e92bb0b92c76cbb9b8352511f741d0c87e7da4cdd18\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a2f2a62315529e148a10a33ccb241e9417fbde0fe42afb61cb7ee2438c74636f\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-12-10T22:50:38Z\\\",\\\"message\\\":\\\"2025-12-10T22:49:52+00:00 [cnibincopy] Successfully copied files in /usr/src/multus-cni/rhel9/bin/ to /host/opt/cni/bin/upgrade_e48d9423-4bca-43c4-9eb9-b55a19a5ffd6\\\\n2025-12-10T22:49:52+00:00 [cnibincopy] Successfully moved files in /host/opt/cni/bin/upgrade_e48d9423-4bca-43c4-9eb9-b55a19a5ffd6 to /host/opt/cni/bin/\\\\n2025-12-10T22:49:52Z [verbose] multus-daemon started\\\\n2025-12-10T22:49:52Z [verbose] Readiness Indicator file check\\\\n2025-12-10T22:50:37Z [error] have you checked that your default network is ready? still waiting for readinessindicatorfile @ /host/run/multus/cni/net.d/10-ovn-kubernetes.conf. pollimmediate error: timed out waiting for the condition\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-10T22:49:52Z\\\"}},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T22:50:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xl62j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-10T22:49:51Z\\\"}}\" for pod \"openshift-multus\"/\"multus-4nwkq\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T22:51:15Z is after 2025-08-24T17:21:41Z" Dec 10 22:51:16 crc kubenswrapper[4791]: I1210 22:51:16.011100 4791 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-zhq64" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3cd47739-0fa9-4321-aff1-220f8721a0b3\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T22:49:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T22:49:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T22:49:52Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T22:49:52Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://68795a29944e56870ffe94fa732e9dcf0bfb344c0d85d1e590f58a876129c61f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T22:49:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wvpw8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://bc721b530c08b433756e978ec0423cc18aabf6771b61eb96f903b4d57ea335ff\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T22:49:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wvpw8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://152a7618be8f38fadf4e564b3948980503cedcdb466c37344ccc20f4a3173a5f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T22:49:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wvpw8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3d18f9cb7207efb1e32de75a30d33e024f656a08f0503db95e5488bc5280971f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T22:49:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wvpw8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://937137ae9d7480cdc7c59bc4a1df0a4c10e72b3b63693af285278789953ed218\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T22:49:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wvpw8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://1a52d19b7671e8b143017c8a981a6f8d3ee3b84650e4b223b814a531ac76f364\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T22:49:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wvpw8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5f299a2b60de0ef5fb17a1e463d8552d6c4d11d166ae32e490104b3d90116d25\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5f299a2b60de0ef5fb17a1e463d8552d6c4d11d166ae32e490104b3d90116d25\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-12-10T22:50:48Z\\\",\\\"message\\\":\\\"11] Stopping reflector *v1.EgressIP (0s) from github.com/openshift/ovn-kubernetes/go-controller/pkg/crd/egressip/v1/apis/informers/externalversions/factory.go:140\\\\nI1210 22:50:47.868727 6852 reflector.go:311] Stopping reflector *v1.Pod (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI1210 22:50:47.868812 6852 reflector.go:311] Stopping reflector *v1.NetworkPolicy (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI1210 22:50:47.869024 6852 reflector.go:311] Stopping reflector *v1.EndpointSlice (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI1210 22:50:47.869223 6852 reflector.go:311] Stopping reflector *v1alpha1.BaselineAdminNetworkPolicy (0s) from sigs.k8s.io/network-policy-api/pkg/client/informers/externalversions/factory.go:141\\\\nI1210 22:50:47.869558 6852 handler.go:190] Sending *v1.Namespace event handler 1 for removal\\\\nI1210 22:50:47.869577 6852 handler.go:190] Sending *v1.Namespace event handler 5 for removal\\\\nI1210 22:50:47.869595 6852 handler.go:208] Removed *v1.Namespace event handler 5\\\\nI1210 22:50:47.869604 6852 handler.go:190] Sending *v1.NetworkPolicy event handler 4 for removal\\\\nI1210 22:50:47.869608 6852 handler.go:208] Removed *v1.Namespace event handler 1\\\\nI1210 22:50:47.869637 6852 factory.go:656] Stopping watch factory\\\\nI1210 22:50:47.869656 6852 ovnkube.go:599] Stopped ovnkube\\\\nI1210 22:50:\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-10T22:50:47Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 40s restarting failed container=ovnkube-controller pod=ovnkube-node-zhq64_openshift-ovn-kubernetes(3cd47739-0fa9-4321-aff1-220f8721a0b3)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wvpw8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://dd3f80e5713221f96ed50da7f2e7fd5b9fabfcd5c1656da8d8a86d818a5fdc3f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T22:49:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wvpw8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3afbd1ffcd2c4b40324aca6e52d6fc822db131c3f35a9af3a6f005c2cffdb7bf\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3afbd1ffcd2c4b40324aca6e52d6fc822db131c3f35a9af3a6f005c2cffdb7bf\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-10T22:49:53Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-10T22:49:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wvpw8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-10T22:49:52Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-zhq64\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T22:51:16Z is after 2025-08-24T17:21:41Z" Dec 10 22:51:16 crc kubenswrapper[4791]: I1210 22:51:16.019308 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 22:51:16 crc kubenswrapper[4791]: I1210 22:51:16.019367 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 22:51:16 crc kubenswrapper[4791]: I1210 22:51:16.019384 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 22:51:16 crc kubenswrapper[4791]: I1210 22:51:16.019405 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 22:51:16 crc kubenswrapper[4791]: I1210 22:51:16.019421 4791 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T22:51:16Z","lastTransitionTime":"2025-12-10T22:51:16Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 22:51:16 crc kubenswrapper[4791]: I1210 22:51:16.025514 4791 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"698bbe9a-fbe5-436d-a835-5e80df77ba24\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T22:49:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T22:49:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T22:49:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T22:49:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T22:49:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://224d73713312376a58ec6cc2bc8fdb6ffa44c4749ae0e7691d5fe27b9afca09f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T22:49:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://372eeaffd40992c6f06efc4df6dfcb44b31de64ee8256c21508cdd7c74ecb96e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T22:49:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://76b6013b062cac5b0b4d12a94f38ce7b43f9c7ea216063455d88a956ac04d26e\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T22:49:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://78e950e122e761d950da1a3c60c06476e53920e8795f6c61db22c01e22d5136e\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T22:49:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-10T22:49:25Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T22:51:16Z is after 2025-08-24T17:21:41Z" Dec 10 22:51:16 crc kubenswrapper[4791]: I1210 22:51:16.038039 4791 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-10T22:49:50Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-10T22:49:50Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T22:51:16Z is after 2025-08-24T17:21:41Z" Dec 10 22:51:16 crc kubenswrapper[4791]: I1210 22:51:16.049978 4791 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-10T22:49:50Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-10T22:49:50Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T22:51:16Z is after 2025-08-24T17:21:41Z" Dec 10 22:51:16 crc kubenswrapper[4791]: I1210 22:51:16.058928 4791 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-tchxk" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"c0ab74e0-42a9-4f42-ac80-16030cf4ffff\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T22:49:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T22:49:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T22:49:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T22:49:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://04e0fd1b53256b7a938e8ccab93e2a4eb73dd1a59e0b9799e09c55e511baf7fe\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T22:49:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zn2qt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-10T22:49:50Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-tchxk\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T22:51:16Z is after 2025-08-24T17:21:41Z" Dec 10 22:51:16 crc kubenswrapper[4791]: I1210 22:51:16.073552 4791 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-rw584" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ff9beb90-69b0-4732-bf37-0b81f58ecc98\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T22:49:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T22:49:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T22:49:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T22:49:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ddb9df501dc6e8e5561fbc2aefd431037a3f28f73bf8e82a6598d261a4224807\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T22:49:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fksqn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ce67e54b696bb53ffbbe48c6ce71476b635e0868993e9c427b22f904891433ee\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ce67e54b696bb53ffbbe48c6ce71476b635e0868993e9c427b22f904891433ee\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-10T22:49:52Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-10T22:49:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fksqn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://818d89bfffb8a110670152c8480684665ea8db578e9768f524fbbe6a92f27c22\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://818d89bfffb8a110670152c8480684665ea8db578e9768f524fbbe6a92f27c22\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-10T22:49:53Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-10T22:49:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fksqn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://9aa9c06e894bb2a2ed3738defaeff68a0a0458ead48177f36b90770944c211a0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://9aa9c06e894bb2a2ed3738defaeff68a0a0458ead48177f36b90770944c211a0\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-10T22:49:54Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-10T22:49:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fksqn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://dfd6d03336dcea7529419a30e8d70d84e2290a0fa5b0e46d09ad341c190638d6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://dfd6d03336dcea7529419a30e8d70d84e2290a0fa5b0e46d09ad341c190638d6\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-10T22:49:55Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-10T22:49:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fksqn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b86738401ece6fbc3083782720c66a515850ed78d4054f34fcabab7ff525b85f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b86738401ece6fbc3083782720c66a515850ed78d4054f34fcabab7ff525b85f\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-10T22:49:56Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-10T22:49:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fksqn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://18b326f572a899ddd10197a745bf9264d4ddd975690ad1cfb0ec576ccd5cf102\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://18b326f572a899ddd10197a745bf9264d4ddd975690ad1cfb0ec576ccd5cf102\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-10T22:49:57Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-10T22:49:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fksqn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-10T22:49:51Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-rw584\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T22:51:16Z is after 2025-08-24T17:21:41Z" Dec 10 22:51:16 crc kubenswrapper[4791]: I1210 22:51:16.085321 4791 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-10T22:49:52Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-10T22:49:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-10T22:49:52Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://20651eb0c16e2c71222f56da1ff755cb0e7b084886acbbb24f0ccaf57041ed85\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T22:49:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T22:51:16Z is after 2025-08-24T17:21:41Z" Dec 10 22:51:16 crc kubenswrapper[4791]: I1210 22:51:16.096945 4791 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"02a949d8-38e2-4be3-989f-40a015407140\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T22:49:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T22:49:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T22:50:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T22:50:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T22:49:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://cd4158bc59b30f0f1716bf43c1ae7c9916a5b7d7164d074987fc1f19ba49b778\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T22:49:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://656065fd3cbdc6a25dab5d06c7baa00af16d8f62fc5da687efe9c6cc91945de0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T22:49:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://90ed7fff816b1772ba9ecd95587a59cbed7c30ac302fff7e7c5de934c912bcf4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T22:49:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9b7a5502803eaad9f76f5ffa91b29fbb2657402065a03a4d6433c096aac7264f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"wait-for-host-port\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://9b7a5502803eaad9f76f5ffa91b29fbb2657402065a03a4d6433c096aac7264f\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-10T22:49:27Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-10T22:49:27Z\\\"}}}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-10T22:49:25Z\\\"}}\" for pod \"openshift-kube-scheduler\"/\"openshift-kube-scheduler-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T22:51:16Z is after 2025-08-24T17:21:41Z" Dec 10 22:51:16 crc kubenswrapper[4791]: I1210 22:51:16.122040 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 22:51:16 crc kubenswrapper[4791]: I1210 22:51:16.122079 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 22:51:16 crc kubenswrapper[4791]: I1210 22:51:16.122094 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 22:51:16 crc kubenswrapper[4791]: I1210 22:51:16.122113 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 22:51:16 crc kubenswrapper[4791]: I1210 22:51:16.122127 4791 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T22:51:16Z","lastTransitionTime":"2025-12-10T22:51:16Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 22:51:16 crc kubenswrapper[4791]: I1210 22:51:16.127476 4791 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"90250e13-6c63-443a-a5e5-6e4e341a2289\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T22:49:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T22:49:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T22:49:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T22:49:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T22:49:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ba54f95b1e00843b854db9a10fe6897592159ccacf30ed0f76002093a13eedd3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T22:49:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5aa8ffa518f3a68b60b275f8df4851b2564aa4984b61efdc9470b73072133e43\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T22:49:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://16f989503edc78d9451638e2cf6cf4b19a2cf2b8809b10b237541cb829303091\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T22:49:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://82556824d63ce6e46df14c45689ff1f81ebf729abe902061815c040b323898e9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T22:49:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://61db531940fa6f2effdae5a005b4c91bd2698bd5d1bff9588c171a65ad550eca\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T22:49:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0f65c864d885319ce2a08d61c30be24b3ffb7d23ea30c82ed7faf039cb2c006d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://0f65c864d885319ce2a08d61c30be24b3ffb7d23ea30c82ed7faf039cb2c006d\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-10T22:49:27Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-10T22:49:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e31a76d1d5b2812ba83e24e38268e25ec8bed98fc3da9adcd06bda5822b6ace1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e31a76d1d5b2812ba83e24e38268e25ec8bed98fc3da9adcd06bda5822b6ace1\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-10T22:49:30Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-10T22:49:29Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://b5b5eceae562f3986b543d41de1e977efb85a91d84c98e09159ca6d092eda691\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b5b5eceae562f3986b543d41de1e977efb85a91d84c98e09159ca6d092eda691\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-10T22:49:31Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-10T22:49:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-10T22:49:25Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T22:51:16Z is after 2025-08-24T17:21:41Z" Dec 10 22:51:16 crc kubenswrapper[4791]: I1210 22:51:16.146431 4791 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-10T22:49:50Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-10T22:49:50Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T22:51:16Z is after 2025-08-24T17:21:41Z" Dec 10 22:51:16 crc kubenswrapper[4791]: I1210 22:51:16.159517 4791 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-10T22:49:54Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-10T22:49:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-10T22:49:54Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://62bfaf5d7a6d1b452cdaf88086447d0c8cb7306e94434e46752be5475124d56a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T22:49:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T22:51:16Z is after 2025-08-24T17:21:41Z" Dec 10 22:51:16 crc kubenswrapper[4791]: I1210 22:51:16.172776 4791 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-b6kf6" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b4527e71-bb58-4a39-a99c-fe614aa4c5f9\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T22:50:04Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T22:50:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T22:50:04Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T22:50:04Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ngpr4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ngpr4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-10T22:50:04Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-b6kf6\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T22:51:16Z is after 2025-08-24T17:21:41Z" Dec 10 22:51:16 crc kubenswrapper[4791]: I1210 22:51:16.186431 4791 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f8349fea-c49b-49a1-b5ee-032c5dffe021\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T22:49:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T22:49:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T22:49:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T22:49:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T22:49:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://acafe12186df99d1059a31260de0eafce35ff7771bcb8f5c6753b72665d6f60b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T22:49:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ea4c346507862a8c12756a3061d6e25e788a8d79f4621b62f81f4322185c82a8\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T22:49:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://2e6e324ec6f4f7bb69fdfab4ea3ac1aa55e480dd8e2a8984b0fe5b4c9f66636e\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T22:49:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://1b2e11ff2bf8789fa1d0d23ae648b3e6ef653106e5bc4bbb5dd6447f7cd6f05d\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d6aa7f8c5d908d90008f2df49860ac8a3480136da567e3d85f86135d779204f9\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-12-10T22:49:44Z\\\",\\\"message\\\":\\\"W1210 22:49:33.845733 1 cmd.go:257] Using insecure, self-signed certificates\\\\nI1210 22:49:33.846108 1 crypto.go:601] Generating new CA for check-endpoints-signer@1765406973 cert, and key in /tmp/serving-cert-2585701456/serving-signer.crt, /tmp/serving-cert-2585701456/serving-signer.key\\\\nI1210 22:49:34.529748 1 observer_polling.go:159] Starting file observer\\\\nW1210 22:49:34.532225 1 builder.go:272] unable to get owner reference (falling back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": dial tcp [::1]:6443: connect: connection refused\\\\nI1210 22:49:34.532599 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1210 22:49:34.533977 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-2585701456/tls.crt::/tmp/serving-cert-2585701456/tls.key\\\\\\\"\\\\nF1210 22:49:44.774710 1 cmd.go:182] error initializing delegating authentication: unable to load configmap based request-header-client-ca-file: Get \\\\\\\"https://localhost:6443/api/v1/namespaces/kube-system/configmaps/extension-apiserver-authentication\\\\\\\": net/http: TLS handshake timeout\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-10T22:49:33Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T22:49:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://0c4d036ab269f62561fe7ef1586012cb3a16e3583a08b9488a0616ec00c8e0ae\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T22:49:32Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c5856d6b26c19fda13b5c9ac2892a7e2f9c6f79feb8173ecfd9b33b673327d67\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c5856d6b26c19fda13b5c9ac2892a7e2f9c6f79feb8173ecfd9b33b673327d67\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-10T22:49:27Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-10T22:49:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-10T22:49:25Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T22:51:16Z is after 2025-08-24T17:21:41Z" Dec 10 22:51:16 crc kubenswrapper[4791]: I1210 22:51:16.225079 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 22:51:16 crc kubenswrapper[4791]: I1210 22:51:16.225115 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 22:51:16 crc kubenswrapper[4791]: I1210 22:51:16.225126 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 22:51:16 crc kubenswrapper[4791]: I1210 22:51:16.225142 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 22:51:16 crc kubenswrapper[4791]: I1210 22:51:16.225153 4791 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T22:51:16Z","lastTransitionTime":"2025-12-10T22:51:16Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 22:51:16 crc kubenswrapper[4791]: I1210 22:51:16.328484 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 22:51:16 crc kubenswrapper[4791]: I1210 22:51:16.328547 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 22:51:16 crc kubenswrapper[4791]: I1210 22:51:16.328567 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 22:51:16 crc kubenswrapper[4791]: I1210 22:51:16.328591 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 22:51:16 crc kubenswrapper[4791]: I1210 22:51:16.328609 4791 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T22:51:16Z","lastTransitionTime":"2025-12-10T22:51:16Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 22:51:16 crc kubenswrapper[4791]: I1210 22:51:16.431241 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 22:51:16 crc kubenswrapper[4791]: I1210 22:51:16.431526 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 22:51:16 crc kubenswrapper[4791]: I1210 22:51:16.431537 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 22:51:16 crc kubenswrapper[4791]: I1210 22:51:16.431551 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 22:51:16 crc kubenswrapper[4791]: I1210 22:51:16.431560 4791 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T22:51:16Z","lastTransitionTime":"2025-12-10T22:51:16Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 22:51:16 crc kubenswrapper[4791]: I1210 22:51:16.534289 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 22:51:16 crc kubenswrapper[4791]: I1210 22:51:16.534326 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 22:51:16 crc kubenswrapper[4791]: I1210 22:51:16.534365 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 22:51:16 crc kubenswrapper[4791]: I1210 22:51:16.534395 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 22:51:16 crc kubenswrapper[4791]: I1210 22:51:16.534408 4791 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T22:51:16Z","lastTransitionTime":"2025-12-10T22:51:16Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 22:51:16 crc kubenswrapper[4791]: I1210 22:51:16.636846 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 22:51:16 crc kubenswrapper[4791]: I1210 22:51:16.636899 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 22:51:16 crc kubenswrapper[4791]: I1210 22:51:16.636913 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 22:51:16 crc kubenswrapper[4791]: I1210 22:51:16.636930 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 22:51:16 crc kubenswrapper[4791]: I1210 22:51:16.636943 4791 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T22:51:16Z","lastTransitionTime":"2025-12-10T22:51:16Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 22:51:16 crc kubenswrapper[4791]: I1210 22:51:16.739534 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 22:51:16 crc kubenswrapper[4791]: I1210 22:51:16.739569 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 22:51:16 crc kubenswrapper[4791]: I1210 22:51:16.739579 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 22:51:16 crc kubenswrapper[4791]: I1210 22:51:16.739592 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 22:51:16 crc kubenswrapper[4791]: I1210 22:51:16.739601 4791 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T22:51:16Z","lastTransitionTime":"2025-12-10T22:51:16Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 22:51:16 crc kubenswrapper[4791]: I1210 22:51:16.842089 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 22:51:16 crc kubenswrapper[4791]: I1210 22:51:16.842135 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 22:51:16 crc kubenswrapper[4791]: I1210 22:51:16.842146 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 22:51:16 crc kubenswrapper[4791]: I1210 22:51:16.842162 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 22:51:16 crc kubenswrapper[4791]: I1210 22:51:16.842172 4791 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T22:51:16Z","lastTransitionTime":"2025-12-10T22:51:16Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 22:51:16 crc kubenswrapper[4791]: I1210 22:51:16.884295 4791 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 10 22:51:16 crc kubenswrapper[4791]: I1210 22:51:16.884363 4791 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 10 22:51:16 crc kubenswrapper[4791]: I1210 22:51:16.884373 4791 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 10 22:51:16 crc kubenswrapper[4791]: I1210 22:51:16.884396 4791 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-b6kf6" Dec 10 22:51:16 crc kubenswrapper[4791]: E1210 22:51:16.884488 4791 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 10 22:51:16 crc kubenswrapper[4791]: E1210 22:51:16.884679 4791 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-b6kf6" podUID="b4527e71-bb58-4a39-a99c-fe614aa4c5f9" Dec 10 22:51:16 crc kubenswrapper[4791]: E1210 22:51:16.885374 4791 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 10 22:51:16 crc kubenswrapper[4791]: E1210 22:51:16.885466 4791 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 10 22:51:16 crc kubenswrapper[4791]: I1210 22:51:16.945475 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 22:51:16 crc kubenswrapper[4791]: I1210 22:51:16.945533 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 22:51:16 crc kubenswrapper[4791]: I1210 22:51:16.945550 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 22:51:16 crc kubenswrapper[4791]: I1210 22:51:16.945574 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 22:51:16 crc kubenswrapper[4791]: I1210 22:51:16.945592 4791 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T22:51:16Z","lastTransitionTime":"2025-12-10T22:51:16Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 22:51:17 crc kubenswrapper[4791]: I1210 22:51:17.048833 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 22:51:17 crc kubenswrapper[4791]: I1210 22:51:17.048896 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 22:51:17 crc kubenswrapper[4791]: I1210 22:51:17.048921 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 22:51:17 crc kubenswrapper[4791]: I1210 22:51:17.048950 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 22:51:17 crc kubenswrapper[4791]: I1210 22:51:17.048975 4791 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T22:51:17Z","lastTransitionTime":"2025-12-10T22:51:17Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 22:51:17 crc kubenswrapper[4791]: I1210 22:51:17.152397 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 22:51:17 crc kubenswrapper[4791]: I1210 22:51:17.152461 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 22:51:17 crc kubenswrapper[4791]: I1210 22:51:17.152478 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 22:51:17 crc kubenswrapper[4791]: I1210 22:51:17.152509 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 22:51:17 crc kubenswrapper[4791]: I1210 22:51:17.152531 4791 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T22:51:17Z","lastTransitionTime":"2025-12-10T22:51:17Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 22:51:17 crc kubenswrapper[4791]: I1210 22:51:17.254525 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 22:51:17 crc kubenswrapper[4791]: I1210 22:51:17.254588 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 22:51:17 crc kubenswrapper[4791]: I1210 22:51:17.254610 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 22:51:17 crc kubenswrapper[4791]: I1210 22:51:17.254629 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 22:51:17 crc kubenswrapper[4791]: I1210 22:51:17.254641 4791 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T22:51:17Z","lastTransitionTime":"2025-12-10T22:51:17Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 22:51:17 crc kubenswrapper[4791]: I1210 22:51:17.356864 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 22:51:17 crc kubenswrapper[4791]: I1210 22:51:17.356899 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 22:51:17 crc kubenswrapper[4791]: I1210 22:51:17.356910 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 22:51:17 crc kubenswrapper[4791]: I1210 22:51:17.356929 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 22:51:17 crc kubenswrapper[4791]: I1210 22:51:17.356941 4791 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T22:51:17Z","lastTransitionTime":"2025-12-10T22:51:17Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 22:51:17 crc kubenswrapper[4791]: I1210 22:51:17.459075 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 22:51:17 crc kubenswrapper[4791]: I1210 22:51:17.459167 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 22:51:17 crc kubenswrapper[4791]: I1210 22:51:17.459225 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 22:51:17 crc kubenswrapper[4791]: I1210 22:51:17.459259 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 22:51:17 crc kubenswrapper[4791]: I1210 22:51:17.459282 4791 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T22:51:17Z","lastTransitionTime":"2025-12-10T22:51:17Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 22:51:17 crc kubenswrapper[4791]: I1210 22:51:17.561611 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 22:51:17 crc kubenswrapper[4791]: I1210 22:51:17.561650 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 22:51:17 crc kubenswrapper[4791]: I1210 22:51:17.561663 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 22:51:17 crc kubenswrapper[4791]: I1210 22:51:17.561683 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 22:51:17 crc kubenswrapper[4791]: I1210 22:51:17.561699 4791 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T22:51:17Z","lastTransitionTime":"2025-12-10T22:51:17Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 22:51:17 crc kubenswrapper[4791]: I1210 22:51:17.663851 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 22:51:17 crc kubenswrapper[4791]: I1210 22:51:17.663892 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 22:51:17 crc kubenswrapper[4791]: I1210 22:51:17.663902 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 22:51:17 crc kubenswrapper[4791]: I1210 22:51:17.663916 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 22:51:17 crc kubenswrapper[4791]: I1210 22:51:17.663927 4791 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T22:51:17Z","lastTransitionTime":"2025-12-10T22:51:17Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 22:51:17 crc kubenswrapper[4791]: I1210 22:51:17.766989 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 22:51:17 crc kubenswrapper[4791]: I1210 22:51:17.767046 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 22:51:17 crc kubenswrapper[4791]: I1210 22:51:17.767067 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 22:51:17 crc kubenswrapper[4791]: I1210 22:51:17.767092 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 22:51:17 crc kubenswrapper[4791]: I1210 22:51:17.767111 4791 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T22:51:17Z","lastTransitionTime":"2025-12-10T22:51:17Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 22:51:17 crc kubenswrapper[4791]: I1210 22:51:17.869737 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 22:51:17 crc kubenswrapper[4791]: I1210 22:51:17.869796 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 22:51:17 crc kubenswrapper[4791]: I1210 22:51:17.869819 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 22:51:17 crc kubenswrapper[4791]: I1210 22:51:17.869850 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 22:51:17 crc kubenswrapper[4791]: I1210 22:51:17.869875 4791 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T22:51:17Z","lastTransitionTime":"2025-12-10T22:51:17Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 22:51:17 crc kubenswrapper[4791]: I1210 22:51:17.973057 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 22:51:17 crc kubenswrapper[4791]: I1210 22:51:17.973120 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 22:51:17 crc kubenswrapper[4791]: I1210 22:51:17.973144 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 22:51:17 crc kubenswrapper[4791]: I1210 22:51:17.973176 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 22:51:17 crc kubenswrapper[4791]: I1210 22:51:17.973198 4791 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T22:51:17Z","lastTransitionTime":"2025-12-10T22:51:17Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 22:51:18 crc kubenswrapper[4791]: I1210 22:51:18.075998 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 22:51:18 crc kubenswrapper[4791]: I1210 22:51:18.076064 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 22:51:18 crc kubenswrapper[4791]: I1210 22:51:18.076086 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 22:51:18 crc kubenswrapper[4791]: I1210 22:51:18.076114 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 22:51:18 crc kubenswrapper[4791]: I1210 22:51:18.076136 4791 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T22:51:18Z","lastTransitionTime":"2025-12-10T22:51:18Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 22:51:18 crc kubenswrapper[4791]: I1210 22:51:18.178418 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 22:51:18 crc kubenswrapper[4791]: I1210 22:51:18.178447 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 22:51:18 crc kubenswrapper[4791]: I1210 22:51:18.178455 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 22:51:18 crc kubenswrapper[4791]: I1210 22:51:18.178469 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 22:51:18 crc kubenswrapper[4791]: I1210 22:51:18.178480 4791 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T22:51:18Z","lastTransitionTime":"2025-12-10T22:51:18Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 22:51:18 crc kubenswrapper[4791]: I1210 22:51:18.281380 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 22:51:18 crc kubenswrapper[4791]: I1210 22:51:18.281453 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 22:51:18 crc kubenswrapper[4791]: I1210 22:51:18.281491 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 22:51:18 crc kubenswrapper[4791]: I1210 22:51:18.281520 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 22:51:18 crc kubenswrapper[4791]: I1210 22:51:18.281541 4791 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T22:51:18Z","lastTransitionTime":"2025-12-10T22:51:18Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 22:51:18 crc kubenswrapper[4791]: I1210 22:51:18.384141 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 22:51:18 crc kubenswrapper[4791]: I1210 22:51:18.384199 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 22:51:18 crc kubenswrapper[4791]: I1210 22:51:18.384211 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 22:51:18 crc kubenswrapper[4791]: I1210 22:51:18.384229 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 22:51:18 crc kubenswrapper[4791]: I1210 22:51:18.384240 4791 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T22:51:18Z","lastTransitionTime":"2025-12-10T22:51:18Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 22:51:18 crc kubenswrapper[4791]: I1210 22:51:18.486696 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 22:51:18 crc kubenswrapper[4791]: I1210 22:51:18.486740 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 22:51:18 crc kubenswrapper[4791]: I1210 22:51:18.486755 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 22:51:18 crc kubenswrapper[4791]: I1210 22:51:18.486770 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 22:51:18 crc kubenswrapper[4791]: I1210 22:51:18.486780 4791 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T22:51:18Z","lastTransitionTime":"2025-12-10T22:51:18Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 22:51:18 crc kubenswrapper[4791]: I1210 22:51:18.589187 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 22:51:18 crc kubenswrapper[4791]: I1210 22:51:18.589278 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 22:51:18 crc kubenswrapper[4791]: I1210 22:51:18.589298 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 22:51:18 crc kubenswrapper[4791]: I1210 22:51:18.589335 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 22:51:18 crc kubenswrapper[4791]: I1210 22:51:18.589404 4791 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T22:51:18Z","lastTransitionTime":"2025-12-10T22:51:18Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 22:51:18 crc kubenswrapper[4791]: I1210 22:51:18.692326 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 22:51:18 crc kubenswrapper[4791]: I1210 22:51:18.692439 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 22:51:18 crc kubenswrapper[4791]: I1210 22:51:18.692492 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 22:51:18 crc kubenswrapper[4791]: I1210 22:51:18.692516 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 22:51:18 crc kubenswrapper[4791]: I1210 22:51:18.692533 4791 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T22:51:18Z","lastTransitionTime":"2025-12-10T22:51:18Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 22:51:18 crc kubenswrapper[4791]: I1210 22:51:18.795546 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 22:51:18 crc kubenswrapper[4791]: I1210 22:51:18.795645 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 22:51:18 crc kubenswrapper[4791]: I1210 22:51:18.795663 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 22:51:18 crc kubenswrapper[4791]: I1210 22:51:18.795684 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 22:51:18 crc kubenswrapper[4791]: I1210 22:51:18.795701 4791 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T22:51:18Z","lastTransitionTime":"2025-12-10T22:51:18Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 22:51:18 crc kubenswrapper[4791]: I1210 22:51:18.883953 4791 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-b6kf6" Dec 10 22:51:18 crc kubenswrapper[4791]: I1210 22:51:18.884015 4791 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 10 22:51:18 crc kubenswrapper[4791]: I1210 22:51:18.883967 4791 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 10 22:51:18 crc kubenswrapper[4791]: E1210 22:51:18.884195 4791 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-b6kf6" podUID="b4527e71-bb58-4a39-a99c-fe614aa4c5f9" Dec 10 22:51:18 crc kubenswrapper[4791]: I1210 22:51:18.884303 4791 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 10 22:51:18 crc kubenswrapper[4791]: E1210 22:51:18.884488 4791 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 10 22:51:18 crc kubenswrapper[4791]: E1210 22:51:18.884618 4791 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 10 22:51:18 crc kubenswrapper[4791]: E1210 22:51:18.884840 4791 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 10 22:51:18 crc kubenswrapper[4791]: I1210 22:51:18.898951 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 22:51:18 crc kubenswrapper[4791]: I1210 22:51:18.899024 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 22:51:18 crc kubenswrapper[4791]: I1210 22:51:18.899038 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 22:51:18 crc kubenswrapper[4791]: I1210 22:51:18.899056 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 22:51:18 crc kubenswrapper[4791]: I1210 22:51:18.899069 4791 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T22:51:18Z","lastTransitionTime":"2025-12-10T22:51:18Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 22:51:19 crc kubenswrapper[4791]: I1210 22:51:19.001655 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 22:51:19 crc kubenswrapper[4791]: I1210 22:51:19.001724 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 22:51:19 crc kubenswrapper[4791]: I1210 22:51:19.002083 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 22:51:19 crc kubenswrapper[4791]: I1210 22:51:19.002123 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 22:51:19 crc kubenswrapper[4791]: I1210 22:51:19.002142 4791 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T22:51:19Z","lastTransitionTime":"2025-12-10T22:51:19Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 22:51:19 crc kubenswrapper[4791]: I1210 22:51:19.106055 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 22:51:19 crc kubenswrapper[4791]: I1210 22:51:19.106114 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 22:51:19 crc kubenswrapper[4791]: I1210 22:51:19.106131 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 22:51:19 crc kubenswrapper[4791]: I1210 22:51:19.106158 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 22:51:19 crc kubenswrapper[4791]: I1210 22:51:19.106175 4791 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T22:51:19Z","lastTransitionTime":"2025-12-10T22:51:19Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 22:51:19 crc kubenswrapper[4791]: I1210 22:51:19.208563 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 22:51:19 crc kubenswrapper[4791]: I1210 22:51:19.208601 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 22:51:19 crc kubenswrapper[4791]: I1210 22:51:19.208612 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 22:51:19 crc kubenswrapper[4791]: I1210 22:51:19.208626 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 22:51:19 crc kubenswrapper[4791]: I1210 22:51:19.208635 4791 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T22:51:19Z","lastTransitionTime":"2025-12-10T22:51:19Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 22:51:19 crc kubenswrapper[4791]: I1210 22:51:19.311498 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 22:51:19 crc kubenswrapper[4791]: I1210 22:51:19.311591 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 22:51:19 crc kubenswrapper[4791]: I1210 22:51:19.311617 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 22:51:19 crc kubenswrapper[4791]: I1210 22:51:19.311649 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 22:51:19 crc kubenswrapper[4791]: I1210 22:51:19.311674 4791 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T22:51:19Z","lastTransitionTime":"2025-12-10T22:51:19Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 22:51:19 crc kubenswrapper[4791]: I1210 22:51:19.414754 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 22:51:19 crc kubenswrapper[4791]: I1210 22:51:19.414825 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 22:51:19 crc kubenswrapper[4791]: I1210 22:51:19.414842 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 22:51:19 crc kubenswrapper[4791]: I1210 22:51:19.414863 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 22:51:19 crc kubenswrapper[4791]: I1210 22:51:19.414876 4791 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T22:51:19Z","lastTransitionTime":"2025-12-10T22:51:19Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 22:51:19 crc kubenswrapper[4791]: I1210 22:51:19.517744 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 22:51:19 crc kubenswrapper[4791]: I1210 22:51:19.517892 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 22:51:19 crc kubenswrapper[4791]: I1210 22:51:19.517919 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 22:51:19 crc kubenswrapper[4791]: I1210 22:51:19.518000 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 22:51:19 crc kubenswrapper[4791]: I1210 22:51:19.518073 4791 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T22:51:19Z","lastTransitionTime":"2025-12-10T22:51:19Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 22:51:19 crc kubenswrapper[4791]: I1210 22:51:19.621547 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 22:51:19 crc kubenswrapper[4791]: I1210 22:51:19.621635 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 22:51:19 crc kubenswrapper[4791]: I1210 22:51:19.621662 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 22:51:19 crc kubenswrapper[4791]: I1210 22:51:19.621694 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 22:51:19 crc kubenswrapper[4791]: I1210 22:51:19.621718 4791 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T22:51:19Z","lastTransitionTime":"2025-12-10T22:51:19Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 22:51:19 crc kubenswrapper[4791]: I1210 22:51:19.697585 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 22:51:19 crc kubenswrapper[4791]: I1210 22:51:19.697637 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 22:51:19 crc kubenswrapper[4791]: I1210 22:51:19.697651 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 22:51:19 crc kubenswrapper[4791]: I1210 22:51:19.697671 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 22:51:19 crc kubenswrapper[4791]: I1210 22:51:19.697686 4791 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T22:51:19Z","lastTransitionTime":"2025-12-10T22:51:19Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 22:51:19 crc kubenswrapper[4791]: E1210 22:51:19.716626 4791 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-12-10T22:51:19Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-10T22:51:19Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-10T22:51:19Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-10T22:51:19Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-10T22:51:19Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-10T22:51:19Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-10T22:51:19Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-10T22:51:19Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"4fab3af6-f657-4dfc-8ef2-12b7c978c94f\\\",\\\"systemUUID\\\":\\\"442c0b68-8dce-47e0-bd68-d5210d7e0493\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T22:51:19Z is after 2025-08-24T17:21:41Z" Dec 10 22:51:19 crc kubenswrapper[4791]: I1210 22:51:19.722394 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 22:51:19 crc kubenswrapper[4791]: I1210 22:51:19.722451 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 22:51:19 crc kubenswrapper[4791]: I1210 22:51:19.722473 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 22:51:19 crc kubenswrapper[4791]: I1210 22:51:19.722499 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 22:51:19 crc kubenswrapper[4791]: I1210 22:51:19.722518 4791 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T22:51:19Z","lastTransitionTime":"2025-12-10T22:51:19Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 22:51:19 crc kubenswrapper[4791]: E1210 22:51:19.741934 4791 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-12-10T22:51:19Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-10T22:51:19Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-10T22:51:19Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-10T22:51:19Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-10T22:51:19Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-10T22:51:19Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-10T22:51:19Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-10T22:51:19Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"4fab3af6-f657-4dfc-8ef2-12b7c978c94f\\\",\\\"systemUUID\\\":\\\"442c0b68-8dce-47e0-bd68-d5210d7e0493\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T22:51:19Z is after 2025-08-24T17:21:41Z" Dec 10 22:51:19 crc kubenswrapper[4791]: I1210 22:51:19.749098 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 22:51:19 crc kubenswrapper[4791]: I1210 22:51:19.749137 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 22:51:19 crc kubenswrapper[4791]: I1210 22:51:19.749145 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 22:51:19 crc kubenswrapper[4791]: I1210 22:51:19.749159 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 22:51:19 crc kubenswrapper[4791]: I1210 22:51:19.749168 4791 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T22:51:19Z","lastTransitionTime":"2025-12-10T22:51:19Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 22:51:19 crc kubenswrapper[4791]: E1210 22:51:19.763571 4791 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-12-10T22:51:19Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-10T22:51:19Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-10T22:51:19Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-10T22:51:19Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-10T22:51:19Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-10T22:51:19Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-10T22:51:19Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-10T22:51:19Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"4fab3af6-f657-4dfc-8ef2-12b7c978c94f\\\",\\\"systemUUID\\\":\\\"442c0b68-8dce-47e0-bd68-d5210d7e0493\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T22:51:19Z is after 2025-08-24T17:21:41Z" Dec 10 22:51:19 crc kubenswrapper[4791]: I1210 22:51:19.766725 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 22:51:19 crc kubenswrapper[4791]: I1210 22:51:19.766786 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 22:51:19 crc kubenswrapper[4791]: I1210 22:51:19.766798 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 22:51:19 crc kubenswrapper[4791]: I1210 22:51:19.766816 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 22:51:19 crc kubenswrapper[4791]: I1210 22:51:19.766828 4791 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T22:51:19Z","lastTransitionTime":"2025-12-10T22:51:19Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 22:51:19 crc kubenswrapper[4791]: E1210 22:51:19.779036 4791 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-12-10T22:51:19Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-10T22:51:19Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-10T22:51:19Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-10T22:51:19Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-10T22:51:19Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-10T22:51:19Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-10T22:51:19Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-10T22:51:19Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"4fab3af6-f657-4dfc-8ef2-12b7c978c94f\\\",\\\"systemUUID\\\":\\\"442c0b68-8dce-47e0-bd68-d5210d7e0493\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T22:51:19Z is after 2025-08-24T17:21:41Z" Dec 10 22:51:19 crc kubenswrapper[4791]: I1210 22:51:19.782325 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 22:51:19 crc kubenswrapper[4791]: I1210 22:51:19.782448 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 22:51:19 crc kubenswrapper[4791]: I1210 22:51:19.782532 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 22:51:19 crc kubenswrapper[4791]: I1210 22:51:19.782621 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 22:51:19 crc kubenswrapper[4791]: I1210 22:51:19.782702 4791 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T22:51:19Z","lastTransitionTime":"2025-12-10T22:51:19Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 22:51:19 crc kubenswrapper[4791]: E1210 22:51:19.800200 4791 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-12-10T22:51:19Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-10T22:51:19Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-10T22:51:19Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-10T22:51:19Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-10T22:51:19Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-10T22:51:19Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-10T22:51:19Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-10T22:51:19Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"4fab3af6-f657-4dfc-8ef2-12b7c978c94f\\\",\\\"systemUUID\\\":\\\"442c0b68-8dce-47e0-bd68-d5210d7e0493\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T22:51:19Z is after 2025-08-24T17:21:41Z" Dec 10 22:51:19 crc kubenswrapper[4791]: E1210 22:51:19.800311 4791 kubelet_node_status.go:572] "Unable to update node status" err="update node status exceeds retry count" Dec 10 22:51:19 crc kubenswrapper[4791]: I1210 22:51:19.801859 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 22:51:19 crc kubenswrapper[4791]: I1210 22:51:19.801992 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 22:51:19 crc kubenswrapper[4791]: I1210 22:51:19.802131 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 22:51:19 crc kubenswrapper[4791]: I1210 22:51:19.802286 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 22:51:19 crc kubenswrapper[4791]: I1210 22:51:19.802422 4791 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T22:51:19Z","lastTransitionTime":"2025-12-10T22:51:19Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 22:51:19 crc kubenswrapper[4791]: I1210 22:51:19.904792 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 22:51:19 crc kubenswrapper[4791]: I1210 22:51:19.904849 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 22:51:19 crc kubenswrapper[4791]: I1210 22:51:19.904873 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 22:51:19 crc kubenswrapper[4791]: I1210 22:51:19.904903 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 22:51:19 crc kubenswrapper[4791]: I1210 22:51:19.904926 4791 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T22:51:19Z","lastTransitionTime":"2025-12-10T22:51:19Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 22:51:20 crc kubenswrapper[4791]: I1210 22:51:20.007301 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 22:51:20 crc kubenswrapper[4791]: I1210 22:51:20.007376 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 22:51:20 crc kubenswrapper[4791]: I1210 22:51:20.007391 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 22:51:20 crc kubenswrapper[4791]: I1210 22:51:20.007408 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 22:51:20 crc kubenswrapper[4791]: I1210 22:51:20.007418 4791 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T22:51:20Z","lastTransitionTime":"2025-12-10T22:51:20Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 22:51:20 crc kubenswrapper[4791]: I1210 22:51:20.110041 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 22:51:20 crc kubenswrapper[4791]: I1210 22:51:20.110164 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 22:51:20 crc kubenswrapper[4791]: I1210 22:51:20.110191 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 22:51:20 crc kubenswrapper[4791]: I1210 22:51:20.110262 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 22:51:20 crc kubenswrapper[4791]: I1210 22:51:20.110293 4791 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T22:51:20Z","lastTransitionTime":"2025-12-10T22:51:20Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 22:51:20 crc kubenswrapper[4791]: I1210 22:51:20.213963 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 22:51:20 crc kubenswrapper[4791]: I1210 22:51:20.214021 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 22:51:20 crc kubenswrapper[4791]: I1210 22:51:20.214032 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 22:51:20 crc kubenswrapper[4791]: I1210 22:51:20.214057 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 22:51:20 crc kubenswrapper[4791]: I1210 22:51:20.214074 4791 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T22:51:20Z","lastTransitionTime":"2025-12-10T22:51:20Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 22:51:20 crc kubenswrapper[4791]: I1210 22:51:20.317593 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 22:51:20 crc kubenswrapper[4791]: I1210 22:51:20.317694 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 22:51:20 crc kubenswrapper[4791]: I1210 22:51:20.317720 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 22:51:20 crc kubenswrapper[4791]: I1210 22:51:20.317760 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 22:51:20 crc kubenswrapper[4791]: I1210 22:51:20.317785 4791 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T22:51:20Z","lastTransitionTime":"2025-12-10T22:51:20Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 22:51:20 crc kubenswrapper[4791]: I1210 22:51:20.420321 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 22:51:20 crc kubenswrapper[4791]: I1210 22:51:20.420432 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 22:51:20 crc kubenswrapper[4791]: I1210 22:51:20.420452 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 22:51:20 crc kubenswrapper[4791]: I1210 22:51:20.420484 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 22:51:20 crc kubenswrapper[4791]: I1210 22:51:20.420507 4791 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T22:51:20Z","lastTransitionTime":"2025-12-10T22:51:20Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 22:51:20 crc kubenswrapper[4791]: I1210 22:51:20.523703 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 22:51:20 crc kubenswrapper[4791]: I1210 22:51:20.523774 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 22:51:20 crc kubenswrapper[4791]: I1210 22:51:20.523797 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 22:51:20 crc kubenswrapper[4791]: I1210 22:51:20.523825 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 22:51:20 crc kubenswrapper[4791]: I1210 22:51:20.523850 4791 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T22:51:20Z","lastTransitionTime":"2025-12-10T22:51:20Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 22:51:20 crc kubenswrapper[4791]: I1210 22:51:20.627068 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 22:51:20 crc kubenswrapper[4791]: I1210 22:51:20.627117 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 22:51:20 crc kubenswrapper[4791]: I1210 22:51:20.627127 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 22:51:20 crc kubenswrapper[4791]: I1210 22:51:20.627145 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 22:51:20 crc kubenswrapper[4791]: I1210 22:51:20.627156 4791 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T22:51:20Z","lastTransitionTime":"2025-12-10T22:51:20Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 22:51:20 crc kubenswrapper[4791]: I1210 22:51:20.729474 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 22:51:20 crc kubenswrapper[4791]: I1210 22:51:20.729526 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 22:51:20 crc kubenswrapper[4791]: I1210 22:51:20.729578 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 22:51:20 crc kubenswrapper[4791]: I1210 22:51:20.729599 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 22:51:20 crc kubenswrapper[4791]: I1210 22:51:20.729612 4791 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T22:51:20Z","lastTransitionTime":"2025-12-10T22:51:20Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 22:51:20 crc kubenswrapper[4791]: I1210 22:51:20.832009 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 22:51:20 crc kubenswrapper[4791]: I1210 22:51:20.832071 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 22:51:20 crc kubenswrapper[4791]: I1210 22:51:20.832084 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 22:51:20 crc kubenswrapper[4791]: I1210 22:51:20.832103 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 22:51:20 crc kubenswrapper[4791]: I1210 22:51:20.832115 4791 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T22:51:20Z","lastTransitionTime":"2025-12-10T22:51:20Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 22:51:20 crc kubenswrapper[4791]: I1210 22:51:20.883772 4791 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-b6kf6" Dec 10 22:51:20 crc kubenswrapper[4791]: I1210 22:51:20.883809 4791 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 10 22:51:20 crc kubenswrapper[4791]: I1210 22:51:20.883836 4791 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 10 22:51:20 crc kubenswrapper[4791]: I1210 22:51:20.883828 4791 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 10 22:51:20 crc kubenswrapper[4791]: E1210 22:51:20.883965 4791 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-b6kf6" podUID="b4527e71-bb58-4a39-a99c-fe614aa4c5f9" Dec 10 22:51:20 crc kubenswrapper[4791]: E1210 22:51:20.884081 4791 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 10 22:51:20 crc kubenswrapper[4791]: E1210 22:51:20.884193 4791 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 10 22:51:20 crc kubenswrapper[4791]: E1210 22:51:20.884288 4791 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 10 22:51:20 crc kubenswrapper[4791]: I1210 22:51:20.934991 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 22:51:20 crc kubenswrapper[4791]: I1210 22:51:20.935035 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 22:51:20 crc kubenswrapper[4791]: I1210 22:51:20.935045 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 22:51:20 crc kubenswrapper[4791]: I1210 22:51:20.935060 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 22:51:20 crc kubenswrapper[4791]: I1210 22:51:20.935073 4791 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T22:51:20Z","lastTransitionTime":"2025-12-10T22:51:20Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 22:51:21 crc kubenswrapper[4791]: I1210 22:51:21.038318 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 22:51:21 crc kubenswrapper[4791]: I1210 22:51:21.038442 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 22:51:21 crc kubenswrapper[4791]: I1210 22:51:21.038468 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 22:51:21 crc kubenswrapper[4791]: I1210 22:51:21.038500 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 22:51:21 crc kubenswrapper[4791]: I1210 22:51:21.038523 4791 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T22:51:21Z","lastTransitionTime":"2025-12-10T22:51:21Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 22:51:21 crc kubenswrapper[4791]: I1210 22:51:21.140553 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 22:51:21 crc kubenswrapper[4791]: I1210 22:51:21.140597 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 22:51:21 crc kubenswrapper[4791]: I1210 22:51:21.140609 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 22:51:21 crc kubenswrapper[4791]: I1210 22:51:21.140625 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 22:51:21 crc kubenswrapper[4791]: I1210 22:51:21.140638 4791 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T22:51:21Z","lastTransitionTime":"2025-12-10T22:51:21Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 22:51:21 crc kubenswrapper[4791]: I1210 22:51:21.243030 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 22:51:21 crc kubenswrapper[4791]: I1210 22:51:21.243077 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 22:51:21 crc kubenswrapper[4791]: I1210 22:51:21.243088 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 22:51:21 crc kubenswrapper[4791]: I1210 22:51:21.243103 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 22:51:21 crc kubenswrapper[4791]: I1210 22:51:21.243116 4791 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T22:51:21Z","lastTransitionTime":"2025-12-10T22:51:21Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 22:51:21 crc kubenswrapper[4791]: I1210 22:51:21.346079 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 22:51:21 crc kubenswrapper[4791]: I1210 22:51:21.346145 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 22:51:21 crc kubenswrapper[4791]: I1210 22:51:21.346162 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 22:51:21 crc kubenswrapper[4791]: I1210 22:51:21.346186 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 22:51:21 crc kubenswrapper[4791]: I1210 22:51:21.346203 4791 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T22:51:21Z","lastTransitionTime":"2025-12-10T22:51:21Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 22:51:21 crc kubenswrapper[4791]: I1210 22:51:21.449121 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 22:51:21 crc kubenswrapper[4791]: I1210 22:51:21.449211 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 22:51:21 crc kubenswrapper[4791]: I1210 22:51:21.449236 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 22:51:21 crc kubenswrapper[4791]: I1210 22:51:21.449268 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 22:51:21 crc kubenswrapper[4791]: I1210 22:51:21.449291 4791 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T22:51:21Z","lastTransitionTime":"2025-12-10T22:51:21Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 22:51:21 crc kubenswrapper[4791]: I1210 22:51:21.553075 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 22:51:21 crc kubenswrapper[4791]: I1210 22:51:21.553153 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 22:51:21 crc kubenswrapper[4791]: I1210 22:51:21.553168 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 22:51:21 crc kubenswrapper[4791]: I1210 22:51:21.553191 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 22:51:21 crc kubenswrapper[4791]: I1210 22:51:21.553206 4791 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T22:51:21Z","lastTransitionTime":"2025-12-10T22:51:21Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 22:51:21 crc kubenswrapper[4791]: I1210 22:51:21.655614 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 22:51:21 crc kubenswrapper[4791]: I1210 22:51:21.655654 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 22:51:21 crc kubenswrapper[4791]: I1210 22:51:21.655663 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 22:51:21 crc kubenswrapper[4791]: I1210 22:51:21.655678 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 22:51:21 crc kubenswrapper[4791]: I1210 22:51:21.655689 4791 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T22:51:21Z","lastTransitionTime":"2025-12-10T22:51:21Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 22:51:21 crc kubenswrapper[4791]: I1210 22:51:21.758371 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 22:51:21 crc kubenswrapper[4791]: I1210 22:51:21.758417 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 22:51:21 crc kubenswrapper[4791]: I1210 22:51:21.758431 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 22:51:21 crc kubenswrapper[4791]: I1210 22:51:21.758451 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 22:51:21 crc kubenswrapper[4791]: I1210 22:51:21.758466 4791 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T22:51:21Z","lastTransitionTime":"2025-12-10T22:51:21Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 22:51:21 crc kubenswrapper[4791]: I1210 22:51:21.860749 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 22:51:21 crc kubenswrapper[4791]: I1210 22:51:21.860784 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 22:51:21 crc kubenswrapper[4791]: I1210 22:51:21.860795 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 22:51:21 crc kubenswrapper[4791]: I1210 22:51:21.860809 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 22:51:21 crc kubenswrapper[4791]: I1210 22:51:21.860835 4791 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T22:51:21Z","lastTransitionTime":"2025-12-10T22:51:21Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 22:51:21 crc kubenswrapper[4791]: I1210 22:51:21.962890 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 22:51:21 crc kubenswrapper[4791]: I1210 22:51:21.962934 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 22:51:21 crc kubenswrapper[4791]: I1210 22:51:21.962947 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 22:51:21 crc kubenswrapper[4791]: I1210 22:51:21.962963 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 22:51:21 crc kubenswrapper[4791]: I1210 22:51:21.962974 4791 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T22:51:21Z","lastTransitionTime":"2025-12-10T22:51:21Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 22:51:22 crc kubenswrapper[4791]: I1210 22:51:22.065535 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 22:51:22 crc kubenswrapper[4791]: I1210 22:51:22.065588 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 22:51:22 crc kubenswrapper[4791]: I1210 22:51:22.065599 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 22:51:22 crc kubenswrapper[4791]: I1210 22:51:22.065618 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 22:51:22 crc kubenswrapper[4791]: I1210 22:51:22.065629 4791 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T22:51:22Z","lastTransitionTime":"2025-12-10T22:51:22Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 22:51:22 crc kubenswrapper[4791]: I1210 22:51:22.168662 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 22:51:22 crc kubenswrapper[4791]: I1210 22:51:22.168709 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 22:51:22 crc kubenswrapper[4791]: I1210 22:51:22.168720 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 22:51:22 crc kubenswrapper[4791]: I1210 22:51:22.168736 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 22:51:22 crc kubenswrapper[4791]: I1210 22:51:22.168746 4791 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T22:51:22Z","lastTransitionTime":"2025-12-10T22:51:22Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 22:51:22 crc kubenswrapper[4791]: I1210 22:51:22.271591 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 22:51:22 crc kubenswrapper[4791]: I1210 22:51:22.271661 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 22:51:22 crc kubenswrapper[4791]: I1210 22:51:22.271679 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 22:51:22 crc kubenswrapper[4791]: I1210 22:51:22.271706 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 22:51:22 crc kubenswrapper[4791]: I1210 22:51:22.271725 4791 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T22:51:22Z","lastTransitionTime":"2025-12-10T22:51:22Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 22:51:22 crc kubenswrapper[4791]: I1210 22:51:22.374256 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 22:51:22 crc kubenswrapper[4791]: I1210 22:51:22.374294 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 22:51:22 crc kubenswrapper[4791]: I1210 22:51:22.374306 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 22:51:22 crc kubenswrapper[4791]: I1210 22:51:22.374325 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 22:51:22 crc kubenswrapper[4791]: I1210 22:51:22.374356 4791 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T22:51:22Z","lastTransitionTime":"2025-12-10T22:51:22Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 22:51:22 crc kubenswrapper[4791]: I1210 22:51:22.477899 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 22:51:22 crc kubenswrapper[4791]: I1210 22:51:22.477974 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 22:51:22 crc kubenswrapper[4791]: I1210 22:51:22.477998 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 22:51:22 crc kubenswrapper[4791]: I1210 22:51:22.478024 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 22:51:22 crc kubenswrapper[4791]: I1210 22:51:22.478044 4791 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T22:51:22Z","lastTransitionTime":"2025-12-10T22:51:22Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 22:51:22 crc kubenswrapper[4791]: I1210 22:51:22.580707 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 22:51:22 crc kubenswrapper[4791]: I1210 22:51:22.580794 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 22:51:22 crc kubenswrapper[4791]: I1210 22:51:22.580816 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 22:51:22 crc kubenswrapper[4791]: I1210 22:51:22.580841 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 22:51:22 crc kubenswrapper[4791]: I1210 22:51:22.580859 4791 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T22:51:22Z","lastTransitionTime":"2025-12-10T22:51:22Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 22:51:22 crc kubenswrapper[4791]: I1210 22:51:22.684423 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 22:51:22 crc kubenswrapper[4791]: I1210 22:51:22.684485 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 22:51:22 crc kubenswrapper[4791]: I1210 22:51:22.684507 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 22:51:22 crc kubenswrapper[4791]: I1210 22:51:22.684535 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 22:51:22 crc kubenswrapper[4791]: I1210 22:51:22.684556 4791 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T22:51:22Z","lastTransitionTime":"2025-12-10T22:51:22Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 22:51:22 crc kubenswrapper[4791]: I1210 22:51:22.788454 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 22:51:22 crc kubenswrapper[4791]: I1210 22:51:22.788512 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 22:51:22 crc kubenswrapper[4791]: I1210 22:51:22.788523 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 22:51:22 crc kubenswrapper[4791]: I1210 22:51:22.788543 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 22:51:22 crc kubenswrapper[4791]: I1210 22:51:22.788559 4791 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T22:51:22Z","lastTransitionTime":"2025-12-10T22:51:22Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 22:51:22 crc kubenswrapper[4791]: I1210 22:51:22.884257 4791 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 10 22:51:22 crc kubenswrapper[4791]: I1210 22:51:22.884336 4791 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 10 22:51:22 crc kubenswrapper[4791]: E1210 22:51:22.884521 4791 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 10 22:51:22 crc kubenswrapper[4791]: I1210 22:51:22.884255 4791 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-b6kf6" Dec 10 22:51:22 crc kubenswrapper[4791]: E1210 22:51:22.884729 4791 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 10 22:51:22 crc kubenswrapper[4791]: I1210 22:51:22.884764 4791 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 10 22:51:22 crc kubenswrapper[4791]: E1210 22:51:22.884929 4791 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-b6kf6" podUID="b4527e71-bb58-4a39-a99c-fe614aa4c5f9" Dec 10 22:51:22 crc kubenswrapper[4791]: E1210 22:51:22.885111 4791 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 10 22:51:22 crc kubenswrapper[4791]: I1210 22:51:22.892078 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 22:51:22 crc kubenswrapper[4791]: I1210 22:51:22.892127 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 22:51:22 crc kubenswrapper[4791]: I1210 22:51:22.892145 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 22:51:22 crc kubenswrapper[4791]: I1210 22:51:22.892167 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 22:51:22 crc kubenswrapper[4791]: I1210 22:51:22.892184 4791 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T22:51:22Z","lastTransitionTime":"2025-12-10T22:51:22Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 22:51:22 crc kubenswrapper[4791]: I1210 22:51:22.995815 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 22:51:22 crc kubenswrapper[4791]: I1210 22:51:22.995900 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 22:51:22 crc kubenswrapper[4791]: I1210 22:51:22.995933 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 22:51:22 crc kubenswrapper[4791]: I1210 22:51:22.995966 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 22:51:22 crc kubenswrapper[4791]: I1210 22:51:22.995987 4791 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T22:51:22Z","lastTransitionTime":"2025-12-10T22:51:22Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 22:51:23 crc kubenswrapper[4791]: I1210 22:51:23.100072 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 22:51:23 crc kubenswrapper[4791]: I1210 22:51:23.100159 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 22:51:23 crc kubenswrapper[4791]: I1210 22:51:23.100181 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 22:51:23 crc kubenswrapper[4791]: I1210 22:51:23.100217 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 22:51:23 crc kubenswrapper[4791]: I1210 22:51:23.100240 4791 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T22:51:23Z","lastTransitionTime":"2025-12-10T22:51:23Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 22:51:23 crc kubenswrapper[4791]: I1210 22:51:23.203771 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 22:51:23 crc kubenswrapper[4791]: I1210 22:51:23.203826 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 22:51:23 crc kubenswrapper[4791]: I1210 22:51:23.203843 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 22:51:23 crc kubenswrapper[4791]: I1210 22:51:23.203879 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 22:51:23 crc kubenswrapper[4791]: I1210 22:51:23.203904 4791 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T22:51:23Z","lastTransitionTime":"2025-12-10T22:51:23Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 22:51:23 crc kubenswrapper[4791]: I1210 22:51:23.307110 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 22:51:23 crc kubenswrapper[4791]: I1210 22:51:23.307188 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 22:51:23 crc kubenswrapper[4791]: I1210 22:51:23.307214 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 22:51:23 crc kubenswrapper[4791]: I1210 22:51:23.307244 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 22:51:23 crc kubenswrapper[4791]: I1210 22:51:23.307268 4791 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T22:51:23Z","lastTransitionTime":"2025-12-10T22:51:23Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 22:51:23 crc kubenswrapper[4791]: I1210 22:51:23.411810 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 22:51:23 crc kubenswrapper[4791]: I1210 22:51:23.411874 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 22:51:23 crc kubenswrapper[4791]: I1210 22:51:23.411895 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 22:51:23 crc kubenswrapper[4791]: I1210 22:51:23.411927 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 22:51:23 crc kubenswrapper[4791]: I1210 22:51:23.411950 4791 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T22:51:23Z","lastTransitionTime":"2025-12-10T22:51:23Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 22:51:23 crc kubenswrapper[4791]: I1210 22:51:23.515003 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 22:51:23 crc kubenswrapper[4791]: I1210 22:51:23.515054 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 22:51:23 crc kubenswrapper[4791]: I1210 22:51:23.515067 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 22:51:23 crc kubenswrapper[4791]: I1210 22:51:23.515085 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 22:51:23 crc kubenswrapper[4791]: I1210 22:51:23.515097 4791 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T22:51:23Z","lastTransitionTime":"2025-12-10T22:51:23Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 22:51:23 crc kubenswrapper[4791]: I1210 22:51:23.618814 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 22:51:23 crc kubenswrapper[4791]: I1210 22:51:23.618907 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 22:51:23 crc kubenswrapper[4791]: I1210 22:51:23.618926 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 22:51:23 crc kubenswrapper[4791]: I1210 22:51:23.618954 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 22:51:23 crc kubenswrapper[4791]: I1210 22:51:23.618973 4791 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T22:51:23Z","lastTransitionTime":"2025-12-10T22:51:23Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 22:51:23 crc kubenswrapper[4791]: I1210 22:51:23.721857 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 22:51:23 crc kubenswrapper[4791]: I1210 22:51:23.721952 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 22:51:23 crc kubenswrapper[4791]: I1210 22:51:23.721974 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 22:51:23 crc kubenswrapper[4791]: I1210 22:51:23.721995 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 22:51:23 crc kubenswrapper[4791]: I1210 22:51:23.722009 4791 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T22:51:23Z","lastTransitionTime":"2025-12-10T22:51:23Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 22:51:23 crc kubenswrapper[4791]: I1210 22:51:23.825783 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 22:51:23 crc kubenswrapper[4791]: I1210 22:51:23.825857 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 22:51:23 crc kubenswrapper[4791]: I1210 22:51:23.825892 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 22:51:23 crc kubenswrapper[4791]: I1210 22:51:23.825922 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 22:51:23 crc kubenswrapper[4791]: I1210 22:51:23.825941 4791 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T22:51:23Z","lastTransitionTime":"2025-12-10T22:51:23Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 22:51:23 crc kubenswrapper[4791]: I1210 22:51:23.929948 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 22:51:23 crc kubenswrapper[4791]: I1210 22:51:23.930018 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 22:51:23 crc kubenswrapper[4791]: I1210 22:51:23.930037 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 22:51:23 crc kubenswrapper[4791]: I1210 22:51:23.930068 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 22:51:23 crc kubenswrapper[4791]: I1210 22:51:23.930085 4791 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T22:51:23Z","lastTransitionTime":"2025-12-10T22:51:23Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 22:51:24 crc kubenswrapper[4791]: I1210 22:51:24.033284 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 22:51:24 crc kubenswrapper[4791]: I1210 22:51:24.033379 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 22:51:24 crc kubenswrapper[4791]: I1210 22:51:24.033392 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 22:51:24 crc kubenswrapper[4791]: I1210 22:51:24.033408 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 22:51:24 crc kubenswrapper[4791]: I1210 22:51:24.033418 4791 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T22:51:24Z","lastTransitionTime":"2025-12-10T22:51:24Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 22:51:24 crc kubenswrapper[4791]: I1210 22:51:24.135543 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 22:51:24 crc kubenswrapper[4791]: I1210 22:51:24.135583 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 22:51:24 crc kubenswrapper[4791]: I1210 22:51:24.135592 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 22:51:24 crc kubenswrapper[4791]: I1210 22:51:24.135607 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 22:51:24 crc kubenswrapper[4791]: I1210 22:51:24.135619 4791 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T22:51:24Z","lastTransitionTime":"2025-12-10T22:51:24Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 22:51:24 crc kubenswrapper[4791]: I1210 22:51:24.239239 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 22:51:24 crc kubenswrapper[4791]: I1210 22:51:24.239313 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 22:51:24 crc kubenswrapper[4791]: I1210 22:51:24.239377 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 22:51:24 crc kubenswrapper[4791]: I1210 22:51:24.239410 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 22:51:24 crc kubenswrapper[4791]: I1210 22:51:24.239433 4791 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T22:51:24Z","lastTransitionTime":"2025-12-10T22:51:24Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 22:51:24 crc kubenswrapper[4791]: I1210 22:51:24.343201 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 22:51:24 crc kubenswrapper[4791]: I1210 22:51:24.343233 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 22:51:24 crc kubenswrapper[4791]: I1210 22:51:24.343241 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 22:51:24 crc kubenswrapper[4791]: I1210 22:51:24.343254 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 22:51:24 crc kubenswrapper[4791]: I1210 22:51:24.343263 4791 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T22:51:24Z","lastTransitionTime":"2025-12-10T22:51:24Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 22:51:24 crc kubenswrapper[4791]: I1210 22:51:24.446314 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 22:51:24 crc kubenswrapper[4791]: I1210 22:51:24.446368 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 22:51:24 crc kubenswrapper[4791]: I1210 22:51:24.446378 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 22:51:24 crc kubenswrapper[4791]: I1210 22:51:24.446393 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 22:51:24 crc kubenswrapper[4791]: I1210 22:51:24.446421 4791 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T22:51:24Z","lastTransitionTime":"2025-12-10T22:51:24Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 22:51:24 crc kubenswrapper[4791]: I1210 22:51:24.548861 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 22:51:24 crc kubenswrapper[4791]: I1210 22:51:24.548902 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 22:51:24 crc kubenswrapper[4791]: I1210 22:51:24.548913 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 22:51:24 crc kubenswrapper[4791]: I1210 22:51:24.548929 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 22:51:24 crc kubenswrapper[4791]: I1210 22:51:24.548941 4791 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T22:51:24Z","lastTransitionTime":"2025-12-10T22:51:24Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 22:51:24 crc kubenswrapper[4791]: I1210 22:51:24.652537 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 22:51:24 crc kubenswrapper[4791]: I1210 22:51:24.652604 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 22:51:24 crc kubenswrapper[4791]: I1210 22:51:24.652627 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 22:51:24 crc kubenswrapper[4791]: I1210 22:51:24.652655 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 22:51:24 crc kubenswrapper[4791]: I1210 22:51:24.652677 4791 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T22:51:24Z","lastTransitionTime":"2025-12-10T22:51:24Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 22:51:24 crc kubenswrapper[4791]: I1210 22:51:24.756127 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 22:51:24 crc kubenswrapper[4791]: I1210 22:51:24.756204 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 22:51:24 crc kubenswrapper[4791]: I1210 22:51:24.756228 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 22:51:24 crc kubenswrapper[4791]: I1210 22:51:24.756254 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 22:51:24 crc kubenswrapper[4791]: I1210 22:51:24.756410 4791 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T22:51:24Z","lastTransitionTime":"2025-12-10T22:51:24Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 22:51:24 crc kubenswrapper[4791]: I1210 22:51:24.859045 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 22:51:24 crc kubenswrapper[4791]: I1210 22:51:24.859111 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 22:51:24 crc kubenswrapper[4791]: I1210 22:51:24.859131 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 22:51:24 crc kubenswrapper[4791]: I1210 22:51:24.859154 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 22:51:24 crc kubenswrapper[4791]: I1210 22:51:24.859170 4791 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T22:51:24Z","lastTransitionTime":"2025-12-10T22:51:24Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 22:51:24 crc kubenswrapper[4791]: I1210 22:51:24.884656 4791 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-b6kf6" Dec 10 22:51:24 crc kubenswrapper[4791]: I1210 22:51:24.884777 4791 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 10 22:51:24 crc kubenswrapper[4791]: E1210 22:51:24.884976 4791 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-b6kf6" podUID="b4527e71-bb58-4a39-a99c-fe614aa4c5f9" Dec 10 22:51:24 crc kubenswrapper[4791]: I1210 22:51:24.885023 4791 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 10 22:51:24 crc kubenswrapper[4791]: I1210 22:51:24.885054 4791 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 10 22:51:24 crc kubenswrapper[4791]: E1210 22:51:24.885197 4791 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 10 22:51:24 crc kubenswrapper[4791]: E1210 22:51:24.885424 4791 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 10 22:51:24 crc kubenswrapper[4791]: E1210 22:51:24.885595 4791 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 10 22:51:24 crc kubenswrapper[4791]: I1210 22:51:24.886580 4791 scope.go:117] "RemoveContainer" containerID="5f299a2b60de0ef5fb17a1e463d8552d6c4d11d166ae32e490104b3d90116d25" Dec 10 22:51:24 crc kubenswrapper[4791]: E1210 22:51:24.886809 4791 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"ovnkube-controller\" with CrashLoopBackOff: \"back-off 40s restarting failed container=ovnkube-controller pod=ovnkube-node-zhq64_openshift-ovn-kubernetes(3cd47739-0fa9-4321-aff1-220f8721a0b3)\"" pod="openshift-ovn-kubernetes/ovnkube-node-zhq64" podUID="3cd47739-0fa9-4321-aff1-220f8721a0b3" Dec 10 22:51:24 crc kubenswrapper[4791]: I1210 22:51:24.961876 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 22:51:24 crc kubenswrapper[4791]: I1210 22:51:24.961945 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 22:51:24 crc kubenswrapper[4791]: I1210 22:51:24.961971 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 22:51:24 crc kubenswrapper[4791]: I1210 22:51:24.962022 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 22:51:24 crc kubenswrapper[4791]: I1210 22:51:24.962047 4791 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T22:51:24Z","lastTransitionTime":"2025-12-10T22:51:24Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 22:51:25 crc kubenswrapper[4791]: I1210 22:51:25.065325 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 22:51:25 crc kubenswrapper[4791]: I1210 22:51:25.065387 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 22:51:25 crc kubenswrapper[4791]: I1210 22:51:25.065397 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 22:51:25 crc kubenswrapper[4791]: I1210 22:51:25.065412 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 22:51:25 crc kubenswrapper[4791]: I1210 22:51:25.065422 4791 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T22:51:25Z","lastTransitionTime":"2025-12-10T22:51:25Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 22:51:25 crc kubenswrapper[4791]: I1210 22:51:25.168322 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 22:51:25 crc kubenswrapper[4791]: I1210 22:51:25.168401 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 22:51:25 crc kubenswrapper[4791]: I1210 22:51:25.168417 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 22:51:25 crc kubenswrapper[4791]: I1210 22:51:25.168440 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 22:51:25 crc kubenswrapper[4791]: I1210 22:51:25.168454 4791 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T22:51:25Z","lastTransitionTime":"2025-12-10T22:51:25Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 22:51:25 crc kubenswrapper[4791]: I1210 22:51:25.271740 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 22:51:25 crc kubenswrapper[4791]: I1210 22:51:25.271789 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 22:51:25 crc kubenswrapper[4791]: I1210 22:51:25.271797 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 22:51:25 crc kubenswrapper[4791]: I1210 22:51:25.271812 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 22:51:25 crc kubenswrapper[4791]: I1210 22:51:25.271821 4791 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T22:51:25Z","lastTransitionTime":"2025-12-10T22:51:25Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 22:51:25 crc kubenswrapper[4791]: I1210 22:51:25.374283 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 22:51:25 crc kubenswrapper[4791]: I1210 22:51:25.374383 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 22:51:25 crc kubenswrapper[4791]: I1210 22:51:25.374423 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 22:51:25 crc kubenswrapper[4791]: I1210 22:51:25.374454 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 22:51:25 crc kubenswrapper[4791]: I1210 22:51:25.374479 4791 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T22:51:25Z","lastTransitionTime":"2025-12-10T22:51:25Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 22:51:25 crc kubenswrapper[4791]: I1210 22:51:25.477939 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 22:51:25 crc kubenswrapper[4791]: I1210 22:51:25.477986 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 22:51:25 crc kubenswrapper[4791]: I1210 22:51:25.477999 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 22:51:25 crc kubenswrapper[4791]: I1210 22:51:25.478019 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 22:51:25 crc kubenswrapper[4791]: I1210 22:51:25.478030 4791 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T22:51:25Z","lastTransitionTime":"2025-12-10T22:51:25Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 22:51:25 crc kubenswrapper[4791]: I1210 22:51:25.581023 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 22:51:25 crc kubenswrapper[4791]: I1210 22:51:25.581059 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 22:51:25 crc kubenswrapper[4791]: I1210 22:51:25.581070 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 22:51:25 crc kubenswrapper[4791]: I1210 22:51:25.581086 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 22:51:25 crc kubenswrapper[4791]: I1210 22:51:25.581096 4791 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T22:51:25Z","lastTransitionTime":"2025-12-10T22:51:25Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 22:51:25 crc kubenswrapper[4791]: I1210 22:51:25.681517 4791 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-multus_multus-4nwkq_672aa28c-8169-49ed-87b8-21187d13a80c/kube-multus/1.log" Dec 10 22:51:25 crc kubenswrapper[4791]: I1210 22:51:25.682241 4791 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-multus_multus-4nwkq_672aa28c-8169-49ed-87b8-21187d13a80c/kube-multus/0.log" Dec 10 22:51:25 crc kubenswrapper[4791]: I1210 22:51:25.682328 4791 generic.go:334] "Generic (PLEG): container finished" podID="672aa28c-8169-49ed-87b8-21187d13a80c" containerID="f337f881ffd71f4a2ee19e92bb0b92c76cbb9b8352511f741d0c87e7da4cdd18" exitCode=1 Dec 10 22:51:25 crc kubenswrapper[4791]: I1210 22:51:25.682410 4791 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-4nwkq" event={"ID":"672aa28c-8169-49ed-87b8-21187d13a80c","Type":"ContainerDied","Data":"f337f881ffd71f4a2ee19e92bb0b92c76cbb9b8352511f741d0c87e7da4cdd18"} Dec 10 22:51:25 crc kubenswrapper[4791]: I1210 22:51:25.682461 4791 scope.go:117] "RemoveContainer" containerID="a2f2a62315529e148a10a33ccb241e9417fbde0fe42afb61cb7ee2438c74636f" Dec 10 22:51:25 crc kubenswrapper[4791]: I1210 22:51:25.683057 4791 scope.go:117] "RemoveContainer" containerID="f337f881ffd71f4a2ee19e92bb0b92c76cbb9b8352511f741d0c87e7da4cdd18" Dec 10 22:51:25 crc kubenswrapper[4791]: E1210 22:51:25.683381 4791 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"kube-multus\" with CrashLoopBackOff: \"back-off 10s restarting failed container=kube-multus pod=multus-4nwkq_openshift-multus(672aa28c-8169-49ed-87b8-21187d13a80c)\"" pod="openshift-multus/multus-4nwkq" podUID="672aa28c-8169-49ed-87b8-21187d13a80c" Dec 10 22:51:25 crc kubenswrapper[4791]: I1210 22:51:25.683613 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 22:51:25 crc kubenswrapper[4791]: I1210 22:51:25.683777 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 22:51:25 crc kubenswrapper[4791]: I1210 22:51:25.683817 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 22:51:25 crc kubenswrapper[4791]: I1210 22:51:25.683902 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 22:51:25 crc kubenswrapper[4791]: I1210 22:51:25.683961 4791 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T22:51:25Z","lastTransitionTime":"2025-12-10T22:51:25Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 22:51:25 crc kubenswrapper[4791]: I1210 22:51:25.723083 4791 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-image-registry/node-ca-vczjq" podStartSLOduration=95.723048273 podStartE2EDuration="1m35.723048273s" podCreationTimestamp="2025-12-10 22:49:50 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-10 22:51:25.706524094 +0000 UTC m=+120.136141747" watchObservedRunningTime="2025-12-10 22:51:25.723048273 +0000 UTC m=+120.152665926" Dec 10 22:51:25 crc kubenswrapper[4791]: I1210 22:51:25.785845 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 22:51:25 crc kubenswrapper[4791]: I1210 22:51:25.785879 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 22:51:25 crc kubenswrapper[4791]: I1210 22:51:25.785890 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 22:51:25 crc kubenswrapper[4791]: I1210 22:51:25.785905 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 22:51:25 crc kubenswrapper[4791]: I1210 22:51:25.785916 4791 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T22:51:25Z","lastTransitionTime":"2025-12-10T22:51:25Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 22:51:25 crc kubenswrapper[4791]: I1210 22:51:25.788761 4791 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-kube-controller-manager/kube-controller-manager-crc" podStartSLOduration=95.788746019 podStartE2EDuration="1m35.788746019s" podCreationTimestamp="2025-12-10 22:49:50 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-10 22:51:25.77261113 +0000 UTC m=+120.202228753" watchObservedRunningTime="2025-12-10 22:51:25.788746019 +0000 UTC m=+120.218363652" Dec 10 22:51:25 crc kubenswrapper[4791]: I1210 22:51:25.789796 4791 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" podStartSLOduration=48.789772105 podStartE2EDuration="48.789772105s" podCreationTimestamp="2025-12-10 22:50:37 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-10 22:51:25.789502008 +0000 UTC m=+120.219119641" watchObservedRunningTime="2025-12-10 22:51:25.789772105 +0000 UTC m=+120.219389738" Dec 10 22:51:25 crc kubenswrapper[4791]: I1210 22:51:25.831859 4791 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-dns/node-resolver-tchxk" podStartSLOduration=96.831835652 podStartE2EDuration="1m36.831835652s" podCreationTimestamp="2025-12-10 22:49:49 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-10 22:51:25.831445362 +0000 UTC m=+120.261063025" watchObservedRunningTime="2025-12-10 22:51:25.831835652 +0000 UTC m=+120.261453265" Dec 10 22:51:25 crc kubenswrapper[4791]: I1210 22:51:25.858755 4791 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-multus/multus-additional-cni-plugins-rw584" podStartSLOduration=94.858733994 podStartE2EDuration="1m34.858733994s" podCreationTimestamp="2025-12-10 22:49:51 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-10 22:51:25.858522428 +0000 UTC m=+120.288140041" watchObservedRunningTime="2025-12-10 22:51:25.858733994 +0000 UTC m=+120.288351617" Dec 10 22:51:25 crc kubenswrapper[4791]: E1210 22:51:25.886284 4791 kubelet_node_status.go:497] "Node not becoming ready in time after startup" Dec 10 22:51:25 crc kubenswrapper[4791]: I1210 22:51:25.955518 4791 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-kube-apiserver/kube-apiserver-crc" podStartSLOduration=95.955496587 podStartE2EDuration="1m35.955496587s" podCreationTimestamp="2025-12-10 22:49:50 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-10 22:51:25.955232441 +0000 UTC m=+120.384850074" watchObservedRunningTime="2025-12-10 22:51:25.955496587 +0000 UTC m=+120.385114210" Dec 10 22:51:25 crc kubenswrapper[4791]: E1210 22:51:25.970307 4791 kubelet.go:2916] "Container runtime network not ready" networkReady="NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" Dec 10 22:51:26 crc kubenswrapper[4791]: I1210 22:51:26.002295 4791 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" podStartSLOduration=66.002269664 podStartE2EDuration="1m6.002269664s" podCreationTimestamp="2025-12-10 22:50:20 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-10 22:51:25.971103073 +0000 UTC m=+120.400720696" watchObservedRunningTime="2025-12-10 22:51:26.002269664 +0000 UTC m=+120.431887287" Dec 10 22:51:26 crc kubenswrapper[4791]: I1210 22:51:26.002804 4791 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-etcd/etcd-crc" podStartSLOduration=94.002796757 podStartE2EDuration="1m34.002796757s" podCreationTimestamp="2025-12-10 22:49:52 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-10 22:51:26.001709439 +0000 UTC m=+120.431327072" watchObservedRunningTime="2025-12-10 22:51:26.002796757 +0000 UTC m=+120.432414380" Dec 10 22:51:26 crc kubenswrapper[4791]: I1210 22:51:26.032071 4791 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-machine-config-operator/machine-config-daemon-5rb5l" podStartSLOduration=96.032049919 podStartE2EDuration="1m36.032049919s" podCreationTimestamp="2025-12-10 22:49:50 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-10 22:51:26.017806178 +0000 UTC m=+120.447423801" watchObservedRunningTime="2025-12-10 22:51:26.032049919 +0000 UTC m=+120.461667532" Dec 10 22:51:26 crc kubenswrapper[4791]: I1210 22:51:26.032729 4791 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-s6rk8" podStartSLOduration=95.032720316 podStartE2EDuration="1m35.032720316s" podCreationTimestamp="2025-12-10 22:49:51 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-10 22:51:26.031603468 +0000 UTC m=+120.461221091" watchObservedRunningTime="2025-12-10 22:51:26.032720316 +0000 UTC m=+120.462337929" Dec 10 22:51:26 crc kubenswrapper[4791]: I1210 22:51:26.687282 4791 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-multus_multus-4nwkq_672aa28c-8169-49ed-87b8-21187d13a80c/kube-multus/1.log" Dec 10 22:51:26 crc kubenswrapper[4791]: I1210 22:51:26.884058 4791 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 10 22:51:26 crc kubenswrapper[4791]: I1210 22:51:26.884113 4791 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-b6kf6" Dec 10 22:51:26 crc kubenswrapper[4791]: I1210 22:51:26.884173 4791 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 10 22:51:26 crc kubenswrapper[4791]: I1210 22:51:26.884058 4791 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 10 22:51:26 crc kubenswrapper[4791]: E1210 22:51:26.884167 4791 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 10 22:51:26 crc kubenswrapper[4791]: E1210 22:51:26.884228 4791 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 10 22:51:26 crc kubenswrapper[4791]: E1210 22:51:26.884269 4791 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 10 22:51:26 crc kubenswrapper[4791]: E1210 22:51:26.884323 4791 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-b6kf6" podUID="b4527e71-bb58-4a39-a99c-fe614aa4c5f9" Dec 10 22:51:28 crc kubenswrapper[4791]: I1210 22:51:28.883720 4791 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-b6kf6" Dec 10 22:51:28 crc kubenswrapper[4791]: I1210 22:51:28.883790 4791 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 10 22:51:28 crc kubenswrapper[4791]: I1210 22:51:28.883806 4791 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 10 22:51:28 crc kubenswrapper[4791]: I1210 22:51:28.883724 4791 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 10 22:51:28 crc kubenswrapper[4791]: E1210 22:51:28.883952 4791 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-b6kf6" podUID="b4527e71-bb58-4a39-a99c-fe614aa4c5f9" Dec 10 22:51:28 crc kubenswrapper[4791]: E1210 22:51:28.884030 4791 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 10 22:51:28 crc kubenswrapper[4791]: E1210 22:51:28.884155 4791 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 10 22:51:28 crc kubenswrapper[4791]: E1210 22:51:28.884298 4791 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 10 22:51:29 crc kubenswrapper[4791]: I1210 22:51:29.885766 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 22:51:29 crc kubenswrapper[4791]: I1210 22:51:29.885794 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 22:51:29 crc kubenswrapper[4791]: I1210 22:51:29.885802 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 22:51:29 crc kubenswrapper[4791]: I1210 22:51:29.885814 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 22:51:29 crc kubenswrapper[4791]: I1210 22:51:29.885822 4791 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T22:51:29Z","lastTransitionTime":"2025-12-10T22:51:29Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 22:51:29 crc kubenswrapper[4791]: I1210 22:51:29.927818 4791 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-cluster-version/cluster-version-operator-5c965bbfc6-rq6tz"] Dec 10 22:51:29 crc kubenswrapper[4791]: I1210 22:51:29.928228 4791 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-rq6tz" Dec 10 22:51:29 crc kubenswrapper[4791]: I1210 22:51:29.930071 4791 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-version"/"kube-root-ca.crt" Dec 10 22:51:29 crc kubenswrapper[4791]: I1210 22:51:29.930099 4791 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-cluster-version"/"cluster-version-operator-serving-cert" Dec 10 22:51:29 crc kubenswrapper[4791]: I1210 22:51:29.930277 4791 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-cluster-version"/"default-dockercfg-gxtc4" Dec 10 22:51:29 crc kubenswrapper[4791]: I1210 22:51:29.931008 4791 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-version"/"openshift-service-ca.crt" Dec 10 22:51:29 crc kubenswrapper[4791]: I1210 22:51:29.986699 4791 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/375c54a9-f930-4c61-b9be-0a87a3df6faf-kube-api-access\") pod \"cluster-version-operator-5c965bbfc6-rq6tz\" (UID: \"375c54a9-f930-4c61-b9be-0a87a3df6faf\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-rq6tz" Dec 10 22:51:29 crc kubenswrapper[4791]: I1210 22:51:29.986762 4791 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-cvo-updatepayloads\" (UniqueName: \"kubernetes.io/host-path/375c54a9-f930-4c61-b9be-0a87a3df6faf-etc-cvo-updatepayloads\") pod \"cluster-version-operator-5c965bbfc6-rq6tz\" (UID: \"375c54a9-f930-4c61-b9be-0a87a3df6faf\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-rq6tz" Dec 10 22:51:29 crc kubenswrapper[4791]: I1210 22:51:29.986791 4791 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/375c54a9-f930-4c61-b9be-0a87a3df6faf-service-ca\") pod \"cluster-version-operator-5c965bbfc6-rq6tz\" (UID: \"375c54a9-f930-4c61-b9be-0a87a3df6faf\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-rq6tz" Dec 10 22:51:29 crc kubenswrapper[4791]: I1210 22:51:29.986823 4791 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-ssl-certs\" (UniqueName: \"kubernetes.io/host-path/375c54a9-f930-4c61-b9be-0a87a3df6faf-etc-ssl-certs\") pod \"cluster-version-operator-5c965bbfc6-rq6tz\" (UID: \"375c54a9-f930-4c61-b9be-0a87a3df6faf\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-rq6tz" Dec 10 22:51:29 crc kubenswrapper[4791]: I1210 22:51:29.986855 4791 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/375c54a9-f930-4c61-b9be-0a87a3df6faf-serving-cert\") pod \"cluster-version-operator-5c965bbfc6-rq6tz\" (UID: \"375c54a9-f930-4c61-b9be-0a87a3df6faf\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-rq6tz" Dec 10 22:51:30 crc kubenswrapper[4791]: I1210 22:51:30.087873 4791 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/375c54a9-f930-4c61-b9be-0a87a3df6faf-kube-api-access\") pod \"cluster-version-operator-5c965bbfc6-rq6tz\" (UID: \"375c54a9-f930-4c61-b9be-0a87a3df6faf\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-rq6tz" Dec 10 22:51:30 crc kubenswrapper[4791]: I1210 22:51:30.087946 4791 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-cvo-updatepayloads\" (UniqueName: \"kubernetes.io/host-path/375c54a9-f930-4c61-b9be-0a87a3df6faf-etc-cvo-updatepayloads\") pod \"cluster-version-operator-5c965bbfc6-rq6tz\" (UID: \"375c54a9-f930-4c61-b9be-0a87a3df6faf\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-rq6tz" Dec 10 22:51:30 crc kubenswrapper[4791]: I1210 22:51:30.087980 4791 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/375c54a9-f930-4c61-b9be-0a87a3df6faf-service-ca\") pod \"cluster-version-operator-5c965bbfc6-rq6tz\" (UID: \"375c54a9-f930-4c61-b9be-0a87a3df6faf\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-rq6tz" Dec 10 22:51:30 crc kubenswrapper[4791]: I1210 22:51:30.088004 4791 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-ssl-certs\" (UniqueName: \"kubernetes.io/host-path/375c54a9-f930-4c61-b9be-0a87a3df6faf-etc-ssl-certs\") pod \"cluster-version-operator-5c965bbfc6-rq6tz\" (UID: \"375c54a9-f930-4c61-b9be-0a87a3df6faf\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-rq6tz" Dec 10 22:51:30 crc kubenswrapper[4791]: I1210 22:51:30.088041 4791 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/375c54a9-f930-4c61-b9be-0a87a3df6faf-serving-cert\") pod \"cluster-version-operator-5c965bbfc6-rq6tz\" (UID: \"375c54a9-f930-4c61-b9be-0a87a3df6faf\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-rq6tz" Dec 10 22:51:30 crc kubenswrapper[4791]: I1210 22:51:30.088109 4791 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-cvo-updatepayloads\" (UniqueName: \"kubernetes.io/host-path/375c54a9-f930-4c61-b9be-0a87a3df6faf-etc-cvo-updatepayloads\") pod \"cluster-version-operator-5c965bbfc6-rq6tz\" (UID: \"375c54a9-f930-4c61-b9be-0a87a3df6faf\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-rq6tz" Dec 10 22:51:30 crc kubenswrapper[4791]: I1210 22:51:30.088166 4791 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-ssl-certs\" (UniqueName: \"kubernetes.io/host-path/375c54a9-f930-4c61-b9be-0a87a3df6faf-etc-ssl-certs\") pod \"cluster-version-operator-5c965bbfc6-rq6tz\" (UID: \"375c54a9-f930-4c61-b9be-0a87a3df6faf\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-rq6tz" Dec 10 22:51:30 crc kubenswrapper[4791]: I1210 22:51:30.089373 4791 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/375c54a9-f930-4c61-b9be-0a87a3df6faf-service-ca\") pod \"cluster-version-operator-5c965bbfc6-rq6tz\" (UID: \"375c54a9-f930-4c61-b9be-0a87a3df6faf\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-rq6tz" Dec 10 22:51:30 crc kubenswrapper[4791]: I1210 22:51:30.096391 4791 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/375c54a9-f930-4c61-b9be-0a87a3df6faf-serving-cert\") pod \"cluster-version-operator-5c965bbfc6-rq6tz\" (UID: \"375c54a9-f930-4c61-b9be-0a87a3df6faf\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-rq6tz" Dec 10 22:51:30 crc kubenswrapper[4791]: I1210 22:51:30.113975 4791 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/375c54a9-f930-4c61-b9be-0a87a3df6faf-kube-api-access\") pod \"cluster-version-operator-5c965bbfc6-rq6tz\" (UID: \"375c54a9-f930-4c61-b9be-0a87a3df6faf\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-rq6tz" Dec 10 22:51:30 crc kubenswrapper[4791]: I1210 22:51:30.244137 4791 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-rq6tz" Dec 10 22:51:30 crc kubenswrapper[4791]: I1210 22:51:30.704929 4791 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-rq6tz" event={"ID":"375c54a9-f930-4c61-b9be-0a87a3df6faf","Type":"ContainerStarted","Data":"78e68382f34b3e35d155ce5eca9c752dab29156a2832de522590ce38848e4a06"} Dec 10 22:51:30 crc kubenswrapper[4791]: I1210 22:51:30.705262 4791 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-rq6tz" event={"ID":"375c54a9-f930-4c61-b9be-0a87a3df6faf","Type":"ContainerStarted","Data":"61d04d886666059b8de23356d92d70fa13c56a64ff43cb2666e024fea306cb72"} Dec 10 22:51:30 crc kubenswrapper[4791]: I1210 22:51:30.726915 4791 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-rq6tz" podStartSLOduration=100.726890883 podStartE2EDuration="1m40.726890883s" podCreationTimestamp="2025-12-10 22:49:50 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-10 22:51:30.725274291 +0000 UTC m=+125.154891934" watchObservedRunningTime="2025-12-10 22:51:30.726890883 +0000 UTC m=+125.156508536" Dec 10 22:51:30 crc kubenswrapper[4791]: I1210 22:51:30.884102 4791 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-b6kf6" Dec 10 22:51:30 crc kubenswrapper[4791]: I1210 22:51:30.884241 4791 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 10 22:51:30 crc kubenswrapper[4791]: I1210 22:51:30.884287 4791 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 10 22:51:30 crc kubenswrapper[4791]: E1210 22:51:30.884481 4791 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-b6kf6" podUID="b4527e71-bb58-4a39-a99c-fe614aa4c5f9" Dec 10 22:51:30 crc kubenswrapper[4791]: I1210 22:51:30.884506 4791 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 10 22:51:30 crc kubenswrapper[4791]: E1210 22:51:30.884575 4791 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 10 22:51:30 crc kubenswrapper[4791]: E1210 22:51:30.884712 4791 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 10 22:51:30 crc kubenswrapper[4791]: E1210 22:51:30.884827 4791 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 10 22:51:30 crc kubenswrapper[4791]: E1210 22:51:30.972143 4791 kubelet.go:2916] "Container runtime network not ready" networkReady="NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" Dec 10 22:51:32 crc kubenswrapper[4791]: I1210 22:51:32.884231 4791 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-b6kf6" Dec 10 22:51:32 crc kubenswrapper[4791]: I1210 22:51:32.884302 4791 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 10 22:51:32 crc kubenswrapper[4791]: I1210 22:51:32.884232 4791 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 10 22:51:32 crc kubenswrapper[4791]: E1210 22:51:32.884383 4791 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-b6kf6" podUID="b4527e71-bb58-4a39-a99c-fe614aa4c5f9" Dec 10 22:51:32 crc kubenswrapper[4791]: E1210 22:51:32.884436 4791 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 10 22:51:32 crc kubenswrapper[4791]: I1210 22:51:32.884303 4791 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 10 22:51:32 crc kubenswrapper[4791]: E1210 22:51:32.884547 4791 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 10 22:51:32 crc kubenswrapper[4791]: E1210 22:51:32.884641 4791 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 10 22:51:34 crc kubenswrapper[4791]: I1210 22:51:34.884037 4791 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 10 22:51:34 crc kubenswrapper[4791]: I1210 22:51:34.884103 4791 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-b6kf6" Dec 10 22:51:34 crc kubenswrapper[4791]: I1210 22:51:34.884049 4791 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 10 22:51:34 crc kubenswrapper[4791]: I1210 22:51:34.884183 4791 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 10 22:51:34 crc kubenswrapper[4791]: E1210 22:51:34.884260 4791 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 10 22:51:34 crc kubenswrapper[4791]: E1210 22:51:34.884391 4791 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-b6kf6" podUID="b4527e71-bb58-4a39-a99c-fe614aa4c5f9" Dec 10 22:51:34 crc kubenswrapper[4791]: E1210 22:51:34.884478 4791 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 10 22:51:34 crc kubenswrapper[4791]: E1210 22:51:34.884647 4791 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 10 22:51:35 crc kubenswrapper[4791]: E1210 22:51:35.973070 4791 kubelet.go:2916] "Container runtime network not ready" networkReady="NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" Dec 10 22:51:36 crc kubenswrapper[4791]: I1210 22:51:36.884702 4791 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-b6kf6" Dec 10 22:51:36 crc kubenswrapper[4791]: I1210 22:51:36.884825 4791 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 10 22:51:36 crc kubenswrapper[4791]: I1210 22:51:36.884755 4791 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 10 22:51:36 crc kubenswrapper[4791]: I1210 22:51:36.884903 4791 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 10 22:51:36 crc kubenswrapper[4791]: E1210 22:51:36.884997 4791 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-b6kf6" podUID="b4527e71-bb58-4a39-a99c-fe614aa4c5f9" Dec 10 22:51:36 crc kubenswrapper[4791]: E1210 22:51:36.885134 4791 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 10 22:51:36 crc kubenswrapper[4791]: E1210 22:51:36.885202 4791 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 10 22:51:36 crc kubenswrapper[4791]: E1210 22:51:36.885443 4791 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 10 22:51:37 crc kubenswrapper[4791]: I1210 22:51:37.885440 4791 scope.go:117] "RemoveContainer" containerID="5f299a2b60de0ef5fb17a1e463d8552d6c4d11d166ae32e490104b3d90116d25" Dec 10 22:51:38 crc kubenswrapper[4791]: I1210 22:51:38.731291 4791 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-zhq64_3cd47739-0fa9-4321-aff1-220f8721a0b3/ovnkube-controller/3.log" Dec 10 22:51:38 crc kubenswrapper[4791]: I1210 22:51:38.733971 4791 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-zhq64" event={"ID":"3cd47739-0fa9-4321-aff1-220f8721a0b3","Type":"ContainerStarted","Data":"8839756dfbc89d902c809e3b2b4c8dd5de8ef52ec1009fb2aac5a06a16de4b66"} Dec 10 22:51:38 crc kubenswrapper[4791]: I1210 22:51:38.734373 4791 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-ovn-kubernetes/ovnkube-node-zhq64" Dec 10 22:51:38 crc kubenswrapper[4791]: I1210 22:51:38.764673 4791 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-ovn-kubernetes/ovnkube-node-zhq64" podStartSLOduration=107.764656923 podStartE2EDuration="1m47.764656923s" podCreationTimestamp="2025-12-10 22:49:51 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-10 22:51:38.763576153 +0000 UTC m=+133.193193776" watchObservedRunningTime="2025-12-10 22:51:38.764656923 +0000 UTC m=+133.194274536" Dec 10 22:51:38 crc kubenswrapper[4791]: I1210 22:51:38.884156 4791 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 10 22:51:38 crc kubenswrapper[4791]: I1210 22:51:38.884144 4791 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 10 22:51:38 crc kubenswrapper[4791]: I1210 22:51:38.884188 4791 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 10 22:51:38 crc kubenswrapper[4791]: I1210 22:51:38.884259 4791 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-b6kf6" Dec 10 22:51:38 crc kubenswrapper[4791]: E1210 22:51:38.884512 4791 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 10 22:51:38 crc kubenswrapper[4791]: E1210 22:51:38.884682 4791 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-b6kf6" podUID="b4527e71-bb58-4a39-a99c-fe614aa4c5f9" Dec 10 22:51:38 crc kubenswrapper[4791]: I1210 22:51:38.884758 4791 scope.go:117] "RemoveContainer" containerID="f337f881ffd71f4a2ee19e92bb0b92c76cbb9b8352511f741d0c87e7da4cdd18" Dec 10 22:51:38 crc kubenswrapper[4791]: E1210 22:51:38.884817 4791 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 10 22:51:38 crc kubenswrapper[4791]: E1210 22:51:38.884766 4791 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 10 22:51:38 crc kubenswrapper[4791]: I1210 22:51:38.953301 4791 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-multus/network-metrics-daemon-b6kf6"] Dec 10 22:51:39 crc kubenswrapper[4791]: I1210 22:51:39.738605 4791 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-multus_multus-4nwkq_672aa28c-8169-49ed-87b8-21187d13a80c/kube-multus/1.log" Dec 10 22:51:39 crc kubenswrapper[4791]: I1210 22:51:39.738698 4791 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-4nwkq" event={"ID":"672aa28c-8169-49ed-87b8-21187d13a80c","Type":"ContainerStarted","Data":"ccdf9cf4480255bc2180847012278cee2a7f7e9392d14b7affe69a361b92cf4b"} Dec 10 22:51:39 crc kubenswrapper[4791]: I1210 22:51:39.738763 4791 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-b6kf6" Dec 10 22:51:39 crc kubenswrapper[4791]: E1210 22:51:39.738932 4791 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-b6kf6" podUID="b4527e71-bb58-4a39-a99c-fe614aa4c5f9" Dec 10 22:51:39 crc kubenswrapper[4791]: I1210 22:51:39.756212 4791 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-multus/multus-4nwkq" podStartSLOduration=108.756177241 podStartE2EDuration="1m48.756177241s" podCreationTimestamp="2025-12-10 22:49:51 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-10 22:51:39.754920126 +0000 UTC m=+134.184537729" watchObservedRunningTime="2025-12-10 22:51:39.756177241 +0000 UTC m=+134.185794894" Dec 10 22:51:40 crc kubenswrapper[4791]: I1210 22:51:40.883946 4791 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 10 22:51:40 crc kubenswrapper[4791]: I1210 22:51:40.884112 4791 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 10 22:51:40 crc kubenswrapper[4791]: E1210 22:51:40.884231 4791 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 10 22:51:40 crc kubenswrapper[4791]: I1210 22:51:40.884120 4791 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 10 22:51:40 crc kubenswrapper[4791]: E1210 22:51:40.884432 4791 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 10 22:51:40 crc kubenswrapper[4791]: E1210 22:51:40.884458 4791 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 10 22:51:40 crc kubenswrapper[4791]: E1210 22:51:40.976100 4791 kubelet.go:2916] "Container runtime network not ready" networkReady="NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" Dec 10 22:51:41 crc kubenswrapper[4791]: I1210 22:51:41.884288 4791 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-b6kf6" Dec 10 22:51:41 crc kubenswrapper[4791]: E1210 22:51:41.884506 4791 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-b6kf6" podUID="b4527e71-bb58-4a39-a99c-fe614aa4c5f9" Dec 10 22:51:42 crc kubenswrapper[4791]: I1210 22:51:42.884092 4791 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 10 22:51:42 crc kubenswrapper[4791]: I1210 22:51:42.884137 4791 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 10 22:51:42 crc kubenswrapper[4791]: I1210 22:51:42.884095 4791 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 10 22:51:42 crc kubenswrapper[4791]: E1210 22:51:42.884270 4791 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 10 22:51:42 crc kubenswrapper[4791]: E1210 22:51:42.884482 4791 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 10 22:51:42 crc kubenswrapper[4791]: E1210 22:51:42.884597 4791 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 10 22:51:43 crc kubenswrapper[4791]: I1210 22:51:43.884065 4791 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-b6kf6" Dec 10 22:51:43 crc kubenswrapper[4791]: E1210 22:51:43.884601 4791 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-b6kf6" podUID="b4527e71-bb58-4a39-a99c-fe614aa4c5f9" Dec 10 22:51:44 crc kubenswrapper[4791]: I1210 22:51:44.884199 4791 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 10 22:51:44 crc kubenswrapper[4791]: I1210 22:51:44.884275 4791 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 10 22:51:44 crc kubenswrapper[4791]: I1210 22:51:44.884284 4791 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 10 22:51:44 crc kubenswrapper[4791]: E1210 22:51:44.884437 4791 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 10 22:51:44 crc kubenswrapper[4791]: E1210 22:51:44.884521 4791 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 10 22:51:44 crc kubenswrapper[4791]: E1210 22:51:44.884606 4791 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 10 22:51:45 crc kubenswrapper[4791]: I1210 22:51:45.884064 4791 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-b6kf6" Dec 10 22:51:45 crc kubenswrapper[4791]: E1210 22:51:45.885128 4791 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-b6kf6" podUID="b4527e71-bb58-4a39-a99c-fe614aa4c5f9" Dec 10 22:51:46 crc kubenswrapper[4791]: I1210 22:51:46.884479 4791 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 10 22:51:46 crc kubenswrapper[4791]: I1210 22:51:46.884553 4791 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 10 22:51:46 crc kubenswrapper[4791]: I1210 22:51:46.884479 4791 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 10 22:51:46 crc kubenswrapper[4791]: I1210 22:51:46.887642 4791 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-diagnostics"/"openshift-service-ca.crt" Dec 10 22:51:46 crc kubenswrapper[4791]: I1210 22:51:46.888189 4791 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-network-console"/"networking-console-plugin-cert" Dec 10 22:51:46 crc kubenswrapper[4791]: I1210 22:51:46.888414 4791 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-diagnostics"/"kube-root-ca.crt" Dec 10 22:51:46 crc kubenswrapper[4791]: I1210 22:51:46.888429 4791 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-console"/"networking-console-plugin" Dec 10 22:51:47 crc kubenswrapper[4791]: I1210 22:51:47.884254 4791 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-b6kf6" Dec 10 22:51:47 crc kubenswrapper[4791]: I1210 22:51:47.887271 4791 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-multus"/"metrics-daemon-sa-dockercfg-d427c" Dec 10 22:51:47 crc kubenswrapper[4791]: I1210 22:51:47.888297 4791 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-multus"/"metrics-daemon-secret" Dec 10 22:51:50 crc kubenswrapper[4791]: I1210 22:51:50.582302 4791 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeReady" Dec 10 22:51:50 crc kubenswrapper[4791]: I1210 22:51:50.633678 4791 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-apiserver/apiserver-76f77b778f-kqtbx"] Dec 10 22:51:50 crc kubenswrapper[4791]: I1210 22:51:50.634694 4791 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-machine-api/machine-api-operator-5694c8668f-cpbhz"] Dec 10 22:51:50 crc kubenswrapper[4791]: I1210 22:51:50.634890 4791 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-apiserver/apiserver-76f77b778f-kqtbx" Dec 10 22:51:50 crc kubenswrapper[4791]: I1210 22:51:50.653694 4791 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-api/machine-api-operator-5694c8668f-cpbhz" Dec 10 22:51:50 crc kubenswrapper[4791]: I1210 22:51:50.654830 4791 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver"/"audit-1" Dec 10 22:51:50 crc kubenswrapper[4791]: I1210 22:51:50.655314 4791 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-apiserver"/"encryption-config-1" Dec 10 22:51:50 crc kubenswrapper[4791]: I1210 22:51:50.655506 4791 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-controller-manager/controller-manager-879f6c89f-9tg4p"] Dec 10 22:51:50 crc kubenswrapper[4791]: I1210 22:51:50.655949 4791 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager/controller-manager-879f6c89f-9tg4p" Dec 10 22:51:50 crc kubenswrapper[4791]: I1210 22:51:50.656318 4791 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-oauth-apiserver/apiserver-7bbb656c7d-nwwpb"] Dec 10 22:51:50 crc kubenswrapper[4791]: I1210 22:51:50.657173 4791 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-route-controller-manager/route-controller-manager-6576b87f9c-vsnsh"] Dec 10 22:51:50 crc kubenswrapper[4791]: I1210 22:51:50.657803 4791 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-vsnsh" Dec 10 22:51:50 crc kubenswrapper[4791]: I1210 22:51:50.657932 4791 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-nwwpb" Dec 10 22:51:50 crc kubenswrapper[4791]: I1210 22:51:50.658603 4791 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-api"/"openshift-service-ca.crt" Dec 10 22:51:50 crc kubenswrapper[4791]: I1210 22:51:50.658734 4791 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver"/"openshift-service-ca.crt" Dec 10 22:51:50 crc kubenswrapper[4791]: I1210 22:51:50.659011 4791 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver"/"image-import-ca" Dec 10 22:51:50 crc kubenswrapper[4791]: I1210 22:51:50.659357 4791 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-api"/"kube-rbac-proxy" Dec 10 22:51:50 crc kubenswrapper[4791]: I1210 22:51:50.659392 4791 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver"/"config" Dec 10 22:51:50 crc kubenswrapper[4791]: I1210 22:51:50.659450 4791 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-api"/"machine-api-operator-images" Dec 10 22:51:50 crc kubenswrapper[4791]: I1210 22:51:50.659585 4791 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver"/"kube-root-ca.crt" Dec 10 22:51:50 crc kubenswrapper[4791]: I1210 22:51:50.659646 4791 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-apiserver"/"etcd-client" Dec 10 22:51:50 crc kubenswrapper[4791]: I1210 22:51:50.659709 4791 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-apiserver"/"openshift-apiserver-sa-dockercfg-djjff" Dec 10 22:51:50 crc kubenswrapper[4791]: I1210 22:51:50.659802 4791 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-apiserver"/"serving-cert" Dec 10 22:51:50 crc kubenswrapper[4791]: I1210 22:51:50.660193 4791 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-api"/"machine-api-operator-dockercfg-mfbb7" Dec 10 22:51:50 crc kubenswrapper[4791]: I1210 22:51:50.660392 4791 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-api"/"machine-api-operator-tls" Dec 10 22:51:50 crc kubenswrapper[4791]: I1210 22:51:50.660536 4791 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver"/"etcd-serving-ca" Dec 10 22:51:50 crc kubenswrapper[4791]: I1210 22:51:50.661088 4791 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-api"/"kube-root-ca.crt" Dec 10 22:51:50 crc kubenswrapper[4791]: I1210 22:51:50.661807 4791 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-t7742"] Dec 10 22:51:50 crc kubenswrapper[4791]: I1210 22:51:50.662661 4791 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-t7742" Dec 10 22:51:50 crc kubenswrapper[4791]: I1210 22:51:50.662801 4791 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-console/downloads-7954f5f757-mlb9p"] Dec 10 22:51:50 crc kubenswrapper[4791]: I1210 22:51:50.663239 4791 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-console/downloads-7954f5f757-mlb9p" Dec 10 22:51:50 crc kubenswrapper[4791]: I1210 22:51:50.664751 4791 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-cluster-machine-approver/machine-approver-56656f9798-m6dwv"] Dec 10 22:51:50 crc kubenswrapper[4791]: I1210 22:51:50.665100 4791 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-cluster-machine-approver/machine-approver-56656f9798-m6dwv" Dec 10 22:51:50 crc kubenswrapper[4791]: I1210 22:51:50.666902 4791 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-authentication/oauth-openshift-558db77b4-ht5t7"] Dec 10 22:51:50 crc kubenswrapper[4791]: I1210 22:51:50.667507 4791 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-console-operator/console-operator-58897d9998-hd6zl"] Dec 10 22:51:50 crc kubenswrapper[4791]: I1210 22:51:50.667898 4791 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-console-operator/console-operator-58897d9998-hd6zl" Dec 10 22:51:50 crc kubenswrapper[4791]: I1210 22:51:50.668134 4791 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-config-operator/openshift-config-operator-7777fb866f-wwpsp"] Dec 10 22:51:50 crc kubenswrapper[4791]: I1210 22:51:50.668325 4791 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-authentication/oauth-openshift-558db77b4-ht5t7" Dec 10 22:51:50 crc kubenswrapper[4791]: I1210 22:51:50.668579 4791 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-config-operator/openshift-config-operator-7777fb866f-wwpsp" Dec 10 22:51:50 crc kubenswrapper[4791]: I1210 22:51:50.672685 4791 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver"/"trusted-ca-bundle" Dec 10 22:51:50 crc kubenswrapper[4791]: I1210 22:51:50.672785 4791 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"config" Dec 10 22:51:50 crc kubenswrapper[4791]: I1210 22:51:50.676519 4791 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-image-registry/image-registry-697d97f7c8-srd52"] Dec 10 22:51:50 crc kubenswrapper[4791]: I1210 22:51:50.677385 4791 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-console/console-f9d7485db-zkfkv"] Dec 10 22:51:50 crc kubenswrapper[4791]: I1210 22:51:50.677949 4791 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-console/console-f9d7485db-zkfkv" Dec 10 22:51:50 crc kubenswrapper[4791]: I1210 22:51:50.678608 4791 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-image-registry/image-registry-697d97f7c8-srd52" Dec 10 22:51:50 crc kubenswrapper[4791]: I1210 22:51:50.685690 4791 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-authentication-operator/authentication-operator-69f744f599-7rhk9"] Dec 10 22:51:50 crc kubenswrapper[4791]: I1210 22:51:50.686677 4791 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-q4cpx"] Dec 10 22:51:50 crc kubenswrapper[4791]: I1210 22:51:50.687043 4791 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-q4cpx" Dec 10 22:51:50 crc kubenswrapper[4791]: I1210 22:51:50.687444 4791 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-authentication-operator/authentication-operator-69f744f599-7rhk9" Dec 10 22:51:50 crc kubenswrapper[4791]: I1210 22:51:50.688617 4791 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-controller-manager"/"openshift-controller-manager-sa-dockercfg-msq4c" Dec 10 22:51:50 crc kubenswrapper[4791]: I1210 22:51:50.688848 4791 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-oauth-apiserver"/"etcd-client" Dec 10 22:51:50 crc kubenswrapper[4791]: I1210 22:51:50.689015 4791 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-oauth-apiserver"/"kube-root-ca.crt" Dec 10 22:51:50 crc kubenswrapper[4791]: I1210 22:51:50.689270 4791 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"kube-root-ca.crt" Dec 10 22:51:50 crc kubenswrapper[4791]: I1210 22:51:50.689392 4791 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-controller-manager"/"serving-cert" Dec 10 22:51:50 crc kubenswrapper[4791]: I1210 22:51:50.689489 4791 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"openshift-service-ca.crt" Dec 10 22:51:50 crc kubenswrapper[4791]: I1210 22:51:50.689608 4791 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"config" Dec 10 22:51:50 crc kubenswrapper[4791]: I1210 22:51:50.689686 4791 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-oauth-apiserver"/"oauth-apiserver-sa-dockercfg-6r2bq" Dec 10 22:51:50 crc kubenswrapper[4791]: I1210 22:51:50.689685 4791 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-oauth-apiserver"/"audit-1" Dec 10 22:51:50 crc kubenswrapper[4791]: I1210 22:51:50.689698 4791 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"openshift-service-ca.crt" Dec 10 22:51:50 crc kubenswrapper[4791]: I1210 22:51:50.689741 4791 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-route-controller-manager"/"serving-cert" Dec 10 22:51:50 crc kubenswrapper[4791]: I1210 22:51:50.689813 4791 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"client-ca" Dec 10 22:51:50 crc kubenswrapper[4791]: I1210 22:51:50.689820 4791 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-oauth-apiserver"/"etcd-serving-ca" Dec 10 22:51:50 crc kubenswrapper[4791]: I1210 22:51:50.689938 4791 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"kube-root-ca.crt" Dec 10 22:51:50 crc kubenswrapper[4791]: I1210 22:51:50.689975 4791 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-oauth-apiserver"/"trusted-ca-bundle" Dec 10 22:51:50 crc kubenswrapper[4791]: I1210 22:51:50.689622 4791 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-oauth-apiserver"/"encryption-config-1" Dec 10 22:51:50 crc kubenswrapper[4791]: I1210 22:51:50.691962 4791 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-oauth-apiserver"/"openshift-service-ca.crt" Dec 10 22:51:50 crc kubenswrapper[4791]: I1210 22:51:50.693660 4791 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-route-controller-manager"/"route-controller-manager-sa-dockercfg-h2zr2" Dec 10 22:51:50 crc kubenswrapper[4791]: I1210 22:51:50.694045 4791 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-oauth-apiserver"/"serving-cert" Dec 10 22:51:50 crc kubenswrapper[4791]: I1210 22:51:50.694136 4791 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-mldvd"] Dec 10 22:51:50 crc kubenswrapper[4791]: I1210 22:51:50.694211 4791 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"client-ca" Dec 10 22:51:50 crc kubenswrapper[4791]: I1210 22:51:50.694871 4791 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-mldvd" Dec 10 22:51:50 crc kubenswrapper[4791]: I1210 22:51:50.694912 4791 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-ingress-operator/ingress-operator-5b745b69d9-8npst"] Dec 10 22:51:50 crc kubenswrapper[4791]: I1210 22:51:50.711899 4791 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-5t9vv"] Dec 10 22:51:50 crc kubenswrapper[4791]: I1210 22:51:50.714178 4791 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-samples-operator"/"openshift-service-ca.crt" Dec 10 22:51:50 crc kubenswrapper[4791]: I1210 22:51:50.715106 4791 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-user-idp-0-file-data" Dec 10 22:51:50 crc kubenswrapper[4791]: I1210 22:51:50.715285 4791 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-machine-approver"/"kube-rbac-proxy" Dec 10 22:51:50 crc kubenswrapper[4791]: I1210 22:51:50.715357 4791 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console-operator"/"console-operator-config" Dec 10 22:51:50 crc kubenswrapper[4791]: I1210 22:51:50.715529 4791 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-machine-approver"/"machine-approver-config" Dec 10 22:51:50 crc kubenswrapper[4791]: I1210 22:51:50.715817 4791 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication"/"openshift-service-ca.crt" Dec 10 22:51:50 crc kubenswrapper[4791]: I1210 22:51:50.716222 4791 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-cluster-machine-approver"/"machine-approver-sa-dockercfg-nl2j4" Dec 10 22:51:50 crc kubenswrapper[4791]: I1210 22:51:50.716382 4791 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-user-template-provider-selection" Dec 10 22:51:50 crc kubenswrapper[4791]: I1210 22:51:50.716606 4791 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console-operator"/"openshift-service-ca.crt" Dec 10 22:51:50 crc kubenswrapper[4791]: I1210 22:51:50.716912 4791 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-console"/"default-dockercfg-chnjx" Dec 10 22:51:50 crc kubenswrapper[4791]: I1210 22:51:50.717441 4791 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication"/"v4-0-config-system-cliconfig" Dec 10 22:51:50 crc kubenswrapper[4791]: I1210 22:51:50.717600 4791 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-system-session" Dec 10 22:51:50 crc kubenswrapper[4791]: I1210 22:51:50.717714 4791 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication"/"audit" Dec 10 22:51:50 crc kubenswrapper[4791]: I1210 22:51:50.717816 4791 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-cluster-machine-approver"/"machine-approver-tls" Dec 10 22:51:50 crc kubenswrapper[4791]: I1210 22:51:50.718076 4791 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-machine-approver"/"openshift-service-ca.crt" Dec 10 22:51:50 crc kubenswrapper[4791]: I1210 22:51:50.718349 4791 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-user-template-error" Dec 10 22:51:50 crc kubenswrapper[4791]: I1210 22:51:50.718471 4791 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-console-operator"/"console-operator-dockercfg-4xjcr" Dec 10 22:51:50 crc kubenswrapper[4791]: I1210 22:51:50.718616 4791 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"oauth-openshift-dockercfg-znhcc" Dec 10 22:51:50 crc kubenswrapper[4791]: I1210 22:51:50.718812 4791 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-ingress/router-default-5444994796-p97fc"] Dec 10 22:51:50 crc kubenswrapper[4791]: I1210 22:51:50.718876 4791 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication"/"v4-0-config-system-service-ca" Dec 10 22:51:50 crc kubenswrapper[4791]: I1210 22:51:50.719215 4791 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-5t9vv" Dec 10 22:51:50 crc kubenswrapper[4791]: I1210 22:51:50.720976 4791 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/3ba17b70-17df-4f8c-87ff-c15e03ce06bc-etcd-client\") pod \"apiserver-7bbb656c7d-nwwpb\" (UID: \"3ba17b70-17df-4f8c-87ff-c15e03ce06bc\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-nwwpb" Dec 10 22:51:50 crc kubenswrapper[4791]: I1210 22:51:50.721054 4791 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-rkmd6\" (UniqueName: \"kubernetes.io/projected/4a1cf441-8a08-4677-a480-947c8cddd6a6-kube-api-access-rkmd6\") pod \"apiserver-76f77b778f-kqtbx\" (UID: \"4a1cf441-8a08-4677-a480-947c8cddd6a6\") " pod="openshift-apiserver/apiserver-76f77b778f-kqtbx" Dec 10 22:51:50 crc kubenswrapper[4791]: I1210 22:51:50.721095 4791 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-xhsjq\" (UniqueName: \"kubernetes.io/projected/a5827e8a-4cb3-494d-87a1-556d920c2a4a-kube-api-access-xhsjq\") pod \"controller-manager-879f6c89f-9tg4p\" (UID: \"a5827e8a-4cb3-494d-87a1-556d920c2a4a\") " pod="openshift-controller-manager/controller-manager-879f6c89f-9tg4p" Dec 10 22:51:50 crc kubenswrapper[4791]: I1210 22:51:50.721115 4791 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-image-registry"/"installation-pull-secrets" Dec 10 22:51:50 crc kubenswrapper[4791]: I1210 22:51:50.721132 4791 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etcd-serving-ca\" (UniqueName: \"kubernetes.io/configmap/4a1cf441-8a08-4677-a480-947c8cddd6a6-etcd-serving-ca\") pod \"apiserver-76f77b778f-kqtbx\" (UID: \"4a1cf441-8a08-4677-a480-947c8cddd6a6\") " pod="openshift-apiserver/apiserver-76f77b778f-kqtbx" Dec 10 22:51:50 crc kubenswrapper[4791]: I1210 22:51:50.721163 4791 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/a5827e8a-4cb3-494d-87a1-556d920c2a4a-proxy-ca-bundles\") pod \"controller-manager-879f6c89f-9tg4p\" (UID: \"a5827e8a-4cb3-494d-87a1-556d920c2a4a\") " pod="openshift-controller-manager/controller-manager-879f6c89f-9tg4p" Dec 10 22:51:50 crc kubenswrapper[4791]: I1210 22:51:50.721198 4791 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"node-pullsecrets\" (UniqueName: \"kubernetes.io/host-path/4a1cf441-8a08-4677-a480-947c8cddd6a6-node-pullsecrets\") pod \"apiserver-76f77b778f-kqtbx\" (UID: \"4a1cf441-8a08-4677-a480-947c8cddd6a6\") " pod="openshift-apiserver/apiserver-76f77b778f-kqtbx" Dec 10 22:51:50 crc kubenswrapper[4791]: I1210 22:51:50.721232 4791 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/cd65e91e-5b02-4f78-829e-93b4596ff8f6-config\") pod \"machine-api-operator-5694c8668f-cpbhz\" (UID: \"cd65e91e-5b02-4f78-829e-93b4596ff8f6\") " pod="openshift-machine-api/machine-api-operator-5694c8668f-cpbhz" Dec 10 22:51:50 crc kubenswrapper[4791]: I1210 22:51:50.721264 4791 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/a5827e8a-4cb3-494d-87a1-556d920c2a4a-config\") pod \"controller-manager-879f6c89f-9tg4p\" (UID: \"a5827e8a-4cb3-494d-87a1-556d920c2a4a\") " pod="openshift-controller-manager/controller-manager-879f6c89f-9tg4p" Dec 10 22:51:50 crc kubenswrapper[4791]: I1210 22:51:50.721531 4791 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console"/"oauth-serving-cert" Dec 10 22:51:50 crc kubenswrapper[4791]: I1210 22:51:50.722944 4791 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console"/"console-config" Dec 10 22:51:50 crc kubenswrapper[4791]: I1210 22:51:50.724458 4791 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/4a1cf441-8a08-4677-a480-947c8cddd6a6-audit-dir\") pod \"apiserver-76f77b778f-kqtbx\" (UID: \"4a1cf441-8a08-4677-a480-947c8cddd6a6\") " pod="openshift-apiserver/apiserver-76f77b778f-kqtbx" Dec 10 22:51:50 crc kubenswrapper[4791]: I1210 22:51:50.724537 4791 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"encryption-config\" (UniqueName: \"kubernetes.io/secret/4a1cf441-8a08-4677-a480-947c8cddd6a6-encryption-config\") pod \"apiserver-76f77b778f-kqtbx\" (UID: \"4a1cf441-8a08-4677-a480-947c8cddd6a6\") " pod="openshift-apiserver/apiserver-76f77b778f-kqtbx" Dec 10 22:51:50 crc kubenswrapper[4791]: I1210 22:51:50.724602 4791 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"encryption-config\" (UniqueName: \"kubernetes.io/secret/3ba17b70-17df-4f8c-87ff-c15e03ce06bc-encryption-config\") pod \"apiserver-7bbb656c7d-nwwpb\" (UID: \"3ba17b70-17df-4f8c-87ff-c15e03ce06bc\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-nwwpb" Dec 10 22:51:50 crc kubenswrapper[4791]: I1210 22:51:50.724611 4791 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console"/"openshift-service-ca.crt" Dec 10 22:51:50 crc kubenswrapper[4791]: I1210 22:51:50.724632 4791 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/4a1cf441-8a08-4677-a480-947c8cddd6a6-config\") pod \"apiserver-76f77b778f-kqtbx\" (UID: \"4a1cf441-8a08-4677-a480-947c8cddd6a6\") " pod="openshift-apiserver/apiserver-76f77b778f-kqtbx" Dec 10 22:51:50 crc kubenswrapper[4791]: I1210 22:51:50.724699 4791 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/a5827e8a-4cb3-494d-87a1-556d920c2a4a-serving-cert\") pod \"controller-manager-879f6c89f-9tg4p\" (UID: \"a5827e8a-4cb3-494d-87a1-556d920c2a4a\") " pod="openshift-controller-manager/controller-manager-879f6c89f-9tg4p" Dec 10 22:51:50 crc kubenswrapper[4791]: I1210 22:51:50.724848 4791 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/4a1cf441-8a08-4677-a480-947c8cddd6a6-serving-cert\") pod \"apiserver-76f77b778f-kqtbx\" (UID: \"4a1cf441-8a08-4677-a480-947c8cddd6a6\") " pod="openshift-apiserver/apiserver-76f77b778f-kqtbx" Dec 10 22:51:50 crc kubenswrapper[4791]: I1210 22:51:50.724872 4791 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-system-serving-cert" Dec 10 22:51:50 crc kubenswrapper[4791]: I1210 22:51:50.722252 4791 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-config-operator"/"openshift-service-ca.crt" Dec 10 22:51:50 crc kubenswrapper[4791]: I1210 22:51:50.724934 4791 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/3ba17b70-17df-4f8c-87ff-c15e03ce06bc-audit-dir\") pod \"apiserver-7bbb656c7d-nwwpb\" (UID: \"3ba17b70-17df-4f8c-87ff-c15e03ce06bc\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-nwwpb" Dec 10 22:51:50 crc kubenswrapper[4791]: I1210 22:51:50.725312 4791 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"images\" (UniqueName: \"kubernetes.io/configmap/cd65e91e-5b02-4f78-829e-93b4596ff8f6-images\") pod \"machine-api-operator-5694c8668f-cpbhz\" (UID: \"cd65e91e-5b02-4f78-829e-93b4596ff8f6\") " pod="openshift-machine-api/machine-api-operator-5694c8668f-cpbhz" Dec 10 22:51:50 crc kubenswrapper[4791]: I1210 22:51:50.725383 4791 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/a5827e8a-4cb3-494d-87a1-556d920c2a4a-client-ca\") pod \"controller-manager-879f6c89f-9tg4p\" (UID: \"a5827e8a-4cb3-494d-87a1-556d920c2a4a\") " pod="openshift-controller-manager/controller-manager-879f6c89f-9tg4p" Dec 10 22:51:50 crc kubenswrapper[4791]: I1210 22:51:50.725437 4791 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"image-import-ca\" (UniqueName: \"kubernetes.io/configmap/4a1cf441-8a08-4677-a480-947c8cddd6a6-image-import-ca\") pod \"apiserver-76f77b778f-kqtbx\" (UID: \"4a1cf441-8a08-4677-a480-947c8cddd6a6\") " pod="openshift-apiserver/apiserver-76f77b778f-kqtbx" Dec 10 22:51:50 crc kubenswrapper[4791]: I1210 22:51:50.725467 4791 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etcd-serving-ca\" (UniqueName: \"kubernetes.io/configmap/3ba17b70-17df-4f8c-87ff-c15e03ce06bc-etcd-serving-ca\") pod \"apiserver-7bbb656c7d-nwwpb\" (UID: \"3ba17b70-17df-4f8c-87ff-c15e03ce06bc\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-nwwpb" Dec 10 22:51:50 crc kubenswrapper[4791]: I1210 22:51:50.725491 4791 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/3ba17b70-17df-4f8c-87ff-c15e03ce06bc-audit-policies\") pod \"apiserver-7bbb656c7d-nwwpb\" (UID: \"3ba17b70-17df-4f8c-87ff-c15e03ce06bc\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-nwwpb" Dec 10 22:51:50 crc kubenswrapper[4791]: I1210 22:51:50.725514 4791 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/3ba17b70-17df-4f8c-87ff-c15e03ce06bc-trusted-ca-bundle\") pod \"apiserver-7bbb656c7d-nwwpb\" (UID: \"3ba17b70-17df-4f8c-87ff-c15e03ce06bc\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-nwwpb" Dec 10 22:51:50 crc kubenswrapper[4791]: I1210 22:51:50.725542 4791 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-k4rbk\" (UniqueName: \"kubernetes.io/projected/3ba17b70-17df-4f8c-87ff-c15e03ce06bc-kube-api-access-k4rbk\") pod \"apiserver-7bbb656c7d-nwwpb\" (UID: \"3ba17b70-17df-4f8c-87ff-c15e03ce06bc\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-nwwpb" Dec 10 22:51:50 crc kubenswrapper[4791]: I1210 22:51:50.725574 4791 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"machine-api-operator-tls\" (UniqueName: \"kubernetes.io/secret/cd65e91e-5b02-4f78-829e-93b4596ff8f6-machine-api-operator-tls\") pod \"machine-api-operator-5694c8668f-cpbhz\" (UID: \"cd65e91e-5b02-4f78-829e-93b4596ff8f6\") " pod="openshift-machine-api/machine-api-operator-5694c8668f-cpbhz" Dec 10 22:51:50 crc kubenswrapper[4791]: I1210 22:51:50.725650 4791 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/3ba17b70-17df-4f8c-87ff-c15e03ce06bc-serving-cert\") pod \"apiserver-7bbb656c7d-nwwpb\" (UID: \"3ba17b70-17df-4f8c-87ff-c15e03ce06bc\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-nwwpb" Dec 10 22:51:50 crc kubenswrapper[4791]: I1210 22:51:50.725720 4791 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/4a1cf441-8a08-4677-a480-947c8cddd6a6-trusted-ca-bundle\") pod \"apiserver-76f77b778f-kqtbx\" (UID: \"4a1cf441-8a08-4677-a480-947c8cddd6a6\") " pod="openshift-apiserver/apiserver-76f77b778f-kqtbx" Dec 10 22:51:50 crc kubenswrapper[4791]: I1210 22:51:50.725783 4791 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"audit\" (UniqueName: \"kubernetes.io/configmap/4a1cf441-8a08-4677-a480-947c8cddd6a6-audit\") pod \"apiserver-76f77b778f-kqtbx\" (UID: \"4a1cf441-8a08-4677-a480-947c8cddd6a6\") " pod="openshift-apiserver/apiserver-76f77b778f-kqtbx" Dec 10 22:51:50 crc kubenswrapper[4791]: I1210 22:51:50.725877 4791 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/4a1cf441-8a08-4677-a480-947c8cddd6a6-etcd-client\") pod \"apiserver-76f77b778f-kqtbx\" (UID: \"4a1cf441-8a08-4677-a480-947c8cddd6a6\") " pod="openshift-apiserver/apiserver-76f77b778f-kqtbx" Dec 10 22:51:50 crc kubenswrapper[4791]: I1210 22:51:50.725982 4791 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-vgqrv\" (UniqueName: \"kubernetes.io/projected/cd65e91e-5b02-4f78-829e-93b4596ff8f6-kube-api-access-vgqrv\") pod \"machine-api-operator-5694c8668f-cpbhz\" (UID: \"cd65e91e-5b02-4f78-829e-93b4596ff8f6\") " pod="openshift-machine-api/machine-api-operator-5694c8668f-cpbhz" Dec 10 22:51:50 crc kubenswrapper[4791]: I1210 22:51:50.727487 4791 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-ingress-operator/ingress-operator-5b745b69d9-8npst" Dec 10 22:51:50 crc kubenswrapper[4791]: I1210 22:51:50.729960 4791 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-console"/"console-oauth-config" Dec 10 22:51:50 crc kubenswrapper[4791]: I1210 22:51:50.732731 4791 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-cluster-samples-operator"/"cluster-samples-operator-dockercfg-xpp9w" Dec 10 22:51:50 crc kubenswrapper[4791]: I1210 22:51:50.745857 4791 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-config-operator"/"config-operator-serving-cert" Dec 10 22:51:50 crc kubenswrapper[4791]: I1210 22:51:50.746209 4791 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-operator-lifecycle-manager/catalog-operator-68c6474976-497m8"] Dec 10 22:51:50 crc kubenswrapper[4791]: I1210 22:51:50.746674 4791 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console"/"service-ca" Dec 10 22:51:50 crc kubenswrapper[4791]: I1210 22:51:50.746796 4791 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-machine-config-operator/machine-config-controller-84d6567774-sc74x"] Dec 10 22:51:50 crc kubenswrapper[4791]: I1210 22:51:50.746986 4791 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication-operator"/"authentication-operator-config" Dec 10 22:51:50 crc kubenswrapper[4791]: I1210 22:51:50.747023 4791 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-ingress/router-default-5444994796-p97fc" Dec 10 22:51:50 crc kubenswrapper[4791]: I1210 22:51:50.747171 4791 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-497m8" Dec 10 22:51:50 crc kubenswrapper[4791]: I1210 22:51:50.747834 4791 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-samples-operator"/"kube-root-ca.crt" Dec 10 22:51:50 crc kubenswrapper[4791]: I1210 22:51:50.747973 4791 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication"/"kube-root-ca.crt" Dec 10 22:51:50 crc kubenswrapper[4791]: I1210 22:51:50.748031 4791 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-bws6n"] Dec 10 22:51:50 crc kubenswrapper[4791]: I1210 22:51:50.748072 4791 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console-operator"/"kube-root-ca.crt" Dec 10 22:51:50 crc kubenswrapper[4791]: I1210 22:51:50.748512 4791 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-bws6n" Dec 10 22:51:50 crc kubenswrapper[4791]: I1210 22:51:50.748593 4791 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-image-registry"/"image-registry-tls" Dec 10 22:51:50 crc kubenswrapper[4791]: I1210 22:51:50.756571 4791 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-config-operator/machine-config-controller-84d6567774-sc74x" Dec 10 22:51:50 crc kubenswrapper[4791]: I1210 22:51:50.758993 4791 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-image-registry"/"image-registry-operator-tls" Dec 10 22:51:50 crc kubenswrapper[4791]: I1210 22:51:50.759223 4791 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-console"/"console-serving-cert" Dec 10 22:51:50 crc kubenswrapper[4791]: I1210 22:51:50.759459 4791 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console"/"kube-root-ca.crt" Dec 10 22:51:50 crc kubenswrapper[4791]: I1210 22:51:50.759513 4791 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-image-registry"/"registry-dockercfg-kzzsd" Dec 10 22:51:50 crc kubenswrapper[4791]: I1210 22:51:50.759557 4791 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-cluster-samples-operator"/"samples-operator-tls" Dec 10 22:51:50 crc kubenswrapper[4791]: I1210 22:51:50.759654 4791 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-machine-approver"/"kube-root-ca.crt" Dec 10 22:51:50 crc kubenswrapper[4791]: I1210 22:51:50.759742 4791 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-image-registry"/"cluster-image-registry-operator-dockercfg-m4qtx" Dec 10 22:51:50 crc kubenswrapper[4791]: I1210 22:51:50.759863 4791 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication-operator"/"authentication-operator-dockercfg-mz9bj" Dec 10 22:51:50 crc kubenswrapper[4791]: I1210 22:51:50.759887 4791 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication-operator"/"service-ca-bundle" Dec 10 22:51:50 crc kubenswrapper[4791]: I1210 22:51:50.759970 4791 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication-operator"/"serving-cert" Dec 10 22:51:50 crc kubenswrapper[4791]: I1210 22:51:50.760321 4791 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-config-operator"/"openshift-config-operator-dockercfg-7pc5z" Dec 10 22:51:50 crc kubenswrapper[4791]: I1210 22:51:50.760714 4791 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-config-operator"/"kube-root-ca.crt" Dec 10 22:51:50 crc kubenswrapper[4791]: I1210 22:51:50.760773 4791 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-console"/"console-dockercfg-f62pw" Dec 10 22:51:50 crc kubenswrapper[4791]: I1210 22:51:50.760892 4791 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-console-operator"/"serving-cert" Dec 10 22:51:50 crc kubenswrapper[4791]: I1210 22:51:50.760905 4791 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication-operator"/"openshift-service-ca.crt" Dec 10 22:51:50 crc kubenswrapper[4791]: I1210 22:51:50.760907 4791 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-system-router-certs" Dec 10 22:51:50 crc kubenswrapper[4791]: I1210 22:51:50.761048 4791 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-controller-manager-operator"/"kube-root-ca.crt" Dec 10 22:51:50 crc kubenswrapper[4791]: I1210 22:51:50.761157 4791 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-controller-manager-operator"/"kube-controller-manager-operator-dockercfg-gkqpw" Dec 10 22:51:50 crc kubenswrapper[4791]: I1210 22:51:50.761207 4791 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-controller-manager-operator"/"kube-controller-manager-operator-serving-cert" Dec 10 22:51:50 crc kubenswrapper[4791]: I1210 22:51:50.761279 4791 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-controller-manager-operator"/"kube-controller-manager-operator-config" Dec 10 22:51:50 crc kubenswrapper[4791]: I1210 22:51:50.765614 4791 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication-operator"/"kube-root-ca.crt" Dec 10 22:51:50 crc kubenswrapper[4791]: I1210 22:51:50.772129 4791 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager-operator"/"openshift-controller-manager-operator-config" Dec 10 22:51:50 crc kubenswrapper[4791]: I1210 22:51:50.772523 4791 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-image-registry"/"trusted-ca" Dec 10 22:51:50 crc kubenswrapper[4791]: I1210 22:51:50.772848 4791 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console-operator"/"trusted-ca" Dec 10 22:51:50 crc kubenswrapper[4791]: I1210 22:51:50.774453 4791 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-dns-operator/dns-operator-744455d44c-cggs2"] Dec 10 22:51:50 crc kubenswrapper[4791]: I1210 22:51:50.775459 4791 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"openshift-global-ca" Dec 10 22:51:50 crc kubenswrapper[4791]: I1210 22:51:50.775511 4791 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-dns-operator/dns-operator-744455d44c-cggs2" Dec 10 22:51:50 crc kubenswrapper[4791]: I1210 22:51:50.776991 4791 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-z6gpf"] Dec 10 22:51:50 crc kubenswrapper[4791]: I1210 22:51:50.777138 4791 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress-operator"/"ingress-operator-dockercfg-7lnqk" Dec 10 22:51:50 crc kubenswrapper[4791]: I1210 22:51:50.777624 4791 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-z6gpf" Dec 10 22:51:50 crc kubenswrapper[4791]: I1210 22:51:50.782223 4791 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-user-template-login" Dec 10 22:51:50 crc kubenswrapper[4791]: I1210 22:51:50.782611 4791 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-rmjd2"] Dec 10 22:51:50 crc kubenswrapper[4791]: I1210 22:51:50.783594 4791 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-rmjd2" Dec 10 22:51:50 crc kubenswrapper[4791]: I1210 22:51:50.786618 4791 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication"/"v4-0-config-system-trusted-ca-bundle" Dec 10 22:51:50 crc kubenswrapper[4791]: I1210 22:51:50.787212 4791 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress-operator"/"kube-root-ca.crt" Dec 10 22:51:50 crc kubenswrapper[4791]: I1210 22:51:50.792434 4791 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console"/"trusted-ca-bundle" Dec 10 22:51:50 crc kubenswrapper[4791]: I1210 22:51:50.799105 4791 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-w5dpr"] Dec 10 22:51:50 crc kubenswrapper[4791]: I1210 22:51:50.801883 4791 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress-operator"/"openshift-service-ca.crt" Dec 10 22:51:50 crc kubenswrapper[4791]: I1210 22:51:50.802892 4791 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-etcd-operator/etcd-operator-b45778765-nkt8g"] Dec 10 22:51:50 crc kubenswrapper[4791]: I1210 22:51:50.804591 4791 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-w5dpr" Dec 10 22:51:50 crc kubenswrapper[4791]: I1210 22:51:50.805564 4791 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-vm842"] Dec 10 22:51:50 crc kubenswrapper[4791]: I1210 22:51:50.805790 4791 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-etcd-operator/etcd-operator-b45778765-nkt8g" Dec 10 22:51:50 crc kubenswrapper[4791]: I1210 22:51:50.806146 4791 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-machine-config-operator/machine-config-operator-74547568cd-rcnrn"] Dec 10 22:51:50 crc kubenswrapper[4791]: I1210 22:51:50.806700 4791 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-vm842" Dec 10 22:51:50 crc kubenswrapper[4791]: I1210 22:51:50.809114 4791 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-config-operator/machine-config-operator-74547568cd-rcnrn" Dec 10 22:51:50 crc kubenswrapper[4791]: I1210 22:51:50.810143 4791 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication-operator"/"trusted-ca-bundle" Dec 10 22:51:50 crc kubenswrapper[4791]: I1210 22:51:50.810209 4791 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-kube-storage-version-migrator/migrator-59844c95c7-vscpm"] Dec 10 22:51:50 crc kubenswrapper[4791]: I1210 22:51:50.810431 4791 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress-operator"/"trusted-ca" Dec 10 22:51:50 crc kubenswrapper[4791]: I1210 22:51:50.814731 4791 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-system-ocp-branding-template" Dec 10 22:51:50 crc kubenswrapper[4791]: I1210 22:51:50.815028 4791 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-storage-version-migrator/migrator-59844c95c7-vscpm" Dec 10 22:51:50 crc kubenswrapper[4791]: I1210 22:51:50.815911 4791 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-multus/multus-admission-controller-857f4d67dd-fbpcr"] Dec 10 22:51:50 crc kubenswrapper[4791]: I1210 22:51:50.816902 4791 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/multus-admission-controller-857f4d67dd-fbpcr" Dec 10 22:51:50 crc kubenswrapper[4791]: I1210 22:51:50.817886 4791 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-k7qsx"] Dec 10 22:51:50 crc kubenswrapper[4791]: I1210 22:51:50.818748 4791 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-k7qsx" Dec 10 22:51:50 crc kubenswrapper[4791]: I1210 22:51:50.819076 4791 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-controller-manager-operator"/"openshift-controller-manager-operator-serving-cert" Dec 10 22:51:50 crc kubenswrapper[4791]: I1210 22:51:50.819673 4791 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-v445v"] Dec 10 22:51:50 crc kubenswrapper[4791]: I1210 22:51:50.820202 4791 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-v445v" Dec 10 22:51:50 crc kubenswrapper[4791]: I1210 22:51:50.820731 4791 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-ms8jw"] Dec 10 22:51:50 crc kubenswrapper[4791]: I1210 22:51:50.821532 4791 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-ms8jw" Dec 10 22:51:50 crc kubenswrapper[4791]: I1210 22:51:50.821782 4791 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/marketplace-operator-79b997595-7g828"] Dec 10 22:51:50 crc kubenswrapper[4791]: I1210 22:51:50.822428 4791 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/marketplace-operator-79b997595-7g828" Dec 10 22:51:50 crc kubenswrapper[4791]: I1210 22:51:50.822786 4791 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-6576b87f9c-vsnsh"] Dec 10 22:51:50 crc kubenswrapper[4791]: I1210 22:51:50.825490 4791 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-controller-manager/controller-manager-879f6c89f-9tg4p"] Dec 10 22:51:50 crc kubenswrapper[4791]: I1210 22:51:50.825538 4791 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-oauth-apiserver/apiserver-7bbb656c7d-nwwpb"] Dec 10 22:51:50 crc kubenswrapper[4791]: I1210 22:51:50.826712 4791 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-apiserver/apiserver-76f77b778f-kqtbx"] Dec 10 22:51:50 crc kubenswrapper[4791]: I1210 22:51:50.827474 4791 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-session\" (UniqueName: \"kubernetes.io/secret/35033f1a-b91b-4e02-ba77-3841a2e0fdab-v4-0-config-system-session\") pod \"oauth-openshift-558db77b4-ht5t7\" (UID: \"35033f1a-b91b-4e02-ba77-3841a2e0fdab\") " pod="openshift-authentication/oauth-openshift-558db77b4-ht5t7" Dec 10 22:51:50 crc kubenswrapper[4791]: I1210 22:51:50.827529 4791 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/33c30ace-bd17-4a3b-a43a-ddfeccb88349-bound-sa-token\") pod \"cluster-image-registry-operator-dc59b4c8b-q4cpx\" (UID: \"33c30ace-bd17-4a3b-a43a-ddfeccb88349\") " pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-q4cpx" Dec 10 22:51:50 crc kubenswrapper[4791]: I1210 22:51:50.827561 4791 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/2f422ac4-b586-4056-8748-94b7de0b0248-config\") pod \"machine-approver-56656f9798-m6dwv\" (UID: \"2f422ac4-b586-4056-8748-94b7de0b0248\") " pod="openshift-cluster-machine-approver/machine-approver-56656f9798-m6dwv" Dec 10 22:51:50 crc kubenswrapper[4791]: I1210 22:51:50.827600 4791 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"encryption-config\" (UniqueName: \"kubernetes.io/secret/3ba17b70-17df-4f8c-87ff-c15e03ce06bc-encryption-config\") pod \"apiserver-7bbb656c7d-nwwpb\" (UID: \"3ba17b70-17df-4f8c-87ff-c15e03ce06bc\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-nwwpb" Dec 10 22:51:50 crc kubenswrapper[4791]: I1210 22:51:50.827625 4791 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"encryption-config\" (UniqueName: \"kubernetes.io/secret/4a1cf441-8a08-4677-a480-947c8cddd6a6-encryption-config\") pod \"apiserver-76f77b778f-kqtbx\" (UID: \"4a1cf441-8a08-4677-a480-947c8cddd6a6\") " pod="openshift-apiserver/apiserver-76f77b778f-kqtbx" Dec 10 22:51:50 crc kubenswrapper[4791]: I1210 22:51:50.827647 4791 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/4a1cf441-8a08-4677-a480-947c8cddd6a6-config\") pod \"apiserver-76f77b778f-kqtbx\" (UID: \"4a1cf441-8a08-4677-a480-947c8cddd6a6\") " pod="openshift-apiserver/apiserver-76f77b778f-kqtbx" Dec 10 22:51:50 crc kubenswrapper[4791]: I1210 22:51:50.827670 4791 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/3d53f039-46f2-478b-b14c-11b723b7dbc1-config\") pod \"console-operator-58897d9998-hd6zl\" (UID: \"3d53f039-46f2-478b-b14c-11b723b7dbc1\") " pod="openshift-console-operator/console-operator-58897d9998-hd6zl" Dec 10 22:51:50 crc kubenswrapper[4791]: I1210 22:51:50.827689 4791 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/e92ca7e3-d360-4e14-aeb9-76738a44c391-metrics-tls\") pod \"ingress-operator-5b745b69d9-8npst\" (UID: \"e92ca7e3-d360-4e14-aeb9-76738a44c391\") " pod="openshift-ingress-operator/ingress-operator-5b745b69d9-8npst" Dec 10 22:51:50 crc kubenswrapper[4791]: I1210 22:51:50.827720 4791 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/5369b8df-3c36-4108-b4e7-cfb11478b3a7-client-ca\") pod \"route-controller-manager-6576b87f9c-vsnsh\" (UID: \"5369b8df-3c36-4108-b4e7-cfb11478b3a7\") " pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-vsnsh" Dec 10 22:51:50 crc kubenswrapper[4791]: I1210 22:51:50.827739 4791 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"console-oauth-config\" (UniqueName: \"kubernetes.io/secret/1b61a973-db5e-4aa7-96e3-7fbdd9dd8cc0-console-oauth-config\") pod \"console-f9d7485db-zkfkv\" (UID: \"1b61a973-db5e-4aa7-96e3-7fbdd9dd8cc0\") " pod="openshift-console/console-f9d7485db-zkfkv" Dec 10 22:51:50 crc kubenswrapper[4791]: I1210 22:51:50.827764 4791 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/a5827e8a-4cb3-494d-87a1-556d920c2a4a-serving-cert\") pod \"controller-manager-879f6c89f-9tg4p\" (UID: \"a5827e8a-4cb3-494d-87a1-556d920c2a4a\") " pod="openshift-controller-manager/controller-manager-879f6c89f-9tg4p" Dec 10 22:51:50 crc kubenswrapper[4791]: I1210 22:51:50.827784 4791 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-dlncw\" (UniqueName: \"kubernetes.io/projected/35033f1a-b91b-4e02-ba77-3841a2e0fdab-kube-api-access-dlncw\") pod \"oauth-openshift-558db77b4-ht5t7\" (UID: \"35033f1a-b91b-4e02-ba77-3841a2e0fdab\") " pod="openshift-authentication/oauth-openshift-558db77b4-ht5t7" Dec 10 22:51:50 crc kubenswrapper[4791]: I1210 22:51:50.827799 4791 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"image-registry-operator-tls\" (UniqueName: \"kubernetes.io/secret/33c30ace-bd17-4a3b-a43a-ddfeccb88349-image-registry-operator-tls\") pod \"cluster-image-registry-operator-dc59b4c8b-q4cpx\" (UID: \"33c30ace-bd17-4a3b-a43a-ddfeccb88349\") " pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-q4cpx" Dec 10 22:51:50 crc kubenswrapper[4791]: I1210 22:51:50.827821 4791 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-jzkqn\" (UniqueName: \"kubernetes.io/projected/3d53f039-46f2-478b-b14c-11b723b7dbc1-kube-api-access-jzkqn\") pod \"console-operator-58897d9998-hd6zl\" (UID: \"3d53f039-46f2-478b-b14c-11b723b7dbc1\") " pod="openshift-console-operator/console-operator-58897d9998-hd6zl" Dec 10 22:51:50 crc kubenswrapper[4791]: I1210 22:51:50.827841 4791 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/5369b8df-3c36-4108-b4e7-cfb11478b3a7-config\") pod \"route-controller-manager-6576b87f9c-vsnsh\" (UID: \"5369b8df-3c36-4108-b4e7-cfb11478b3a7\") " pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-vsnsh" Dec 10 22:51:50 crc kubenswrapper[4791]: I1210 22:51:50.827858 4791 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-mtfjz\" (UniqueName: \"kubernetes.io/projected/e92ca7e3-d360-4e14-aeb9-76738a44c391-kube-api-access-mtfjz\") pod \"ingress-operator-5b745b69d9-8npst\" (UID: \"e92ca7e3-d360-4e14-aeb9-76738a44c391\") " pod="openshift-ingress-operator/ingress-operator-5b745b69d9-8npst" Dec 10 22:51:50 crc kubenswrapper[4791]: I1210 22:51:50.827879 4791 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/cff60894-8deb-4354-9ad1-3a83baa98714-trusted-ca-bundle\") pod \"authentication-operator-69f744f599-7rhk9\" (UID: \"cff60894-8deb-4354-9ad1-3a83baa98714\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-7rhk9" Dec 10 22:51:50 crc kubenswrapper[4791]: I1210 22:51:50.827898 4791 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/35033f1a-b91b-4e02-ba77-3841a2e0fdab-audit-dir\") pod \"oauth-openshift-558db77b4-ht5t7\" (UID: \"35033f1a-b91b-4e02-ba77-3841a2e0fdab\") " pod="openshift-authentication/oauth-openshift-558db77b4-ht5t7" Dec 10 22:51:50 crc kubenswrapper[4791]: I1210 22:51:50.827917 4791 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/1b61a973-db5e-4aa7-96e3-7fbdd9dd8cc0-service-ca\") pod \"console-f9d7485db-zkfkv\" (UID: \"1b61a973-db5e-4aa7-96e3-7fbdd9dd8cc0\") " pod="openshift-console/console-f9d7485db-zkfkv" Dec 10 22:51:50 crc kubenswrapper[4791]: I1210 22:51:50.827936 4791 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/4a1cf441-8a08-4677-a480-947c8cddd6a6-serving-cert\") pod \"apiserver-76f77b778f-kqtbx\" (UID: \"4a1cf441-8a08-4677-a480-947c8cddd6a6\") " pod="openshift-apiserver/apiserver-76f77b778f-kqtbx" Dec 10 22:51:50 crc kubenswrapper[4791]: I1210 22:51:50.827952 4791 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-f2tld\" (UniqueName: \"kubernetes.io/projected/1b61a973-db5e-4aa7-96e3-7fbdd9dd8cc0-kube-api-access-f2tld\") pod \"console-f9d7485db-zkfkv\" (UID: \"1b61a973-db5e-4aa7-96e3-7fbdd9dd8cc0\") " pod="openshift-console/console-f9d7485db-zkfkv" Dec 10 22:51:50 crc kubenswrapper[4791]: I1210 22:51:50.827972 4791 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/e92ca7e3-d360-4e14-aeb9-76738a44c391-bound-sa-token\") pod \"ingress-operator-5b745b69d9-8npst\" (UID: \"e92ca7e3-d360-4e14-aeb9-76738a44c391\") " pod="openshift-ingress-operator/ingress-operator-5b745b69d9-8npst" Dec 10 22:51:50 crc kubenswrapper[4791]: I1210 22:51:50.827990 4791 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/3d53f039-46f2-478b-b14c-11b723b7dbc1-serving-cert\") pod \"console-operator-58897d9998-hd6zl\" (UID: \"3d53f039-46f2-478b-b14c-11b723b7dbc1\") " pod="openshift-console-operator/console-operator-58897d9998-hd6zl" Dec 10 22:51:50 crc kubenswrapper[4791]: I1210 22:51:50.828009 4791 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"machine-approver-tls\" (UniqueName: \"kubernetes.io/secret/2f422ac4-b586-4056-8748-94b7de0b0248-machine-approver-tls\") pod \"machine-approver-56656f9798-m6dwv\" (UID: \"2f422ac4-b586-4056-8748-94b7de0b0248\") " pod="openshift-cluster-machine-approver/machine-approver-56656f9798-m6dwv" Dec 10 22:51:50 crc kubenswrapper[4791]: I1210 22:51:50.828028 4791 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"srv-cert\" (UniqueName: \"kubernetes.io/secret/8f0772d5-db77-42b2-b712-0293c0422161-srv-cert\") pod \"catalog-operator-68c6474976-497m8\" (UID: \"8f0772d5-db77-42b2-b712-0293c0422161\") " pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-497m8" Dec 10 22:51:50 crc kubenswrapper[4791]: I1210 22:51:50.828048 4791 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-w87xw\" (UniqueName: \"kubernetes.io/projected/2f422ac4-b586-4056-8748-94b7de0b0248-kube-api-access-w87xw\") pod \"machine-approver-56656f9798-m6dwv\" (UID: \"2f422ac4-b586-4056-8748-94b7de0b0248\") " pod="openshift-cluster-machine-approver/machine-approver-56656f9798-m6dwv" Dec 10 22:51:50 crc kubenswrapper[4791]: I1210 22:51:50.828066 4791 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-7m4mz\" (UniqueName: \"kubernetes.io/projected/cff60894-8deb-4354-9ad1-3a83baa98714-kube-api-access-7m4mz\") pod \"authentication-operator-69f744f599-7rhk9\" (UID: \"cff60894-8deb-4354-9ad1-3a83baa98714\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-7rhk9" Dec 10 22:51:50 crc kubenswrapper[4791]: I1210 22:51:50.828084 4791 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-serving-cert\" (UniqueName: \"kubernetes.io/secret/35033f1a-b91b-4e02-ba77-3841a2e0fdab-v4-0-config-system-serving-cert\") pod \"oauth-openshift-558db77b4-ht5t7\" (UID: \"35033f1a-b91b-4e02-ba77-3841a2e0fdab\") " pod="openshift-authentication/oauth-openshift-558db77b4-ht5t7" Dec 10 22:51:50 crc kubenswrapper[4791]: I1210 22:51:50.828105 4791 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-user-template-login\" (UniqueName: \"kubernetes.io/secret/35033f1a-b91b-4e02-ba77-3841a2e0fdab-v4-0-config-user-template-login\") pod \"oauth-openshift-558db77b4-ht5t7\" (UID: \"35033f1a-b91b-4e02-ba77-3841a2e0fdab\") " pod="openshift-authentication/oauth-openshift-558db77b4-ht5t7" Dec 10 22:51:50 crc kubenswrapper[4791]: I1210 22:51:50.828127 4791 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/3ba17b70-17df-4f8c-87ff-c15e03ce06bc-audit-dir\") pod \"apiserver-7bbb656c7d-nwwpb\" (UID: \"3ba17b70-17df-4f8c-87ff-c15e03ce06bc\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-nwwpb" Dec 10 22:51:50 crc kubenswrapper[4791]: I1210 22:51:50.828145 4791 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"console-serving-cert\" (UniqueName: \"kubernetes.io/secret/1b61a973-db5e-4aa7-96e3-7fbdd9dd8cc0-console-serving-cert\") pod \"console-f9d7485db-zkfkv\" (UID: \"1b61a973-db5e-4aa7-96e3-7fbdd9dd8cc0\") " pod="openshift-console/console-f9d7485db-zkfkv" Dec 10 22:51:50 crc kubenswrapper[4791]: I1210 22:51:50.828164 4791 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/d3de1ba0-b014-4f98-b597-cc89f5a70364-metrics-certs\") pod \"router-default-5444994796-p97fc\" (UID: \"d3de1ba0-b014-4f98-b597-cc89f5a70364\") " pod="openshift-ingress/router-default-5444994796-p97fc" Dec 10 22:51:50 crc kubenswrapper[4791]: I1210 22:51:50.828185 4791 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"mcc-auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/e2b778bb-d80f-49f9-804b-0550c648a02f-mcc-auth-proxy-config\") pod \"machine-config-controller-84d6567774-sc74x\" (UID: \"e2b778bb-d80f-49f9-804b-0550c648a02f\") " pod="openshift-machine-config-operator/machine-config-controller-84d6567774-sc74x" Dec 10 22:51:50 crc kubenswrapper[4791]: I1210 22:51:50.828203 4791 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"console-config\" (UniqueName: \"kubernetes.io/configmap/1b61a973-db5e-4aa7-96e3-7fbdd9dd8cc0-console-config\") pod \"console-f9d7485db-zkfkv\" (UID: \"1b61a973-db5e-4aa7-96e3-7fbdd9dd8cc0\") " pod="openshift-console/console-f9d7485db-zkfkv" Dec 10 22:51:50 crc kubenswrapper[4791]: I1210 22:51:50.828227 4791 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"images\" (UniqueName: \"kubernetes.io/configmap/cd65e91e-5b02-4f78-829e-93b4596ff8f6-images\") pod \"machine-api-operator-5694c8668f-cpbhz\" (UID: \"cd65e91e-5b02-4f78-829e-93b4596ff8f6\") " pod="openshift-machine-api/machine-api-operator-5694c8668f-cpbhz" Dec 10 22:51:50 crc kubenswrapper[4791]: I1210 22:51:50.828253 4791 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-tss7t\" (UniqueName: \"kubernetes.io/projected/680d2296-dd48-4786-bea7-64c32edc94c7-kube-api-access-tss7t\") pod \"downloads-7954f5f757-mlb9p\" (UID: \"680d2296-dd48-4786-bea7-64c32edc94c7\") " pod="openshift-console/downloads-7954f5f757-mlb9p" Dec 10 22:51:50 crc kubenswrapper[4791]: I1210 22:51:50.828276 4791 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/35033f1a-b91b-4e02-ba77-3841a2e0fdab-v4-0-config-system-trusted-ca-bundle\") pod \"oauth-openshift-558db77b4-ht5t7\" (UID: \"35033f1a-b91b-4e02-ba77-3841a2e0fdab\") " pod="openshift-authentication/oauth-openshift-558db77b4-ht5t7" Dec 10 22:51:50 crc kubenswrapper[4791]: I1210 22:51:50.828297 4791 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"profile-collector-cert\" (UniqueName: \"kubernetes.io/secret/8f0772d5-db77-42b2-b712-0293c0422161-profile-collector-cert\") pod \"catalog-operator-68c6474976-497m8\" (UID: \"8f0772d5-db77-42b2-b712-0293c0422161\") " pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-497m8" Dec 10 22:51:50 crc kubenswrapper[4791]: I1210 22:51:50.828319 4791 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/a5827e8a-4cb3-494d-87a1-556d920c2a4a-client-ca\") pod \"controller-manager-879f6c89f-9tg4p\" (UID: \"a5827e8a-4cb3-494d-87a1-556d920c2a4a\") " pod="openshift-controller-manager/controller-manager-879f6c89f-9tg4p" Dec 10 22:51:50 crc kubenswrapper[4791]: I1210 22:51:50.828353 4791 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etcd-serving-ca\" (UniqueName: \"kubernetes.io/configmap/3ba17b70-17df-4f8c-87ff-c15e03ce06bc-etcd-serving-ca\") pod \"apiserver-7bbb656c7d-nwwpb\" (UID: \"3ba17b70-17df-4f8c-87ff-c15e03ce06bc\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-nwwpb" Dec 10 22:51:50 crc kubenswrapper[4791]: I1210 22:51:50.828376 4791 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/5be73604-c6c9-4ef6-a53d-6c3f6df27202-config\") pod \"kube-controller-manager-operator-78b949d7b-mldvd\" (UID: \"5be73604-c6c9-4ef6-a53d-6c3f6df27202\") " pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-mldvd" Dec 10 22:51:50 crc kubenswrapper[4791]: I1210 22:51:50.828396 4791 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"image-import-ca\" (UniqueName: \"kubernetes.io/configmap/4a1cf441-8a08-4677-a480-947c8cddd6a6-image-import-ca\") pod \"apiserver-76f77b778f-kqtbx\" (UID: \"4a1cf441-8a08-4677-a480-947c8cddd6a6\") " pod="openshift-apiserver/apiserver-76f77b778f-kqtbx" Dec 10 22:51:50 crc kubenswrapper[4791]: I1210 22:51:50.828418 4791 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-k4rbk\" (UniqueName: \"kubernetes.io/projected/3ba17b70-17df-4f8c-87ff-c15e03ce06bc-kube-api-access-k4rbk\") pod \"apiserver-7bbb656c7d-nwwpb\" (UID: \"3ba17b70-17df-4f8c-87ff-c15e03ce06bc\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-nwwpb" Dec 10 22:51:50 crc kubenswrapper[4791]: I1210 22:51:50.828437 4791 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-user-idp-0-file-data\" (UniqueName: \"kubernetes.io/secret/35033f1a-b91b-4e02-ba77-3841a2e0fdab-v4-0-config-user-idp-0-file-data\") pod \"oauth-openshift-558db77b4-ht5t7\" (UID: \"35033f1a-b91b-4e02-ba77-3841a2e0fdab\") " pod="openshift-authentication/oauth-openshift-558db77b4-ht5t7" Dec 10 22:51:50 crc kubenswrapper[4791]: I1210 22:51:50.828455 4791 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"oauth-serving-cert\" (UniqueName: \"kubernetes.io/configmap/1b61a973-db5e-4aa7-96e3-7fbdd9dd8cc0-oauth-serving-cert\") pod \"console-f9d7485db-zkfkv\" (UID: \"1b61a973-db5e-4aa7-96e3-7fbdd9dd8cc0\") " pod="openshift-console/console-f9d7485db-zkfkv" Dec 10 22:51:50 crc kubenswrapper[4791]: I1210 22:51:50.828475 4791 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/3ba17b70-17df-4f8c-87ff-c15e03ce06bc-audit-policies\") pod \"apiserver-7bbb656c7d-nwwpb\" (UID: \"3ba17b70-17df-4f8c-87ff-c15e03ce06bc\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-nwwpb" Dec 10 22:51:50 crc kubenswrapper[4791]: I1210 22:51:50.828496 4791 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/3ba17b70-17df-4f8c-87ff-c15e03ce06bc-trusted-ca-bundle\") pod \"apiserver-7bbb656c7d-nwwpb\" (UID: \"3ba17b70-17df-4f8c-87ff-c15e03ce06bc\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-nwwpb" Dec 10 22:51:50 crc kubenswrapper[4791]: I1210 22:51:50.828528 4791 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"machine-api-operator-tls\" (UniqueName: \"kubernetes.io/secret/cd65e91e-5b02-4f78-829e-93b4596ff8f6-machine-api-operator-tls\") pod \"machine-api-operator-5694c8668f-cpbhz\" (UID: \"cd65e91e-5b02-4f78-829e-93b4596ff8f6\") " pod="openshift-machine-api/machine-api-operator-5694c8668f-cpbhz" Dec 10 22:51:50 crc kubenswrapper[4791]: I1210 22:51:50.828548 4791 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/35033f1a-b91b-4e02-ba77-3841a2e0fdab-audit-policies\") pod \"oauth-openshift-558db77b4-ht5t7\" (UID: \"35033f1a-b91b-4e02-ba77-3841a2e0fdab\") " pod="openshift-authentication/oauth-openshift-558db77b4-ht5t7" Dec 10 22:51:50 crc kubenswrapper[4791]: I1210 22:51:50.828566 4791 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-user-template-provider-selection\" (UniqueName: \"kubernetes.io/secret/35033f1a-b91b-4e02-ba77-3841a2e0fdab-v4-0-config-user-template-provider-selection\") pod \"oauth-openshift-558db77b4-ht5t7\" (UID: \"35033f1a-b91b-4e02-ba77-3841a2e0fdab\") " pod="openshift-authentication/oauth-openshift-558db77b4-ht5t7" Dec 10 22:51:50 crc kubenswrapper[4791]: I1210 22:51:50.828588 4791 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-user-template-error\" (UniqueName: \"kubernetes.io/secret/35033f1a-b91b-4e02-ba77-3841a2e0fdab-v4-0-config-user-template-error\") pod \"oauth-openshift-558db77b4-ht5t7\" (UID: \"35033f1a-b91b-4e02-ba77-3841a2e0fdab\") " pod="openshift-authentication/oauth-openshift-558db77b4-ht5t7" Dec 10 22:51:50 crc kubenswrapper[4791]: I1210 22:51:50.828610 4791 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-fltgv\" (UniqueName: \"kubernetes.io/projected/4205f348-3276-40f3-a35b-69306b331d80-kube-api-access-fltgv\") pod \"cluster-samples-operator-665b6dd947-t7742\" (UID: \"4205f348-3276-40f3-a35b-69306b331d80\") " pod="openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-t7742" Dec 10 22:51:50 crc kubenswrapper[4791]: I1210 22:51:50.828562 4791 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29423445-b9cxb"] Dec 10 22:51:50 crc kubenswrapper[4791]: I1210 22:51:50.828777 4791 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/3ba17b70-17df-4f8c-87ff-c15e03ce06bc-audit-dir\") pod \"apiserver-7bbb656c7d-nwwpb\" (UID: \"3ba17b70-17df-4f8c-87ff-c15e03ce06bc\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-nwwpb" Dec 10 22:51:50 crc kubenswrapper[4791]: I1210 22:51:50.828811 4791 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/3ba17b70-17df-4f8c-87ff-c15e03ce06bc-serving-cert\") pod \"apiserver-7bbb656c7d-nwwpb\" (UID: \"3ba17b70-17df-4f8c-87ff-c15e03ce06bc\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-nwwpb" Dec 10 22:51:50 crc kubenswrapper[4791]: I1210 22:51:50.828849 4791 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-cliconfig\" (UniqueName: \"kubernetes.io/configmap/35033f1a-b91b-4e02-ba77-3841a2e0fdab-v4-0-config-system-cliconfig\") pod \"oauth-openshift-558db77b4-ht5t7\" (UID: \"35033f1a-b91b-4e02-ba77-3841a2e0fdab\") " pod="openshift-authentication/oauth-openshift-558db77b4-ht5t7" Dec 10 22:51:50 crc kubenswrapper[4791]: I1210 22:51:50.828877 4791 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/4a1cf441-8a08-4677-a480-947c8cddd6a6-trusted-ca-bundle\") pod \"apiserver-76f77b778f-kqtbx\" (UID: \"4a1cf441-8a08-4677-a480-947c8cddd6a6\") " pod="openshift-apiserver/apiserver-76f77b778f-kqtbx" Dec 10 22:51:50 crc kubenswrapper[4791]: I1210 22:51:50.828899 4791 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/cff60894-8deb-4354-9ad1-3a83baa98714-serving-cert\") pod \"authentication-operator-69f744f599-7rhk9\" (UID: \"cff60894-8deb-4354-9ad1-3a83baa98714\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-7rhk9" Dec 10 22:51:50 crc kubenswrapper[4791]: I1210 22:51:50.828939 4791 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"audit\" (UniqueName: \"kubernetes.io/configmap/4a1cf441-8a08-4677-a480-947c8cddd6a6-audit\") pod \"apiserver-76f77b778f-kqtbx\" (UID: \"4a1cf441-8a08-4677-a480-947c8cddd6a6\") " pod="openshift-apiserver/apiserver-76f77b778f-kqtbx" Dec 10 22:51:50 crc kubenswrapper[4791]: I1210 22:51:50.828959 4791 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-vwnrt\" (UniqueName: \"kubernetes.io/projected/e2b778bb-d80f-49f9-804b-0550c648a02f-kube-api-access-vwnrt\") pod \"machine-config-controller-84d6567774-sc74x\" (UID: \"e2b778bb-d80f-49f9-804b-0550c648a02f\") " pod="openshift-machine-config-operator/machine-config-controller-84d6567774-sc74x" Dec 10 22:51:50 crc kubenswrapper[4791]: I1210 22:51:50.828979 4791 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/b419e5d6-d6a4-4258-a1bd-a462bd7f7a12-serving-cert\") pod \"openshift-config-operator-7777fb866f-wwpsp\" (UID: \"b419e5d6-d6a4-4258-a1bd-a462bd7f7a12\") " pod="openshift-config-operator/openshift-config-operator-7777fb866f-wwpsp" Dec 10 22:51:50 crc kubenswrapper[4791]: I1210 22:51:50.828998 4791 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/33c30ace-bd17-4a3b-a43a-ddfeccb88349-trusted-ca\") pod \"cluster-image-registry-operator-dc59b4c8b-q4cpx\" (UID: \"33c30ace-bd17-4a3b-a43a-ddfeccb88349\") " pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-q4cpx" Dec 10 22:51:50 crc kubenswrapper[4791]: I1210 22:51:50.829014 4791 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/e92ca7e3-d360-4e14-aeb9-76738a44c391-trusted-ca\") pod \"ingress-operator-5b745b69d9-8npst\" (UID: \"e92ca7e3-d360-4e14-aeb9-76738a44c391\") " pod="openshift-ingress-operator/ingress-operator-5b745b69d9-8npst" Dec 10 22:51:50 crc kubenswrapper[4791]: I1210 22:51:50.829035 4791 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-z5dwf\" (UniqueName: \"kubernetes.io/projected/5369b8df-3c36-4108-b4e7-cfb11478b3a7-kube-api-access-z5dwf\") pod \"route-controller-manager-6576b87f9c-vsnsh\" (UID: \"5369b8df-3c36-4108-b4e7-cfb11478b3a7\") " pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-vsnsh" Dec 10 22:51:50 crc kubenswrapper[4791]: I1210 22:51:50.829052 4791 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/2f422ac4-b586-4056-8748-94b7de0b0248-auth-proxy-config\") pod \"machine-approver-56656f9798-m6dwv\" (UID: \"2f422ac4-b586-4056-8748-94b7de0b0248\") " pod="openshift-cluster-machine-approver/machine-approver-56656f9798-m6dwv" Dec 10 22:51:50 crc kubenswrapper[4791]: I1210 22:51:50.829072 4791 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/4a1cf441-8a08-4677-a480-947c8cddd6a6-etcd-client\") pod \"apiserver-76f77b778f-kqtbx\" (UID: \"4a1cf441-8a08-4677-a480-947c8cddd6a6\") " pod="openshift-apiserver/apiserver-76f77b778f-kqtbx" Dec 10 22:51:50 crc kubenswrapper[4791]: I1210 22:51:50.829174 4791 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/4a1cf441-8a08-4677-a480-947c8cddd6a6-config\") pod \"apiserver-76f77b778f-kqtbx\" (UID: \"4a1cf441-8a08-4677-a480-947c8cddd6a6\") " pod="openshift-apiserver/apiserver-76f77b778f-kqtbx" Dec 10 22:51:50 crc kubenswrapper[4791]: I1210 22:51:50.829177 4791 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-ocp-branding-template\" (UniqueName: \"kubernetes.io/secret/35033f1a-b91b-4e02-ba77-3841a2e0fdab-v4-0-config-system-ocp-branding-template\") pod \"oauth-openshift-558db77b4-ht5t7\" (UID: \"35033f1a-b91b-4e02-ba77-3841a2e0fdab\") " pod="openshift-authentication/oauth-openshift-558db77b4-ht5t7" Dec 10 22:51:50 crc kubenswrapper[4791]: I1210 22:51:50.829228 4791 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-vn4qx\" (UniqueName: \"kubernetes.io/projected/d3de1ba0-b014-4f98-b597-cc89f5a70364-kube-api-access-vn4qx\") pod \"router-default-5444994796-p97fc\" (UID: \"d3de1ba0-b014-4f98-b597-cc89f5a70364\") " pod="openshift-ingress/router-default-5444994796-p97fc" Dec 10 22:51:50 crc kubenswrapper[4791]: I1210 22:51:50.829264 4791 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-vgqrv\" (UniqueName: \"kubernetes.io/projected/cd65e91e-5b02-4f78-829e-93b4596ff8f6-kube-api-access-vgqrv\") pod \"machine-api-operator-5694c8668f-cpbhz\" (UID: \"cd65e91e-5b02-4f78-829e-93b4596ff8f6\") " pod="openshift-machine-api/machine-api-operator-5694c8668f-cpbhz" Dec 10 22:51:50 crc kubenswrapper[4791]: I1210 22:51:50.829286 4791 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/3ba17b70-17df-4f8c-87ff-c15e03ce06bc-etcd-client\") pod \"apiserver-7bbb656c7d-nwwpb\" (UID: \"3ba17b70-17df-4f8c-87ff-c15e03ce06bc\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-nwwpb" Dec 10 22:51:50 crc kubenswrapper[4791]: I1210 22:51:50.829306 4791 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/5be73604-c6c9-4ef6-a53d-6c3f6df27202-kube-api-access\") pod \"kube-controller-manager-operator-78b949d7b-mldvd\" (UID: \"5be73604-c6c9-4ef6-a53d-6c3f6df27202\") " pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-mldvd" Dec 10 22:51:50 crc kubenswrapper[4791]: I1210 22:51:50.829398 4791 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-rkmd6\" (UniqueName: \"kubernetes.io/projected/4a1cf441-8a08-4677-a480-947c8cddd6a6-kube-api-access-rkmd6\") pod \"apiserver-76f77b778f-kqtbx\" (UID: \"4a1cf441-8a08-4677-a480-947c8cddd6a6\") " pod="openshift-apiserver/apiserver-76f77b778f-kqtbx" Dec 10 22:51:50 crc kubenswrapper[4791]: I1210 22:51:50.829426 4791 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/3d53f039-46f2-478b-b14c-11b723b7dbc1-trusted-ca\") pod \"console-operator-58897d9998-hd6zl\" (UID: \"3d53f039-46f2-478b-b14c-11b723b7dbc1\") " pod="openshift-console-operator/console-operator-58897d9998-hd6zl" Dec 10 22:51:50 crc kubenswrapper[4791]: I1210 22:51:50.829443 4791 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"service-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/cff60894-8deb-4354-9ad1-3a83baa98714-service-ca-bundle\") pod \"authentication-operator-69f744f599-7rhk9\" (UID: \"cff60894-8deb-4354-9ad1-3a83baa98714\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-7rhk9" Dec 10 22:51:50 crc kubenswrapper[4791]: I1210 22:51:50.829460 4791 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/5be73604-c6c9-4ef6-a53d-6c3f6df27202-serving-cert\") pod \"kube-controller-manager-operator-78b949d7b-mldvd\" (UID: \"5be73604-c6c9-4ef6-a53d-6c3f6df27202\") " pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-mldvd" Dec 10 22:51:50 crc kubenswrapper[4791]: I1210 22:51:50.829497 4791 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-xhsjq\" (UniqueName: \"kubernetes.io/projected/a5827e8a-4cb3-494d-87a1-556d920c2a4a-kube-api-access-xhsjq\") pod \"controller-manager-879f6c89f-9tg4p\" (UID: \"a5827e8a-4cb3-494d-87a1-556d920c2a4a\") " pod="openshift-controller-manager/controller-manager-879f6c89f-9tg4p" Dec 10 22:51:50 crc kubenswrapper[4791]: I1210 22:51:50.829517 4791 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"samples-operator-tls\" (UniqueName: \"kubernetes.io/secret/4205f348-3276-40f3-a35b-69306b331d80-samples-operator-tls\") pod \"cluster-samples-operator-665b6dd947-t7742\" (UID: \"4205f348-3276-40f3-a35b-69306b331d80\") " pod="openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-t7742" Dec 10 22:51:50 crc kubenswrapper[4791]: I1210 22:51:50.829542 4791 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etcd-serving-ca\" (UniqueName: \"kubernetes.io/configmap/4a1cf441-8a08-4677-a480-947c8cddd6a6-etcd-serving-ca\") pod \"apiserver-76f77b778f-kqtbx\" (UID: \"4a1cf441-8a08-4677-a480-947c8cddd6a6\") " pod="openshift-apiserver/apiserver-76f77b778f-kqtbx" Dec 10 22:51:50 crc kubenswrapper[4791]: I1210 22:51:50.829561 4791 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/a5827e8a-4cb3-494d-87a1-556d920c2a4a-proxy-ca-bundles\") pod \"controller-manager-879f6c89f-9tg4p\" (UID: \"a5827e8a-4cb3-494d-87a1-556d920c2a4a\") " pod="openshift-controller-manager/controller-manager-879f6c89f-9tg4p" Dec 10 22:51:50 crc kubenswrapper[4791]: I1210 22:51:50.829588 4791 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-ggn6x\" (UniqueName: \"kubernetes.io/projected/33c30ace-bd17-4a3b-a43a-ddfeccb88349-kube-api-access-ggn6x\") pod \"cluster-image-registry-operator-dc59b4c8b-q4cpx\" (UID: \"33c30ace-bd17-4a3b-a43a-ddfeccb88349\") " pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-q4cpx" Dec 10 22:51:50 crc kubenswrapper[4791]: I1210 22:51:50.829607 4791 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/465ba82f-1142-4cf1-b2ec-1f73e565ec4d-config\") pod \"openshift-controller-manager-operator-756b6f6bc6-5t9vv\" (UID: \"465ba82f-1142-4cf1-b2ec-1f73e565ec4d\") " pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-5t9vv" Dec 10 22:51:50 crc kubenswrapper[4791]: I1210 22:51:50.829626 4791 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/e2b778bb-d80f-49f9-804b-0550c648a02f-proxy-tls\") pod \"machine-config-controller-84d6567774-sc74x\" (UID: \"e2b778bb-d80f-49f9-804b-0550c648a02f\") " pod="openshift-machine-config-operator/machine-config-controller-84d6567774-sc74x" Dec 10 22:51:50 crc kubenswrapper[4791]: I1210 22:51:50.829644 4791 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/1b61a973-db5e-4aa7-96e3-7fbdd9dd8cc0-trusted-ca-bundle\") pod \"console-f9d7485db-zkfkv\" (UID: \"1b61a973-db5e-4aa7-96e3-7fbdd9dd8cc0\") " pod="openshift-console/console-f9d7485db-zkfkv" Dec 10 22:51:50 crc kubenswrapper[4791]: I1210 22:51:50.829666 4791 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"node-pullsecrets\" (UniqueName: \"kubernetes.io/host-path/4a1cf441-8a08-4677-a480-947c8cddd6a6-node-pullsecrets\") pod \"apiserver-76f77b778f-kqtbx\" (UID: \"4a1cf441-8a08-4677-a480-947c8cddd6a6\") " pod="openshift-apiserver/apiserver-76f77b778f-kqtbx" Dec 10 22:51:50 crc kubenswrapper[4791]: I1210 22:51:50.829456 4791 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-dns/dns-default-qs77n"] Dec 10 22:51:50 crc kubenswrapper[4791]: I1210 22:51:50.829892 4791 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/a5827e8a-4cb3-494d-87a1-556d920c2a4a-client-ca\") pod \"controller-manager-879f6c89f-9tg4p\" (UID: \"a5827e8a-4cb3-494d-87a1-556d920c2a4a\") " pod="openshift-controller-manager/controller-manager-879f6c89f-9tg4p" Dec 10 22:51:50 crc kubenswrapper[4791]: I1210 22:51:50.830443 4791 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-dns/dns-default-qs77n" Dec 10 22:51:50 crc kubenswrapper[4791]: I1210 22:51:50.830752 4791 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-router-certs\" (UniqueName: \"kubernetes.io/secret/35033f1a-b91b-4e02-ba77-3841a2e0fdab-v4-0-config-system-router-certs\") pod \"oauth-openshift-558db77b4-ht5t7\" (UID: \"35033f1a-b91b-4e02-ba77-3841a2e0fdab\") " pod="openshift-authentication/oauth-openshift-558db77b4-ht5t7" Dec 10 22:51:50 crc kubenswrapper[4791]: I1210 22:51:50.830787 4791 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"available-featuregates\" (UniqueName: \"kubernetes.io/empty-dir/b419e5d6-d6a4-4258-a1bd-a462bd7f7a12-available-featuregates\") pod \"openshift-config-operator-7777fb866f-wwpsp\" (UID: \"b419e5d6-d6a4-4258-a1bd-a462bd7f7a12\") " pod="openshift-config-operator/openshift-config-operator-7777fb866f-wwpsp" Dec 10 22:51:50 crc kubenswrapper[4791]: I1210 22:51:50.830813 4791 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/cd65e91e-5b02-4f78-829e-93b4596ff8f6-config\") pod \"machine-api-operator-5694c8668f-cpbhz\" (UID: \"cd65e91e-5b02-4f78-829e-93b4596ff8f6\") " pod="openshift-machine-api/machine-api-operator-5694c8668f-cpbhz" Dec 10 22:51:50 crc kubenswrapper[4791]: I1210 22:51:50.830840 4791 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/a5827e8a-4cb3-494d-87a1-556d920c2a4a-config\") pod \"controller-manager-879f6c89f-9tg4p\" (UID: \"a5827e8a-4cb3-494d-87a1-556d920c2a4a\") " pod="openshift-controller-manager/controller-manager-879f6c89f-9tg4p" Dec 10 22:51:50 crc kubenswrapper[4791]: I1210 22:51:50.830858 4791 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/5369b8df-3c36-4108-b4e7-cfb11478b3a7-serving-cert\") pod \"route-controller-manager-6576b87f9c-vsnsh\" (UID: \"5369b8df-3c36-4108-b4e7-cfb11478b3a7\") " pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-vsnsh" Dec 10 22:51:50 crc kubenswrapper[4791]: I1210 22:51:50.830894 4791 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-f2cbn\" (UniqueName: \"kubernetes.io/projected/465ba82f-1142-4cf1-b2ec-1f73e565ec4d-kube-api-access-f2cbn\") pod \"openshift-controller-manager-operator-756b6f6bc6-5t9vv\" (UID: \"465ba82f-1142-4cf1-b2ec-1f73e565ec4d\") " pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-5t9vv" Dec 10 22:51:50 crc kubenswrapper[4791]: I1210 22:51:50.830910 4791 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29423445-b9cxb" Dec 10 22:51:50 crc kubenswrapper[4791]: I1210 22:51:50.830962 4791 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"node-pullsecrets\" (UniqueName: \"kubernetes.io/host-path/4a1cf441-8a08-4677-a480-947c8cddd6a6-node-pullsecrets\") pod \"apiserver-76f77b778f-kqtbx\" (UID: \"4a1cf441-8a08-4677-a480-947c8cddd6a6\") " pod="openshift-apiserver/apiserver-76f77b778f-kqtbx" Dec 10 22:51:50 crc kubenswrapper[4791]: I1210 22:51:50.831029 4791 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"image-import-ca\" (UniqueName: \"kubernetes.io/configmap/4a1cf441-8a08-4677-a480-947c8cddd6a6-image-import-ca\") pod \"apiserver-76f77b778f-kqtbx\" (UID: \"4a1cf441-8a08-4677-a480-947c8cddd6a6\") " pod="openshift-apiserver/apiserver-76f77b778f-kqtbx" Dec 10 22:51:50 crc kubenswrapper[4791]: I1210 22:51:50.831080 4791 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-service-ca-operator/service-ca-operator-777779d784-b8v4c"] Dec 10 22:51:50 crc kubenswrapper[4791]: I1210 22:51:50.831675 4791 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-service-ca-operator/service-ca-operator-777779d784-b8v4c" Dec 10 22:51:50 crc kubenswrapper[4791]: I1210 22:51:50.831704 4791 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-machine-config-operator/machine-config-server-ktcmn"] Dec 10 22:51:50 crc kubenswrapper[4791]: I1210 22:51:50.830914 4791 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-692qq\" (UniqueName: \"kubernetes.io/projected/b419e5d6-d6a4-4258-a1bd-a462bd7f7a12-kube-api-access-692qq\") pod \"openshift-config-operator-7777fb866f-wwpsp\" (UID: \"b419e5d6-d6a4-4258-a1bd-a462bd7f7a12\") " pod="openshift-config-operator/openshift-config-operator-7777fb866f-wwpsp" Dec 10 22:51:50 crc kubenswrapper[4791]: I1210 22:51:50.832060 4791 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"images\" (UniqueName: \"kubernetes.io/configmap/cd65e91e-5b02-4f78-829e-93b4596ff8f6-images\") pod \"machine-api-operator-5694c8668f-cpbhz\" (UID: \"cd65e91e-5b02-4f78-829e-93b4596ff8f6\") " pod="openshift-machine-api/machine-api-operator-5694c8668f-cpbhz" Dec 10 22:51:50 crc kubenswrapper[4791]: I1210 22:51:50.832084 4791 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"default-certificate\" (UniqueName: \"kubernetes.io/secret/d3de1ba0-b014-4f98-b597-cc89f5a70364-default-certificate\") pod \"router-default-5444994796-p97fc\" (UID: \"d3de1ba0-b014-4f98-b597-cc89f5a70364\") " pod="openshift-ingress/router-default-5444994796-p97fc" Dec 10 22:51:50 crc kubenswrapper[4791]: I1210 22:51:50.832111 4791 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"service-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/d3de1ba0-b014-4f98-b597-cc89f5a70364-service-ca-bundle\") pod \"router-default-5444994796-p97fc\" (UID: \"d3de1ba0-b014-4f98-b597-cc89f5a70364\") " pod="openshift-ingress/router-default-5444994796-p97fc" Dec 10 22:51:50 crc kubenswrapper[4791]: I1210 22:51:50.832138 4791 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/cff60894-8deb-4354-9ad1-3a83baa98714-config\") pod \"authentication-operator-69f744f599-7rhk9\" (UID: \"cff60894-8deb-4354-9ad1-3a83baa98714\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-7rhk9" Dec 10 22:51:50 crc kubenswrapper[4791]: I1210 22:51:50.832164 4791 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-z492g\" (UniqueName: \"kubernetes.io/projected/8f0772d5-db77-42b2-b712-0293c0422161-kube-api-access-z492g\") pod \"catalog-operator-68c6474976-497m8\" (UID: \"8f0772d5-db77-42b2-b712-0293c0422161\") " pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-497m8" Dec 10 22:51:50 crc kubenswrapper[4791]: I1210 22:51:50.832182 4791 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-service-ca\" (UniqueName: \"kubernetes.io/configmap/35033f1a-b91b-4e02-ba77-3841a2e0fdab-v4-0-config-system-service-ca\") pod \"oauth-openshift-558db77b4-ht5t7\" (UID: \"35033f1a-b91b-4e02-ba77-3841a2e0fdab\") " pod="openshift-authentication/oauth-openshift-558db77b4-ht5t7" Dec 10 22:51:50 crc kubenswrapper[4791]: I1210 22:51:50.832198 4791 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"stats-auth\" (UniqueName: \"kubernetes.io/secret/d3de1ba0-b014-4f98-b597-cc89f5a70364-stats-auth\") pod \"router-default-5444994796-p97fc\" (UID: \"d3de1ba0-b014-4f98-b597-cc89f5a70364\") " pod="openshift-ingress/router-default-5444994796-p97fc" Dec 10 22:51:50 crc kubenswrapper[4791]: I1210 22:51:50.832220 4791 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/465ba82f-1142-4cf1-b2ec-1f73e565ec4d-serving-cert\") pod \"openshift-controller-manager-operator-756b6f6bc6-5t9vv\" (UID: \"465ba82f-1142-4cf1-b2ec-1f73e565ec4d\") " pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-5t9vv" Dec 10 22:51:50 crc kubenswrapper[4791]: I1210 22:51:50.832247 4791 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/4a1cf441-8a08-4677-a480-947c8cddd6a6-audit-dir\") pod \"apiserver-76f77b778f-kqtbx\" (UID: \"4a1cf441-8a08-4677-a480-947c8cddd6a6\") " pod="openshift-apiserver/apiserver-76f77b778f-kqtbx" Dec 10 22:51:50 crc kubenswrapper[4791]: I1210 22:51:50.832250 4791 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-config-operator/machine-config-server-ktcmn" Dec 10 22:51:50 crc kubenswrapper[4791]: I1210 22:51:50.832515 4791 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"encryption-config\" (UniqueName: \"kubernetes.io/secret/3ba17b70-17df-4f8c-87ff-c15e03ce06bc-encryption-config\") pod \"apiserver-7bbb656c7d-nwwpb\" (UID: \"3ba17b70-17df-4f8c-87ff-c15e03ce06bc\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-nwwpb" Dec 10 22:51:50 crc kubenswrapper[4791]: I1210 22:51:50.833654 4791 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"audit\" (UniqueName: \"kubernetes.io/configmap/4a1cf441-8a08-4677-a480-947c8cddd6a6-audit\") pod \"apiserver-76f77b778f-kqtbx\" (UID: \"4a1cf441-8a08-4677-a480-947c8cddd6a6\") " pod="openshift-apiserver/apiserver-76f77b778f-kqtbx" Dec 10 22:51:50 crc kubenswrapper[4791]: I1210 22:51:50.833995 4791 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/4a1cf441-8a08-4677-a480-947c8cddd6a6-audit-dir\") pod \"apiserver-76f77b778f-kqtbx\" (UID: \"4a1cf441-8a08-4677-a480-947c8cddd6a6\") " pod="openshift-apiserver/apiserver-76f77b778f-kqtbx" Dec 10 22:51:50 crc kubenswrapper[4791]: I1210 22:51:50.834768 4791 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/3ba17b70-17df-4f8c-87ff-c15e03ce06bc-audit-policies\") pod \"apiserver-7bbb656c7d-nwwpb\" (UID: \"3ba17b70-17df-4f8c-87ff-c15e03ce06bc\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-nwwpb" Dec 10 22:51:50 crc kubenswrapper[4791]: I1210 22:51:50.834909 4791 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/cd65e91e-5b02-4f78-829e-93b4596ff8f6-config\") pod \"machine-api-operator-5694c8668f-cpbhz\" (UID: \"cd65e91e-5b02-4f78-829e-93b4596ff8f6\") " pod="openshift-machine-api/machine-api-operator-5694c8668f-cpbhz" Dec 10 22:51:50 crc kubenswrapper[4791]: I1210 22:51:50.835290 4791 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/3ba17b70-17df-4f8c-87ff-c15e03ce06bc-trusted-ca-bundle\") pod \"apiserver-7bbb656c7d-nwwpb\" (UID: \"3ba17b70-17df-4f8c-87ff-c15e03ce06bc\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-nwwpb" Dec 10 22:51:50 crc kubenswrapper[4791]: I1210 22:51:50.835421 4791 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-service-ca/service-ca-9c57cc56f-x9pw8"] Dec 10 22:51:50 crc kubenswrapper[4791]: I1210 22:51:50.835753 4791 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"encryption-config\" (UniqueName: \"kubernetes.io/secret/4a1cf441-8a08-4677-a480-947c8cddd6a6-encryption-config\") pod \"apiserver-76f77b778f-kqtbx\" (UID: \"4a1cf441-8a08-4677-a480-947c8cddd6a6\") " pod="openshift-apiserver/apiserver-76f77b778f-kqtbx" Dec 10 22:51:50 crc kubenswrapper[4791]: I1210 22:51:50.836211 4791 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/a5827e8a-4cb3-494d-87a1-556d920c2a4a-proxy-ca-bundles\") pod \"controller-manager-879f6c89f-9tg4p\" (UID: \"a5827e8a-4cb3-494d-87a1-556d920c2a4a\") " pod="openshift-controller-manager/controller-manager-879f6c89f-9tg4p" Dec 10 22:51:50 crc kubenswrapper[4791]: I1210 22:51:50.836875 4791 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/3ba17b70-17df-4f8c-87ff-c15e03ce06bc-etcd-client\") pod \"apiserver-7bbb656c7d-nwwpb\" (UID: \"3ba17b70-17df-4f8c-87ff-c15e03ce06bc\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-nwwpb" Dec 10 22:51:50 crc kubenswrapper[4791]: I1210 22:51:50.837027 4791 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/a5827e8a-4cb3-494d-87a1-556d920c2a4a-config\") pod \"controller-manager-879f6c89f-9tg4p\" (UID: \"a5827e8a-4cb3-494d-87a1-556d920c2a4a\") " pod="openshift-controller-manager/controller-manager-879f6c89f-9tg4p" Dec 10 22:51:50 crc kubenswrapper[4791]: I1210 22:51:50.837485 4791 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etcd-serving-ca\" (UniqueName: \"kubernetes.io/configmap/3ba17b70-17df-4f8c-87ff-c15e03ce06bc-etcd-serving-ca\") pod \"apiserver-7bbb656c7d-nwwpb\" (UID: \"3ba17b70-17df-4f8c-87ff-c15e03ce06bc\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-nwwpb" Dec 10 22:51:50 crc kubenswrapper[4791]: I1210 22:51:50.838549 4791 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/a5827e8a-4cb3-494d-87a1-556d920c2a4a-serving-cert\") pod \"controller-manager-879f6c89f-9tg4p\" (UID: \"a5827e8a-4cb3-494d-87a1-556d920c2a4a\") " pod="openshift-controller-manager/controller-manager-879f6c89f-9tg4p" Dec 10 22:51:50 crc kubenswrapper[4791]: I1210 22:51:50.838557 4791 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etcd-serving-ca\" (UniqueName: \"kubernetes.io/configmap/4a1cf441-8a08-4677-a480-947c8cddd6a6-etcd-serving-ca\") pod \"apiserver-76f77b778f-kqtbx\" (UID: \"4a1cf441-8a08-4677-a480-947c8cddd6a6\") " pod="openshift-apiserver/apiserver-76f77b778f-kqtbx" Dec 10 22:51:50 crc kubenswrapper[4791]: I1210 22:51:50.838616 4791 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-console/downloads-7954f5f757-mlb9p"] Dec 10 22:51:50 crc kubenswrapper[4791]: I1210 22:51:50.838641 4791 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-t7742"] Dec 10 22:51:50 crc kubenswrapper[4791]: I1210 22:51:50.838751 4791 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-service-ca/service-ca-9c57cc56f-x9pw8" Dec 10 22:51:50 crc kubenswrapper[4791]: I1210 22:51:50.838974 4791 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-console-operator/console-operator-58897d9998-hd6zl"] Dec 10 22:51:50 crc kubenswrapper[4791]: I1210 22:51:50.839910 4791 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager-operator"/"openshift-service-ca.crt" Dec 10 22:51:50 crc kubenswrapper[4791]: I1210 22:51:50.840110 4791 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/4a1cf441-8a08-4677-a480-947c8cddd6a6-trusted-ca-bundle\") pod \"apiserver-76f77b778f-kqtbx\" (UID: \"4a1cf441-8a08-4677-a480-947c8cddd6a6\") " pod="openshift-apiserver/apiserver-76f77b778f-kqtbx" Dec 10 22:51:50 crc kubenswrapper[4791]: I1210 22:51:50.840320 4791 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-q4cpx"] Dec 10 22:51:50 crc kubenswrapper[4791]: I1210 22:51:50.842435 4791 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-5t9vv"] Dec 10 22:51:50 crc kubenswrapper[4791]: I1210 22:51:50.842914 4791 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/3ba17b70-17df-4f8c-87ff-c15e03ce06bc-serving-cert\") pod \"apiserver-7bbb656c7d-nwwpb\" (UID: \"3ba17b70-17df-4f8c-87ff-c15e03ce06bc\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-nwwpb" Dec 10 22:51:50 crc kubenswrapper[4791]: I1210 22:51:50.843294 4791 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-machine-api/machine-api-operator-5694c8668f-cpbhz"] Dec 10 22:51:50 crc kubenswrapper[4791]: I1210 22:51:50.844152 4791 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/4a1cf441-8a08-4677-a480-947c8cddd6a6-etcd-client\") pod \"apiserver-76f77b778f-kqtbx\" (UID: \"4a1cf441-8a08-4677-a480-947c8cddd6a6\") " pod="openshift-apiserver/apiserver-76f77b778f-kqtbx" Dec 10 22:51:50 crc kubenswrapper[4791]: I1210 22:51:50.844640 4791 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-bws6n"] Dec 10 22:51:50 crc kubenswrapper[4791]: I1210 22:51:50.845138 4791 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/4a1cf441-8a08-4677-a480-947c8cddd6a6-serving-cert\") pod \"apiserver-76f77b778f-kqtbx\" (UID: \"4a1cf441-8a08-4677-a480-947c8cddd6a6\") " pod="openshift-apiserver/apiserver-76f77b778f-kqtbx" Dec 10 22:51:50 crc kubenswrapper[4791]: I1210 22:51:50.845887 4791 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-k7qsx"] Dec 10 22:51:50 crc kubenswrapper[4791]: I1210 22:51:50.847196 4791 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-z6gpf"] Dec 10 22:51:50 crc kubenswrapper[4791]: I1210 22:51:50.848020 4791 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"machine-api-operator-tls\" (UniqueName: \"kubernetes.io/secret/cd65e91e-5b02-4f78-829e-93b4596ff8f6-machine-api-operator-tls\") pod \"machine-api-operator-5694c8668f-cpbhz\" (UID: \"cd65e91e-5b02-4f78-829e-93b4596ff8f6\") " pod="openshift-machine-api/machine-api-operator-5694c8668f-cpbhz" Dec 10 22:51:50 crc kubenswrapper[4791]: I1210 22:51:50.848629 4791 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-config-operator/openshift-config-operator-7777fb866f-wwpsp"] Dec 10 22:51:50 crc kubenswrapper[4791]: I1210 22:51:50.850076 4791 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-ingress-operator/ingress-operator-5b745b69d9-8npst"] Dec 10 22:51:50 crc kubenswrapper[4791]: I1210 22:51:50.851466 4791 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-ms8jw"] Dec 10 22:51:50 crc kubenswrapper[4791]: I1210 22:51:50.853435 4791 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/catalog-operator-68c6474976-497m8"] Dec 10 22:51:50 crc kubenswrapper[4791]: I1210 22:51:50.854917 4791 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-storage-version-migrator/migrator-59844c95c7-vscpm"] Dec 10 22:51:50 crc kubenswrapper[4791]: I1210 22:51:50.856272 4791 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-authentication-operator/authentication-operator-69f744f599-7rhk9"] Dec 10 22:51:50 crc kubenswrapper[4791]: I1210 22:51:50.856704 4791 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager-operator"/"kube-root-ca.crt" Dec 10 22:51:50 crc kubenswrapper[4791]: I1210 22:51:50.857664 4791 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-mldvd"] Dec 10 22:51:50 crc kubenswrapper[4791]: I1210 22:51:50.859117 4791 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-authentication/oauth-openshift-558db77b4-ht5t7"] Dec 10 22:51:50 crc kubenswrapper[4791]: I1210 22:51:50.861793 4791 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-machine-config-operator/machine-config-controller-84d6567774-sc74x"] Dec 10 22:51:50 crc kubenswrapper[4791]: I1210 22:51:50.863240 4791 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-ingress-canary/ingress-canary-9dwf5"] Dec 10 22:51:50 crc kubenswrapper[4791]: I1210 22:51:50.865036 4791 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["hostpath-provisioner/csi-hostpathplugin-mkz5r"] Dec 10 22:51:50 crc kubenswrapper[4791]: I1210 22:51:50.865273 4791 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-ingress-canary/ingress-canary-9dwf5" Dec 10 22:51:50 crc kubenswrapper[4791]: I1210 22:51:50.866432 4791 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-v445v"] Dec 10 22:51:50 crc kubenswrapper[4791]: I1210 22:51:50.866523 4791 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="hostpath-provisioner/csi-hostpathplugin-mkz5r" Dec 10 22:51:50 crc kubenswrapper[4791]: I1210 22:51:50.873822 4791 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-w5dpr"] Dec 10 22:51:50 crc kubenswrapper[4791]: I1210 22:51:50.875870 4791 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-machine-config-operator/machine-config-operator-74547568cd-rcnrn"] Dec 10 22:51:50 crc kubenswrapper[4791]: I1210 22:51:50.878138 4791 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-controller-manager-operator"/"openshift-controller-manager-operator-dockercfg-vw8fw" Dec 10 22:51:50 crc kubenswrapper[4791]: I1210 22:51:50.878957 4791 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-image-registry/image-registry-697d97f7c8-srd52"] Dec 10 22:51:50 crc kubenswrapper[4791]: I1210 22:51:50.881972 4791 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-multus/multus-admission-controller-857f4d67dd-fbpcr"] Dec 10 22:51:50 crc kubenswrapper[4791]: I1210 22:51:50.883789 4791 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-dns-operator/dns-operator-744455d44c-cggs2"] Dec 10 22:51:50 crc kubenswrapper[4791]: I1210 22:51:50.884769 4791 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-rmjd2"] Dec 10 22:51:50 crc kubenswrapper[4791]: I1210 22:51:50.885921 4791 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-vm842"] Dec 10 22:51:50 crc kubenswrapper[4791]: I1210 22:51:50.887056 4791 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-dns/dns-default-qs77n"] Dec 10 22:51:50 crc kubenswrapper[4791]: I1210 22:51:50.888313 4791 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-etcd-operator/etcd-operator-b45778765-nkt8g"] Dec 10 22:51:50 crc kubenswrapper[4791]: I1210 22:51:50.889534 4791 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-console/console-f9d7485db-zkfkv"] Dec 10 22:51:50 crc kubenswrapper[4791]: I1210 22:51:50.890607 4791 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["hostpath-provisioner/csi-hostpathplugin-mkz5r"] Dec 10 22:51:50 crc kubenswrapper[4791]: I1210 22:51:50.891692 4791 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-ingress-canary/ingress-canary-9dwf5"] Dec 10 22:51:50 crc kubenswrapper[4791]: I1210 22:51:50.892845 4791 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/marketplace-operator-79b997595-7g828"] Dec 10 22:51:50 crc kubenswrapper[4791]: I1210 22:51:50.893854 4791 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29423445-b9cxb"] Dec 10 22:51:50 crc kubenswrapper[4791]: I1210 22:51:50.894932 4791 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-service-ca/service-ca-9c57cc56f-x9pw8"] Dec 10 22:51:50 crc kubenswrapper[4791]: I1210 22:51:50.895998 4791 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-service-ca-operator/service-ca-operator-777779d784-b8v4c"] Dec 10 22:51:50 crc kubenswrapper[4791]: I1210 22:51:50.897222 4791 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress-operator"/"metrics-tls" Dec 10 22:51:50 crc kubenswrapper[4791]: I1210 22:51:50.916423 4791 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress"/"openshift-service-ca.crt" Dec 10 22:51:50 crc kubenswrapper[4791]: I1210 22:51:50.932794 4791 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-ggn6x\" (UniqueName: \"kubernetes.io/projected/33c30ace-bd17-4a3b-a43a-ddfeccb88349-kube-api-access-ggn6x\") pod \"cluster-image-registry-operator-dc59b4c8b-q4cpx\" (UID: \"33c30ace-bd17-4a3b-a43a-ddfeccb88349\") " pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-q4cpx" Dec 10 22:51:50 crc kubenswrapper[4791]: I1210 22:51:50.932842 4791 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/465ba82f-1142-4cf1-b2ec-1f73e565ec4d-config\") pod \"openshift-controller-manager-operator-756b6f6bc6-5t9vv\" (UID: \"465ba82f-1142-4cf1-b2ec-1f73e565ec4d\") " pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-5t9vv" Dec 10 22:51:50 crc kubenswrapper[4791]: I1210 22:51:50.932871 4791 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/e2b778bb-d80f-49f9-804b-0550c648a02f-proxy-tls\") pod \"machine-config-controller-84d6567774-sc74x\" (UID: \"e2b778bb-d80f-49f9-804b-0550c648a02f\") " pod="openshift-machine-config-operator/machine-config-controller-84d6567774-sc74x" Dec 10 22:51:50 crc kubenswrapper[4791]: I1210 22:51:50.932894 4791 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/1b61a973-db5e-4aa7-96e3-7fbdd9dd8cc0-trusted-ca-bundle\") pod \"console-f9d7485db-zkfkv\" (UID: \"1b61a973-db5e-4aa7-96e3-7fbdd9dd8cc0\") " pod="openshift-console/console-f9d7485db-zkfkv" Dec 10 22:51:50 crc kubenswrapper[4791]: I1210 22:51:50.932917 4791 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-router-certs\" (UniqueName: \"kubernetes.io/secret/35033f1a-b91b-4e02-ba77-3841a2e0fdab-v4-0-config-system-router-certs\") pod \"oauth-openshift-558db77b4-ht5t7\" (UID: \"35033f1a-b91b-4e02-ba77-3841a2e0fdab\") " pod="openshift-authentication/oauth-openshift-558db77b4-ht5t7" Dec 10 22:51:50 crc kubenswrapper[4791]: I1210 22:51:50.932942 4791 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"available-featuregates\" (UniqueName: \"kubernetes.io/empty-dir/b419e5d6-d6a4-4258-a1bd-a462bd7f7a12-available-featuregates\") pod \"openshift-config-operator-7777fb866f-wwpsp\" (UID: \"b419e5d6-d6a4-4258-a1bd-a462bd7f7a12\") " pod="openshift-config-operator/openshift-config-operator-7777fb866f-wwpsp" Dec 10 22:51:50 crc kubenswrapper[4791]: I1210 22:51:50.932965 4791 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/5369b8df-3c36-4108-b4e7-cfb11478b3a7-serving-cert\") pod \"route-controller-manager-6576b87f9c-vsnsh\" (UID: \"5369b8df-3c36-4108-b4e7-cfb11478b3a7\") " pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-vsnsh" Dec 10 22:51:50 crc kubenswrapper[4791]: I1210 22:51:50.933036 4791 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-f2cbn\" (UniqueName: \"kubernetes.io/projected/465ba82f-1142-4cf1-b2ec-1f73e565ec4d-kube-api-access-f2cbn\") pod \"openshift-controller-manager-operator-756b6f6bc6-5t9vv\" (UID: \"465ba82f-1142-4cf1-b2ec-1f73e565ec4d\") " pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-5t9vv" Dec 10 22:51:50 crc kubenswrapper[4791]: I1210 22:51:50.933062 4791 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-692qq\" (UniqueName: \"kubernetes.io/projected/b419e5d6-d6a4-4258-a1bd-a462bd7f7a12-kube-api-access-692qq\") pod \"openshift-config-operator-7777fb866f-wwpsp\" (UID: \"b419e5d6-d6a4-4258-a1bd-a462bd7f7a12\") " pod="openshift-config-operator/openshift-config-operator-7777fb866f-wwpsp" Dec 10 22:51:50 crc kubenswrapper[4791]: I1210 22:51:50.933083 4791 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"default-certificate\" (UniqueName: \"kubernetes.io/secret/d3de1ba0-b014-4f98-b597-cc89f5a70364-default-certificate\") pod \"router-default-5444994796-p97fc\" (UID: \"d3de1ba0-b014-4f98-b597-cc89f5a70364\") " pod="openshift-ingress/router-default-5444994796-p97fc" Dec 10 22:51:50 crc kubenswrapper[4791]: I1210 22:51:50.933106 4791 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"service-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/d3de1ba0-b014-4f98-b597-cc89f5a70364-service-ca-bundle\") pod \"router-default-5444994796-p97fc\" (UID: \"d3de1ba0-b014-4f98-b597-cc89f5a70364\") " pod="openshift-ingress/router-default-5444994796-p97fc" Dec 10 22:51:50 crc kubenswrapper[4791]: I1210 22:51:50.933129 4791 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/cff60894-8deb-4354-9ad1-3a83baa98714-config\") pod \"authentication-operator-69f744f599-7rhk9\" (UID: \"cff60894-8deb-4354-9ad1-3a83baa98714\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-7rhk9" Dec 10 22:51:50 crc kubenswrapper[4791]: I1210 22:51:50.933154 4791 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-z492g\" (UniqueName: \"kubernetes.io/projected/8f0772d5-db77-42b2-b712-0293c0422161-kube-api-access-z492g\") pod \"catalog-operator-68c6474976-497m8\" (UID: \"8f0772d5-db77-42b2-b712-0293c0422161\") " pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-497m8" Dec 10 22:51:50 crc kubenswrapper[4791]: I1210 22:51:50.933178 4791 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-service-ca\" (UniqueName: \"kubernetes.io/configmap/35033f1a-b91b-4e02-ba77-3841a2e0fdab-v4-0-config-system-service-ca\") pod \"oauth-openshift-558db77b4-ht5t7\" (UID: \"35033f1a-b91b-4e02-ba77-3841a2e0fdab\") " pod="openshift-authentication/oauth-openshift-558db77b4-ht5t7" Dec 10 22:51:50 crc kubenswrapper[4791]: I1210 22:51:50.933199 4791 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"stats-auth\" (UniqueName: \"kubernetes.io/secret/d3de1ba0-b014-4f98-b597-cc89f5a70364-stats-auth\") pod \"router-default-5444994796-p97fc\" (UID: \"d3de1ba0-b014-4f98-b597-cc89f5a70364\") " pod="openshift-ingress/router-default-5444994796-p97fc" Dec 10 22:51:50 crc kubenswrapper[4791]: I1210 22:51:50.933221 4791 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/465ba82f-1142-4cf1-b2ec-1f73e565ec4d-serving-cert\") pod \"openshift-controller-manager-operator-756b6f6bc6-5t9vv\" (UID: \"465ba82f-1142-4cf1-b2ec-1f73e565ec4d\") " pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-5t9vv" Dec 10 22:51:50 crc kubenswrapper[4791]: I1210 22:51:50.933248 4791 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-session\" (UniqueName: \"kubernetes.io/secret/35033f1a-b91b-4e02-ba77-3841a2e0fdab-v4-0-config-system-session\") pod \"oauth-openshift-558db77b4-ht5t7\" (UID: \"35033f1a-b91b-4e02-ba77-3841a2e0fdab\") " pod="openshift-authentication/oauth-openshift-558db77b4-ht5t7" Dec 10 22:51:50 crc kubenswrapper[4791]: I1210 22:51:50.933271 4791 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/33c30ace-bd17-4a3b-a43a-ddfeccb88349-bound-sa-token\") pod \"cluster-image-registry-operator-dc59b4c8b-q4cpx\" (UID: \"33c30ace-bd17-4a3b-a43a-ddfeccb88349\") " pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-q4cpx" Dec 10 22:51:50 crc kubenswrapper[4791]: I1210 22:51:50.933297 4791 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/2f422ac4-b586-4056-8748-94b7de0b0248-config\") pod \"machine-approver-56656f9798-m6dwv\" (UID: \"2f422ac4-b586-4056-8748-94b7de0b0248\") " pod="openshift-cluster-machine-approver/machine-approver-56656f9798-m6dwv" Dec 10 22:51:50 crc kubenswrapper[4791]: I1210 22:51:50.933324 4791 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/3d53f039-46f2-478b-b14c-11b723b7dbc1-config\") pod \"console-operator-58897d9998-hd6zl\" (UID: \"3d53f039-46f2-478b-b14c-11b723b7dbc1\") " pod="openshift-console-operator/console-operator-58897d9998-hd6zl" Dec 10 22:51:50 crc kubenswrapper[4791]: I1210 22:51:50.933359 4791 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/e92ca7e3-d360-4e14-aeb9-76738a44c391-metrics-tls\") pod \"ingress-operator-5b745b69d9-8npst\" (UID: \"e92ca7e3-d360-4e14-aeb9-76738a44c391\") " pod="openshift-ingress-operator/ingress-operator-5b745b69d9-8npst" Dec 10 22:51:50 crc kubenswrapper[4791]: I1210 22:51:50.933399 4791 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/5369b8df-3c36-4108-b4e7-cfb11478b3a7-client-ca\") pod \"route-controller-manager-6576b87f9c-vsnsh\" (UID: \"5369b8df-3c36-4108-b4e7-cfb11478b3a7\") " pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-vsnsh" Dec 10 22:51:50 crc kubenswrapper[4791]: I1210 22:51:50.933426 4791 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"console-oauth-config\" (UniqueName: \"kubernetes.io/secret/1b61a973-db5e-4aa7-96e3-7fbdd9dd8cc0-console-oauth-config\") pod \"console-f9d7485db-zkfkv\" (UID: \"1b61a973-db5e-4aa7-96e3-7fbdd9dd8cc0\") " pod="openshift-console/console-f9d7485db-zkfkv" Dec 10 22:51:50 crc kubenswrapper[4791]: I1210 22:51:50.933450 4791 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-dlncw\" (UniqueName: \"kubernetes.io/projected/35033f1a-b91b-4e02-ba77-3841a2e0fdab-kube-api-access-dlncw\") pod \"oauth-openshift-558db77b4-ht5t7\" (UID: \"35033f1a-b91b-4e02-ba77-3841a2e0fdab\") " pod="openshift-authentication/oauth-openshift-558db77b4-ht5t7" Dec 10 22:51:50 crc kubenswrapper[4791]: I1210 22:51:50.933475 4791 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"image-registry-operator-tls\" (UniqueName: \"kubernetes.io/secret/33c30ace-bd17-4a3b-a43a-ddfeccb88349-image-registry-operator-tls\") pod \"cluster-image-registry-operator-dc59b4c8b-q4cpx\" (UID: \"33c30ace-bd17-4a3b-a43a-ddfeccb88349\") " pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-q4cpx" Dec 10 22:51:50 crc kubenswrapper[4791]: I1210 22:51:50.933504 4791 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-jzkqn\" (UniqueName: \"kubernetes.io/projected/3d53f039-46f2-478b-b14c-11b723b7dbc1-kube-api-access-jzkqn\") pod \"console-operator-58897d9998-hd6zl\" (UID: \"3d53f039-46f2-478b-b14c-11b723b7dbc1\") " pod="openshift-console-operator/console-operator-58897d9998-hd6zl" Dec 10 22:51:50 crc kubenswrapper[4791]: I1210 22:51:50.933528 4791 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/5369b8df-3c36-4108-b4e7-cfb11478b3a7-config\") pod \"route-controller-manager-6576b87f9c-vsnsh\" (UID: \"5369b8df-3c36-4108-b4e7-cfb11478b3a7\") " pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-vsnsh" Dec 10 22:51:50 crc kubenswrapper[4791]: I1210 22:51:50.933552 4791 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-mtfjz\" (UniqueName: \"kubernetes.io/projected/e92ca7e3-d360-4e14-aeb9-76738a44c391-kube-api-access-mtfjz\") pod \"ingress-operator-5b745b69d9-8npst\" (UID: \"e92ca7e3-d360-4e14-aeb9-76738a44c391\") " pod="openshift-ingress-operator/ingress-operator-5b745b69d9-8npst" Dec 10 22:51:50 crc kubenswrapper[4791]: I1210 22:51:50.933576 4791 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/cff60894-8deb-4354-9ad1-3a83baa98714-trusted-ca-bundle\") pod \"authentication-operator-69f744f599-7rhk9\" (UID: \"cff60894-8deb-4354-9ad1-3a83baa98714\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-7rhk9" Dec 10 22:51:50 crc kubenswrapper[4791]: I1210 22:51:50.933600 4791 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/35033f1a-b91b-4e02-ba77-3841a2e0fdab-audit-dir\") pod \"oauth-openshift-558db77b4-ht5t7\" (UID: \"35033f1a-b91b-4e02-ba77-3841a2e0fdab\") " pod="openshift-authentication/oauth-openshift-558db77b4-ht5t7" Dec 10 22:51:50 crc kubenswrapper[4791]: I1210 22:51:50.933622 4791 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/1b61a973-db5e-4aa7-96e3-7fbdd9dd8cc0-service-ca\") pod \"console-f9d7485db-zkfkv\" (UID: \"1b61a973-db5e-4aa7-96e3-7fbdd9dd8cc0\") " pod="openshift-console/console-f9d7485db-zkfkv" Dec 10 22:51:50 crc kubenswrapper[4791]: I1210 22:51:50.933950 4791 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-f2tld\" (UniqueName: \"kubernetes.io/projected/1b61a973-db5e-4aa7-96e3-7fbdd9dd8cc0-kube-api-access-f2tld\") pod \"console-f9d7485db-zkfkv\" (UID: \"1b61a973-db5e-4aa7-96e3-7fbdd9dd8cc0\") " pod="openshift-console/console-f9d7485db-zkfkv" Dec 10 22:51:50 crc kubenswrapper[4791]: I1210 22:51:50.933980 4791 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/e92ca7e3-d360-4e14-aeb9-76738a44c391-bound-sa-token\") pod \"ingress-operator-5b745b69d9-8npst\" (UID: \"e92ca7e3-d360-4e14-aeb9-76738a44c391\") " pod="openshift-ingress-operator/ingress-operator-5b745b69d9-8npst" Dec 10 22:51:50 crc kubenswrapper[4791]: I1210 22:51:50.934002 4791 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/3d53f039-46f2-478b-b14c-11b723b7dbc1-serving-cert\") pod \"console-operator-58897d9998-hd6zl\" (UID: \"3d53f039-46f2-478b-b14c-11b723b7dbc1\") " pod="openshift-console-operator/console-operator-58897d9998-hd6zl" Dec 10 22:51:50 crc kubenswrapper[4791]: I1210 22:51:50.934100 4791 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"machine-approver-tls\" (UniqueName: \"kubernetes.io/secret/2f422ac4-b586-4056-8748-94b7de0b0248-machine-approver-tls\") pod \"machine-approver-56656f9798-m6dwv\" (UID: \"2f422ac4-b586-4056-8748-94b7de0b0248\") " pod="openshift-cluster-machine-approver/machine-approver-56656f9798-m6dwv" Dec 10 22:51:50 crc kubenswrapper[4791]: I1210 22:51:50.934124 4791 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"srv-cert\" (UniqueName: \"kubernetes.io/secret/8f0772d5-db77-42b2-b712-0293c0422161-srv-cert\") pod \"catalog-operator-68c6474976-497m8\" (UID: \"8f0772d5-db77-42b2-b712-0293c0422161\") " pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-497m8" Dec 10 22:51:50 crc kubenswrapper[4791]: I1210 22:51:50.934147 4791 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-w87xw\" (UniqueName: \"kubernetes.io/projected/2f422ac4-b586-4056-8748-94b7de0b0248-kube-api-access-w87xw\") pod \"machine-approver-56656f9798-m6dwv\" (UID: \"2f422ac4-b586-4056-8748-94b7de0b0248\") " pod="openshift-cluster-machine-approver/machine-approver-56656f9798-m6dwv" Dec 10 22:51:50 crc kubenswrapper[4791]: I1210 22:51:50.934167 4791 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-7m4mz\" (UniqueName: \"kubernetes.io/projected/cff60894-8deb-4354-9ad1-3a83baa98714-kube-api-access-7m4mz\") pod \"authentication-operator-69f744f599-7rhk9\" (UID: \"cff60894-8deb-4354-9ad1-3a83baa98714\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-7rhk9" Dec 10 22:51:50 crc kubenswrapper[4791]: I1210 22:51:50.934190 4791 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-serving-cert\" (UniqueName: \"kubernetes.io/secret/35033f1a-b91b-4e02-ba77-3841a2e0fdab-v4-0-config-system-serving-cert\") pod \"oauth-openshift-558db77b4-ht5t7\" (UID: \"35033f1a-b91b-4e02-ba77-3841a2e0fdab\") " pod="openshift-authentication/oauth-openshift-558db77b4-ht5t7" Dec 10 22:51:50 crc kubenswrapper[4791]: I1210 22:51:50.934212 4791 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-user-template-login\" (UniqueName: \"kubernetes.io/secret/35033f1a-b91b-4e02-ba77-3841a2e0fdab-v4-0-config-user-template-login\") pod \"oauth-openshift-558db77b4-ht5t7\" (UID: \"35033f1a-b91b-4e02-ba77-3841a2e0fdab\") " pod="openshift-authentication/oauth-openshift-558db77b4-ht5t7" Dec 10 22:51:50 crc kubenswrapper[4791]: I1210 22:51:50.934234 4791 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"console-serving-cert\" (UniqueName: \"kubernetes.io/secret/1b61a973-db5e-4aa7-96e3-7fbdd9dd8cc0-console-serving-cert\") pod \"console-f9d7485db-zkfkv\" (UID: \"1b61a973-db5e-4aa7-96e3-7fbdd9dd8cc0\") " pod="openshift-console/console-f9d7485db-zkfkv" Dec 10 22:51:50 crc kubenswrapper[4791]: I1210 22:51:50.934257 4791 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/d3de1ba0-b014-4f98-b597-cc89f5a70364-metrics-certs\") pod \"router-default-5444994796-p97fc\" (UID: \"d3de1ba0-b014-4f98-b597-cc89f5a70364\") " pod="openshift-ingress/router-default-5444994796-p97fc" Dec 10 22:51:50 crc kubenswrapper[4791]: I1210 22:51:50.934278 4791 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"mcc-auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/e2b778bb-d80f-49f9-804b-0550c648a02f-mcc-auth-proxy-config\") pod \"machine-config-controller-84d6567774-sc74x\" (UID: \"e2b778bb-d80f-49f9-804b-0550c648a02f\") " pod="openshift-machine-config-operator/machine-config-controller-84d6567774-sc74x" Dec 10 22:51:50 crc kubenswrapper[4791]: I1210 22:51:50.934280 4791 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"available-featuregates\" (UniqueName: \"kubernetes.io/empty-dir/b419e5d6-d6a4-4258-a1bd-a462bd7f7a12-available-featuregates\") pod \"openshift-config-operator-7777fb866f-wwpsp\" (UID: \"b419e5d6-d6a4-4258-a1bd-a462bd7f7a12\") " pod="openshift-config-operator/openshift-config-operator-7777fb866f-wwpsp" Dec 10 22:51:50 crc kubenswrapper[4791]: I1210 22:51:50.934301 4791 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"console-config\" (UniqueName: \"kubernetes.io/configmap/1b61a973-db5e-4aa7-96e3-7fbdd9dd8cc0-console-config\") pod \"console-f9d7485db-zkfkv\" (UID: \"1b61a973-db5e-4aa7-96e3-7fbdd9dd8cc0\") " pod="openshift-console/console-f9d7485db-zkfkv" Dec 10 22:51:50 crc kubenswrapper[4791]: I1210 22:51:50.934327 4791 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-tss7t\" (UniqueName: \"kubernetes.io/projected/680d2296-dd48-4786-bea7-64c32edc94c7-kube-api-access-tss7t\") pod \"downloads-7954f5f757-mlb9p\" (UID: \"680d2296-dd48-4786-bea7-64c32edc94c7\") " pod="openshift-console/downloads-7954f5f757-mlb9p" Dec 10 22:51:50 crc kubenswrapper[4791]: I1210 22:51:50.934378 4791 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/465ba82f-1142-4cf1-b2ec-1f73e565ec4d-config\") pod \"openshift-controller-manager-operator-756b6f6bc6-5t9vv\" (UID: \"465ba82f-1142-4cf1-b2ec-1f73e565ec4d\") " pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-5t9vv" Dec 10 22:51:50 crc kubenswrapper[4791]: I1210 22:51:50.934383 4791 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/35033f1a-b91b-4e02-ba77-3841a2e0fdab-v4-0-config-system-trusted-ca-bundle\") pod \"oauth-openshift-558db77b4-ht5t7\" (UID: \"35033f1a-b91b-4e02-ba77-3841a2e0fdab\") " pod="openshift-authentication/oauth-openshift-558db77b4-ht5t7" Dec 10 22:51:50 crc kubenswrapper[4791]: I1210 22:51:50.934485 4791 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"profile-collector-cert\" (UniqueName: \"kubernetes.io/secret/8f0772d5-db77-42b2-b712-0293c0422161-profile-collector-cert\") pod \"catalog-operator-68c6474976-497m8\" (UID: \"8f0772d5-db77-42b2-b712-0293c0422161\") " pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-497m8" Dec 10 22:51:50 crc kubenswrapper[4791]: I1210 22:51:50.934575 4791 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/5be73604-c6c9-4ef6-a53d-6c3f6df27202-config\") pod \"kube-controller-manager-operator-78b949d7b-mldvd\" (UID: \"5be73604-c6c9-4ef6-a53d-6c3f6df27202\") " pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-mldvd" Dec 10 22:51:50 crc kubenswrapper[4791]: I1210 22:51:50.934722 4791 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-user-idp-0-file-data\" (UniqueName: \"kubernetes.io/secret/35033f1a-b91b-4e02-ba77-3841a2e0fdab-v4-0-config-user-idp-0-file-data\") pod \"oauth-openshift-558db77b4-ht5t7\" (UID: \"35033f1a-b91b-4e02-ba77-3841a2e0fdab\") " pod="openshift-authentication/oauth-openshift-558db77b4-ht5t7" Dec 10 22:51:50 crc kubenswrapper[4791]: I1210 22:51:50.934749 4791 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"oauth-serving-cert\" (UniqueName: \"kubernetes.io/configmap/1b61a973-db5e-4aa7-96e3-7fbdd9dd8cc0-oauth-serving-cert\") pod \"console-f9d7485db-zkfkv\" (UID: \"1b61a973-db5e-4aa7-96e3-7fbdd9dd8cc0\") " pod="openshift-console/console-f9d7485db-zkfkv" Dec 10 22:51:50 crc kubenswrapper[4791]: I1210 22:51:50.934798 4791 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/1b61a973-db5e-4aa7-96e3-7fbdd9dd8cc0-trusted-ca-bundle\") pod \"console-f9d7485db-zkfkv\" (UID: \"1b61a973-db5e-4aa7-96e3-7fbdd9dd8cc0\") " pod="openshift-console/console-f9d7485db-zkfkv" Dec 10 22:51:50 crc kubenswrapper[4791]: I1210 22:51:50.934820 4791 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/35033f1a-b91b-4e02-ba77-3841a2e0fdab-audit-policies\") pod \"oauth-openshift-558db77b4-ht5t7\" (UID: \"35033f1a-b91b-4e02-ba77-3841a2e0fdab\") " pod="openshift-authentication/oauth-openshift-558db77b4-ht5t7" Dec 10 22:51:50 crc kubenswrapper[4791]: I1210 22:51:50.934882 4791 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-user-template-provider-selection\" (UniqueName: \"kubernetes.io/secret/35033f1a-b91b-4e02-ba77-3841a2e0fdab-v4-0-config-user-template-provider-selection\") pod \"oauth-openshift-558db77b4-ht5t7\" (UID: \"35033f1a-b91b-4e02-ba77-3841a2e0fdab\") " pod="openshift-authentication/oauth-openshift-558db77b4-ht5t7" Dec 10 22:51:50 crc kubenswrapper[4791]: I1210 22:51:50.934972 4791 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-user-template-error\" (UniqueName: \"kubernetes.io/secret/35033f1a-b91b-4e02-ba77-3841a2e0fdab-v4-0-config-user-template-error\") pod \"oauth-openshift-558db77b4-ht5t7\" (UID: \"35033f1a-b91b-4e02-ba77-3841a2e0fdab\") " pod="openshift-authentication/oauth-openshift-558db77b4-ht5t7" Dec 10 22:51:50 crc kubenswrapper[4791]: I1210 22:51:50.934997 4791 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-fltgv\" (UniqueName: \"kubernetes.io/projected/4205f348-3276-40f3-a35b-69306b331d80-kube-api-access-fltgv\") pod \"cluster-samples-operator-665b6dd947-t7742\" (UID: \"4205f348-3276-40f3-a35b-69306b331d80\") " pod="openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-t7742" Dec 10 22:51:50 crc kubenswrapper[4791]: I1210 22:51:50.935050 4791 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-cliconfig\" (UniqueName: \"kubernetes.io/configmap/35033f1a-b91b-4e02-ba77-3841a2e0fdab-v4-0-config-system-cliconfig\") pod \"oauth-openshift-558db77b4-ht5t7\" (UID: \"35033f1a-b91b-4e02-ba77-3841a2e0fdab\") " pod="openshift-authentication/oauth-openshift-558db77b4-ht5t7" Dec 10 22:51:50 crc kubenswrapper[4791]: I1210 22:51:50.935079 4791 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/cff60894-8deb-4354-9ad1-3a83baa98714-serving-cert\") pod \"authentication-operator-69f744f599-7rhk9\" (UID: \"cff60894-8deb-4354-9ad1-3a83baa98714\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-7rhk9" Dec 10 22:51:50 crc kubenswrapper[4791]: I1210 22:51:50.935140 4791 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-vwnrt\" (UniqueName: \"kubernetes.io/projected/e2b778bb-d80f-49f9-804b-0550c648a02f-kube-api-access-vwnrt\") pod \"machine-config-controller-84d6567774-sc74x\" (UID: \"e2b778bb-d80f-49f9-804b-0550c648a02f\") " pod="openshift-machine-config-operator/machine-config-controller-84d6567774-sc74x" Dec 10 22:51:50 crc kubenswrapper[4791]: I1210 22:51:50.935165 4791 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/b419e5d6-d6a4-4258-a1bd-a462bd7f7a12-serving-cert\") pod \"openshift-config-operator-7777fb866f-wwpsp\" (UID: \"b419e5d6-d6a4-4258-a1bd-a462bd7f7a12\") " pod="openshift-config-operator/openshift-config-operator-7777fb866f-wwpsp" Dec 10 22:51:50 crc kubenswrapper[4791]: I1210 22:51:50.935223 4791 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/33c30ace-bd17-4a3b-a43a-ddfeccb88349-trusted-ca\") pod \"cluster-image-registry-operator-dc59b4c8b-q4cpx\" (UID: \"33c30ace-bd17-4a3b-a43a-ddfeccb88349\") " pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-q4cpx" Dec 10 22:51:50 crc kubenswrapper[4791]: I1210 22:51:50.935249 4791 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/e92ca7e3-d360-4e14-aeb9-76738a44c391-trusted-ca\") pod \"ingress-operator-5b745b69d9-8npst\" (UID: \"e92ca7e3-d360-4e14-aeb9-76738a44c391\") " pod="openshift-ingress-operator/ingress-operator-5b745b69d9-8npst" Dec 10 22:51:50 crc kubenswrapper[4791]: I1210 22:51:50.935299 4791 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-z5dwf\" (UniqueName: \"kubernetes.io/projected/5369b8df-3c36-4108-b4e7-cfb11478b3a7-kube-api-access-z5dwf\") pod \"route-controller-manager-6576b87f9c-vsnsh\" (UID: \"5369b8df-3c36-4108-b4e7-cfb11478b3a7\") " pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-vsnsh" Dec 10 22:51:50 crc kubenswrapper[4791]: I1210 22:51:50.935326 4791 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/2f422ac4-b586-4056-8748-94b7de0b0248-auth-proxy-config\") pod \"machine-approver-56656f9798-m6dwv\" (UID: \"2f422ac4-b586-4056-8748-94b7de0b0248\") " pod="openshift-cluster-machine-approver/machine-approver-56656f9798-m6dwv" Dec 10 22:51:50 crc kubenswrapper[4791]: I1210 22:51:50.935369 4791 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-ocp-branding-template\" (UniqueName: \"kubernetes.io/secret/35033f1a-b91b-4e02-ba77-3841a2e0fdab-v4-0-config-system-ocp-branding-template\") pod \"oauth-openshift-558db77b4-ht5t7\" (UID: \"35033f1a-b91b-4e02-ba77-3841a2e0fdab\") " pod="openshift-authentication/oauth-openshift-558db77b4-ht5t7" Dec 10 22:51:50 crc kubenswrapper[4791]: I1210 22:51:50.935394 4791 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-vn4qx\" (UniqueName: \"kubernetes.io/projected/d3de1ba0-b014-4f98-b597-cc89f5a70364-kube-api-access-vn4qx\") pod \"router-default-5444994796-p97fc\" (UID: \"d3de1ba0-b014-4f98-b597-cc89f5a70364\") " pod="openshift-ingress/router-default-5444994796-p97fc" Dec 10 22:51:50 crc kubenswrapper[4791]: I1210 22:51:50.935460 4791 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/5be73604-c6c9-4ef6-a53d-6c3f6df27202-kube-api-access\") pod \"kube-controller-manager-operator-78b949d7b-mldvd\" (UID: \"5be73604-c6c9-4ef6-a53d-6c3f6df27202\") " pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-mldvd" Dec 10 22:51:50 crc kubenswrapper[4791]: I1210 22:51:50.935491 4791 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/3d53f039-46f2-478b-b14c-11b723b7dbc1-trusted-ca\") pod \"console-operator-58897d9998-hd6zl\" (UID: \"3d53f039-46f2-478b-b14c-11b723b7dbc1\") " pod="openshift-console-operator/console-operator-58897d9998-hd6zl" Dec 10 22:51:50 crc kubenswrapper[4791]: I1210 22:51:50.935513 4791 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"service-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/cff60894-8deb-4354-9ad1-3a83baa98714-service-ca-bundle\") pod \"authentication-operator-69f744f599-7rhk9\" (UID: \"cff60894-8deb-4354-9ad1-3a83baa98714\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-7rhk9" Dec 10 22:51:50 crc kubenswrapper[4791]: I1210 22:51:50.935518 4791 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/5be73604-c6c9-4ef6-a53d-6c3f6df27202-config\") pod \"kube-controller-manager-operator-78b949d7b-mldvd\" (UID: \"5be73604-c6c9-4ef6-a53d-6c3f6df27202\") " pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-mldvd" Dec 10 22:51:50 crc kubenswrapper[4791]: I1210 22:51:50.935538 4791 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/5be73604-c6c9-4ef6-a53d-6c3f6df27202-serving-cert\") pod \"kube-controller-manager-operator-78b949d7b-mldvd\" (UID: \"5be73604-c6c9-4ef6-a53d-6c3f6df27202\") " pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-mldvd" Dec 10 22:51:50 crc kubenswrapper[4791]: I1210 22:51:50.935587 4791 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"samples-operator-tls\" (UniqueName: \"kubernetes.io/secret/4205f348-3276-40f3-a35b-69306b331d80-samples-operator-tls\") pod \"cluster-samples-operator-665b6dd947-t7742\" (UID: \"4205f348-3276-40f3-a35b-69306b331d80\") " pod="openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-t7742" Dec 10 22:51:50 crc kubenswrapper[4791]: I1210 22:51:50.936402 4791 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/5369b8df-3c36-4108-b4e7-cfb11478b3a7-serving-cert\") pod \"route-controller-manager-6576b87f9c-vsnsh\" (UID: \"5369b8df-3c36-4108-b4e7-cfb11478b3a7\") " pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-vsnsh" Dec 10 22:51:50 crc kubenswrapper[4791]: I1210 22:51:50.937579 4791 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/5369b8df-3c36-4108-b4e7-cfb11478b3a7-config\") pod \"route-controller-manager-6576b87f9c-vsnsh\" (UID: \"5369b8df-3c36-4108-b4e7-cfb11478b3a7\") " pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-vsnsh" Dec 10 22:51:50 crc kubenswrapper[4791]: I1210 22:51:50.937725 4791 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/5369b8df-3c36-4108-b4e7-cfb11478b3a7-client-ca\") pod \"route-controller-manager-6576b87f9c-vsnsh\" (UID: \"5369b8df-3c36-4108-b4e7-cfb11478b3a7\") " pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-vsnsh" Dec 10 22:51:50 crc kubenswrapper[4791]: I1210 22:51:50.937939 4791 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/2f422ac4-b586-4056-8748-94b7de0b0248-config\") pod \"machine-approver-56656f9798-m6dwv\" (UID: \"2f422ac4-b586-4056-8748-94b7de0b0248\") " pod="openshift-cluster-machine-approver/machine-approver-56656f9798-m6dwv" Dec 10 22:51:50 crc kubenswrapper[4791]: I1210 22:51:50.938039 4791 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-operator-lifecycle-manager"/"openshift-service-ca.crt" Dec 10 22:51:50 crc kubenswrapper[4791]: I1210 22:51:50.938142 4791 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/cff60894-8deb-4354-9ad1-3a83baa98714-config\") pod \"authentication-operator-69f744f599-7rhk9\" (UID: \"cff60894-8deb-4354-9ad1-3a83baa98714\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-7rhk9" Dec 10 22:51:50 crc kubenswrapper[4791]: I1210 22:51:50.938291 4791 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/33c30ace-bd17-4a3b-a43a-ddfeccb88349-trusted-ca\") pod \"cluster-image-registry-operator-dc59b4c8b-q4cpx\" (UID: \"33c30ace-bd17-4a3b-a43a-ddfeccb88349\") " pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-q4cpx" Dec 10 22:51:50 crc kubenswrapper[4791]: I1210 22:51:50.938848 4791 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/cff60894-8deb-4354-9ad1-3a83baa98714-trusted-ca-bundle\") pod \"authentication-operator-69f744f599-7rhk9\" (UID: \"cff60894-8deb-4354-9ad1-3a83baa98714\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-7rhk9" Dec 10 22:51:50 crc kubenswrapper[4791]: I1210 22:51:50.935538 4791 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/35033f1a-b91b-4e02-ba77-3841a2e0fdab-v4-0-config-system-trusted-ca-bundle\") pod \"oauth-openshift-558db77b4-ht5t7\" (UID: \"35033f1a-b91b-4e02-ba77-3841a2e0fdab\") " pod="openshift-authentication/oauth-openshift-558db77b4-ht5t7" Dec 10 22:51:50 crc kubenswrapper[4791]: I1210 22:51:50.939071 4791 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/2f422ac4-b586-4056-8748-94b7de0b0248-auth-proxy-config\") pod \"machine-approver-56656f9798-m6dwv\" (UID: \"2f422ac4-b586-4056-8748-94b7de0b0248\") " pod="openshift-cluster-machine-approver/machine-approver-56656f9798-m6dwv" Dec 10 22:51:50 crc kubenswrapper[4791]: I1210 22:51:50.939167 4791 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/e92ca7e3-d360-4e14-aeb9-76738a44c391-trusted-ca\") pod \"ingress-operator-5b745b69d9-8npst\" (UID: \"e92ca7e3-d360-4e14-aeb9-76738a44c391\") " pod="openshift-ingress-operator/ingress-operator-5b745b69d9-8npst" Dec 10 22:51:50 crc kubenswrapper[4791]: I1210 22:51:50.939545 4791 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"mcc-auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/e2b778bb-d80f-49f9-804b-0550c648a02f-mcc-auth-proxy-config\") pod \"machine-config-controller-84d6567774-sc74x\" (UID: \"e2b778bb-d80f-49f9-804b-0550c648a02f\") " pod="openshift-machine-config-operator/machine-config-controller-84d6567774-sc74x" Dec 10 22:51:50 crc kubenswrapper[4791]: I1210 22:51:50.939613 4791 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-service-ca\" (UniqueName: \"kubernetes.io/configmap/35033f1a-b91b-4e02-ba77-3841a2e0fdab-v4-0-config-system-service-ca\") pod \"oauth-openshift-558db77b4-ht5t7\" (UID: \"35033f1a-b91b-4e02-ba77-3841a2e0fdab\") " pod="openshift-authentication/oauth-openshift-558db77b4-ht5t7" Dec 10 22:51:50 crc kubenswrapper[4791]: I1210 22:51:50.939664 4791 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-cliconfig\" (UniqueName: \"kubernetes.io/configmap/35033f1a-b91b-4e02-ba77-3841a2e0fdab-v4-0-config-system-cliconfig\") pod \"oauth-openshift-558db77b4-ht5t7\" (UID: \"35033f1a-b91b-4e02-ba77-3841a2e0fdab\") " pod="openshift-authentication/oauth-openshift-558db77b4-ht5t7" Dec 10 22:51:50 crc kubenswrapper[4791]: I1210 22:51:50.939724 4791 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"console-oauth-config\" (UniqueName: \"kubernetes.io/secret/1b61a973-db5e-4aa7-96e3-7fbdd9dd8cc0-console-oauth-config\") pod \"console-f9d7485db-zkfkv\" (UID: \"1b61a973-db5e-4aa7-96e3-7fbdd9dd8cc0\") " pod="openshift-console/console-f9d7485db-zkfkv" Dec 10 22:51:50 crc kubenswrapper[4791]: I1210 22:51:50.939747 4791 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/3d53f039-46f2-478b-b14c-11b723b7dbc1-trusted-ca\") pod \"console-operator-58897d9998-hd6zl\" (UID: \"3d53f039-46f2-478b-b14c-11b723b7dbc1\") " pod="openshift-console-operator/console-operator-58897d9998-hd6zl" Dec 10 22:51:50 crc kubenswrapper[4791]: I1210 22:51:50.939902 4791 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"oauth-serving-cert\" (UniqueName: \"kubernetes.io/configmap/1b61a973-db5e-4aa7-96e3-7fbdd9dd8cc0-oauth-serving-cert\") pod \"console-f9d7485db-zkfkv\" (UID: \"1b61a973-db5e-4aa7-96e3-7fbdd9dd8cc0\") " pod="openshift-console/console-f9d7485db-zkfkv" Dec 10 22:51:50 crc kubenswrapper[4791]: I1210 22:51:50.939906 4791 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/35033f1a-b91b-4e02-ba77-3841a2e0fdab-audit-dir\") pod \"oauth-openshift-558db77b4-ht5t7\" (UID: \"35033f1a-b91b-4e02-ba77-3841a2e0fdab\") " pod="openshift-authentication/oauth-openshift-558db77b4-ht5t7" Dec 10 22:51:50 crc kubenswrapper[4791]: I1210 22:51:50.940137 4791 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"console-config\" (UniqueName: \"kubernetes.io/configmap/1b61a973-db5e-4aa7-96e3-7fbdd9dd8cc0-console-config\") pod \"console-f9d7485db-zkfkv\" (UID: \"1b61a973-db5e-4aa7-96e3-7fbdd9dd8cc0\") " pod="openshift-console/console-f9d7485db-zkfkv" Dec 10 22:51:50 crc kubenswrapper[4791]: I1210 22:51:50.940142 4791 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/3d53f039-46f2-478b-b14c-11b723b7dbc1-config\") pod \"console-operator-58897d9998-hd6zl\" (UID: \"3d53f039-46f2-478b-b14c-11b723b7dbc1\") " pod="openshift-console-operator/console-operator-58897d9998-hd6zl" Dec 10 22:51:50 crc kubenswrapper[4791]: I1210 22:51:50.940215 4791 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/35033f1a-b91b-4e02-ba77-3841a2e0fdab-audit-policies\") pod \"oauth-openshift-558db77b4-ht5t7\" (UID: \"35033f1a-b91b-4e02-ba77-3841a2e0fdab\") " pod="openshift-authentication/oauth-openshift-558db77b4-ht5t7" Dec 10 22:51:50 crc kubenswrapper[4791]: I1210 22:51:50.940384 4791 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/1b61a973-db5e-4aa7-96e3-7fbdd9dd8cc0-service-ca\") pod \"console-f9d7485db-zkfkv\" (UID: \"1b61a973-db5e-4aa7-96e3-7fbdd9dd8cc0\") " pod="openshift-console/console-f9d7485db-zkfkv" Dec 10 22:51:50 crc kubenswrapper[4791]: I1210 22:51:50.940668 4791 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-user-template-error\" (UniqueName: \"kubernetes.io/secret/35033f1a-b91b-4e02-ba77-3841a2e0fdab-v4-0-config-user-template-error\") pod \"oauth-openshift-558db77b4-ht5t7\" (UID: \"35033f1a-b91b-4e02-ba77-3841a2e0fdab\") " pod="openshift-authentication/oauth-openshift-558db77b4-ht5t7" Dec 10 22:51:50 crc kubenswrapper[4791]: I1210 22:51:50.941163 4791 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"samples-operator-tls\" (UniqueName: \"kubernetes.io/secret/4205f348-3276-40f3-a35b-69306b331d80-samples-operator-tls\") pod \"cluster-samples-operator-665b6dd947-t7742\" (UID: \"4205f348-3276-40f3-a35b-69306b331d80\") " pod="openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-t7742" Dec 10 22:51:50 crc kubenswrapper[4791]: I1210 22:51:50.941512 4791 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-serving-cert\" (UniqueName: \"kubernetes.io/secret/35033f1a-b91b-4e02-ba77-3841a2e0fdab-v4-0-config-system-serving-cert\") pod \"oauth-openshift-558db77b4-ht5t7\" (UID: \"35033f1a-b91b-4e02-ba77-3841a2e0fdab\") " pod="openshift-authentication/oauth-openshift-558db77b4-ht5t7" Dec 10 22:51:50 crc kubenswrapper[4791]: I1210 22:51:50.942004 4791 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-router-certs\" (UniqueName: \"kubernetes.io/secret/35033f1a-b91b-4e02-ba77-3841a2e0fdab-v4-0-config-system-router-certs\") pod \"oauth-openshift-558db77b4-ht5t7\" (UID: \"35033f1a-b91b-4e02-ba77-3841a2e0fdab\") " pod="openshift-authentication/oauth-openshift-558db77b4-ht5t7" Dec 10 22:51:50 crc kubenswrapper[4791]: I1210 22:51:50.942014 4791 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"service-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/cff60894-8deb-4354-9ad1-3a83baa98714-service-ca-bundle\") pod \"authentication-operator-69f744f599-7rhk9\" (UID: \"cff60894-8deb-4354-9ad1-3a83baa98714\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-7rhk9" Dec 10 22:51:50 crc kubenswrapper[4791]: I1210 22:51:50.942645 4791 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-user-idp-0-file-data\" (UniqueName: \"kubernetes.io/secret/35033f1a-b91b-4e02-ba77-3841a2e0fdab-v4-0-config-user-idp-0-file-data\") pod \"oauth-openshift-558db77b4-ht5t7\" (UID: \"35033f1a-b91b-4e02-ba77-3841a2e0fdab\") " pod="openshift-authentication/oauth-openshift-558db77b4-ht5t7" Dec 10 22:51:50 crc kubenswrapper[4791]: I1210 22:51:50.942907 4791 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"console-serving-cert\" (UniqueName: \"kubernetes.io/secret/1b61a973-db5e-4aa7-96e3-7fbdd9dd8cc0-console-serving-cert\") pod \"console-f9d7485db-zkfkv\" (UID: \"1b61a973-db5e-4aa7-96e3-7fbdd9dd8cc0\") " pod="openshift-console/console-f9d7485db-zkfkv" Dec 10 22:51:50 crc kubenswrapper[4791]: I1210 22:51:50.942968 4791 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-session\" (UniqueName: \"kubernetes.io/secret/35033f1a-b91b-4e02-ba77-3841a2e0fdab-v4-0-config-system-session\") pod \"oauth-openshift-558db77b4-ht5t7\" (UID: \"35033f1a-b91b-4e02-ba77-3841a2e0fdab\") " pod="openshift-authentication/oauth-openshift-558db77b4-ht5t7" Dec 10 22:51:50 crc kubenswrapper[4791]: I1210 22:51:50.943487 4791 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/5be73604-c6c9-4ef6-a53d-6c3f6df27202-serving-cert\") pod \"kube-controller-manager-operator-78b949d7b-mldvd\" (UID: \"5be73604-c6c9-4ef6-a53d-6c3f6df27202\") " pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-mldvd" Dec 10 22:51:50 crc kubenswrapper[4791]: I1210 22:51:50.943955 4791 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/cff60894-8deb-4354-9ad1-3a83baa98714-serving-cert\") pod \"authentication-operator-69f744f599-7rhk9\" (UID: \"cff60894-8deb-4354-9ad1-3a83baa98714\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-7rhk9" Dec 10 22:51:50 crc kubenswrapper[4791]: I1210 22:51:50.944018 4791 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-ocp-branding-template\" (UniqueName: \"kubernetes.io/secret/35033f1a-b91b-4e02-ba77-3841a2e0fdab-v4-0-config-system-ocp-branding-template\") pod \"oauth-openshift-558db77b4-ht5t7\" (UID: \"35033f1a-b91b-4e02-ba77-3841a2e0fdab\") " pod="openshift-authentication/oauth-openshift-558db77b4-ht5t7" Dec 10 22:51:50 crc kubenswrapper[4791]: I1210 22:51:50.944136 4791 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/3d53f039-46f2-478b-b14c-11b723b7dbc1-serving-cert\") pod \"console-operator-58897d9998-hd6zl\" (UID: \"3d53f039-46f2-478b-b14c-11b723b7dbc1\") " pod="openshift-console-operator/console-operator-58897d9998-hd6zl" Dec 10 22:51:50 crc kubenswrapper[4791]: I1210 22:51:50.944426 4791 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"machine-approver-tls\" (UniqueName: \"kubernetes.io/secret/2f422ac4-b586-4056-8748-94b7de0b0248-machine-approver-tls\") pod \"machine-approver-56656f9798-m6dwv\" (UID: \"2f422ac4-b586-4056-8748-94b7de0b0248\") " pod="openshift-cluster-machine-approver/machine-approver-56656f9798-m6dwv" Dec 10 22:51:50 crc kubenswrapper[4791]: I1210 22:51:50.944772 4791 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"image-registry-operator-tls\" (UniqueName: \"kubernetes.io/secret/33c30ace-bd17-4a3b-a43a-ddfeccb88349-image-registry-operator-tls\") pod \"cluster-image-registry-operator-dc59b4c8b-q4cpx\" (UID: \"33c30ace-bd17-4a3b-a43a-ddfeccb88349\") " pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-q4cpx" Dec 10 22:51:50 crc kubenswrapper[4791]: I1210 22:51:50.945236 4791 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-user-template-login\" (UniqueName: \"kubernetes.io/secret/35033f1a-b91b-4e02-ba77-3841a2e0fdab-v4-0-config-user-template-login\") pod \"oauth-openshift-558db77b4-ht5t7\" (UID: \"35033f1a-b91b-4e02-ba77-3841a2e0fdab\") " pod="openshift-authentication/oauth-openshift-558db77b4-ht5t7" Dec 10 22:51:50 crc kubenswrapper[4791]: I1210 22:51:50.945479 4791 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/b419e5d6-d6a4-4258-a1bd-a462bd7f7a12-serving-cert\") pod \"openshift-config-operator-7777fb866f-wwpsp\" (UID: \"b419e5d6-d6a4-4258-a1bd-a462bd7f7a12\") " pod="openshift-config-operator/openshift-config-operator-7777fb866f-wwpsp" Dec 10 22:51:50 crc kubenswrapper[4791]: I1210 22:51:50.946459 4791 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-user-template-provider-selection\" (UniqueName: \"kubernetes.io/secret/35033f1a-b91b-4e02-ba77-3841a2e0fdab-v4-0-config-user-template-provider-selection\") pod \"oauth-openshift-558db77b4-ht5t7\" (UID: \"35033f1a-b91b-4e02-ba77-3841a2e0fdab\") " pod="openshift-authentication/oauth-openshift-558db77b4-ht5t7" Dec 10 22:51:50 crc kubenswrapper[4791]: I1210 22:51:50.946468 4791 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/465ba82f-1142-4cf1-b2ec-1f73e565ec4d-serving-cert\") pod \"openshift-controller-manager-operator-756b6f6bc6-5t9vv\" (UID: \"465ba82f-1142-4cf1-b2ec-1f73e565ec4d\") " pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-5t9vv" Dec 10 22:51:50 crc kubenswrapper[4791]: I1210 22:51:50.948691 4791 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/e92ca7e3-d360-4e14-aeb9-76738a44c391-metrics-tls\") pod \"ingress-operator-5b745b69d9-8npst\" (UID: \"e92ca7e3-d360-4e14-aeb9-76738a44c391\") " pod="openshift-ingress-operator/ingress-operator-5b745b69d9-8npst" Dec 10 22:51:50 crc kubenswrapper[4791]: I1210 22:51:50.956994 4791 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress"/"router-certs-default" Dec 10 22:51:50 crc kubenswrapper[4791]: I1210 22:51:50.968824 4791 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"default-certificate\" (UniqueName: \"kubernetes.io/secret/d3de1ba0-b014-4f98-b597-cc89f5a70364-default-certificate\") pod \"router-default-5444994796-p97fc\" (UID: \"d3de1ba0-b014-4f98-b597-cc89f5a70364\") " pod="openshift-ingress/router-default-5444994796-p97fc" Dec 10 22:51:50 crc kubenswrapper[4791]: I1210 22:51:50.977481 4791 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress"/"router-stats-default" Dec 10 22:51:50 crc kubenswrapper[4791]: I1210 22:51:50.982694 4791 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"stats-auth\" (UniqueName: \"kubernetes.io/secret/d3de1ba0-b014-4f98-b597-cc89f5a70364-stats-auth\") pod \"router-default-5444994796-p97fc\" (UID: \"d3de1ba0-b014-4f98-b597-cc89f5a70364\") " pod="openshift-ingress/router-default-5444994796-p97fc" Dec 10 22:51:51 crc kubenswrapper[4791]: I1210 22:51:51.014499 4791 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress"/"router-metrics-certs-default" Dec 10 22:51:51 crc kubenswrapper[4791]: I1210 22:51:51.016487 4791 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress"/"router-dockercfg-zdk86" Dec 10 22:51:51 crc kubenswrapper[4791]: I1210 22:51:51.021760 4791 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/d3de1ba0-b014-4f98-b597-cc89f5a70364-metrics-certs\") pod \"router-default-5444994796-p97fc\" (UID: \"d3de1ba0-b014-4f98-b597-cc89f5a70364\") " pod="openshift-ingress/router-default-5444994796-p97fc" Dec 10 22:51:51 crc kubenswrapper[4791]: I1210 22:51:51.036697 4791 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress"/"service-ca-bundle" Dec 10 22:51:51 crc kubenswrapper[4791]: I1210 22:51:51.046137 4791 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"service-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/d3de1ba0-b014-4f98-b597-cc89f5a70364-service-ca-bundle\") pod \"router-default-5444994796-p97fc\" (UID: \"d3de1ba0-b014-4f98-b597-cc89f5a70364\") " pod="openshift-ingress/router-default-5444994796-p97fc" Dec 10 22:51:51 crc kubenswrapper[4791]: I1210 22:51:51.056970 4791 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress"/"kube-root-ca.crt" Dec 10 22:51:51 crc kubenswrapper[4791]: I1210 22:51:51.077173 4791 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"catalog-operator-serving-cert" Dec 10 22:51:51 crc kubenswrapper[4791]: I1210 22:51:51.089697 4791 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"srv-cert\" (UniqueName: \"kubernetes.io/secret/8f0772d5-db77-42b2-b712-0293c0422161-srv-cert\") pod \"catalog-operator-68c6474976-497m8\" (UID: \"8f0772d5-db77-42b2-b712-0293c0422161\") " pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-497m8" Dec 10 22:51:51 crc kubenswrapper[4791]: I1210 22:51:51.097158 4791 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"olm-operator-serviceaccount-dockercfg-rq7zk" Dec 10 22:51:51 crc kubenswrapper[4791]: I1210 22:51:51.116453 4791 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-operator-lifecycle-manager"/"kube-root-ca.crt" Dec 10 22:51:51 crc kubenswrapper[4791]: I1210 22:51:51.136922 4791 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"olm-operator-serving-cert" Dec 10 22:51:51 crc kubenswrapper[4791]: I1210 22:51:51.156476 4791 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"pprof-cert" Dec 10 22:51:51 crc kubenswrapper[4791]: I1210 22:51:51.169105 4791 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"profile-collector-cert\" (UniqueName: \"kubernetes.io/secret/8f0772d5-db77-42b2-b712-0293c0422161-profile-collector-cert\") pod \"catalog-operator-68c6474976-497m8\" (UID: \"8f0772d5-db77-42b2-b712-0293c0422161\") " pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-497m8" Dec 10 22:51:51 crc kubenswrapper[4791]: I1210 22:51:51.178027 4791 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"mcc-proxy-tls" Dec 10 22:51:51 crc kubenswrapper[4791]: I1210 22:51:51.186714 4791 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/e2b778bb-d80f-49f9-804b-0550c648a02f-proxy-tls\") pod \"machine-config-controller-84d6567774-sc74x\" (UID: \"e2b778bb-d80f-49f9-804b-0550c648a02f\") " pod="openshift-machine-config-operator/machine-config-controller-84d6567774-sc74x" Dec 10 22:51:51 crc kubenswrapper[4791]: I1210 22:51:51.197918 4791 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"machine-config-controller-dockercfg-c2lfx" Dec 10 22:51:51 crc kubenswrapper[4791]: I1210 22:51:51.258045 4791 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-dns-operator"/"dns-operator-dockercfg-9mqw5" Dec 10 22:51:51 crc kubenswrapper[4791]: I1210 22:51:51.278207 4791 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-dns-operator"/"kube-root-ca.crt" Dec 10 22:51:51 crc kubenswrapper[4791]: I1210 22:51:51.297961 4791 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-dns-operator"/"openshift-service-ca.crt" Dec 10 22:51:51 crc kubenswrapper[4791]: I1210 22:51:51.318103 4791 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-dns-operator"/"metrics-tls" Dec 10 22:51:51 crc kubenswrapper[4791]: I1210 22:51:51.338285 4791 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-scheduler-operator"/"kube-root-ca.crt" Dec 10 22:51:51 crc kubenswrapper[4791]: I1210 22:51:51.356964 4791 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-scheduler-operator"/"openshift-kube-scheduler-operator-dockercfg-qt55r" Dec 10 22:51:51 crc kubenswrapper[4791]: I1210 22:51:51.377216 4791 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-scheduler-operator"/"kube-scheduler-operator-serving-cert" Dec 10 22:51:51 crc kubenswrapper[4791]: I1210 22:51:51.396659 4791 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-scheduler-operator"/"openshift-kube-scheduler-operator-config" Dec 10 22:51:51 crc kubenswrapper[4791]: I1210 22:51:51.418636 4791 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver-operator"/"openshift-service-ca.crt" Dec 10 22:51:51 crc kubenswrapper[4791]: I1210 22:51:51.437193 4791 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-apiserver-operator"/"openshift-apiserver-operator-serving-cert" Dec 10 22:51:51 crc kubenswrapper[4791]: I1210 22:51:51.457074 4791 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver-operator"/"openshift-apiserver-operator-config" Dec 10 22:51:51 crc kubenswrapper[4791]: I1210 22:51:51.478928 4791 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-apiserver-operator"/"openshift-apiserver-operator-dockercfg-xtcjv" Dec 10 22:51:51 crc kubenswrapper[4791]: I1210 22:51:51.498018 4791 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver-operator"/"kube-root-ca.crt" Dec 10 22:51:51 crc kubenswrapper[4791]: I1210 22:51:51.518565 4791 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"package-server-manager-serving-cert" Dec 10 22:51:51 crc kubenswrapper[4791]: I1210 22:51:51.537763 4791 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-etcd-operator"/"openshift-service-ca.crt" Dec 10 22:51:51 crc kubenswrapper[4791]: I1210 22:51:51.558738 4791 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-etcd-operator"/"etcd-operator-config" Dec 10 22:51:51 crc kubenswrapper[4791]: I1210 22:51:51.577303 4791 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-etcd-operator"/"etcd-ca-bundle" Dec 10 22:51:51 crc kubenswrapper[4791]: I1210 22:51:51.597422 4791 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-etcd-operator"/"etcd-operator-dockercfg-r9srn" Dec 10 22:51:51 crc kubenswrapper[4791]: I1210 22:51:51.618105 4791 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-etcd-operator"/"etcd-service-ca-bundle" Dec 10 22:51:51 crc kubenswrapper[4791]: I1210 22:51:51.639452 4791 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-etcd-operator"/"etcd-client" Dec 10 22:51:51 crc kubenswrapper[4791]: I1210 22:51:51.658205 4791 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-etcd-operator"/"kube-root-ca.crt" Dec 10 22:51:51 crc kubenswrapper[4791]: I1210 22:51:51.678577 4791 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-etcd-operator"/"etcd-operator-serving-cert" Dec 10 22:51:51 crc kubenswrapper[4791]: I1210 22:51:51.698214 4791 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-apiserver-operator"/"kube-apiserver-operator-dockercfg-x57mr" Dec 10 22:51:51 crc kubenswrapper[4791]: I1210 22:51:51.716598 4791 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-apiserver-operator"/"kube-apiserver-operator-config" Dec 10 22:51:51 crc kubenswrapper[4791]: I1210 22:51:51.737257 4791 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-apiserver-operator"/"kube-apiserver-operator-serving-cert" Dec 10 22:51:51 crc kubenswrapper[4791]: I1210 22:51:51.759607 4791 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-apiserver-operator"/"kube-root-ca.crt" Dec 10 22:51:51 crc kubenswrapper[4791]: I1210 22:51:51.779971 4791 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"machine-config-operator-dockercfg-98p87" Dec 10 22:51:51 crc kubenswrapper[4791]: I1210 22:51:51.797555 4791 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"mco-proxy-tls" Dec 10 22:51:51 crc kubenswrapper[4791]: I1210 22:51:51.815633 4791 request.go:700] Waited for 1.005365908s due to client-side throttling, not priority and fairness, request: GET:https://api-int.crc.testing:6443/api/v1/namespaces/openshift-machine-config-operator/configmaps?fieldSelector=metadata.name%3Dmachine-config-operator-images&limit=500&resourceVersion=0 Dec 10 22:51:51 crc kubenswrapper[4791]: I1210 22:51:51.818316 4791 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-config-operator"/"machine-config-operator-images" Dec 10 22:51:51 crc kubenswrapper[4791]: I1210 22:51:51.838492 4791 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-storage-version-migrator"/"openshift-service-ca.crt" Dec 10 22:51:51 crc kubenswrapper[4791]: I1210 22:51:51.858627 4791 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-storage-version-migrator"/"kube-storage-version-migrator-sa-dockercfg-5xfcg" Dec 10 22:51:51 crc kubenswrapper[4791]: I1210 22:51:51.878152 4791 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-storage-version-migrator"/"kube-root-ca.crt" Dec 10 22:51:51 crc kubenswrapper[4791]: I1210 22:51:51.897817 4791 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-multus"/"multus-ac-dockercfg-9lkdf" Dec 10 22:51:51 crc kubenswrapper[4791]: I1210 22:51:51.917859 4791 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-multus"/"multus-admission-controller-secret" Dec 10 22:51:51 crc kubenswrapper[4791]: I1210 22:51:51.937872 4791 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-storage-version-migrator-operator"/"openshift-service-ca.crt" Dec 10 22:51:51 crc kubenswrapper[4791]: I1210 22:51:51.957495 4791 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-storage-version-migrator-operator"/"kube-storage-version-migrator-operator-dockercfg-2bh8d" Dec 10 22:51:51 crc kubenswrapper[4791]: I1210 22:51:51.977377 4791 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-storage-version-migrator-operator"/"config" Dec 10 22:51:51 crc kubenswrapper[4791]: I1210 22:51:51.997053 4791 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-storage-version-migrator-operator"/"kube-root-ca.crt" Dec 10 22:51:52 crc kubenswrapper[4791]: I1210 22:51:52.017127 4791 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-storage-version-migrator-operator"/"serving-cert" Dec 10 22:51:52 crc kubenswrapper[4791]: I1210 22:51:52.038942 4791 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"packageserver-service-cert" Dec 10 22:51:52 crc kubenswrapper[4791]: I1210 22:51:52.057389 4791 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-api"/"control-plane-machine-set-operator-tls" Dec 10 22:51:52 crc kubenswrapper[4791]: I1210 22:51:52.078738 4791 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-api"/"control-plane-machine-set-operator-dockercfg-k9rxt" Dec 10 22:51:52 crc kubenswrapper[4791]: I1210 22:51:52.097787 4791 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-marketplace"/"openshift-service-ca.crt" Dec 10 22:51:52 crc kubenswrapper[4791]: I1210 22:51:52.116930 4791 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"marketplace-operator-dockercfg-5nsgg" Dec 10 22:51:52 crc kubenswrapper[4791]: I1210 22:51:52.137961 4791 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"marketplace-operator-metrics" Dec 10 22:51:52 crc kubenswrapper[4791]: I1210 22:51:52.167763 4791 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-marketplace"/"marketplace-trusted-ca" Dec 10 22:51:52 crc kubenswrapper[4791]: I1210 22:51:52.176905 4791 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-marketplace"/"kube-root-ca.crt" Dec 10 22:51:52 crc kubenswrapper[4791]: I1210 22:51:52.197899 4791 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-dns"/"dns-default-metrics-tls" Dec 10 22:51:52 crc kubenswrapper[4791]: I1210 22:51:52.238010 4791 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-dns"/"dns-dockercfg-jwfmh" Dec 10 22:51:52 crc kubenswrapper[4791]: I1210 22:51:52.238186 4791 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-vgqrv\" (UniqueName: \"kubernetes.io/projected/cd65e91e-5b02-4f78-829e-93b4596ff8f6-kube-api-access-vgqrv\") pod \"machine-api-operator-5694c8668f-cpbhz\" (UID: \"cd65e91e-5b02-4f78-829e-93b4596ff8f6\") " pod="openshift-machine-api/machine-api-operator-5694c8668f-cpbhz" Dec 10 22:51:52 crc kubenswrapper[4791]: I1210 22:51:52.257619 4791 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"collect-profiles-dockercfg-kzf4t" Dec 10 22:51:52 crc kubenswrapper[4791]: I1210 22:51:52.277719 4791 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-service-ca-operator"/"serving-cert" Dec 10 22:51:52 crc kubenswrapper[4791]: I1210 22:51:52.297198 4791 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-service-ca-operator"/"service-ca-operator-config" Dec 10 22:51:52 crc kubenswrapper[4791]: I1210 22:51:52.318385 4791 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-service-ca-operator"/"service-ca-operator-dockercfg-rg9jl" Dec 10 22:51:52 crc kubenswrapper[4791]: I1210 22:51:52.337143 4791 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-service-ca-operator"/"kube-root-ca.crt" Dec 10 22:51:52 crc kubenswrapper[4791]: I1210 22:51:52.384459 4791 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-rkmd6\" (UniqueName: \"kubernetes.io/projected/4a1cf441-8a08-4677-a480-947c8cddd6a6-kube-api-access-rkmd6\") pod \"apiserver-76f77b778f-kqtbx\" (UID: \"4a1cf441-8a08-4677-a480-947c8cddd6a6\") " pod="openshift-apiserver/apiserver-76f77b778f-kqtbx" Dec 10 22:51:52 crc kubenswrapper[4791]: I1210 22:51:52.397286 4791 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"machine-config-server-dockercfg-qx5rd" Dec 10 22:51:52 crc kubenswrapper[4791]: I1210 22:51:52.403382 4791 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-k4rbk\" (UniqueName: \"kubernetes.io/projected/3ba17b70-17df-4f8c-87ff-c15e03ce06bc-kube-api-access-k4rbk\") pod \"apiserver-7bbb656c7d-nwwpb\" (UID: \"3ba17b70-17df-4f8c-87ff-c15e03ce06bc\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-nwwpb" Dec 10 22:51:52 crc kubenswrapper[4791]: I1210 22:51:52.417080 4791 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-operator-lifecycle-manager"/"collect-profiles-config" Dec 10 22:51:52 crc kubenswrapper[4791]: I1210 22:51:52.446314 4791 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"node-bootstrapper-token" Dec 10 22:51:52 crc kubenswrapper[4791]: I1210 22:51:52.458741 4791 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"machine-config-server-tls" Dec 10 22:51:52 crc kubenswrapper[4791]: I1210 22:51:52.478396 4791 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-dns"/"dns-default" Dec 10 22:51:52 crc kubenswrapper[4791]: I1210 22:51:52.485635 4791 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-apiserver/apiserver-76f77b778f-kqtbx" Dec 10 22:51:52 crc kubenswrapper[4791]: I1210 22:51:52.499587 4791 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-service-ca-operator"/"openshift-service-ca.crt" Dec 10 22:51:52 crc kubenswrapper[4791]: I1210 22:51:52.507581 4791 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-api/machine-api-operator-5694c8668f-cpbhz" Dec 10 22:51:52 crc kubenswrapper[4791]: I1210 22:51:52.537919 4791 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-service-ca"/"signing-key" Dec 10 22:51:52 crc kubenswrapper[4791]: I1210 22:51:52.547142 4791 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-xhsjq\" (UniqueName: \"kubernetes.io/projected/a5827e8a-4cb3-494d-87a1-556d920c2a4a-kube-api-access-xhsjq\") pod \"controller-manager-879f6c89f-9tg4p\" (UID: \"a5827e8a-4cb3-494d-87a1-556d920c2a4a\") " pod="openshift-controller-manager/controller-manager-879f6c89f-9tg4p" Dec 10 22:51:52 crc kubenswrapper[4791]: I1210 22:51:52.557227 4791 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-service-ca"/"service-ca-dockercfg-pn86c" Dec 10 22:51:52 crc kubenswrapper[4791]: I1210 22:51:52.573160 4791 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-nwwpb" Dec 10 22:51:52 crc kubenswrapper[4791]: I1210 22:51:52.576695 4791 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-service-ca"/"kube-root-ca.crt" Dec 10 22:51:52 crc kubenswrapper[4791]: I1210 22:51:52.599662 4791 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-service-ca"/"signing-cabundle" Dec 10 22:51:52 crc kubenswrapper[4791]: I1210 22:51:52.618120 4791 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-service-ca"/"openshift-service-ca.crt" Dec 10 22:51:52 crc kubenswrapper[4791]: I1210 22:51:52.637701 4791 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress-canary"/"openshift-service-ca.crt" Dec 10 22:51:52 crc kubenswrapper[4791]: I1210 22:51:52.657693 4791 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress-canary"/"canary-serving-cert" Dec 10 22:51:52 crc kubenswrapper[4791]: I1210 22:51:52.677764 4791 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress-canary"/"kube-root-ca.crt" Dec 10 22:51:52 crc kubenswrapper[4791]: I1210 22:51:52.700798 4791 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress-canary"/"default-dockercfg-2llfx" Dec 10 22:51:52 crc kubenswrapper[4791]: I1210 22:51:52.717380 4791 reflector.go:368] Caches populated for *v1.ConfigMap from object-"hostpath-provisioner"/"kube-root-ca.crt" Dec 10 22:51:52 crc kubenswrapper[4791]: I1210 22:51:52.718139 4791 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-apiserver/apiserver-76f77b778f-kqtbx"] Dec 10 22:51:52 crc kubenswrapper[4791]: W1210 22:51:52.728186 4791 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod4a1cf441_8a08_4677_a480_947c8cddd6a6.slice/crio-6b50a263a3b0ecb398128c3585459b03003223b95f28088cd9c205965769c946 WatchSource:0}: Error finding container 6b50a263a3b0ecb398128c3585459b03003223b95f28088cd9c205965769c946: Status 404 returned error can't find the container with id 6b50a263a3b0ecb398128c3585459b03003223b95f28088cd9c205965769c946 Dec 10 22:51:52 crc kubenswrapper[4791]: I1210 22:51:52.729734 4791 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-machine-api/machine-api-operator-5694c8668f-cpbhz"] Dec 10 22:51:52 crc kubenswrapper[4791]: I1210 22:51:52.736552 4791 reflector.go:368] Caches populated for *v1.Secret from object-"hostpath-provisioner"/"csi-hostpath-provisioner-sa-dockercfg-qd74k" Dec 10 22:51:52 crc kubenswrapper[4791]: W1210 22:51:52.739499 4791 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podcd65e91e_5b02_4f78_829e_93b4596ff8f6.slice/crio-4f916d9078d82074d4047f6cd60b543e936241991cd48374b4c950bae6964d25 WatchSource:0}: Error finding container 4f916d9078d82074d4047f6cd60b543e936241991cd48374b4c950bae6964d25: Status 404 returned error can't find the container with id 4f916d9078d82074d4047f6cd60b543e936241991cd48374b4c950bae6964d25 Dec 10 22:51:52 crc kubenswrapper[4791]: I1210 22:51:52.758942 4791 reflector.go:368] Caches populated for *v1.ConfigMap from object-"hostpath-provisioner"/"openshift-service-ca.crt" Dec 10 22:51:52 crc kubenswrapper[4791]: I1210 22:51:52.771662 4791 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-oauth-apiserver/apiserver-7bbb656c7d-nwwpb"] Dec 10 22:51:52 crc kubenswrapper[4791]: I1210 22:51:52.789669 4791 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-api/machine-api-operator-5694c8668f-cpbhz" event={"ID":"cd65e91e-5b02-4f78-829e-93b4596ff8f6","Type":"ContainerStarted","Data":"4f916d9078d82074d4047f6cd60b543e936241991cd48374b4c950bae6964d25"} Dec 10 22:51:52 crc kubenswrapper[4791]: I1210 22:51:52.792107 4791 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-ggn6x\" (UniqueName: \"kubernetes.io/projected/33c30ace-bd17-4a3b-a43a-ddfeccb88349-kube-api-access-ggn6x\") pod \"cluster-image-registry-operator-dc59b4c8b-q4cpx\" (UID: \"33c30ace-bd17-4a3b-a43a-ddfeccb88349\") " pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-q4cpx" Dec 10 22:51:52 crc kubenswrapper[4791]: I1210 22:51:52.792582 4791 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-apiserver/apiserver-76f77b778f-kqtbx" event={"ID":"4a1cf441-8a08-4677-a480-947c8cddd6a6","Type":"ContainerStarted","Data":"6b50a263a3b0ecb398128c3585459b03003223b95f28088cd9c205965769c946"} Dec 10 22:51:52 crc kubenswrapper[4791]: I1210 22:51:52.810826 4791 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-jzkqn\" (UniqueName: \"kubernetes.io/projected/3d53f039-46f2-478b-b14c-11b723b7dbc1-kube-api-access-jzkqn\") pod \"console-operator-58897d9998-hd6zl\" (UID: \"3d53f039-46f2-478b-b14c-11b723b7dbc1\") " pod="openshift-console-operator/console-operator-58897d9998-hd6zl" Dec 10 22:51:52 crc kubenswrapper[4791]: I1210 22:51:52.827351 4791 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager/controller-manager-879f6c89f-9tg4p" Dec 10 22:51:52 crc kubenswrapper[4791]: I1210 22:51:52.830575 4791 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-f2cbn\" (UniqueName: \"kubernetes.io/projected/465ba82f-1142-4cf1-b2ec-1f73e565ec4d-kube-api-access-f2cbn\") pod \"openshift-controller-manager-operator-756b6f6bc6-5t9vv\" (UID: \"465ba82f-1142-4cf1-b2ec-1f73e565ec4d\") " pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-5t9vv" Dec 10 22:51:52 crc kubenswrapper[4791]: I1210 22:51:52.835986 4791 request.go:700] Waited for 1.9007606s due to client-side throttling, not priority and fairness, request: POST:https://api-int.crc.testing:6443/api/v1/namespaces/openshift-config-operator/serviceaccounts/openshift-config-operator/token Dec 10 22:51:52 crc kubenswrapper[4791]: I1210 22:51:52.854415 4791 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-692qq\" (UniqueName: \"kubernetes.io/projected/b419e5d6-d6a4-4258-a1bd-a462bd7f7a12-kube-api-access-692qq\") pod \"openshift-config-operator-7777fb866f-wwpsp\" (UID: \"b419e5d6-d6a4-4258-a1bd-a462bd7f7a12\") " pod="openshift-config-operator/openshift-config-operator-7777fb866f-wwpsp" Dec 10 22:51:52 crc kubenswrapper[4791]: I1210 22:51:52.871990 4791 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-dlncw\" (UniqueName: \"kubernetes.io/projected/35033f1a-b91b-4e02-ba77-3841a2e0fdab-kube-api-access-dlncw\") pod \"oauth-openshift-558db77b4-ht5t7\" (UID: \"35033f1a-b91b-4e02-ba77-3841a2e0fdab\") " pod="openshift-authentication/oauth-openshift-558db77b4-ht5t7" Dec 10 22:51:52 crc kubenswrapper[4791]: I1210 22:51:52.892133 4791 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-w87xw\" (UniqueName: \"kubernetes.io/projected/2f422ac4-b586-4056-8748-94b7de0b0248-kube-api-access-w87xw\") pod \"machine-approver-56656f9798-m6dwv\" (UID: \"2f422ac4-b586-4056-8748-94b7de0b0248\") " pod="openshift-cluster-machine-approver/machine-approver-56656f9798-m6dwv" Dec 10 22:51:52 crc kubenswrapper[4791]: I1210 22:51:52.911231 4791 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-7m4mz\" (UniqueName: \"kubernetes.io/projected/cff60894-8deb-4354-9ad1-3a83baa98714-kube-api-access-7m4mz\") pod \"authentication-operator-69f744f599-7rhk9\" (UID: \"cff60894-8deb-4354-9ad1-3a83baa98714\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-7rhk9" Dec 10 22:51:52 crc kubenswrapper[4791]: I1210 22:51:52.917651 4791 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-cluster-machine-approver/machine-approver-56656f9798-m6dwv" Dec 10 22:51:52 crc kubenswrapper[4791]: I1210 22:51:52.925755 4791 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-console-operator/console-operator-58897d9998-hd6zl" Dec 10 22:51:52 crc kubenswrapper[4791]: I1210 22:51:52.932303 4791 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-mtfjz\" (UniqueName: \"kubernetes.io/projected/e92ca7e3-d360-4e14-aeb9-76738a44c391-kube-api-access-mtfjz\") pod \"ingress-operator-5b745b69d9-8npst\" (UID: \"e92ca7e3-d360-4e14-aeb9-76738a44c391\") " pod="openshift-ingress-operator/ingress-operator-5b745b69d9-8npst" Dec 10 22:51:52 crc kubenswrapper[4791]: I1210 22:51:52.952447 4791 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-z492g\" (UniqueName: \"kubernetes.io/projected/8f0772d5-db77-42b2-b712-0293c0422161-kube-api-access-z492g\") pod \"catalog-operator-68c6474976-497m8\" (UID: \"8f0772d5-db77-42b2-b712-0293c0422161\") " pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-497m8" Dec 10 22:51:52 crc kubenswrapper[4791]: I1210 22:51:52.965410 4791 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-authentication/oauth-openshift-558db77b4-ht5t7" Dec 10 22:51:52 crc kubenswrapper[4791]: I1210 22:51:52.969459 4791 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/5be73604-c6c9-4ef6-a53d-6c3f6df27202-kube-api-access\") pod \"kube-controller-manager-operator-78b949d7b-mldvd\" (UID: \"5be73604-c6c9-4ef6-a53d-6c3f6df27202\") " pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-mldvd" Dec 10 22:51:52 crc kubenswrapper[4791]: I1210 22:51:52.977447 4791 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-config-operator/openshift-config-operator-7777fb866f-wwpsp" Dec 10 22:51:52 crc kubenswrapper[4791]: I1210 22:51:52.994154 4791 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-z5dwf\" (UniqueName: \"kubernetes.io/projected/5369b8df-3c36-4108-b4e7-cfb11478b3a7-kube-api-access-z5dwf\") pod \"route-controller-manager-6576b87f9c-vsnsh\" (UID: \"5369b8df-3c36-4108-b4e7-cfb11478b3a7\") " pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-vsnsh" Dec 10 22:51:53 crc kubenswrapper[4791]: I1210 22:51:53.010174 4791 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-vn4qx\" (UniqueName: \"kubernetes.io/projected/d3de1ba0-b014-4f98-b597-cc89f5a70364-kube-api-access-vn4qx\") pod \"router-default-5444994796-p97fc\" (UID: \"d3de1ba0-b014-4f98-b597-cc89f5a70364\") " pod="openshift-ingress/router-default-5444994796-p97fc" Dec 10 22:51:53 crc kubenswrapper[4791]: I1210 22:51:53.014871 4791 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-authentication-operator/authentication-operator-69f744f599-7rhk9" Dec 10 22:51:53 crc kubenswrapper[4791]: I1210 22:51:53.023879 4791 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-mldvd" Dec 10 22:51:53 crc kubenswrapper[4791]: I1210 22:51:53.030161 4791 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-tss7t\" (UniqueName: \"kubernetes.io/projected/680d2296-dd48-4786-bea7-64c32edc94c7-kube-api-access-tss7t\") pod \"downloads-7954f5f757-mlb9p\" (UID: \"680d2296-dd48-4786-bea7-64c32edc94c7\") " pod="openshift-console/downloads-7954f5f757-mlb9p" Dec 10 22:51:53 crc kubenswrapper[4791]: I1210 22:51:53.050914 4791 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-f2tld\" (UniqueName: \"kubernetes.io/projected/1b61a973-db5e-4aa7-96e3-7fbdd9dd8cc0-kube-api-access-f2tld\") pod \"console-f9d7485db-zkfkv\" (UID: \"1b61a973-db5e-4aa7-96e3-7fbdd9dd8cc0\") " pod="openshift-console/console-f9d7485db-zkfkv" Dec 10 22:51:53 crc kubenswrapper[4791]: I1210 22:51:53.071695 4791 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/33c30ace-bd17-4a3b-a43a-ddfeccb88349-bound-sa-token\") pod \"cluster-image-registry-operator-dc59b4c8b-q4cpx\" (UID: \"33c30ace-bd17-4a3b-a43a-ddfeccb88349\") " pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-q4cpx" Dec 10 22:51:53 crc kubenswrapper[4791]: I1210 22:51:53.077672 4791 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-5t9vv" Dec 10 22:51:53 crc kubenswrapper[4791]: I1210 22:51:53.085514 4791 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-ingress/router-default-5444994796-p97fc" Dec 10 22:51:53 crc kubenswrapper[4791]: I1210 22:51:53.091119 4791 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-fltgv\" (UniqueName: \"kubernetes.io/projected/4205f348-3276-40f3-a35b-69306b331d80-kube-api-access-fltgv\") pod \"cluster-samples-operator-665b6dd947-t7742\" (UID: \"4205f348-3276-40f3-a35b-69306b331d80\") " pod="openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-t7742" Dec 10 22:51:53 crc kubenswrapper[4791]: I1210 22:51:53.111857 4791 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-vwnrt\" (UniqueName: \"kubernetes.io/projected/e2b778bb-d80f-49f9-804b-0550c648a02f-kube-api-access-vwnrt\") pod \"machine-config-controller-84d6567774-sc74x\" (UID: \"e2b778bb-d80f-49f9-804b-0550c648a02f\") " pod="openshift-machine-config-operator/machine-config-controller-84d6567774-sc74x" Dec 10 22:51:53 crc kubenswrapper[4791]: I1210 22:51:53.117288 4791 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-497m8" Dec 10 22:51:53 crc kubenswrapper[4791]: I1210 22:51:53.128951 4791 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-config-operator/machine-config-controller-84d6567774-sc74x" Dec 10 22:51:53 crc kubenswrapper[4791]: I1210 22:51:53.129881 4791 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/e92ca7e3-d360-4e14-aeb9-76738a44c391-bound-sa-token\") pod \"ingress-operator-5b745b69d9-8npst\" (UID: \"e92ca7e3-d360-4e14-aeb9-76738a44c391\") " pod="openshift-ingress-operator/ingress-operator-5b745b69d9-8npst" Dec 10 22:51:53 crc kubenswrapper[4791]: I1210 22:51:53.517060 4791 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-vsnsh" Dec 10 22:51:53 crc kubenswrapper[4791]: I1210 22:51:53.517705 4791 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-console/downloads-7954f5f757-mlb9p" Dec 10 22:51:53 crc kubenswrapper[4791]: I1210 22:51:53.518170 4791 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-t7742" Dec 10 22:51:53 crc kubenswrapper[4791]: I1210 22:51:53.518184 4791 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-q4cpx" Dec 10 22:51:53 crc kubenswrapper[4791]: I1210 22:51:53.518633 4791 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-ingress-operator/ingress-operator-5b745b69d9-8npst" Dec 10 22:51:53 crc kubenswrapper[4791]: I1210 22:51:53.518682 4791 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-console/console-f9d7485db-zkfkv" Dec 10 22:51:53 crc kubenswrapper[4791]: I1210 22:51:53.519485 4791 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"installation-pull-secrets\" (UniqueName: \"kubernetes.io/secret/e93dbd5d-0ca7-4a15-9eb3-0d71595fada0-installation-pull-secrets\") pod \"image-registry-697d97f7c8-srd52\" (UID: \"e93dbd5d-0ca7-4a15-9eb3-0d71595fada0\") " pod="openshift-image-registry/image-registry-697d97f7c8-srd52" Dec 10 22:51:53 crc kubenswrapper[4791]: I1210 22:51:53.520001 4791 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"registry-certificates\" (UniqueName: \"kubernetes.io/configmap/e93dbd5d-0ca7-4a15-9eb3-0d71595fada0-registry-certificates\") pod \"image-registry-697d97f7c8-srd52\" (UID: \"e93dbd5d-0ca7-4a15-9eb3-0d71595fada0\") " pod="openshift-image-registry/image-registry-697d97f7c8-srd52" Dec 10 22:51:53 crc kubenswrapper[4791]: I1210 22:51:53.520190 4791 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/e93dbd5d-0ca7-4a15-9eb3-0d71595fada0-trusted-ca\") pod \"image-registry-697d97f7c8-srd52\" (UID: \"e93dbd5d-0ca7-4a15-9eb3-0d71595fada0\") " pod="openshift-image-registry/image-registry-697d97f7c8-srd52" Dec 10 22:51:53 crc kubenswrapper[4791]: I1210 22:51:53.520476 4791 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ca-trust-extracted\" (UniqueName: \"kubernetes.io/empty-dir/e93dbd5d-0ca7-4a15-9eb3-0d71595fada0-ca-trust-extracted\") pod \"image-registry-697d97f7c8-srd52\" (UID: \"e93dbd5d-0ca7-4a15-9eb3-0d71595fada0\") " pod="openshift-image-registry/image-registry-697d97f7c8-srd52" Dec 10 22:51:53 crc kubenswrapper[4791]: I1210 22:51:53.520547 4791 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-srd52\" (UID: \"e93dbd5d-0ca7-4a15-9eb3-0d71595fada0\") " pod="openshift-image-registry/image-registry-697d97f7c8-srd52" Dec 10 22:51:53 crc kubenswrapper[4791]: I1210 22:51:53.520583 4791 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"registry-tls\" (UniqueName: \"kubernetes.io/projected/e93dbd5d-0ca7-4a15-9eb3-0d71595fada0-registry-tls\") pod \"image-registry-697d97f7c8-srd52\" (UID: \"e93dbd5d-0ca7-4a15-9eb3-0d71595fada0\") " pod="openshift-image-registry/image-registry-697d97f7c8-srd52" Dec 10 22:51:53 crc kubenswrapper[4791]: I1210 22:51:53.520678 4791 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/e93dbd5d-0ca7-4a15-9eb3-0d71595fada0-bound-sa-token\") pod \"image-registry-697d97f7c8-srd52\" (UID: \"e93dbd5d-0ca7-4a15-9eb3-0d71595fada0\") " pod="openshift-image-registry/image-registry-697d97f7c8-srd52" Dec 10 22:51:53 crc kubenswrapper[4791]: I1210 22:51:53.520739 4791 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-97w86\" (UniqueName: \"kubernetes.io/projected/e93dbd5d-0ca7-4a15-9eb3-0d71595fada0-kube-api-access-97w86\") pod \"image-registry-697d97f7c8-srd52\" (UID: \"e93dbd5d-0ca7-4a15-9eb3-0d71595fada0\") " pod="openshift-image-registry/image-registry-697d97f7c8-srd52" Dec 10 22:51:53 crc kubenswrapper[4791]: E1210 22:51:53.521501 4791 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-10 22:51:54.021470578 +0000 UTC m=+148.451088221 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-srd52" (UID: "e93dbd5d-0ca7-4a15-9eb3-0d71595fada0") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 10 22:51:53 crc kubenswrapper[4791]: I1210 22:51:53.622968 4791 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 10 22:51:53 crc kubenswrapper[4791]: I1210 22:51:53.623278 4791 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"installation-pull-secrets\" (UniqueName: \"kubernetes.io/secret/e93dbd5d-0ca7-4a15-9eb3-0d71595fada0-installation-pull-secrets\") pod \"image-registry-697d97f7c8-srd52\" (UID: \"e93dbd5d-0ca7-4a15-9eb3-0d71595fada0\") " pod="openshift-image-registry/image-registry-697d97f7c8-srd52" Dec 10 22:51:53 crc kubenswrapper[4791]: I1210 22:51:53.623330 4791 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"profile-collector-cert\" (UniqueName: \"kubernetes.io/secret/06eb2b28-22d6-4116-9ce4-1abf7a21db36-profile-collector-cert\") pod \"olm-operator-6b444d44fb-bws6n\" (UID: \"06eb2b28-22d6-4116-9ce4-1abf7a21db36\") " pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-bws6n" Dec 10 22:51:53 crc kubenswrapper[4791]: I1210 22:51:53.623406 4791 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"srv-cert\" (UniqueName: \"kubernetes.io/secret/06eb2b28-22d6-4116-9ce4-1abf7a21db36-srv-cert\") pod \"olm-operator-6b444d44fb-bws6n\" (UID: \"06eb2b28-22d6-4116-9ce4-1abf7a21db36\") " pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-bws6n" Dec 10 22:51:53 crc kubenswrapper[4791]: I1210 22:51:53.623456 4791 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"registry-certificates\" (UniqueName: \"kubernetes.io/configmap/e93dbd5d-0ca7-4a15-9eb3-0d71595fada0-registry-certificates\") pod \"image-registry-697d97f7c8-srd52\" (UID: \"e93dbd5d-0ca7-4a15-9eb3-0d71595fada0\") " pod="openshift-image-registry/image-registry-697d97f7c8-srd52" Dec 10 22:51:53 crc kubenswrapper[4791]: I1210 22:51:53.623483 4791 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/e93dbd5d-0ca7-4a15-9eb3-0d71595fada0-trusted-ca\") pod \"image-registry-697d97f7c8-srd52\" (UID: \"e93dbd5d-0ca7-4a15-9eb3-0d71595fada0\") " pod="openshift-image-registry/image-registry-697d97f7c8-srd52" Dec 10 22:51:53 crc kubenswrapper[4791]: I1210 22:51:53.623514 4791 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ca-trust-extracted\" (UniqueName: \"kubernetes.io/empty-dir/e93dbd5d-0ca7-4a15-9eb3-0d71595fada0-ca-trust-extracted\") pod \"image-registry-697d97f7c8-srd52\" (UID: \"e93dbd5d-0ca7-4a15-9eb3-0d71595fada0\") " pod="openshift-image-registry/image-registry-697d97f7c8-srd52" Dec 10 22:51:53 crc kubenswrapper[4791]: I1210 22:51:53.623577 4791 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"registry-tls\" (UniqueName: \"kubernetes.io/projected/e93dbd5d-0ca7-4a15-9eb3-0d71595fada0-registry-tls\") pod \"image-registry-697d97f7c8-srd52\" (UID: \"e93dbd5d-0ca7-4a15-9eb3-0d71595fada0\") " pod="openshift-image-registry/image-registry-697d97f7c8-srd52" Dec 10 22:51:53 crc kubenswrapper[4791]: I1210 22:51:53.623607 4791 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-g7ll5\" (UniqueName: \"kubernetes.io/projected/06eb2b28-22d6-4116-9ce4-1abf7a21db36-kube-api-access-g7ll5\") pod \"olm-operator-6b444d44fb-bws6n\" (UID: \"06eb2b28-22d6-4116-9ce4-1abf7a21db36\") " pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-bws6n" Dec 10 22:51:53 crc kubenswrapper[4791]: I1210 22:51:53.623652 4791 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/e93dbd5d-0ca7-4a15-9eb3-0d71595fada0-bound-sa-token\") pod \"image-registry-697d97f7c8-srd52\" (UID: \"e93dbd5d-0ca7-4a15-9eb3-0d71595fada0\") " pod="openshift-image-registry/image-registry-697d97f7c8-srd52" Dec 10 22:51:53 crc kubenswrapper[4791]: I1210 22:51:53.623691 4791 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-97w86\" (UniqueName: \"kubernetes.io/projected/e93dbd5d-0ca7-4a15-9eb3-0d71595fada0-kube-api-access-97w86\") pod \"image-registry-697d97f7c8-srd52\" (UID: \"e93dbd5d-0ca7-4a15-9eb3-0d71595fada0\") " pod="openshift-image-registry/image-registry-697d97f7c8-srd52" Dec 10 22:51:53 crc kubenswrapper[4791]: I1210 22:51:53.626249 4791 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/e93dbd5d-0ca7-4a15-9eb3-0d71595fada0-trusted-ca\") pod \"image-registry-697d97f7c8-srd52\" (UID: \"e93dbd5d-0ca7-4a15-9eb3-0d71595fada0\") " pod="openshift-image-registry/image-registry-697d97f7c8-srd52" Dec 10 22:51:53 crc kubenswrapper[4791]: I1210 22:51:53.626897 4791 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ca-trust-extracted\" (UniqueName: \"kubernetes.io/empty-dir/e93dbd5d-0ca7-4a15-9eb3-0d71595fada0-ca-trust-extracted\") pod \"image-registry-697d97f7c8-srd52\" (UID: \"e93dbd5d-0ca7-4a15-9eb3-0d71595fada0\") " pod="openshift-image-registry/image-registry-697d97f7c8-srd52" Dec 10 22:51:53 crc kubenswrapper[4791]: E1210 22:51:53.629668 4791 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-10 22:51:54.129636325 +0000 UTC m=+148.559253938 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 10 22:51:53 crc kubenswrapper[4791]: I1210 22:51:53.629893 4791 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"registry-certificates\" (UniqueName: \"kubernetes.io/configmap/e93dbd5d-0ca7-4a15-9eb3-0d71595fada0-registry-certificates\") pod \"image-registry-697d97f7c8-srd52\" (UID: \"e93dbd5d-0ca7-4a15-9eb3-0d71595fada0\") " pod="openshift-image-registry/image-registry-697d97f7c8-srd52" Dec 10 22:51:53 crc kubenswrapper[4791]: I1210 22:51:53.644216 4791 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"registry-tls\" (UniqueName: \"kubernetes.io/projected/e93dbd5d-0ca7-4a15-9eb3-0d71595fada0-registry-tls\") pod \"image-registry-697d97f7c8-srd52\" (UID: \"e93dbd5d-0ca7-4a15-9eb3-0d71595fada0\") " pod="openshift-image-registry/image-registry-697d97f7c8-srd52" Dec 10 22:51:53 crc kubenswrapper[4791]: I1210 22:51:53.645521 4791 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"installation-pull-secrets\" (UniqueName: \"kubernetes.io/secret/e93dbd5d-0ca7-4a15-9eb3-0d71595fada0-installation-pull-secrets\") pod \"image-registry-697d97f7c8-srd52\" (UID: \"e93dbd5d-0ca7-4a15-9eb3-0d71595fada0\") " pod="openshift-image-registry/image-registry-697d97f7c8-srd52" Dec 10 22:51:53 crc kubenswrapper[4791]: I1210 22:51:53.663439 4791 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-97w86\" (UniqueName: \"kubernetes.io/projected/e93dbd5d-0ca7-4a15-9eb3-0d71595fada0-kube-api-access-97w86\") pod \"image-registry-697d97f7c8-srd52\" (UID: \"e93dbd5d-0ca7-4a15-9eb3-0d71595fada0\") " pod="openshift-image-registry/image-registry-697d97f7c8-srd52" Dec 10 22:51:53 crc kubenswrapper[4791]: I1210 22:51:53.663612 4791 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/e93dbd5d-0ca7-4a15-9eb3-0d71595fada0-bound-sa-token\") pod \"image-registry-697d97f7c8-srd52\" (UID: \"e93dbd5d-0ca7-4a15-9eb3-0d71595fada0\") " pod="openshift-image-registry/image-registry-697d97f7c8-srd52" Dec 10 22:51:53 crc kubenswrapper[4791]: I1210 22:51:53.725132 4791 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-6mhcw\" (UniqueName: \"kubernetes.io/projected/3f132df5-3de5-4455-a3da-7848b0de69cc-kube-api-access-6mhcw\") pod \"package-server-manager-789f6589d5-w5dpr\" (UID: \"3f132df5-3de5-4455-a3da-7848b0de69cc\") " pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-w5dpr" Dec 10 22:51:53 crc kubenswrapper[4791]: I1210 22:51:53.725177 4791 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/90640468-53d7-42bc-a835-f7acf48a486e-config-volume\") pod \"collect-profiles-29423445-b9cxb\" (UID: \"90640468-53d7-42bc-a835-f7acf48a486e\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29423445-b9cxb" Dec 10 22:51:53 crc kubenswrapper[4791]: I1210 22:51:53.725243 4791 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/8f993103-9ee4-446f-a91a-a03dcd3ed9ea-kube-api-access\") pod \"openshift-kube-scheduler-operator-5fdd9b5758-z6gpf\" (UID: \"8f993103-9ee4-446f-a91a-a03dcd3ed9ea\") " pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-z6gpf" Dec 10 22:51:53 crc kubenswrapper[4791]: I1210 22:51:53.725294 4791 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/a05f3579-4703-4c46-bb11-5de60eedbc02-apiservice-cert\") pod \"packageserver-d55dfcdfc-v445v\" (UID: \"a05f3579-4703-4c46-bb11-5de60eedbc02\") " pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-v445v" Dec 10 22:51:53 crc kubenswrapper[4791]: I1210 22:51:53.725331 4791 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/e9978f48-c33a-4765-9b05-16e37f5cdb24-config-volume\") pod \"dns-default-qs77n\" (UID: \"e9978f48-c33a-4765-9b05-16e37f5cdb24\") " pod="openshift-dns/dns-default-qs77n" Dec 10 22:51:53 crc kubenswrapper[4791]: I1210 22:51:53.725414 4791 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-z25fj\" (UniqueName: \"kubernetes.io/projected/0921ef8b-8842-4d66-890b-3ad15bd48e70-kube-api-access-z25fj\") pod \"ingress-canary-9dwf5\" (UID: \"0921ef8b-8842-4d66-890b-3ad15bd48e70\") " pod="openshift-ingress-canary/ingress-canary-9dwf5" Dec 10 22:51:53 crc kubenswrapper[4791]: I1210 22:51:53.725458 4791 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"csi-data-dir\" (UniqueName: \"kubernetes.io/host-path/1666d5bb-295d-4828-868e-b1b17b5741bb-csi-data-dir\") pod \"csi-hostpathplugin-mkz5r\" (UID: \"1666d5bb-295d-4828-868e-b1b17b5741bb\") " pod="hostpath-provisioner/csi-hostpathplugin-mkz5r" Dec 10 22:51:53 crc kubenswrapper[4791]: I1210 22:51:53.725479 4791 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-dk24r\" (UniqueName: \"kubernetes.io/projected/1666d5bb-295d-4828-868e-b1b17b5741bb-kube-api-access-dk24r\") pod \"csi-hostpathplugin-mkz5r\" (UID: \"1666d5bb-295d-4828-868e-b1b17b5741bb\") " pod="hostpath-provisioner/csi-hostpathplugin-mkz5r" Dec 10 22:51:53 crc kubenswrapper[4791]: I1210 22:51:53.725510 4791 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/e9978f48-c33a-4765-9b05-16e37f5cdb24-metrics-tls\") pod \"dns-default-qs77n\" (UID: \"e9978f48-c33a-4765-9b05-16e37f5cdb24\") " pod="openshift-dns/dns-default-qs77n" Dec 10 22:51:53 crc kubenswrapper[4791]: I1210 22:51:53.725561 4791 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-pbhrm\" (UniqueName: \"kubernetes.io/projected/2fee22d4-2e9b-4768-9603-22e11d9dad9a-kube-api-access-pbhrm\") pod \"machine-config-operator-74547568cd-rcnrn\" (UID: \"2fee22d4-2e9b-4768-9603-22e11d9dad9a\") " pod="openshift-machine-config-operator/machine-config-operator-74547568cd-rcnrn" Dec 10 22:51:53 crc kubenswrapper[4791]: I1210 22:51:53.725604 4791 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"signing-cabundle\" (UniqueName: \"kubernetes.io/configmap/34da0cc5-3e59-4c25-af94-e755e772b89a-signing-cabundle\") pod \"service-ca-9c57cc56f-x9pw8\" (UID: \"34da0cc5-3e59-4c25-af94-e755e772b89a\") " pod="openshift-service-ca/service-ca-9c57cc56f-x9pw8" Dec 10 22:51:53 crc kubenswrapper[4791]: I1210 22:51:53.725651 4791 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/97c517e6-0af1-4c17-a2b9-f7a7db3368b3-serving-cert\") pod \"etcd-operator-b45778765-nkt8g\" (UID: \"97c517e6-0af1-4c17-a2b9-f7a7db3368b3\") " pod="openshift-etcd-operator/etcd-operator-b45778765-nkt8g" Dec 10 22:51:53 crc kubenswrapper[4791]: I1210 22:51:53.725696 4791 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/8f993103-9ee4-446f-a91a-a03dcd3ed9ea-serving-cert\") pod \"openshift-kube-scheduler-operator-5fdd9b5758-z6gpf\" (UID: \"8f993103-9ee4-446f-a91a-a03dcd3ed9ea\") " pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-z6gpf" Dec 10 22:51:53 crc kubenswrapper[4791]: I1210 22:51:53.725732 4791 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-tm8cv\" (UniqueName: \"kubernetes.io/projected/97c517e6-0af1-4c17-a2b9-f7a7db3368b3-kube-api-access-tm8cv\") pod \"etcd-operator-b45778765-nkt8g\" (UID: \"97c517e6-0af1-4c17-a2b9-f7a7db3368b3\") " pod="openshift-etcd-operator/etcd-operator-b45778765-nkt8g" Dec 10 22:51:53 crc kubenswrapper[4791]: I1210 22:51:53.725753 4791 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-t6768\" (UniqueName: \"kubernetes.io/projected/ded98820-5308-46e3-b87b-18d0a81c0519-kube-api-access-t6768\") pod \"multus-admission-controller-857f4d67dd-fbpcr\" (UID: \"ded98820-5308-46e3-b87b-18d0a81c0519\") " pod="openshift-multus/multus-admission-controller-857f4d67dd-fbpcr" Dec 10 22:51:53 crc kubenswrapper[4791]: I1210 22:51:53.725796 4791 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-bdjr4\" (UniqueName: \"kubernetes.io/projected/a05f3579-4703-4c46-bb11-5de60eedbc02-kube-api-access-bdjr4\") pod \"packageserver-d55dfcdfc-v445v\" (UID: \"a05f3579-4703-4c46-bb11-5de60eedbc02\") " pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-v445v" Dec 10 22:51:53 crc kubenswrapper[4791]: I1210 22:51:53.725825 4791 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"tmpfs\" (UniqueName: \"kubernetes.io/empty-dir/a05f3579-4703-4c46-bb11-5de60eedbc02-tmpfs\") pod \"packageserver-d55dfcdfc-v445v\" (UID: \"a05f3579-4703-4c46-bb11-5de60eedbc02\") " pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-v445v" Dec 10 22:51:53 crc kubenswrapper[4791]: I1210 22:51:53.725843 4791 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/90640468-53d7-42bc-a835-f7acf48a486e-secret-volume\") pod \"collect-profiles-29423445-b9cxb\" (UID: \"90640468-53d7-42bc-a835-f7acf48a486e\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29423445-b9cxb" Dec 10 22:51:53 crc kubenswrapper[4791]: I1210 22:51:53.725865 4791 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-n8jwh\" (UniqueName: \"kubernetes.io/projected/868d6095-79bd-41fc-ae14-f369fdf68fde-kube-api-access-n8jwh\") pod \"migrator-59844c95c7-vscpm\" (UID: \"868d6095-79bd-41fc-ae14-f369fdf68fde\") " pod="openshift-kube-storage-version-migrator/migrator-59844c95c7-vscpm" Dec 10 22:51:53 crc kubenswrapper[4791]: I1210 22:51:53.725884 4791 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"marketplace-trusted-ca\" (UniqueName: \"kubernetes.io/configmap/a9185ee1-adaf-48e8-b9a9-90ef658f8212-marketplace-trusted-ca\") pod \"marketplace-operator-79b997595-7g828\" (UID: \"a9185ee1-adaf-48e8-b9a9-90ef658f8212\") " pod="openshift-marketplace/marketplace-operator-79b997595-7g828" Dec 10 22:51:53 crc kubenswrapper[4791]: I1210 22:51:53.725906 4791 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/93a38a79-5c1c-4296-849e-94137de8f974-config\") pod \"openshift-apiserver-operator-796bbdcf4f-rmjd2\" (UID: \"93a38a79-5c1c-4296-849e-94137de8f974\") " pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-rmjd2" Dec 10 22:51:53 crc kubenswrapper[4791]: I1210 22:51:53.725939 4791 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/6e58f465-ef21-4b20-9eb8-5ade4810688a-serving-cert\") pod \"kube-apiserver-operator-766d6c64bb-vm842\" (UID: \"6e58f465-ef21-4b20-9eb8-5ade4810688a\") " pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-vm842" Dec 10 22:51:53 crc kubenswrapper[4791]: I1210 22:51:53.725960 4791 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"signing-key\" (UniqueName: \"kubernetes.io/secret/34da0cc5-3e59-4c25-af94-e755e772b89a-signing-key\") pod \"service-ca-9c57cc56f-x9pw8\" (UID: \"34da0cc5-3e59-4c25-af94-e755e772b89a\") " pod="openshift-service-ca/service-ca-9c57cc56f-x9pw8" Dec 10 22:51:53 crc kubenswrapper[4791]: I1210 22:51:53.726033 4791 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"srv-cert\" (UniqueName: \"kubernetes.io/secret/06eb2b28-22d6-4116-9ce4-1abf7a21db36-srv-cert\") pod \"olm-operator-6b444d44fb-bws6n\" (UID: \"06eb2b28-22d6-4116-9ce4-1abf7a21db36\") " pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-bws6n" Dec 10 22:51:53 crc kubenswrapper[4791]: I1210 22:51:53.726071 4791 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-jpwld\" (UniqueName: \"kubernetes.io/projected/c6c773c2-29e4-477d-91fb-f63052272f06-kube-api-access-jpwld\") pod \"control-plane-machine-set-operator-78cbb6b69f-ms8jw\" (UID: \"c6c773c2-29e4-477d-91fb-f63052272f06\") " pod="openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-ms8jw" Dec 10 22:51:53 crc kubenswrapper[4791]: I1210 22:51:53.726091 4791 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/5a2896d3-fca8-4c64-abeb-6320f0b78d4e-metrics-tls\") pod \"dns-operator-744455d44c-cggs2\" (UID: \"5a2896d3-fca8-4c64-abeb-6320f0b78d4e\") " pod="openshift-dns-operator/dns-operator-744455d44c-cggs2" Dec 10 22:51:53 crc kubenswrapper[4791]: I1210 22:51:53.726127 4791 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/97c517e6-0af1-4c17-a2b9-f7a7db3368b3-etcd-client\") pod \"etcd-operator-b45778765-nkt8g\" (UID: \"97c517e6-0af1-4c17-a2b9-f7a7db3368b3\") " pod="openshift-etcd-operator/etcd-operator-b45778765-nkt8g" Dec 10 22:51:53 crc kubenswrapper[4791]: I1210 22:51:53.726149 4791 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"node-bootstrap-token\" (UniqueName: \"kubernetes.io/secret/36532144-441a-4c12-a436-59d93d5ec3b7-node-bootstrap-token\") pod \"machine-config-server-ktcmn\" (UID: \"36532144-441a-4c12-a436-59d93d5ec3b7\") " pod="openshift-machine-config-operator/machine-config-server-ktcmn" Dec 10 22:51:53 crc kubenswrapper[4791]: I1210 22:51:53.726209 4791 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/64199728-979b-4628-8b37-2dd792f86439-serving-cert\") pod \"service-ca-operator-777779d784-b8v4c\" (UID: \"64199728-979b-4628-8b37-2dd792f86439\") " pod="openshift-service-ca-operator/service-ca-operator-777779d784-b8v4c" Dec 10 22:51:53 crc kubenswrapper[4791]: I1210 22:51:53.726245 4791 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"certs\" (UniqueName: \"kubernetes.io/secret/36532144-441a-4c12-a436-59d93d5ec3b7-certs\") pod \"machine-config-server-ktcmn\" (UID: \"36532144-441a-4c12-a436-59d93d5ec3b7\") " pod="openshift-machine-config-operator/machine-config-server-ktcmn" Dec 10 22:51:53 crc kubenswrapper[4791]: I1210 22:51:53.726265 4791 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"mountpoint-dir\" (UniqueName: \"kubernetes.io/host-path/1666d5bb-295d-4828-868e-b1b17b5741bb-mountpoint-dir\") pod \"csi-hostpathplugin-mkz5r\" (UID: \"1666d5bb-295d-4828-868e-b1b17b5741bb\") " pod="hostpath-provisioner/csi-hostpathplugin-mkz5r" Dec 10 22:51:53 crc kubenswrapper[4791]: I1210 22:51:53.726288 4791 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-mnnjz\" (UniqueName: \"kubernetes.io/projected/36532144-441a-4c12-a436-59d93d5ec3b7-kube-api-access-mnnjz\") pod \"machine-config-server-ktcmn\" (UID: \"36532144-441a-4c12-a436-59d93d5ec3b7\") " pod="openshift-machine-config-operator/machine-config-server-ktcmn" Dec 10 22:51:53 crc kubenswrapper[4791]: I1210 22:51:53.726379 4791 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"webhook-certs\" (UniqueName: \"kubernetes.io/secret/ded98820-5308-46e3-b87b-18d0a81c0519-webhook-certs\") pod \"multus-admission-controller-857f4d67dd-fbpcr\" (UID: \"ded98820-5308-46e3-b87b-18d0a81c0519\") " pod="openshift-multus/multus-admission-controller-857f4d67dd-fbpcr" Dec 10 22:51:53 crc kubenswrapper[4791]: I1210 22:51:53.726419 4791 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/93a38a79-5c1c-4296-849e-94137de8f974-serving-cert\") pod \"openshift-apiserver-operator-796bbdcf4f-rmjd2\" (UID: \"93a38a79-5c1c-4296-849e-94137de8f974\") " pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-rmjd2" Dec 10 22:51:53 crc kubenswrapper[4791]: I1210 22:51:53.726499 4791 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-gnhrf\" (UniqueName: \"kubernetes.io/projected/90640468-53d7-42bc-a835-f7acf48a486e-kube-api-access-gnhrf\") pod \"collect-profiles-29423445-b9cxb\" (UID: \"90640468-53d7-42bc-a835-f7acf48a486e\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29423445-b9cxb" Dec 10 22:51:53 crc kubenswrapper[4791]: I1210 22:51:53.726536 4791 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-srd52\" (UID: \"e93dbd5d-0ca7-4a15-9eb3-0d71595fada0\") " pod="openshift-image-registry/image-registry-697d97f7c8-srd52" Dec 10 22:51:53 crc kubenswrapper[4791]: I1210 22:51:53.726564 4791 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-hsmgd\" (UniqueName: \"kubernetes.io/projected/a9185ee1-adaf-48e8-b9a9-90ef658f8212-kube-api-access-hsmgd\") pod \"marketplace-operator-79b997595-7g828\" (UID: \"a9185ee1-adaf-48e8-b9a9-90ef658f8212\") " pod="openshift-marketplace/marketplace-operator-79b997595-7g828" Dec 10 22:51:53 crc kubenswrapper[4791]: I1210 22:51:53.726614 4791 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/6e58f465-ef21-4b20-9eb8-5ade4810688a-kube-api-access\") pod \"kube-apiserver-operator-766d6c64bb-vm842\" (UID: \"6e58f465-ef21-4b20-9eb8-5ade4810688a\") " pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-vm842" Dec 10 22:51:53 crc kubenswrapper[4791]: I1210 22:51:53.726677 4791 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-g7ll5\" (UniqueName: \"kubernetes.io/projected/06eb2b28-22d6-4116-9ce4-1abf7a21db36-kube-api-access-g7ll5\") pod \"olm-operator-6b444d44fb-bws6n\" (UID: \"06eb2b28-22d6-4116-9ce4-1abf7a21db36\") " pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-bws6n" Dec 10 22:51:53 crc kubenswrapper[4791]: I1210 22:51:53.726706 4791 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-xcxbr\" (UniqueName: \"kubernetes.io/projected/e9978f48-c33a-4765-9b05-16e37f5cdb24-kube-api-access-xcxbr\") pod \"dns-default-qs77n\" (UID: \"e9978f48-c33a-4765-9b05-16e37f5cdb24\") " pod="openshift-dns/dns-default-qs77n" Dec 10 22:51:53 crc kubenswrapper[4791]: I1210 22:51:53.726730 4791 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"marketplace-operator-metrics\" (UniqueName: \"kubernetes.io/secret/a9185ee1-adaf-48e8-b9a9-90ef658f8212-marketplace-operator-metrics\") pod \"marketplace-operator-79b997595-7g828\" (UID: \"a9185ee1-adaf-48e8-b9a9-90ef658f8212\") " pod="openshift-marketplace/marketplace-operator-79b997595-7g828" Dec 10 22:51:53 crc kubenswrapper[4791]: I1210 22:51:53.726780 4791 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"control-plane-machine-set-operator-tls\" (UniqueName: \"kubernetes.io/secret/c6c773c2-29e4-477d-91fb-f63052272f06-control-plane-machine-set-operator-tls\") pod \"control-plane-machine-set-operator-78cbb6b69f-ms8jw\" (UID: \"c6c773c2-29e4-477d-91fb-f63052272f06\") " pod="openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-ms8jw" Dec 10 22:51:53 crc kubenswrapper[4791]: I1210 22:51:53.726852 4791 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-fwbdr\" (UniqueName: \"kubernetes.io/projected/5a2896d3-fca8-4c64-abeb-6320f0b78d4e-kube-api-access-fwbdr\") pod \"dns-operator-744455d44c-cggs2\" (UID: \"5a2896d3-fca8-4c64-abeb-6320f0b78d4e\") " pod="openshift-dns-operator/dns-operator-744455d44c-cggs2" Dec 10 22:51:53 crc kubenswrapper[4791]: I1210 22:51:53.726890 4791 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/a05f3579-4703-4c46-bb11-5de60eedbc02-webhook-cert\") pod \"packageserver-d55dfcdfc-v445v\" (UID: \"a05f3579-4703-4c46-bb11-5de60eedbc02\") " pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-v445v" Dec 10 22:51:53 crc kubenswrapper[4791]: I1210 22:51:53.727971 4791 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-87746\" (UniqueName: \"kubernetes.io/projected/34da0cc5-3e59-4c25-af94-e755e772b89a-kube-api-access-87746\") pod \"service-ca-9c57cc56f-x9pw8\" (UID: \"34da0cc5-3e59-4c25-af94-e755e772b89a\") " pod="openshift-service-ca/service-ca-9c57cc56f-x9pw8" Dec 10 22:51:53 crc kubenswrapper[4791]: I1210 22:51:53.728099 4791 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"socket-dir\" (UniqueName: \"kubernetes.io/host-path/1666d5bb-295d-4828-868e-b1b17b5741bb-socket-dir\") pod \"csi-hostpathplugin-mkz5r\" (UID: \"1666d5bb-295d-4828-868e-b1b17b5741bb\") " pod="hostpath-provisioner/csi-hostpathplugin-mkz5r" Dec 10 22:51:53 crc kubenswrapper[4791]: I1210 22:51:53.728139 4791 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etcd-service-ca\" (UniqueName: \"kubernetes.io/configmap/97c517e6-0af1-4c17-a2b9-f7a7db3368b3-etcd-service-ca\") pod \"etcd-operator-b45778765-nkt8g\" (UID: \"97c517e6-0af1-4c17-a2b9-f7a7db3368b3\") " pod="openshift-etcd-operator/etcd-operator-b45778765-nkt8g" Dec 10 22:51:53 crc kubenswrapper[4791]: I1210 22:51:53.728193 4791 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/169ebb73-7b32-4f87-a8d7-9272bff04856-serving-cert\") pod \"kube-storage-version-migrator-operator-b67b599dd-k7qsx\" (UID: \"169ebb73-7b32-4f87-a8d7-9272bff04856\") " pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-k7qsx" Dec 10 22:51:53 crc kubenswrapper[4791]: I1210 22:51:53.728230 4791 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/2fee22d4-2e9b-4768-9603-22e11d9dad9a-proxy-tls\") pod \"machine-config-operator-74547568cd-rcnrn\" (UID: \"2fee22d4-2e9b-4768-9603-22e11d9dad9a\") " pod="openshift-machine-config-operator/machine-config-operator-74547568cd-rcnrn" Dec 10 22:51:53 crc kubenswrapper[4791]: I1210 22:51:53.728310 4791 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"package-server-manager-serving-cert\" (UniqueName: \"kubernetes.io/secret/3f132df5-3de5-4455-a3da-7848b0de69cc-package-server-manager-serving-cert\") pod \"package-server-manager-789f6589d5-w5dpr\" (UID: \"3f132df5-3de5-4455-a3da-7848b0de69cc\") " pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-w5dpr" Dec 10 22:51:53 crc kubenswrapper[4791]: I1210 22:51:53.728405 4791 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-tjc7n\" (UniqueName: \"kubernetes.io/projected/169ebb73-7b32-4f87-a8d7-9272bff04856-kube-api-access-tjc7n\") pod \"kube-storage-version-migrator-operator-b67b599dd-k7qsx\" (UID: \"169ebb73-7b32-4f87-a8d7-9272bff04856\") " pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-k7qsx" Dec 10 22:51:53 crc kubenswrapper[4791]: I1210 22:51:53.728434 4791 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/0921ef8b-8842-4d66-890b-3ad15bd48e70-cert\") pod \"ingress-canary-9dwf5\" (UID: \"0921ef8b-8842-4d66-890b-3ad15bd48e70\") " pod="openshift-ingress-canary/ingress-canary-9dwf5" Dec 10 22:51:53 crc kubenswrapper[4791]: I1210 22:51:53.728478 4791 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"profile-collector-cert\" (UniqueName: \"kubernetes.io/secret/06eb2b28-22d6-4116-9ce4-1abf7a21db36-profile-collector-cert\") pod \"olm-operator-6b444d44fb-bws6n\" (UID: \"06eb2b28-22d6-4116-9ce4-1abf7a21db36\") " pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-bws6n" Dec 10 22:51:53 crc kubenswrapper[4791]: I1210 22:51:53.728553 4791 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"images\" (UniqueName: \"kubernetes.io/configmap/2fee22d4-2e9b-4768-9603-22e11d9dad9a-images\") pod \"machine-config-operator-74547568cd-rcnrn\" (UID: \"2fee22d4-2e9b-4768-9603-22e11d9dad9a\") " pod="openshift-machine-config-operator/machine-config-operator-74547568cd-rcnrn" Dec 10 22:51:53 crc kubenswrapper[4791]: I1210 22:51:53.728578 4791 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/8f993103-9ee4-446f-a91a-a03dcd3ed9ea-config\") pod \"openshift-kube-scheduler-operator-5fdd9b5758-z6gpf\" (UID: \"8f993103-9ee4-446f-a91a-a03dcd3ed9ea\") " pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-z6gpf" Dec 10 22:51:53 crc kubenswrapper[4791]: I1210 22:51:53.728663 4791 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/64199728-979b-4628-8b37-2dd792f86439-config\") pod \"service-ca-operator-777779d784-b8v4c\" (UID: \"64199728-979b-4628-8b37-2dd792f86439\") " pod="openshift-service-ca-operator/service-ca-operator-777779d784-b8v4c" Dec 10 22:51:53 crc kubenswrapper[4791]: I1210 22:51:53.728705 4791 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/97c517e6-0af1-4c17-a2b9-f7a7db3368b3-config\") pod \"etcd-operator-b45778765-nkt8g\" (UID: \"97c517e6-0af1-4c17-a2b9-f7a7db3368b3\") " pod="openshift-etcd-operator/etcd-operator-b45778765-nkt8g" Dec 10 22:51:53 crc kubenswrapper[4791]: I1210 22:51:53.728725 4791 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etcd-ca\" (UniqueName: \"kubernetes.io/configmap/97c517e6-0af1-4c17-a2b9-f7a7db3368b3-etcd-ca\") pod \"etcd-operator-b45778765-nkt8g\" (UID: \"97c517e6-0af1-4c17-a2b9-f7a7db3368b3\") " pod="openshift-etcd-operator/etcd-operator-b45778765-nkt8g" Dec 10 22:51:53 crc kubenswrapper[4791]: I1210 22:51:53.728747 4791 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/169ebb73-7b32-4f87-a8d7-9272bff04856-config\") pod \"kube-storage-version-migrator-operator-b67b599dd-k7qsx\" (UID: \"169ebb73-7b32-4f87-a8d7-9272bff04856\") " pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-k7qsx" Dec 10 22:51:53 crc kubenswrapper[4791]: I1210 22:51:53.728769 4791 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/6e58f465-ef21-4b20-9eb8-5ade4810688a-config\") pod \"kube-apiserver-operator-766d6c64bb-vm842\" (UID: \"6e58f465-ef21-4b20-9eb8-5ade4810688a\") " pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-vm842" Dec 10 22:51:53 crc kubenswrapper[4791]: E1210 22:51:53.730845 4791 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-10 22:51:54.230827964 +0000 UTC m=+148.660445637 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-srd52" (UID: "e93dbd5d-0ca7-4a15-9eb3-0d71595fada0") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 10 22:51:53 crc kubenswrapper[4791]: I1210 22:51:53.731655 4791 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-gb99d\" (UniqueName: \"kubernetes.io/projected/64199728-979b-4628-8b37-2dd792f86439-kube-api-access-gb99d\") pod \"service-ca-operator-777779d784-b8v4c\" (UID: \"64199728-979b-4628-8b37-2dd792f86439\") " pod="openshift-service-ca-operator/service-ca-operator-777779d784-b8v4c" Dec 10 22:51:53 crc kubenswrapper[4791]: I1210 22:51:53.731706 4791 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"registration-dir\" (UniqueName: \"kubernetes.io/host-path/1666d5bb-295d-4828-868e-b1b17b5741bb-registration-dir\") pod \"csi-hostpathplugin-mkz5r\" (UID: \"1666d5bb-295d-4828-868e-b1b17b5741bb\") " pod="hostpath-provisioner/csi-hostpathplugin-mkz5r" Dec 10 22:51:53 crc kubenswrapper[4791]: I1210 22:51:53.731768 4791 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"plugins-dir\" (UniqueName: \"kubernetes.io/host-path/1666d5bb-295d-4828-868e-b1b17b5741bb-plugins-dir\") pod \"csi-hostpathplugin-mkz5r\" (UID: \"1666d5bb-295d-4828-868e-b1b17b5741bb\") " pod="hostpath-provisioner/csi-hostpathplugin-mkz5r" Dec 10 22:51:53 crc kubenswrapper[4791]: I1210 22:51:53.731816 4791 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/2fee22d4-2e9b-4768-9603-22e11d9dad9a-auth-proxy-config\") pod \"machine-config-operator-74547568cd-rcnrn\" (UID: \"2fee22d4-2e9b-4768-9603-22e11d9dad9a\") " pod="openshift-machine-config-operator/machine-config-operator-74547568cd-rcnrn" Dec 10 22:51:53 crc kubenswrapper[4791]: I1210 22:51:53.731930 4791 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-4jd6j\" (UniqueName: \"kubernetes.io/projected/93a38a79-5c1c-4296-849e-94137de8f974-kube-api-access-4jd6j\") pod \"openshift-apiserver-operator-796bbdcf4f-rmjd2\" (UID: \"93a38a79-5c1c-4296-849e-94137de8f974\") " pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-rmjd2" Dec 10 22:51:53 crc kubenswrapper[4791]: I1210 22:51:53.738765 4791 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"srv-cert\" (UniqueName: \"kubernetes.io/secret/06eb2b28-22d6-4116-9ce4-1abf7a21db36-srv-cert\") pod \"olm-operator-6b444d44fb-bws6n\" (UID: \"06eb2b28-22d6-4116-9ce4-1abf7a21db36\") " pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-bws6n" Dec 10 22:51:53 crc kubenswrapper[4791]: I1210 22:51:53.739348 4791 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"profile-collector-cert\" (UniqueName: \"kubernetes.io/secret/06eb2b28-22d6-4116-9ce4-1abf7a21db36-profile-collector-cert\") pod \"olm-operator-6b444d44fb-bws6n\" (UID: \"06eb2b28-22d6-4116-9ce4-1abf7a21db36\") " pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-bws6n" Dec 10 22:51:53 crc kubenswrapper[4791]: I1210 22:51:53.783777 4791 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-g7ll5\" (UniqueName: \"kubernetes.io/projected/06eb2b28-22d6-4116-9ce4-1abf7a21db36-kube-api-access-g7ll5\") pod \"olm-operator-6b444d44fb-bws6n\" (UID: \"06eb2b28-22d6-4116-9ce4-1abf7a21db36\") " pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-bws6n" Dec 10 22:51:53 crc kubenswrapper[4791]: I1210 22:51:53.806394 4791 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-cluster-machine-approver/machine-approver-56656f9798-m6dwv" event={"ID":"2f422ac4-b586-4056-8748-94b7de0b0248","Type":"ContainerStarted","Data":"25ae148598b40f33979b7a099c88f67ce6b4d0ebdd60a1c1cb14c3b696f6dd81"} Dec 10 22:51:53 crc kubenswrapper[4791]: I1210 22:51:53.808358 4791 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ingress/router-default-5444994796-p97fc" event={"ID":"d3de1ba0-b014-4f98-b597-cc89f5a70364","Type":"ContainerStarted","Data":"48223eab798967c7f251641f333f95856cbefeae4287c40c23a77a5354ee96a9"} Dec 10 22:51:53 crc kubenswrapper[4791]: I1210 22:51:53.810567 4791 generic.go:334] "Generic (PLEG): container finished" podID="4a1cf441-8a08-4677-a480-947c8cddd6a6" containerID="dab9bc926a4ba6260270be4cec6e18a38afb4f63e8c07a708b0e9875ed9d8c95" exitCode=0 Dec 10 22:51:53 crc kubenswrapper[4791]: I1210 22:51:53.810654 4791 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-apiserver/apiserver-76f77b778f-kqtbx" event={"ID":"4a1cf441-8a08-4677-a480-947c8cddd6a6","Type":"ContainerDied","Data":"dab9bc926a4ba6260270be4cec6e18a38afb4f63e8c07a708b0e9875ed9d8c95"} Dec 10 22:51:53 crc kubenswrapper[4791]: I1210 22:51:53.814034 4791 generic.go:334] "Generic (PLEG): container finished" podID="3ba17b70-17df-4f8c-87ff-c15e03ce06bc" containerID="cd2654939aa182ad2d5abfcab63a9879e3d20b86f94ee31691bda805cd2a6b13" exitCode=0 Dec 10 22:51:53 crc kubenswrapper[4791]: I1210 22:51:53.814171 4791 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-nwwpb" event={"ID":"3ba17b70-17df-4f8c-87ff-c15e03ce06bc","Type":"ContainerDied","Data":"cd2654939aa182ad2d5abfcab63a9879e3d20b86f94ee31691bda805cd2a6b13"} Dec 10 22:51:53 crc kubenswrapper[4791]: I1210 22:51:53.814201 4791 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-nwwpb" event={"ID":"3ba17b70-17df-4f8c-87ff-c15e03ce06bc","Type":"ContainerStarted","Data":"7d64cfcb802f72469e71158be1a33b1e7b5d15b33e104489a30c1f005b01d663"} Dec 10 22:51:53 crc kubenswrapper[4791]: I1210 22:51:53.815675 4791 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-api/machine-api-operator-5694c8668f-cpbhz" event={"ID":"cd65e91e-5b02-4f78-829e-93b4596ff8f6","Type":"ContainerStarted","Data":"2ea06dd446ec5e27cfcd9fa306ba103075c97dd1f774881d4c41865972a77bbe"} Dec 10 22:51:53 crc kubenswrapper[4791]: I1210 22:51:53.832922 4791 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 10 22:51:53 crc kubenswrapper[4791]: I1210 22:51:53.833123 4791 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/e9978f48-c33a-4765-9b05-16e37f5cdb24-config-volume\") pod \"dns-default-qs77n\" (UID: \"e9978f48-c33a-4765-9b05-16e37f5cdb24\") " pod="openshift-dns/dns-default-qs77n" Dec 10 22:51:53 crc kubenswrapper[4791]: I1210 22:51:53.833162 4791 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-z25fj\" (UniqueName: \"kubernetes.io/projected/0921ef8b-8842-4d66-890b-3ad15bd48e70-kube-api-access-z25fj\") pod \"ingress-canary-9dwf5\" (UID: \"0921ef8b-8842-4d66-890b-3ad15bd48e70\") " pod="openshift-ingress-canary/ingress-canary-9dwf5" Dec 10 22:51:53 crc kubenswrapper[4791]: I1210 22:51:53.833186 4791 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"csi-data-dir\" (UniqueName: \"kubernetes.io/host-path/1666d5bb-295d-4828-868e-b1b17b5741bb-csi-data-dir\") pod \"csi-hostpathplugin-mkz5r\" (UID: \"1666d5bb-295d-4828-868e-b1b17b5741bb\") " pod="hostpath-provisioner/csi-hostpathplugin-mkz5r" Dec 10 22:51:53 crc kubenswrapper[4791]: I1210 22:51:53.833209 4791 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-dk24r\" (UniqueName: \"kubernetes.io/projected/1666d5bb-295d-4828-868e-b1b17b5741bb-kube-api-access-dk24r\") pod \"csi-hostpathplugin-mkz5r\" (UID: \"1666d5bb-295d-4828-868e-b1b17b5741bb\") " pod="hostpath-provisioner/csi-hostpathplugin-mkz5r" Dec 10 22:51:53 crc kubenswrapper[4791]: I1210 22:51:53.833236 4791 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/e9978f48-c33a-4765-9b05-16e37f5cdb24-metrics-tls\") pod \"dns-default-qs77n\" (UID: \"e9978f48-c33a-4765-9b05-16e37f5cdb24\") " pod="openshift-dns/dns-default-qs77n" Dec 10 22:51:53 crc kubenswrapper[4791]: E1210 22:51:53.833267 4791 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-10 22:51:54.333232898 +0000 UTC m=+148.762850531 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 10 22:51:53 crc kubenswrapper[4791]: I1210 22:51:53.833316 4791 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-pbhrm\" (UniqueName: \"kubernetes.io/projected/2fee22d4-2e9b-4768-9603-22e11d9dad9a-kube-api-access-pbhrm\") pod \"machine-config-operator-74547568cd-rcnrn\" (UID: \"2fee22d4-2e9b-4768-9603-22e11d9dad9a\") " pod="openshift-machine-config-operator/machine-config-operator-74547568cd-rcnrn" Dec 10 22:51:53 crc kubenswrapper[4791]: I1210 22:51:53.833395 4791 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"signing-cabundle\" (UniqueName: \"kubernetes.io/configmap/34da0cc5-3e59-4c25-af94-e755e772b89a-signing-cabundle\") pod \"service-ca-9c57cc56f-x9pw8\" (UID: \"34da0cc5-3e59-4c25-af94-e755e772b89a\") " pod="openshift-service-ca/service-ca-9c57cc56f-x9pw8" Dec 10 22:51:53 crc kubenswrapper[4791]: I1210 22:51:53.833433 4791 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/97c517e6-0af1-4c17-a2b9-f7a7db3368b3-serving-cert\") pod \"etcd-operator-b45778765-nkt8g\" (UID: \"97c517e6-0af1-4c17-a2b9-f7a7db3368b3\") " pod="openshift-etcd-operator/etcd-operator-b45778765-nkt8g" Dec 10 22:51:53 crc kubenswrapper[4791]: I1210 22:51:53.833450 4791 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"csi-data-dir\" (UniqueName: \"kubernetes.io/host-path/1666d5bb-295d-4828-868e-b1b17b5741bb-csi-data-dir\") pod \"csi-hostpathplugin-mkz5r\" (UID: \"1666d5bb-295d-4828-868e-b1b17b5741bb\") " pod="hostpath-provisioner/csi-hostpathplugin-mkz5r" Dec 10 22:51:53 crc kubenswrapper[4791]: I1210 22:51:53.833480 4791 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/8f993103-9ee4-446f-a91a-a03dcd3ed9ea-serving-cert\") pod \"openshift-kube-scheduler-operator-5fdd9b5758-z6gpf\" (UID: \"8f993103-9ee4-446f-a91a-a03dcd3ed9ea\") " pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-z6gpf" Dec 10 22:51:53 crc kubenswrapper[4791]: I1210 22:51:53.833582 4791 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-tm8cv\" (UniqueName: \"kubernetes.io/projected/97c517e6-0af1-4c17-a2b9-f7a7db3368b3-kube-api-access-tm8cv\") pod \"etcd-operator-b45778765-nkt8g\" (UID: \"97c517e6-0af1-4c17-a2b9-f7a7db3368b3\") " pod="openshift-etcd-operator/etcd-operator-b45778765-nkt8g" Dec 10 22:51:53 crc kubenswrapper[4791]: I1210 22:51:53.833666 4791 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-t6768\" (UniqueName: \"kubernetes.io/projected/ded98820-5308-46e3-b87b-18d0a81c0519-kube-api-access-t6768\") pod \"multus-admission-controller-857f4d67dd-fbpcr\" (UID: \"ded98820-5308-46e3-b87b-18d0a81c0519\") " pod="openshift-multus/multus-admission-controller-857f4d67dd-fbpcr" Dec 10 22:51:53 crc kubenswrapper[4791]: I1210 22:51:53.833737 4791 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-bdjr4\" (UniqueName: \"kubernetes.io/projected/a05f3579-4703-4c46-bb11-5de60eedbc02-kube-api-access-bdjr4\") pod \"packageserver-d55dfcdfc-v445v\" (UID: \"a05f3579-4703-4c46-bb11-5de60eedbc02\") " pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-v445v" Dec 10 22:51:53 crc kubenswrapper[4791]: I1210 22:51:53.833795 4791 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"tmpfs\" (UniqueName: \"kubernetes.io/empty-dir/a05f3579-4703-4c46-bb11-5de60eedbc02-tmpfs\") pod \"packageserver-d55dfcdfc-v445v\" (UID: \"a05f3579-4703-4c46-bb11-5de60eedbc02\") " pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-v445v" Dec 10 22:51:53 crc kubenswrapper[4791]: I1210 22:51:53.835585 4791 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/90640468-53d7-42bc-a835-f7acf48a486e-secret-volume\") pod \"collect-profiles-29423445-b9cxb\" (UID: \"90640468-53d7-42bc-a835-f7acf48a486e\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29423445-b9cxb" Dec 10 22:51:53 crc kubenswrapper[4791]: I1210 22:51:53.835628 4791 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-n8jwh\" (UniqueName: \"kubernetes.io/projected/868d6095-79bd-41fc-ae14-f369fdf68fde-kube-api-access-n8jwh\") pod \"migrator-59844c95c7-vscpm\" (UID: \"868d6095-79bd-41fc-ae14-f369fdf68fde\") " pod="openshift-kube-storage-version-migrator/migrator-59844c95c7-vscpm" Dec 10 22:51:53 crc kubenswrapper[4791]: I1210 22:51:53.835658 4791 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"marketplace-trusted-ca\" (UniqueName: \"kubernetes.io/configmap/a9185ee1-adaf-48e8-b9a9-90ef658f8212-marketplace-trusted-ca\") pod \"marketplace-operator-79b997595-7g828\" (UID: \"a9185ee1-adaf-48e8-b9a9-90ef658f8212\") " pod="openshift-marketplace/marketplace-operator-79b997595-7g828" Dec 10 22:51:53 crc kubenswrapper[4791]: I1210 22:51:53.834262 4791 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-controller-manager/controller-manager-879f6c89f-9tg4p"] Dec 10 22:51:53 crc kubenswrapper[4791]: I1210 22:51:53.835685 4791 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/93a38a79-5c1c-4296-849e-94137de8f974-config\") pod \"openshift-apiserver-operator-796bbdcf4f-rmjd2\" (UID: \"93a38a79-5c1c-4296-849e-94137de8f974\") " pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-rmjd2" Dec 10 22:51:53 crc kubenswrapper[4791]: I1210 22:51:53.835716 4791 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/6e58f465-ef21-4b20-9eb8-5ade4810688a-serving-cert\") pod \"kube-apiserver-operator-766d6c64bb-vm842\" (UID: \"6e58f465-ef21-4b20-9eb8-5ade4810688a\") " pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-vm842" Dec 10 22:51:53 crc kubenswrapper[4791]: I1210 22:51:53.834066 4791 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/e9978f48-c33a-4765-9b05-16e37f5cdb24-config-volume\") pod \"dns-default-qs77n\" (UID: \"e9978f48-c33a-4765-9b05-16e37f5cdb24\") " pod="openshift-dns/dns-default-qs77n" Dec 10 22:51:53 crc kubenswrapper[4791]: I1210 22:51:53.834483 4791 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"tmpfs\" (UniqueName: \"kubernetes.io/empty-dir/a05f3579-4703-4c46-bb11-5de60eedbc02-tmpfs\") pod \"packageserver-d55dfcdfc-v445v\" (UID: \"a05f3579-4703-4c46-bb11-5de60eedbc02\") " pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-v445v" Dec 10 22:51:53 crc kubenswrapper[4791]: I1210 22:51:53.839423 4791 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/97c517e6-0af1-4c17-a2b9-f7a7db3368b3-serving-cert\") pod \"etcd-operator-b45778765-nkt8g\" (UID: \"97c517e6-0af1-4c17-a2b9-f7a7db3368b3\") " pod="openshift-etcd-operator/etcd-operator-b45778765-nkt8g" Dec 10 22:51:53 crc kubenswrapper[4791]: I1210 22:51:53.834778 4791 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"signing-cabundle\" (UniqueName: \"kubernetes.io/configmap/34da0cc5-3e59-4c25-af94-e755e772b89a-signing-cabundle\") pod \"service-ca-9c57cc56f-x9pw8\" (UID: \"34da0cc5-3e59-4c25-af94-e755e772b89a\") " pod="openshift-service-ca/service-ca-9c57cc56f-x9pw8" Dec 10 22:51:53 crc kubenswrapper[4791]: I1210 22:51:53.839843 4791 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"signing-key\" (UniqueName: \"kubernetes.io/secret/34da0cc5-3e59-4c25-af94-e755e772b89a-signing-key\") pod \"service-ca-9c57cc56f-x9pw8\" (UID: \"34da0cc5-3e59-4c25-af94-e755e772b89a\") " pod="openshift-service-ca/service-ca-9c57cc56f-x9pw8" Dec 10 22:51:53 crc kubenswrapper[4791]: I1210 22:51:53.839921 4791 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-jpwld\" (UniqueName: \"kubernetes.io/projected/c6c773c2-29e4-477d-91fb-f63052272f06-kube-api-access-jpwld\") pod \"control-plane-machine-set-operator-78cbb6b69f-ms8jw\" (UID: \"c6c773c2-29e4-477d-91fb-f63052272f06\") " pod="openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-ms8jw" Dec 10 22:51:53 crc kubenswrapper[4791]: I1210 22:51:53.839958 4791 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/5a2896d3-fca8-4c64-abeb-6320f0b78d4e-metrics-tls\") pod \"dns-operator-744455d44c-cggs2\" (UID: \"5a2896d3-fca8-4c64-abeb-6320f0b78d4e\") " pod="openshift-dns-operator/dns-operator-744455d44c-cggs2" Dec 10 22:51:53 crc kubenswrapper[4791]: I1210 22:51:53.840012 4791 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/97c517e6-0af1-4c17-a2b9-f7a7db3368b3-etcd-client\") pod \"etcd-operator-b45778765-nkt8g\" (UID: \"97c517e6-0af1-4c17-a2b9-f7a7db3368b3\") " pod="openshift-etcd-operator/etcd-operator-b45778765-nkt8g" Dec 10 22:51:53 crc kubenswrapper[4791]: I1210 22:51:53.840045 4791 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"node-bootstrap-token\" (UniqueName: \"kubernetes.io/secret/36532144-441a-4c12-a436-59d93d5ec3b7-node-bootstrap-token\") pod \"machine-config-server-ktcmn\" (UID: \"36532144-441a-4c12-a436-59d93d5ec3b7\") " pod="openshift-machine-config-operator/machine-config-server-ktcmn" Dec 10 22:51:53 crc kubenswrapper[4791]: I1210 22:51:53.840083 4791 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/64199728-979b-4628-8b37-2dd792f86439-serving-cert\") pod \"service-ca-operator-777779d784-b8v4c\" (UID: \"64199728-979b-4628-8b37-2dd792f86439\") " pod="openshift-service-ca-operator/service-ca-operator-777779d784-b8v4c" Dec 10 22:51:53 crc kubenswrapper[4791]: I1210 22:51:53.840118 4791 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"certs\" (UniqueName: \"kubernetes.io/secret/36532144-441a-4c12-a436-59d93d5ec3b7-certs\") pod \"machine-config-server-ktcmn\" (UID: \"36532144-441a-4c12-a436-59d93d5ec3b7\") " pod="openshift-machine-config-operator/machine-config-server-ktcmn" Dec 10 22:51:53 crc kubenswrapper[4791]: I1210 22:51:53.840149 4791 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"mountpoint-dir\" (UniqueName: \"kubernetes.io/host-path/1666d5bb-295d-4828-868e-b1b17b5741bb-mountpoint-dir\") pod \"csi-hostpathplugin-mkz5r\" (UID: \"1666d5bb-295d-4828-868e-b1b17b5741bb\") " pod="hostpath-provisioner/csi-hostpathplugin-mkz5r" Dec 10 22:51:53 crc kubenswrapper[4791]: I1210 22:51:53.840181 4791 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-mnnjz\" (UniqueName: \"kubernetes.io/projected/36532144-441a-4c12-a436-59d93d5ec3b7-kube-api-access-mnnjz\") pod \"machine-config-server-ktcmn\" (UID: \"36532144-441a-4c12-a436-59d93d5ec3b7\") " pod="openshift-machine-config-operator/machine-config-server-ktcmn" Dec 10 22:51:53 crc kubenswrapper[4791]: I1210 22:51:53.840217 4791 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"webhook-certs\" (UniqueName: \"kubernetes.io/secret/ded98820-5308-46e3-b87b-18d0a81c0519-webhook-certs\") pod \"multus-admission-controller-857f4d67dd-fbpcr\" (UID: \"ded98820-5308-46e3-b87b-18d0a81c0519\") " pod="openshift-multus/multus-admission-controller-857f4d67dd-fbpcr" Dec 10 22:51:53 crc kubenswrapper[4791]: I1210 22:51:53.840257 4791 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/93a38a79-5c1c-4296-849e-94137de8f974-serving-cert\") pod \"openshift-apiserver-operator-796bbdcf4f-rmjd2\" (UID: \"93a38a79-5c1c-4296-849e-94137de8f974\") " pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-rmjd2" Dec 10 22:51:53 crc kubenswrapper[4791]: I1210 22:51:53.840375 4791 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-gnhrf\" (UniqueName: \"kubernetes.io/projected/90640468-53d7-42bc-a835-f7acf48a486e-kube-api-access-gnhrf\") pod \"collect-profiles-29423445-b9cxb\" (UID: \"90640468-53d7-42bc-a835-f7acf48a486e\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29423445-b9cxb" Dec 10 22:51:53 crc kubenswrapper[4791]: I1210 22:51:53.840415 4791 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-srd52\" (UID: \"e93dbd5d-0ca7-4a15-9eb3-0d71595fada0\") " pod="openshift-image-registry/image-registry-697d97f7c8-srd52" Dec 10 22:51:53 crc kubenswrapper[4791]: I1210 22:51:53.840441 4791 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-hsmgd\" (UniqueName: \"kubernetes.io/projected/a9185ee1-adaf-48e8-b9a9-90ef658f8212-kube-api-access-hsmgd\") pod \"marketplace-operator-79b997595-7g828\" (UID: \"a9185ee1-adaf-48e8-b9a9-90ef658f8212\") " pod="openshift-marketplace/marketplace-operator-79b997595-7g828" Dec 10 22:51:53 crc kubenswrapper[4791]: I1210 22:51:53.840464 4791 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/6e58f465-ef21-4b20-9eb8-5ade4810688a-kube-api-access\") pod \"kube-apiserver-operator-766d6c64bb-vm842\" (UID: \"6e58f465-ef21-4b20-9eb8-5ade4810688a\") " pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-vm842" Dec 10 22:51:53 crc kubenswrapper[4791]: I1210 22:51:53.840466 4791 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/93a38a79-5c1c-4296-849e-94137de8f974-config\") pod \"openshift-apiserver-operator-796bbdcf4f-rmjd2\" (UID: \"93a38a79-5c1c-4296-849e-94137de8f974\") " pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-rmjd2" Dec 10 22:51:53 crc kubenswrapper[4791]: I1210 22:51:53.840501 4791 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-xcxbr\" (UniqueName: \"kubernetes.io/projected/e9978f48-c33a-4765-9b05-16e37f5cdb24-kube-api-access-xcxbr\") pod \"dns-default-qs77n\" (UID: \"e9978f48-c33a-4765-9b05-16e37f5cdb24\") " pod="openshift-dns/dns-default-qs77n" Dec 10 22:51:53 crc kubenswrapper[4791]: I1210 22:51:53.840524 4791 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"marketplace-operator-metrics\" (UniqueName: \"kubernetes.io/secret/a9185ee1-adaf-48e8-b9a9-90ef658f8212-marketplace-operator-metrics\") pod \"marketplace-operator-79b997595-7g828\" (UID: \"a9185ee1-adaf-48e8-b9a9-90ef658f8212\") " pod="openshift-marketplace/marketplace-operator-79b997595-7g828" Dec 10 22:51:53 crc kubenswrapper[4791]: I1210 22:51:53.840524 4791 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"marketplace-trusted-ca\" (UniqueName: \"kubernetes.io/configmap/a9185ee1-adaf-48e8-b9a9-90ef658f8212-marketplace-trusted-ca\") pod \"marketplace-operator-79b997595-7g828\" (UID: \"a9185ee1-adaf-48e8-b9a9-90ef658f8212\") " pod="openshift-marketplace/marketplace-operator-79b997595-7g828" Dec 10 22:51:53 crc kubenswrapper[4791]: I1210 22:51:53.840583 4791 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"control-plane-machine-set-operator-tls\" (UniqueName: \"kubernetes.io/secret/c6c773c2-29e4-477d-91fb-f63052272f06-control-plane-machine-set-operator-tls\") pod \"control-plane-machine-set-operator-78cbb6b69f-ms8jw\" (UID: \"c6c773c2-29e4-477d-91fb-f63052272f06\") " pod="openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-ms8jw" Dec 10 22:51:53 crc kubenswrapper[4791]: I1210 22:51:53.840692 4791 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-fwbdr\" (UniqueName: \"kubernetes.io/projected/5a2896d3-fca8-4c64-abeb-6320f0b78d4e-kube-api-access-fwbdr\") pod \"dns-operator-744455d44c-cggs2\" (UID: \"5a2896d3-fca8-4c64-abeb-6320f0b78d4e\") " pod="openshift-dns-operator/dns-operator-744455d44c-cggs2" Dec 10 22:51:53 crc kubenswrapper[4791]: I1210 22:51:53.840762 4791 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/a05f3579-4703-4c46-bb11-5de60eedbc02-webhook-cert\") pod \"packageserver-d55dfcdfc-v445v\" (UID: \"a05f3579-4703-4c46-bb11-5de60eedbc02\") " pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-v445v" Dec 10 22:51:53 crc kubenswrapper[4791]: I1210 22:51:53.840806 4791 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-87746\" (UniqueName: \"kubernetes.io/projected/34da0cc5-3e59-4c25-af94-e755e772b89a-kube-api-access-87746\") pod \"service-ca-9c57cc56f-x9pw8\" (UID: \"34da0cc5-3e59-4c25-af94-e755e772b89a\") " pod="openshift-service-ca/service-ca-9c57cc56f-x9pw8" Dec 10 22:51:53 crc kubenswrapper[4791]: I1210 22:51:53.840836 4791 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"socket-dir\" (UniqueName: \"kubernetes.io/host-path/1666d5bb-295d-4828-868e-b1b17b5741bb-socket-dir\") pod \"csi-hostpathplugin-mkz5r\" (UID: \"1666d5bb-295d-4828-868e-b1b17b5741bb\") " pod="hostpath-provisioner/csi-hostpathplugin-mkz5r" Dec 10 22:51:53 crc kubenswrapper[4791]: I1210 22:51:53.840874 4791 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etcd-service-ca\" (UniqueName: \"kubernetes.io/configmap/97c517e6-0af1-4c17-a2b9-f7a7db3368b3-etcd-service-ca\") pod \"etcd-operator-b45778765-nkt8g\" (UID: \"97c517e6-0af1-4c17-a2b9-f7a7db3368b3\") " pod="openshift-etcd-operator/etcd-operator-b45778765-nkt8g" Dec 10 22:51:53 crc kubenswrapper[4791]: I1210 22:51:53.840913 4791 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/169ebb73-7b32-4f87-a8d7-9272bff04856-serving-cert\") pod \"kube-storage-version-migrator-operator-b67b599dd-k7qsx\" (UID: \"169ebb73-7b32-4f87-a8d7-9272bff04856\") " pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-k7qsx" Dec 10 22:51:53 crc kubenswrapper[4791]: I1210 22:51:53.840986 4791 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/2fee22d4-2e9b-4768-9603-22e11d9dad9a-proxy-tls\") pod \"machine-config-operator-74547568cd-rcnrn\" (UID: \"2fee22d4-2e9b-4768-9603-22e11d9dad9a\") " pod="openshift-machine-config-operator/machine-config-operator-74547568cd-rcnrn" Dec 10 22:51:53 crc kubenswrapper[4791]: I1210 22:51:53.841021 4791 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"package-server-manager-serving-cert\" (UniqueName: \"kubernetes.io/secret/3f132df5-3de5-4455-a3da-7848b0de69cc-package-server-manager-serving-cert\") pod \"package-server-manager-789f6589d5-w5dpr\" (UID: \"3f132df5-3de5-4455-a3da-7848b0de69cc\") " pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-w5dpr" Dec 10 22:51:53 crc kubenswrapper[4791]: I1210 22:51:53.841287 4791 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-tjc7n\" (UniqueName: \"kubernetes.io/projected/169ebb73-7b32-4f87-a8d7-9272bff04856-kube-api-access-tjc7n\") pod \"kube-storage-version-migrator-operator-b67b599dd-k7qsx\" (UID: \"169ebb73-7b32-4f87-a8d7-9272bff04856\") " pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-k7qsx" Dec 10 22:51:53 crc kubenswrapper[4791]: I1210 22:51:53.841327 4791 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/0921ef8b-8842-4d66-890b-3ad15bd48e70-cert\") pod \"ingress-canary-9dwf5\" (UID: \"0921ef8b-8842-4d66-890b-3ad15bd48e70\") " pod="openshift-ingress-canary/ingress-canary-9dwf5" Dec 10 22:51:53 crc kubenswrapper[4791]: I1210 22:51:53.841513 4791 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/8f993103-9ee4-446f-a91a-a03dcd3ed9ea-serving-cert\") pod \"openshift-kube-scheduler-operator-5fdd9b5758-z6gpf\" (UID: \"8f993103-9ee4-446f-a91a-a03dcd3ed9ea\") " pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-z6gpf" Dec 10 22:51:53 crc kubenswrapper[4791]: I1210 22:51:53.841783 4791 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"images\" (UniqueName: \"kubernetes.io/configmap/2fee22d4-2e9b-4768-9603-22e11d9dad9a-images\") pod \"machine-config-operator-74547568cd-rcnrn\" (UID: \"2fee22d4-2e9b-4768-9603-22e11d9dad9a\") " pod="openshift-machine-config-operator/machine-config-operator-74547568cd-rcnrn" Dec 10 22:51:53 crc kubenswrapper[4791]: I1210 22:51:53.841824 4791 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/8f993103-9ee4-446f-a91a-a03dcd3ed9ea-config\") pod \"openshift-kube-scheduler-operator-5fdd9b5758-z6gpf\" (UID: \"8f993103-9ee4-446f-a91a-a03dcd3ed9ea\") " pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-z6gpf" Dec 10 22:51:53 crc kubenswrapper[4791]: I1210 22:51:53.841876 4791 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/64199728-979b-4628-8b37-2dd792f86439-config\") pod \"service-ca-operator-777779d784-b8v4c\" (UID: \"64199728-979b-4628-8b37-2dd792f86439\") " pod="openshift-service-ca-operator/service-ca-operator-777779d784-b8v4c" Dec 10 22:51:53 crc kubenswrapper[4791]: I1210 22:51:53.841910 4791 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/97c517e6-0af1-4c17-a2b9-f7a7db3368b3-config\") pod \"etcd-operator-b45778765-nkt8g\" (UID: \"97c517e6-0af1-4c17-a2b9-f7a7db3368b3\") " pod="openshift-etcd-operator/etcd-operator-b45778765-nkt8g" Dec 10 22:51:53 crc kubenswrapper[4791]: I1210 22:51:53.842124 4791 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etcd-ca\" (UniqueName: \"kubernetes.io/configmap/97c517e6-0af1-4c17-a2b9-f7a7db3368b3-etcd-ca\") pod \"etcd-operator-b45778765-nkt8g\" (UID: \"97c517e6-0af1-4c17-a2b9-f7a7db3368b3\") " pod="openshift-etcd-operator/etcd-operator-b45778765-nkt8g" Dec 10 22:51:53 crc kubenswrapper[4791]: I1210 22:51:53.842157 4791 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/169ebb73-7b32-4f87-a8d7-9272bff04856-config\") pod \"kube-storage-version-migrator-operator-b67b599dd-k7qsx\" (UID: \"169ebb73-7b32-4f87-a8d7-9272bff04856\") " pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-k7qsx" Dec 10 22:51:53 crc kubenswrapper[4791]: I1210 22:51:53.842190 4791 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/6e58f465-ef21-4b20-9eb8-5ade4810688a-config\") pod \"kube-apiserver-operator-766d6c64bb-vm842\" (UID: \"6e58f465-ef21-4b20-9eb8-5ade4810688a\") " pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-vm842" Dec 10 22:51:53 crc kubenswrapper[4791]: I1210 22:51:53.842234 4791 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-gb99d\" (UniqueName: \"kubernetes.io/projected/64199728-979b-4628-8b37-2dd792f86439-kube-api-access-gb99d\") pod \"service-ca-operator-777779d784-b8v4c\" (UID: \"64199728-979b-4628-8b37-2dd792f86439\") " pod="openshift-service-ca-operator/service-ca-operator-777779d784-b8v4c" Dec 10 22:51:53 crc kubenswrapper[4791]: I1210 22:51:53.842266 4791 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"registration-dir\" (UniqueName: \"kubernetes.io/host-path/1666d5bb-295d-4828-868e-b1b17b5741bb-registration-dir\") pod \"csi-hostpathplugin-mkz5r\" (UID: \"1666d5bb-295d-4828-868e-b1b17b5741bb\") " pod="hostpath-provisioner/csi-hostpathplugin-mkz5r" Dec 10 22:51:53 crc kubenswrapper[4791]: I1210 22:51:53.842297 4791 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"plugins-dir\" (UniqueName: \"kubernetes.io/host-path/1666d5bb-295d-4828-868e-b1b17b5741bb-plugins-dir\") pod \"csi-hostpathplugin-mkz5r\" (UID: \"1666d5bb-295d-4828-868e-b1b17b5741bb\") " pod="hostpath-provisioner/csi-hostpathplugin-mkz5r" Dec 10 22:51:53 crc kubenswrapper[4791]: I1210 22:51:53.842421 4791 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/2fee22d4-2e9b-4768-9603-22e11d9dad9a-auth-proxy-config\") pod \"machine-config-operator-74547568cd-rcnrn\" (UID: \"2fee22d4-2e9b-4768-9603-22e11d9dad9a\") " pod="openshift-machine-config-operator/machine-config-operator-74547568cd-rcnrn" Dec 10 22:51:53 crc kubenswrapper[4791]: I1210 22:51:53.842453 4791 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-4jd6j\" (UniqueName: \"kubernetes.io/projected/93a38a79-5c1c-4296-849e-94137de8f974-kube-api-access-4jd6j\") pod \"openshift-apiserver-operator-796bbdcf4f-rmjd2\" (UID: \"93a38a79-5c1c-4296-849e-94137de8f974\") " pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-rmjd2" Dec 10 22:51:53 crc kubenswrapper[4791]: I1210 22:51:53.842626 4791 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-6mhcw\" (UniqueName: \"kubernetes.io/projected/3f132df5-3de5-4455-a3da-7848b0de69cc-kube-api-access-6mhcw\") pod \"package-server-manager-789f6589d5-w5dpr\" (UID: \"3f132df5-3de5-4455-a3da-7848b0de69cc\") " pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-w5dpr" Dec 10 22:51:53 crc kubenswrapper[4791]: I1210 22:51:53.842688 4791 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/90640468-53d7-42bc-a835-f7acf48a486e-config-volume\") pod \"collect-profiles-29423445-b9cxb\" (UID: \"90640468-53d7-42bc-a835-f7acf48a486e\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29423445-b9cxb" Dec 10 22:51:53 crc kubenswrapper[4791]: I1210 22:51:53.842698 4791 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"mountpoint-dir\" (UniqueName: \"kubernetes.io/host-path/1666d5bb-295d-4828-868e-b1b17b5741bb-mountpoint-dir\") pod \"csi-hostpathplugin-mkz5r\" (UID: \"1666d5bb-295d-4828-868e-b1b17b5741bb\") " pod="hostpath-provisioner/csi-hostpathplugin-mkz5r" Dec 10 22:51:53 crc kubenswrapper[4791]: I1210 22:51:53.842873 4791 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/8f993103-9ee4-446f-a91a-a03dcd3ed9ea-kube-api-access\") pod \"openshift-kube-scheduler-operator-5fdd9b5758-z6gpf\" (UID: \"8f993103-9ee4-446f-a91a-a03dcd3ed9ea\") " pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-z6gpf" Dec 10 22:51:53 crc kubenswrapper[4791]: I1210 22:51:53.842920 4791 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/a05f3579-4703-4c46-bb11-5de60eedbc02-apiservice-cert\") pod \"packageserver-d55dfcdfc-v445v\" (UID: \"a05f3579-4703-4c46-bb11-5de60eedbc02\") " pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-v445v" Dec 10 22:51:53 crc kubenswrapper[4791]: I1210 22:51:53.844527 4791 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"socket-dir\" (UniqueName: \"kubernetes.io/host-path/1666d5bb-295d-4828-868e-b1b17b5741bb-socket-dir\") pod \"csi-hostpathplugin-mkz5r\" (UID: \"1666d5bb-295d-4828-868e-b1b17b5741bb\") " pod="hostpath-provisioner/csi-hostpathplugin-mkz5r" Dec 10 22:51:53 crc kubenswrapper[4791]: E1210 22:51:53.844995 4791 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-10 22:51:54.34497842 +0000 UTC m=+148.774596033 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-srd52" (UID: "e93dbd5d-0ca7-4a15-9eb3-0d71595fada0") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 10 22:51:53 crc kubenswrapper[4791]: I1210 22:51:53.845908 4791 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etcd-service-ca\" (UniqueName: \"kubernetes.io/configmap/97c517e6-0af1-4c17-a2b9-f7a7db3368b3-etcd-service-ca\") pod \"etcd-operator-b45778765-nkt8g\" (UID: \"97c517e6-0af1-4c17-a2b9-f7a7db3368b3\") " pod="openshift-etcd-operator/etcd-operator-b45778765-nkt8g" Dec 10 22:51:53 crc kubenswrapper[4791]: I1210 22:51:53.846634 4791 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/6e58f465-ef21-4b20-9eb8-5ade4810688a-serving-cert\") pod \"kube-apiserver-operator-766d6c64bb-vm842\" (UID: \"6e58f465-ef21-4b20-9eb8-5ade4810688a\") " pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-vm842" Dec 10 22:51:53 crc kubenswrapper[4791]: I1210 22:51:53.846789 4791 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/6e58f465-ef21-4b20-9eb8-5ade4810688a-config\") pod \"kube-apiserver-operator-766d6c64bb-vm842\" (UID: \"6e58f465-ef21-4b20-9eb8-5ade4810688a\") " pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-vm842" Dec 10 22:51:53 crc kubenswrapper[4791]: I1210 22:51:53.847228 4791 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"plugins-dir\" (UniqueName: \"kubernetes.io/host-path/1666d5bb-295d-4828-868e-b1b17b5741bb-plugins-dir\") pod \"csi-hostpathplugin-mkz5r\" (UID: \"1666d5bb-295d-4828-868e-b1b17b5741bb\") " pod="hostpath-provisioner/csi-hostpathplugin-mkz5r" Dec 10 22:51:53 crc kubenswrapper[4791]: I1210 22:51:53.847592 4791 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/64199728-979b-4628-8b37-2dd792f86439-config\") pod \"service-ca-operator-777779d784-b8v4c\" (UID: \"64199728-979b-4628-8b37-2dd792f86439\") " pod="openshift-service-ca-operator/service-ca-operator-777779d784-b8v4c" Dec 10 22:51:53 crc kubenswrapper[4791]: I1210 22:51:53.848995 4791 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/169ebb73-7b32-4f87-a8d7-9272bff04856-config\") pod \"kube-storage-version-migrator-operator-b67b599dd-k7qsx\" (UID: \"169ebb73-7b32-4f87-a8d7-9272bff04856\") " pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-k7qsx" Dec 10 22:51:53 crc kubenswrapper[4791]: I1210 22:51:53.849044 4791 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etcd-ca\" (UniqueName: \"kubernetes.io/configmap/97c517e6-0af1-4c17-a2b9-f7a7db3368b3-etcd-ca\") pod \"etcd-operator-b45778765-nkt8g\" (UID: \"97c517e6-0af1-4c17-a2b9-f7a7db3368b3\") " pod="openshift-etcd-operator/etcd-operator-b45778765-nkt8g" Dec 10 22:51:53 crc kubenswrapper[4791]: I1210 22:51:53.849101 4791 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/2fee22d4-2e9b-4768-9603-22e11d9dad9a-auth-proxy-config\") pod \"machine-config-operator-74547568cd-rcnrn\" (UID: \"2fee22d4-2e9b-4768-9603-22e11d9dad9a\") " pod="openshift-machine-config-operator/machine-config-operator-74547568cd-rcnrn" Dec 10 22:51:53 crc kubenswrapper[4791]: I1210 22:51:53.849489 4791 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/8f993103-9ee4-446f-a91a-a03dcd3ed9ea-config\") pod \"openshift-kube-scheduler-operator-5fdd9b5758-z6gpf\" (UID: \"8f993103-9ee4-446f-a91a-a03dcd3ed9ea\") " pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-z6gpf" Dec 10 22:51:53 crc kubenswrapper[4791]: I1210 22:51:53.849617 4791 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/90640468-53d7-42bc-a835-f7acf48a486e-secret-volume\") pod \"collect-profiles-29423445-b9cxb\" (UID: \"90640468-53d7-42bc-a835-f7acf48a486e\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29423445-b9cxb" Dec 10 22:51:53 crc kubenswrapper[4791]: I1210 22:51:53.849672 4791 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"registration-dir\" (UniqueName: \"kubernetes.io/host-path/1666d5bb-295d-4828-868e-b1b17b5741bb-registration-dir\") pod \"csi-hostpathplugin-mkz5r\" (UID: \"1666d5bb-295d-4828-868e-b1b17b5741bb\") " pod="hostpath-provisioner/csi-hostpathplugin-mkz5r" Dec 10 22:51:53 crc kubenswrapper[4791]: I1210 22:51:53.849809 4791 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"images\" (UniqueName: \"kubernetes.io/configmap/2fee22d4-2e9b-4768-9603-22e11d9dad9a-images\") pod \"machine-config-operator-74547568cd-rcnrn\" (UID: \"2fee22d4-2e9b-4768-9603-22e11d9dad9a\") " pod="openshift-machine-config-operator/machine-config-operator-74547568cd-rcnrn" Dec 10 22:51:53 crc kubenswrapper[4791]: I1210 22:51:53.850844 4791 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/97c517e6-0af1-4c17-a2b9-f7a7db3368b3-config\") pod \"etcd-operator-b45778765-nkt8g\" (UID: \"97c517e6-0af1-4c17-a2b9-f7a7db3368b3\") " pod="openshift-etcd-operator/etcd-operator-b45778765-nkt8g" Dec 10 22:51:53 crc kubenswrapper[4791]: I1210 22:51:53.851978 4791 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/90640468-53d7-42bc-a835-f7acf48a486e-config-volume\") pod \"collect-profiles-29423445-b9cxb\" (UID: \"90640468-53d7-42bc-a835-f7acf48a486e\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29423445-b9cxb" Dec 10 22:51:53 crc kubenswrapper[4791]: I1210 22:51:53.856513 4791 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"certs\" (UniqueName: \"kubernetes.io/secret/36532144-441a-4c12-a436-59d93d5ec3b7-certs\") pod \"machine-config-server-ktcmn\" (UID: \"36532144-441a-4c12-a436-59d93d5ec3b7\") " pod="openshift-machine-config-operator/machine-config-server-ktcmn" Dec 10 22:51:53 crc kubenswrapper[4791]: I1210 22:51:53.857529 4791 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"webhook-certs\" (UniqueName: \"kubernetes.io/secret/ded98820-5308-46e3-b87b-18d0a81c0519-webhook-certs\") pod \"multus-admission-controller-857f4d67dd-fbpcr\" (UID: \"ded98820-5308-46e3-b87b-18d0a81c0519\") " pod="openshift-multus/multus-admission-controller-857f4d67dd-fbpcr" Dec 10 22:51:53 crc kubenswrapper[4791]: I1210 22:51:53.857575 4791 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/97c517e6-0af1-4c17-a2b9-f7a7db3368b3-etcd-client\") pod \"etcd-operator-b45778765-nkt8g\" (UID: \"97c517e6-0af1-4c17-a2b9-f7a7db3368b3\") " pod="openshift-etcd-operator/etcd-operator-b45778765-nkt8g" Dec 10 22:51:53 crc kubenswrapper[4791]: I1210 22:51:53.858148 4791 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"marketplace-operator-metrics\" (UniqueName: \"kubernetes.io/secret/a9185ee1-adaf-48e8-b9a9-90ef658f8212-marketplace-operator-metrics\") pod \"marketplace-operator-79b997595-7g828\" (UID: \"a9185ee1-adaf-48e8-b9a9-90ef658f8212\") " pod="openshift-marketplace/marketplace-operator-79b997595-7g828" Dec 10 22:51:53 crc kubenswrapper[4791]: I1210 22:51:53.858459 4791 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/a05f3579-4703-4c46-bb11-5de60eedbc02-webhook-cert\") pod \"packageserver-d55dfcdfc-v445v\" (UID: \"a05f3579-4703-4c46-bb11-5de60eedbc02\") " pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-v445v" Dec 10 22:51:53 crc kubenswrapper[4791]: I1210 22:51:53.858535 4791 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"control-plane-machine-set-operator-tls\" (UniqueName: \"kubernetes.io/secret/c6c773c2-29e4-477d-91fb-f63052272f06-control-plane-machine-set-operator-tls\") pod \"control-plane-machine-set-operator-78cbb6b69f-ms8jw\" (UID: \"c6c773c2-29e4-477d-91fb-f63052272f06\") " pod="openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-ms8jw" Dec 10 22:51:53 crc kubenswrapper[4791]: I1210 22:51:53.858845 4791 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"signing-key\" (UniqueName: \"kubernetes.io/secret/34da0cc5-3e59-4c25-af94-e755e772b89a-signing-key\") pod \"service-ca-9c57cc56f-x9pw8\" (UID: \"34da0cc5-3e59-4c25-af94-e755e772b89a\") " pod="openshift-service-ca/service-ca-9c57cc56f-x9pw8" Dec 10 22:51:53 crc kubenswrapper[4791]: I1210 22:51:53.859182 4791 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/2fee22d4-2e9b-4768-9603-22e11d9dad9a-proxy-tls\") pod \"machine-config-operator-74547568cd-rcnrn\" (UID: \"2fee22d4-2e9b-4768-9603-22e11d9dad9a\") " pod="openshift-machine-config-operator/machine-config-operator-74547568cd-rcnrn" Dec 10 22:51:53 crc kubenswrapper[4791]: I1210 22:51:53.860441 4791 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/93a38a79-5c1c-4296-849e-94137de8f974-serving-cert\") pod \"openshift-apiserver-operator-796bbdcf4f-rmjd2\" (UID: \"93a38a79-5c1c-4296-849e-94137de8f974\") " pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-rmjd2" Dec 10 22:51:53 crc kubenswrapper[4791]: I1210 22:51:53.862242 4791 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/5a2896d3-fca8-4c64-abeb-6320f0b78d4e-metrics-tls\") pod \"dns-operator-744455d44c-cggs2\" (UID: \"5a2896d3-fca8-4c64-abeb-6320f0b78d4e\") " pod="openshift-dns-operator/dns-operator-744455d44c-cggs2" Dec 10 22:51:53 crc kubenswrapper[4791]: I1210 22:51:53.862319 4791 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/64199728-979b-4628-8b37-2dd792f86439-serving-cert\") pod \"service-ca-operator-777779d784-b8v4c\" (UID: \"64199728-979b-4628-8b37-2dd792f86439\") " pod="openshift-service-ca-operator/service-ca-operator-777779d784-b8v4c" Dec 10 22:51:53 crc kubenswrapper[4791]: I1210 22:51:53.863248 4791 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cert\" (UniqueName: \"kubernetes.io/secret/0921ef8b-8842-4d66-890b-3ad15bd48e70-cert\") pod \"ingress-canary-9dwf5\" (UID: \"0921ef8b-8842-4d66-890b-3ad15bd48e70\") " pod="openshift-ingress-canary/ingress-canary-9dwf5" Dec 10 22:51:53 crc kubenswrapper[4791]: I1210 22:51:53.866547 4791 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/169ebb73-7b32-4f87-a8d7-9272bff04856-serving-cert\") pod \"kube-storage-version-migrator-operator-b67b599dd-k7qsx\" (UID: \"169ebb73-7b32-4f87-a8d7-9272bff04856\") " pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-k7qsx" Dec 10 22:51:53 crc kubenswrapper[4791]: I1210 22:51:53.867664 4791 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"node-bootstrap-token\" (UniqueName: \"kubernetes.io/secret/36532144-441a-4c12-a436-59d93d5ec3b7-node-bootstrap-token\") pod \"machine-config-server-ktcmn\" (UID: \"36532144-441a-4c12-a436-59d93d5ec3b7\") " pod="openshift-machine-config-operator/machine-config-server-ktcmn" Dec 10 22:51:53 crc kubenswrapper[4791]: I1210 22:51:53.867779 4791 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"package-server-manager-serving-cert\" (UniqueName: \"kubernetes.io/secret/3f132df5-3de5-4455-a3da-7848b0de69cc-package-server-manager-serving-cert\") pod \"package-server-manager-789f6589d5-w5dpr\" (UID: \"3f132df5-3de5-4455-a3da-7848b0de69cc\") " pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-w5dpr" Dec 10 22:51:53 crc kubenswrapper[4791]: I1210 22:51:53.869195 4791 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/e9978f48-c33a-4765-9b05-16e37f5cdb24-metrics-tls\") pod \"dns-default-qs77n\" (UID: \"e9978f48-c33a-4765-9b05-16e37f5cdb24\") " pod="openshift-dns/dns-default-qs77n" Dec 10 22:51:53 crc kubenswrapper[4791]: I1210 22:51:53.871790 4791 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/a05f3579-4703-4c46-bb11-5de60eedbc02-apiservice-cert\") pod \"packageserver-d55dfcdfc-v445v\" (UID: \"a05f3579-4703-4c46-bb11-5de60eedbc02\") " pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-v445v" Dec 10 22:51:53 crc kubenswrapper[4791]: I1210 22:51:53.883092 4791 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-z25fj\" (UniqueName: \"kubernetes.io/projected/0921ef8b-8842-4d66-890b-3ad15bd48e70-kube-api-access-z25fj\") pod \"ingress-canary-9dwf5\" (UID: \"0921ef8b-8842-4d66-890b-3ad15bd48e70\") " pod="openshift-ingress-canary/ingress-canary-9dwf5" Dec 10 22:51:53 crc kubenswrapper[4791]: I1210 22:51:53.921429 4791 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-pbhrm\" (UniqueName: \"kubernetes.io/projected/2fee22d4-2e9b-4768-9603-22e11d9dad9a-kube-api-access-pbhrm\") pod \"machine-config-operator-74547568cd-rcnrn\" (UID: \"2fee22d4-2e9b-4768-9603-22e11d9dad9a\") " pod="openshift-machine-config-operator/machine-config-operator-74547568cd-rcnrn" Dec 10 22:51:53 crc kubenswrapper[4791]: I1210 22:51:53.925263 4791 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-dk24r\" (UniqueName: \"kubernetes.io/projected/1666d5bb-295d-4828-868e-b1b17b5741bb-kube-api-access-dk24r\") pod \"csi-hostpathplugin-mkz5r\" (UID: \"1666d5bb-295d-4828-868e-b1b17b5741bb\") " pod="hostpath-provisioner/csi-hostpathplugin-mkz5r" Dec 10 22:51:53 crc kubenswrapper[4791]: I1210 22:51:53.940578 4791 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-t6768\" (UniqueName: \"kubernetes.io/projected/ded98820-5308-46e3-b87b-18d0a81c0519-kube-api-access-t6768\") pod \"multus-admission-controller-857f4d67dd-fbpcr\" (UID: \"ded98820-5308-46e3-b87b-18d0a81c0519\") " pod="openshift-multus/multus-admission-controller-857f4d67dd-fbpcr" Dec 10 22:51:53 crc kubenswrapper[4791]: I1210 22:51:53.945712 4791 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 10 22:51:53 crc kubenswrapper[4791]: E1210 22:51:53.945836 4791 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-10 22:51:54.445815719 +0000 UTC m=+148.875433332 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 10 22:51:53 crc kubenswrapper[4791]: I1210 22:51:53.946030 4791 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-srd52\" (UID: \"e93dbd5d-0ca7-4a15-9eb3-0d71595fada0\") " pod="openshift-image-registry/image-registry-697d97f7c8-srd52" Dec 10 22:51:53 crc kubenswrapper[4791]: E1210 22:51:53.948661 4791 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-10 22:51:54.448621399 +0000 UTC m=+148.878239092 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-srd52" (UID: "e93dbd5d-0ca7-4a15-9eb3-0d71595fada0") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 10 22:51:53 crc kubenswrapper[4791]: I1210 22:51:53.975878 4791 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-tm8cv\" (UniqueName: \"kubernetes.io/projected/97c517e6-0af1-4c17-a2b9-f7a7db3368b3-kube-api-access-tm8cv\") pod \"etcd-operator-b45778765-nkt8g\" (UID: \"97c517e6-0af1-4c17-a2b9-f7a7db3368b3\") " pod="openshift-etcd-operator/etcd-operator-b45778765-nkt8g" Dec 10 22:51:53 crc kubenswrapper[4791]: I1210 22:51:53.981884 4791 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-bdjr4\" (UniqueName: \"kubernetes.io/projected/a05f3579-4703-4c46-bb11-5de60eedbc02-kube-api-access-bdjr4\") pod \"packageserver-d55dfcdfc-v445v\" (UID: \"a05f3579-4703-4c46-bb11-5de60eedbc02\") " pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-v445v" Dec 10 22:51:54 crc kubenswrapper[4791]: I1210 22:51:54.012571 4791 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-n8jwh\" (UniqueName: \"kubernetes.io/projected/868d6095-79bd-41fc-ae14-f369fdf68fde-kube-api-access-n8jwh\") pod \"migrator-59844c95c7-vscpm\" (UID: \"868d6095-79bd-41fc-ae14-f369fdf68fde\") " pod="openshift-kube-storage-version-migrator/migrator-59844c95c7-vscpm" Dec 10 22:51:54 crc kubenswrapper[4791]: I1210 22:51:54.022463 4791 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-bws6n" Dec 10 22:51:54 crc kubenswrapper[4791]: I1210 22:51:54.033924 4791 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-jpwld\" (UniqueName: \"kubernetes.io/projected/c6c773c2-29e4-477d-91fb-f63052272f06-kube-api-access-jpwld\") pod \"control-plane-machine-set-operator-78cbb6b69f-ms8jw\" (UID: \"c6c773c2-29e4-477d-91fb-f63052272f06\") " pod="openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-ms8jw" Dec 10 22:51:54 crc kubenswrapper[4791]: I1210 22:51:54.049047 4791 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 10 22:51:54 crc kubenswrapper[4791]: E1210 22:51:54.049556 4791 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-10 22:51:54.54953266 +0000 UTC m=+148.979150273 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 10 22:51:54 crc kubenswrapper[4791]: I1210 22:51:54.052073 4791 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-87746\" (UniqueName: \"kubernetes.io/projected/34da0cc5-3e59-4c25-af94-e755e772b89a-kube-api-access-87746\") pod \"service-ca-9c57cc56f-x9pw8\" (UID: \"34da0cc5-3e59-4c25-af94-e755e772b89a\") " pod="openshift-service-ca/service-ca-9c57cc56f-x9pw8" Dec 10 22:51:54 crc kubenswrapper[4791]: I1210 22:51:54.065512 4791 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-etcd-operator/etcd-operator-b45778765-nkt8g" Dec 10 22:51:54 crc kubenswrapper[4791]: I1210 22:51:54.078890 4791 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-mnnjz\" (UniqueName: \"kubernetes.io/projected/36532144-441a-4c12-a436-59d93d5ec3b7-kube-api-access-mnnjz\") pod \"machine-config-server-ktcmn\" (UID: \"36532144-441a-4c12-a436-59d93d5ec3b7\") " pod="openshift-machine-config-operator/machine-config-server-ktcmn" Dec 10 22:51:54 crc kubenswrapper[4791]: I1210 22:51:54.079558 4791 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-config-operator/machine-config-operator-74547568cd-rcnrn" Dec 10 22:51:54 crc kubenswrapper[4791]: I1210 22:51:54.089629 4791 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-storage-version-migrator/migrator-59844c95c7-vscpm" Dec 10 22:51:54 crc kubenswrapper[4791]: I1210 22:51:54.096211 4791 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/multus-admission-controller-857f4d67dd-fbpcr" Dec 10 22:51:54 crc kubenswrapper[4791]: I1210 22:51:54.096359 4791 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-gnhrf\" (UniqueName: \"kubernetes.io/projected/90640468-53d7-42bc-a835-f7acf48a486e-kube-api-access-gnhrf\") pod \"collect-profiles-29423445-b9cxb\" (UID: \"90640468-53d7-42bc-a835-f7acf48a486e\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29423445-b9cxb" Dec 10 22:51:54 crc kubenswrapper[4791]: I1210 22:51:54.108980 4791 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-v445v" Dec 10 22:51:54 crc kubenswrapper[4791]: I1210 22:51:54.112487 4791 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-fwbdr\" (UniqueName: \"kubernetes.io/projected/5a2896d3-fca8-4c64-abeb-6320f0b78d4e-kube-api-access-fwbdr\") pod \"dns-operator-744455d44c-cggs2\" (UID: \"5a2896d3-fca8-4c64-abeb-6320f0b78d4e\") " pod="openshift-dns-operator/dns-operator-744455d44c-cggs2" Dec 10 22:51:54 crc kubenswrapper[4791]: I1210 22:51:54.116112 4791 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-ms8jw" Dec 10 22:51:54 crc kubenswrapper[4791]: I1210 22:51:54.134609 4791 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-hsmgd\" (UniqueName: \"kubernetes.io/projected/a9185ee1-adaf-48e8-b9a9-90ef658f8212-kube-api-access-hsmgd\") pod \"marketplace-operator-79b997595-7g828\" (UID: \"a9185ee1-adaf-48e8-b9a9-90ef658f8212\") " pod="openshift-marketplace/marketplace-operator-79b997595-7g828" Dec 10 22:51:54 crc kubenswrapper[4791]: I1210 22:51:54.144909 4791 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-config-operator/machine-config-server-ktcmn" Dec 10 22:51:54 crc kubenswrapper[4791]: I1210 22:51:54.145758 4791 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29423445-b9cxb" Dec 10 22:51:54 crc kubenswrapper[4791]: I1210 22:51:54.152530 4791 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-srd52\" (UID: \"e93dbd5d-0ca7-4a15-9eb3-0d71595fada0\") " pod="openshift-image-registry/image-registry-697d97f7c8-srd52" Dec 10 22:51:54 crc kubenswrapper[4791]: E1210 22:51:54.153012 4791 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-10 22:51:54.652996954 +0000 UTC m=+149.082614567 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-srd52" (UID: "e93dbd5d-0ca7-4a15-9eb3-0d71595fada0") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 10 22:51:54 crc kubenswrapper[4791]: I1210 22:51:54.159230 4791 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-service-ca/service-ca-9c57cc56f-x9pw8" Dec 10 22:51:54 crc kubenswrapper[4791]: I1210 22:51:54.161866 4791 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/6e58f465-ef21-4b20-9eb8-5ade4810688a-kube-api-access\") pod \"kube-apiserver-operator-766d6c64bb-vm842\" (UID: \"6e58f465-ef21-4b20-9eb8-5ade4810688a\") " pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-vm842" Dec 10 22:51:54 crc kubenswrapper[4791]: I1210 22:51:54.166828 4791 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-ingress-canary/ingress-canary-9dwf5" Dec 10 22:51:54 crc kubenswrapper[4791]: I1210 22:51:54.174925 4791 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-xcxbr\" (UniqueName: \"kubernetes.io/projected/e9978f48-c33a-4765-9b05-16e37f5cdb24-kube-api-access-xcxbr\") pod \"dns-default-qs77n\" (UID: \"e9978f48-c33a-4765-9b05-16e37f5cdb24\") " pod="openshift-dns/dns-default-qs77n" Dec 10 22:51:54 crc kubenswrapper[4791]: I1210 22:51:54.188732 4791 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="hostpath-provisioner/csi-hostpathplugin-mkz5r" Dec 10 22:51:54 crc kubenswrapper[4791]: I1210 22:51:54.196836 4791 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/8f993103-9ee4-446f-a91a-a03dcd3ed9ea-kube-api-access\") pod \"openshift-kube-scheduler-operator-5fdd9b5758-z6gpf\" (UID: \"8f993103-9ee4-446f-a91a-a03dcd3ed9ea\") " pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-z6gpf" Dec 10 22:51:54 crc kubenswrapper[4791]: I1210 22:51:54.212865 4791 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-tjc7n\" (UniqueName: \"kubernetes.io/projected/169ebb73-7b32-4f87-a8d7-9272bff04856-kube-api-access-tjc7n\") pod \"kube-storage-version-migrator-operator-b67b599dd-k7qsx\" (UID: \"169ebb73-7b32-4f87-a8d7-9272bff04856\") " pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-k7qsx" Dec 10 22:51:54 crc kubenswrapper[4791]: I1210 22:51:54.238392 4791 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-console-operator/console-operator-58897d9998-hd6zl"] Dec 10 22:51:54 crc kubenswrapper[4791]: I1210 22:51:54.244326 4791 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-gb99d\" (UniqueName: \"kubernetes.io/projected/64199728-979b-4628-8b37-2dd792f86439-kube-api-access-gb99d\") pod \"service-ca-operator-777779d784-b8v4c\" (UID: \"64199728-979b-4628-8b37-2dd792f86439\") " pod="openshift-service-ca-operator/service-ca-operator-777779d784-b8v4c" Dec 10 22:51:54 crc kubenswrapper[4791]: I1210 22:51:54.259739 4791 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 10 22:51:54 crc kubenswrapper[4791]: E1210 22:51:54.260067 4791 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-10 22:51:54.760053469 +0000 UTC m=+149.189671082 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 10 22:51:54 crc kubenswrapper[4791]: I1210 22:51:54.263786 4791 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-6mhcw\" (UniqueName: \"kubernetes.io/projected/3f132df5-3de5-4455-a3da-7848b0de69cc-kube-api-access-6mhcw\") pod \"package-server-manager-789f6589d5-w5dpr\" (UID: \"3f132df5-3de5-4455-a3da-7848b0de69cc\") " pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-w5dpr" Dec 10 22:51:54 crc kubenswrapper[4791]: I1210 22:51:54.271475 4791 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-machine-config-operator/machine-config-controller-84d6567774-sc74x"] Dec 10 22:51:54 crc kubenswrapper[4791]: I1210 22:51:54.287201 4791 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-4jd6j\" (UniqueName: \"kubernetes.io/projected/93a38a79-5c1c-4296-849e-94137de8f974-kube-api-access-4jd6j\") pod \"openshift-apiserver-operator-796bbdcf4f-rmjd2\" (UID: \"93a38a79-5c1c-4296-849e-94137de8f974\") " pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-rmjd2" Dec 10 22:51:54 crc kubenswrapper[4791]: I1210 22:51:54.335971 4791 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-dns-operator/dns-operator-744455d44c-cggs2" Dec 10 22:51:54 crc kubenswrapper[4791]: I1210 22:51:54.343297 4791 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-z6gpf" Dec 10 22:51:54 crc kubenswrapper[4791]: I1210 22:51:54.350390 4791 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-rmjd2" Dec 10 22:51:54 crc kubenswrapper[4791]: I1210 22:51:54.357604 4791 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-w5dpr" Dec 10 22:51:54 crc kubenswrapper[4791]: I1210 22:51:54.361307 4791 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-srd52\" (UID: \"e93dbd5d-0ca7-4a15-9eb3-0d71595fada0\") " pod="openshift-image-registry/image-registry-697d97f7c8-srd52" Dec 10 22:51:54 crc kubenswrapper[4791]: E1210 22:51:54.361636 4791 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-10 22:51:54.861619889 +0000 UTC m=+149.291237502 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-srd52" (UID: "e93dbd5d-0ca7-4a15-9eb3-0d71595fada0") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 10 22:51:54 crc kubenswrapper[4791]: I1210 22:51:54.372062 4791 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-vm842" Dec 10 22:51:54 crc kubenswrapper[4791]: I1210 22:51:54.401429 4791 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-k7qsx" Dec 10 22:51:54 crc kubenswrapper[4791]: I1210 22:51:54.427140 4791 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/marketplace-operator-79b997595-7g828" Dec 10 22:51:54 crc kubenswrapper[4791]: I1210 22:51:54.437097 4791 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-dns/dns-default-qs77n" Dec 10 22:51:54 crc kubenswrapper[4791]: I1210 22:51:54.455397 4791 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-service-ca-operator/service-ca-operator-777779d784-b8v4c" Dec 10 22:51:54 crc kubenswrapper[4791]: I1210 22:51:54.465884 4791 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 10 22:51:54 crc kubenswrapper[4791]: E1210 22:51:54.466256 4791 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-10 22:51:54.966224015 +0000 UTC m=+149.395841628 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 10 22:51:54 crc kubenswrapper[4791]: I1210 22:51:54.466441 4791 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-srd52\" (UID: \"e93dbd5d-0ca7-4a15-9eb3-0d71595fada0\") " pod="openshift-image-registry/image-registry-697d97f7c8-srd52" Dec 10 22:51:54 crc kubenswrapper[4791]: E1210 22:51:54.466890 4791 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-10 22:51:54.966878753 +0000 UTC m=+149.396496366 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-srd52" (UID: "e93dbd5d-0ca7-4a15-9eb3-0d71595fada0") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 10 22:51:54 crc kubenswrapper[4791]: I1210 22:51:54.567721 4791 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 10 22:51:54 crc kubenswrapper[4791]: E1210 22:51:54.567914 4791 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-10 22:51:55.067881067 +0000 UTC m=+149.497498700 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 10 22:51:54 crc kubenswrapper[4791]: I1210 22:51:54.568155 4791 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-srd52\" (UID: \"e93dbd5d-0ca7-4a15-9eb3-0d71595fada0\") " pod="openshift-image-registry/image-registry-697d97f7c8-srd52" Dec 10 22:51:54 crc kubenswrapper[4791]: E1210 22:51:54.568542 4791 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-10 22:51:55.068528375 +0000 UTC m=+149.498145988 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-srd52" (UID: "e93dbd5d-0ca7-4a15-9eb3-0d71595fada0") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 10 22:51:54 crc kubenswrapper[4791]: I1210 22:51:54.669610 4791 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 10 22:51:54 crc kubenswrapper[4791]: E1210 22:51:54.669761 4791 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-10 22:51:55.169736925 +0000 UTC m=+149.599354538 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 10 22:51:54 crc kubenswrapper[4791]: I1210 22:51:54.670822 4791 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-srd52\" (UID: \"e93dbd5d-0ca7-4a15-9eb3-0d71595fada0\") " pod="openshift-image-registry/image-registry-697d97f7c8-srd52" Dec 10 22:51:54 crc kubenswrapper[4791]: E1210 22:51:54.671658 4791 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-10 22:51:55.171635339 +0000 UTC m=+149.601252952 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-srd52" (UID: "e93dbd5d-0ca7-4a15-9eb3-0d71595fada0") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 10 22:51:54 crc kubenswrapper[4791]: I1210 22:51:54.731844 4791 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-authentication-operator/authentication-operator-69f744f599-7rhk9"] Dec 10 22:51:54 crc kubenswrapper[4791]: I1210 22:51:54.760351 4791 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-mldvd"] Dec 10 22:51:54 crc kubenswrapper[4791]: I1210 22:51:54.767026 4791 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-config-operator/openshift-config-operator-7777fb866f-wwpsp"] Dec 10 22:51:54 crc kubenswrapper[4791]: I1210 22:51:54.772225 4791 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 10 22:51:54 crc kubenswrapper[4791]: E1210 22:51:54.772721 4791 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-10 22:51:55.272705914 +0000 UTC m=+149.702323527 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 10 22:51:54 crc kubenswrapper[4791]: I1210 22:51:54.797978 4791 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-5t9vv"] Dec 10 22:51:54 crc kubenswrapper[4791]: I1210 22:51:54.799920 4791 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-authentication/oauth-openshift-558db77b4-ht5t7"] Dec 10 22:51:54 crc kubenswrapper[4791]: I1210 22:51:54.842639 4791 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-authentication-operator/authentication-operator-69f744f599-7rhk9" event={"ID":"cff60894-8deb-4354-9ad1-3a83baa98714","Type":"ContainerStarted","Data":"2f7051b89e277b9d1ddd33a9a52b06583e702421b36678d472ed52572745bedb"} Dec 10 22:51:54 crc kubenswrapper[4791]: I1210 22:51:54.850692 4791 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-cluster-machine-approver/machine-approver-56656f9798-m6dwv" event={"ID":"2f422ac4-b586-4056-8748-94b7de0b0248","Type":"ContainerStarted","Data":"216289f6e26428d410b37596e2ebc58464da2014211824b45debfaa4ca2f8a93"} Dec 10 22:51:54 crc kubenswrapper[4791]: I1210 22:51:54.850731 4791 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-cluster-machine-approver/machine-approver-56656f9798-m6dwv" event={"ID":"2f422ac4-b586-4056-8748-94b7de0b0248","Type":"ContainerStarted","Data":"bb91c5124506e027cab275b01aee6b921fdf5fd958b05249e0bcbb5899b26083"} Dec 10 22:51:54 crc kubenswrapper[4791]: I1210 22:51:54.854312 4791 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-controller-84d6567774-sc74x" event={"ID":"e2b778bb-d80f-49f9-804b-0550c648a02f","Type":"ContainerStarted","Data":"6c7154a11796111da93db56e58033d8d70b5a4b47600f45bbed4acf87d578df5"} Dec 10 22:51:54 crc kubenswrapper[4791]: I1210 22:51:54.854359 4791 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-controller-84d6567774-sc74x" event={"ID":"e2b778bb-d80f-49f9-804b-0550c648a02f","Type":"ContainerStarted","Data":"e5326c17c01e257998fd995e65ed8c54e4b7d9e9521704e7afe05037df8b59c7"} Dec 10 22:51:54 crc kubenswrapper[4791]: I1210 22:51:54.858812 4791 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-server-ktcmn" event={"ID":"36532144-441a-4c12-a436-59d93d5ec3b7","Type":"ContainerStarted","Data":"8c7771afc7c55c91b537811551c106b7d55c0f088b63bc9191355376a6e5178c"} Dec 10 22:51:54 crc kubenswrapper[4791]: I1210 22:51:54.858848 4791 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-server-ktcmn" event={"ID":"36532144-441a-4c12-a436-59d93d5ec3b7","Type":"ContainerStarted","Data":"082aa029b0f6c1217a4ae991defe391a17ad11946e1f83d431fec1c6ebe092e6"} Dec 10 22:51:54 crc kubenswrapper[4791]: I1210 22:51:54.864157 4791 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager/controller-manager-879f6c89f-9tg4p" event={"ID":"a5827e8a-4cb3-494d-87a1-556d920c2a4a","Type":"ContainerStarted","Data":"1b567163092c2059ebeb396c427945f52d9972a1c85abdccb0bd739aac9e995f"} Dec 10 22:51:54 crc kubenswrapper[4791]: I1210 22:51:54.864207 4791 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager/controller-manager-879f6c89f-9tg4p" event={"ID":"a5827e8a-4cb3-494d-87a1-556d920c2a4a","Type":"ContainerStarted","Data":"f323620800ff77b675c44d18a09436762bdf76b08f3551d85e0a8cd459422268"} Dec 10 22:51:54 crc kubenswrapper[4791]: I1210 22:51:54.864418 4791 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-controller-manager/controller-manager-879f6c89f-9tg4p" Dec 10 22:51:54 crc kubenswrapper[4791]: I1210 22:51:54.876421 4791 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-controller-manager/controller-manager-879f6c89f-9tg4p" Dec 10 22:51:54 crc kubenswrapper[4791]: I1210 22:51:54.876969 4791 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-srd52\" (UID: \"e93dbd5d-0ca7-4a15-9eb3-0d71595fada0\") " pod="openshift-image-registry/image-registry-697d97f7c8-srd52" Dec 10 22:51:54 crc kubenswrapper[4791]: E1210 22:51:54.878384 4791 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-10 22:51:55.37837225 +0000 UTC m=+149.807989863 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-srd52" (UID: "e93dbd5d-0ca7-4a15-9eb3-0d71595fada0") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 10 22:51:54 crc kubenswrapper[4791]: I1210 22:51:54.878702 4791 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ingress/router-default-5444994796-p97fc" event={"ID":"d3de1ba0-b014-4f98-b597-cc89f5a70364","Type":"ContainerStarted","Data":"4cbc3e1726058120732dab462ed8f362b446d301eb0bf3164f1785bc7ade6241"} Dec 10 22:51:54 crc kubenswrapper[4791]: I1210 22:51:54.896799 4791 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-apiserver/apiserver-76f77b778f-kqtbx" event={"ID":"4a1cf441-8a08-4677-a480-947c8cddd6a6","Type":"ContainerStarted","Data":"09bbcf60b3729871e4b4bfb1adf074453e24f40a2e8914d2f1e7c16e248baff8"} Dec 10 22:51:54 crc kubenswrapper[4791]: I1210 22:51:54.914089 4791 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-nwwpb" event={"ID":"3ba17b70-17df-4f8c-87ff-c15e03ce06bc","Type":"ContainerStarted","Data":"23e255dd457d1ccf6b995b81010a0a533e1fed292a0568dd74d3c61a12b669f4"} Dec 10 22:51:54 crc kubenswrapper[4791]: I1210 22:51:54.917606 4791 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-console-operator/console-operator-58897d9998-hd6zl" event={"ID":"3d53f039-46f2-478b-b14c-11b723b7dbc1","Type":"ContainerStarted","Data":"246b2589c96fe1c008810df9ce63165e38317d972483233c5c78aa426734583a"} Dec 10 22:51:54 crc kubenswrapper[4791]: I1210 22:51:54.917642 4791 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-console-operator/console-operator-58897d9998-hd6zl" event={"ID":"3d53f039-46f2-478b-b14c-11b723b7dbc1","Type":"ContainerStarted","Data":"3ede587ebe2deab09dc44da418c6348856d1452514b7550af687e3a3c4737d64"} Dec 10 22:51:54 crc kubenswrapper[4791]: I1210 22:51:54.918114 4791 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-console-operator/console-operator-58897d9998-hd6zl" Dec 10 22:51:54 crc kubenswrapper[4791]: I1210 22:51:54.923099 4791 patch_prober.go:28] interesting pod/console-operator-58897d9998-hd6zl container/console-operator namespace/openshift-console-operator: Readiness probe status=failure output="Get \"https://10.217.0.12:8443/readyz\": dial tcp 10.217.0.12:8443: connect: connection refused" start-of-body= Dec 10 22:51:54 crc kubenswrapper[4791]: I1210 22:51:54.923184 4791 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-console-operator/console-operator-58897d9998-hd6zl" podUID="3d53f039-46f2-478b-b14c-11b723b7dbc1" containerName="console-operator" probeResult="failure" output="Get \"https://10.217.0.12:8443/readyz\": dial tcp 10.217.0.12:8443: connect: connection refused" Dec 10 22:51:54 crc kubenswrapper[4791]: I1210 22:51:54.928679 4791 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-api/machine-api-operator-5694c8668f-cpbhz" event={"ID":"cd65e91e-5b02-4f78-829e-93b4596ff8f6","Type":"ContainerStarted","Data":"f3b2be05d6a0c8586ad6ae6c6ae1c770ba06eb612c2975ffd843836874a11b1c"} Dec 10 22:51:54 crc kubenswrapper[4791]: I1210 22:51:54.946399 4791 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-t7742"] Dec 10 22:51:54 crc kubenswrapper[4791]: I1210 22:51:54.948665 4791 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-ingress-operator/ingress-operator-5b745b69d9-8npst"] Dec 10 22:51:54 crc kubenswrapper[4791]: W1210 22:51:54.950998 4791 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pode92ca7e3_d360_4e14_aeb9_76738a44c391.slice/crio-675018a16342172a763fe858fc105c5f5bdf54b0e66377e94f08ce5647604dba WatchSource:0}: Error finding container 675018a16342172a763fe858fc105c5f5bdf54b0e66377e94f08ce5647604dba: Status 404 returned error can't find the container with id 675018a16342172a763fe858fc105c5f5bdf54b0e66377e94f08ce5647604dba Dec 10 22:51:54 crc kubenswrapper[4791]: I1210 22:51:54.956654 4791 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-console/downloads-7954f5f757-mlb9p"] Dec 10 22:51:54 crc kubenswrapper[4791]: W1210 22:51:54.965148 4791 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podb419e5d6_d6a4_4258_a1bd_a462bd7f7a12.slice/crio-c3ff611a802a93d7cd48c0a382a4a4f10ace5b0955f9431c78652ae4feed1aef WatchSource:0}: Error finding container c3ff611a802a93d7cd48c0a382a4a4f10ace5b0955f9431c78652ae4feed1aef: Status 404 returned error can't find the container with id c3ff611a802a93d7cd48c0a382a4a4f10ace5b0955f9431c78652ae4feed1aef Dec 10 22:51:54 crc kubenswrapper[4791]: W1210 22:51:54.965914 4791 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod5be73604_c6c9_4ef6_a53d_6c3f6df27202.slice/crio-fdf99ce46ea76b994f89654fee996e46a4c14b5a6a4dab95849768db51e90894 WatchSource:0}: Error finding container fdf99ce46ea76b994f89654fee996e46a4c14b5a6a4dab95849768db51e90894: Status 404 returned error can't find the container with id fdf99ce46ea76b994f89654fee996e46a4c14b5a6a4dab95849768db51e90894 Dec 10 22:51:54 crc kubenswrapper[4791]: I1210 22:51:54.979857 4791 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 10 22:51:54 crc kubenswrapper[4791]: E1210 22:51:54.980026 4791 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-10 22:51:55.479995921 +0000 UTC m=+149.909613544 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 10 22:51:54 crc kubenswrapper[4791]: I1210 22:51:54.980609 4791 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-srd52\" (UID: \"e93dbd5d-0ca7-4a15-9eb3-0d71595fada0\") " pod="openshift-image-registry/image-registry-697d97f7c8-srd52" Dec 10 22:51:54 crc kubenswrapper[4791]: E1210 22:51:54.983475 4791 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-10 22:51:55.483454579 +0000 UTC m=+149.913072252 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-srd52" (UID: "e93dbd5d-0ca7-4a15-9eb3-0d71595fada0") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 10 22:51:55 crc kubenswrapper[4791]: W1210 22:51:55.043221 4791 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod35033f1a_b91b_4e02_ba77_3841a2e0fdab.slice/crio-6100b3fe0d820fdba5310d4b39a19699ff6808a46c982f7731462ffd7f68c89c WatchSource:0}: Error finding container 6100b3fe0d820fdba5310d4b39a19699ff6808a46c982f7731462ffd7f68c89c: Status 404 returned error can't find the container with id 6100b3fe0d820fdba5310d4b39a19699ff6808a46c982f7731462ffd7f68c89c Dec 10 22:51:55 crc kubenswrapper[4791]: I1210 22:51:55.044003 4791 patch_prober.go:28] interesting pod/machine-config-daemon-5rb5l container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 10 22:51:55 crc kubenswrapper[4791]: I1210 22:51:55.044048 4791 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-5rb5l" podUID="ba35653c-6e06-4cee-a4d6-137764090d18" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 10 22:51:55 crc kubenswrapper[4791]: I1210 22:51:55.082177 4791 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 10 22:51:55 crc kubenswrapper[4791]: E1210 22:51:55.083772 4791 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-10 22:51:55.583747853 +0000 UTC m=+150.013365466 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 10 22:51:55 crc kubenswrapper[4791]: I1210 22:51:55.090164 4791 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-ingress/router-default-5444994796-p97fc" Dec 10 22:51:55 crc kubenswrapper[4791]: I1210 22:51:55.186626 4791 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-srd52\" (UID: \"e93dbd5d-0ca7-4a15-9eb3-0d71595fada0\") " pod="openshift-image-registry/image-registry-697d97f7c8-srd52" Dec 10 22:51:55 crc kubenswrapper[4791]: E1210 22:51:55.187150 4791 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-10 22:51:55.687135695 +0000 UTC m=+150.116753308 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-srd52" (UID: "e93dbd5d-0ca7-4a15-9eb3-0d71595fada0") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 10 22:51:55 crc kubenswrapper[4791]: I1210 22:51:55.289165 4791 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 10 22:51:55 crc kubenswrapper[4791]: E1210 22:51:55.289765 4791 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-10 22:51:55.789710803 +0000 UTC m=+150.219328416 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 10 22:51:55 crc kubenswrapper[4791]: I1210 22:51:55.352050 4791 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-machine-config-operator/machine-config-server-ktcmn" podStartSLOduration=5.352031644 podStartE2EDuration="5.352031644s" podCreationTimestamp="2025-12-10 22:51:50 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-10 22:51:55.349703718 +0000 UTC m=+149.779321331" watchObservedRunningTime="2025-12-10 22:51:55.352031644 +0000 UTC m=+149.781649257" Dec 10 22:51:55 crc kubenswrapper[4791]: I1210 22:51:55.392227 4791 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-srd52\" (UID: \"e93dbd5d-0ca7-4a15-9eb3-0d71595fada0\") " pod="openshift-image-registry/image-registry-697d97f7c8-srd52" Dec 10 22:51:55 crc kubenswrapper[4791]: E1210 22:51:55.392705 4791 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-10 22:51:55.892691323 +0000 UTC m=+150.322308936 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-srd52" (UID: "e93dbd5d-0ca7-4a15-9eb3-0d71595fada0") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 10 22:51:55 crc kubenswrapper[4791]: I1210 22:51:55.393452 4791 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-machine-api/machine-api-operator-5694c8668f-cpbhz" podStartSLOduration=124.393428504 podStartE2EDuration="2m4.393428504s" podCreationTimestamp="2025-12-10 22:49:51 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-10 22:51:55.386723294 +0000 UTC m=+149.816340907" watchObservedRunningTime="2025-12-10 22:51:55.393428504 +0000 UTC m=+149.823046117" Dec 10 22:51:55 crc kubenswrapper[4791]: I1210 22:51:55.439025 4791 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-controller-manager/controller-manager-879f6c89f-9tg4p" podStartSLOduration=124.438995241 podStartE2EDuration="2m4.438995241s" podCreationTimestamp="2025-12-10 22:49:51 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-10 22:51:55.43220076 +0000 UTC m=+149.861818393" watchObservedRunningTime="2025-12-10 22:51:55.438995241 +0000 UTC m=+149.868612894" Dec 10 22:51:55 crc kubenswrapper[4791]: I1210 22:51:55.439104 4791 patch_prober.go:28] interesting pod/router-default-5444994796-p97fc container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Dec 10 22:51:55 crc kubenswrapper[4791]: [-]has-synced failed: reason withheld Dec 10 22:51:55 crc kubenswrapper[4791]: [+]process-running ok Dec 10 22:51:55 crc kubenswrapper[4791]: healthz check failed Dec 10 22:51:55 crc kubenswrapper[4791]: I1210 22:51:55.439151 4791 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-p97fc" podUID="d3de1ba0-b014-4f98-b597-cc89f5a70364" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Dec 10 22:51:55 crc kubenswrapper[4791]: I1210 22:51:55.481933 4791 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-cluster-machine-approver/machine-approver-56656f9798-m6dwv" podStartSLOduration=126.481896014 podStartE2EDuration="2m6.481896014s" podCreationTimestamp="2025-12-10 22:49:49 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-10 22:51:55.479691372 +0000 UTC m=+149.909308985" watchObservedRunningTime="2025-12-10 22:51:55.481896014 +0000 UTC m=+149.911513627" Dec 10 22:51:55 crc kubenswrapper[4791]: I1210 22:51:55.494607 4791 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 10 22:51:55 crc kubenswrapper[4791]: E1210 22:51:55.495044 4791 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-10 22:51:55.995023085 +0000 UTC m=+150.424640698 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 10 22:51:55 crc kubenswrapper[4791]: I1210 22:51:55.525477 4791 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-console-operator/console-operator-58897d9998-hd6zl" podStartSLOduration=124.525462385 podStartE2EDuration="2m4.525462385s" podCreationTimestamp="2025-12-10 22:49:51 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-10 22:51:55.523036106 +0000 UTC m=+149.952653719" watchObservedRunningTime="2025-12-10 22:51:55.525462385 +0000 UTC m=+149.955079998" Dec 10 22:51:55 crc kubenswrapper[4791]: I1210 22:51:55.568357 4791 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-nwwpb" podStartSLOduration=124.568331086 podStartE2EDuration="2m4.568331086s" podCreationTimestamp="2025-12-10 22:49:51 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-10 22:51:55.56812652 +0000 UTC m=+149.997744133" watchObservedRunningTime="2025-12-10 22:51:55.568331086 +0000 UTC m=+149.997948699" Dec 10 22:51:55 crc kubenswrapper[4791]: I1210 22:51:55.596904 4791 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-srd52\" (UID: \"e93dbd5d-0ca7-4a15-9eb3-0d71595fada0\") " pod="openshift-image-registry/image-registry-697d97f7c8-srd52" Dec 10 22:51:55 crc kubenswrapper[4791]: E1210 22:51:55.597399 4791 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-10 22:51:56.097381727 +0000 UTC m=+150.526999340 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-srd52" (UID: "e93dbd5d-0ca7-4a15-9eb3-0d71595fada0") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 10 22:51:55 crc kubenswrapper[4791]: I1210 22:51:55.639079 4791 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-ingress/router-default-5444994796-p97fc" podStartSLOduration=124.639044064 podStartE2EDuration="2m4.639044064s" podCreationTimestamp="2025-12-10 22:49:51 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-10 22:51:55.600458094 +0000 UTC m=+150.030075727" watchObservedRunningTime="2025-12-10 22:51:55.639044064 +0000 UTC m=+150.068661677" Dec 10 22:51:55 crc kubenswrapper[4791]: I1210 22:51:55.640488 4791 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-6576b87f9c-vsnsh"] Dec 10 22:51:55 crc kubenswrapper[4791]: I1210 22:51:55.642714 4791 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-console/console-f9d7485db-zkfkv"] Dec 10 22:51:55 crc kubenswrapper[4791]: I1210 22:51:55.659066 4791 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-bws6n"] Dec 10 22:51:55 crc kubenswrapper[4791]: I1210 22:51:55.669101 4791 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-q4cpx"] Dec 10 22:51:55 crc kubenswrapper[4791]: I1210 22:51:55.690450 4791 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-storage-version-migrator/migrator-59844c95c7-vscpm"] Dec 10 22:51:55 crc kubenswrapper[4791]: I1210 22:51:55.697430 4791 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 10 22:51:55 crc kubenswrapper[4791]: E1210 22:51:55.697846 4791 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-10 22:51:56.197831136 +0000 UTC m=+150.627448749 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 10 22:51:55 crc kubenswrapper[4791]: I1210 22:51:55.704203 4791 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/catalog-operator-68c6474976-497m8"] Dec 10 22:51:55 crc kubenswrapper[4791]: W1210 22:51:55.716941 4791 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod1b61a973_db5e_4aa7_96e3_7fbdd9dd8cc0.slice/crio-8be5e83a023b172a495e79f8fe719fc8e54ba3500d683e4fce82952ba90d0523 WatchSource:0}: Error finding container 8be5e83a023b172a495e79f8fe719fc8e54ba3500d683e4fce82952ba90d0523: Status 404 returned error can't find the container with id 8be5e83a023b172a495e79f8fe719fc8e54ba3500d683e4fce82952ba90d0523 Dec 10 22:51:55 crc kubenswrapper[4791]: W1210 22:51:55.749268 4791 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod06eb2b28_22d6_4116_9ce4_1abf7a21db36.slice/crio-101a0b0df93291fe6a85b2048d936ab30807d60a36b8c2f7c67f36f431326f31 WatchSource:0}: Error finding container 101a0b0df93291fe6a85b2048d936ab30807d60a36b8c2f7c67f36f431326f31: Status 404 returned error can't find the container with id 101a0b0df93291fe6a85b2048d936ab30807d60a36b8c2f7c67f36f431326f31 Dec 10 22:51:55 crc kubenswrapper[4791]: I1210 22:51:55.798382 4791 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-machine-config-operator/machine-config-operator-74547568cd-rcnrn"] Dec 10 22:51:55 crc kubenswrapper[4791]: I1210 22:51:55.798936 4791 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-srd52\" (UID: \"e93dbd5d-0ca7-4a15-9eb3-0d71595fada0\") " pod="openshift-image-registry/image-registry-697d97f7c8-srd52" Dec 10 22:51:55 crc kubenswrapper[4791]: E1210 22:51:55.799199 4791 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-10 22:51:56.29918937 +0000 UTC m=+150.728806983 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-srd52" (UID: "e93dbd5d-0ca7-4a15-9eb3-0d71595fada0") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 10 22:51:55 crc kubenswrapper[4791]: W1210 22:51:55.800547 4791 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod8f0772d5_db77_42b2_b712_0293c0422161.slice/crio-2a673925b6a9b5369c5dec28c58733217e066e08890de9acac1a4b057a70ff83 WatchSource:0}: Error finding container 2a673925b6a9b5369c5dec28c58733217e066e08890de9acac1a4b057a70ff83: Status 404 returned error can't find the container with id 2a673925b6a9b5369c5dec28c58733217e066e08890de9acac1a4b057a70ff83 Dec 10 22:51:55 crc kubenswrapper[4791]: I1210 22:51:55.855904 4791 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-v445v"] Dec 10 22:51:55 crc kubenswrapper[4791]: I1210 22:51:55.901609 4791 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 10 22:51:55 crc kubenswrapper[4791]: E1210 22:51:55.902124 4791 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-10 22:51:56.402108488 +0000 UTC m=+150.831726101 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 10 22:51:56 crc kubenswrapper[4791]: I1210 22:51:56.009117 4791 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-authentication/oauth-openshift-558db77b4-ht5t7" event={"ID":"35033f1a-b91b-4e02-ba77-3841a2e0fdab","Type":"ContainerStarted","Data":"74fcc1d511780523d69f3aea9165e4405a91db746d4444a4393c970da15aa078"} Dec 10 22:51:56 crc kubenswrapper[4791]: I1210 22:51:56.009160 4791 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-authentication/oauth-openshift-558db77b4-ht5t7" event={"ID":"35033f1a-b91b-4e02-ba77-3841a2e0fdab","Type":"ContainerStarted","Data":"6100b3fe0d820fdba5310d4b39a19699ff6808a46c982f7731462ffd7f68c89c"} Dec 10 22:51:56 crc kubenswrapper[4791]: I1210 22:51:56.010223 4791 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-authentication/oauth-openshift-558db77b4-ht5t7" Dec 10 22:51:56 crc kubenswrapper[4791]: I1210 22:51:56.010618 4791 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-srd52\" (UID: \"e93dbd5d-0ca7-4a15-9eb3-0d71595fada0\") " pod="openshift-image-registry/image-registry-697d97f7c8-srd52" Dec 10 22:51:56 crc kubenswrapper[4791]: E1210 22:51:56.011010 4791 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-10 22:51:56.510998915 +0000 UTC m=+150.940616528 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-srd52" (UID: "e93dbd5d-0ca7-4a15-9eb3-0d71595fada0") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 10 22:51:56 crc kubenswrapper[4791]: I1210 22:51:56.031734 4791 patch_prober.go:28] interesting pod/oauth-openshift-558db77b4-ht5t7 container/oauth-openshift namespace/openshift-authentication: Readiness probe status=failure output="Get \"https://10.217.0.11:6443/healthz\": dial tcp 10.217.0.11:6443: connect: connection refused" start-of-body= Dec 10 22:51:56 crc kubenswrapper[4791]: I1210 22:51:56.031794 4791 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-authentication/oauth-openshift-558db77b4-ht5t7" podUID="35033f1a-b91b-4e02-ba77-3841a2e0fdab" containerName="oauth-openshift" probeResult="failure" output="Get \"https://10.217.0.11:6443/healthz\": dial tcp 10.217.0.11:6443: connect: connection refused" Dec 10 22:51:56 crc kubenswrapper[4791]: I1210 22:51:56.032710 4791 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-controller-84d6567774-sc74x" event={"ID":"e2b778bb-d80f-49f9-804b-0550c648a02f","Type":"ContainerStarted","Data":"349ad29718168aed310aabe928c0ac26032fd2e3f57a980b824d60fec0222a70"} Dec 10 22:51:56 crc kubenswrapper[4791]: I1210 22:51:56.035842 4791 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-multus/multus-admission-controller-857f4d67dd-fbpcr"] Dec 10 22:51:56 crc kubenswrapper[4791]: I1210 22:51:56.086970 4791 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-5t9vv" event={"ID":"465ba82f-1142-4cf1-b2ec-1f73e565ec4d","Type":"ContainerStarted","Data":"b61eeeac663296e3f8441b56a3e48a1a72ca14ed1b7f3f515e066db8eca833f6"} Dec 10 22:51:56 crc kubenswrapper[4791]: I1210 22:51:56.087280 4791 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-etcd-operator/etcd-operator-b45778765-nkt8g"] Dec 10 22:51:56 crc kubenswrapper[4791]: I1210 22:51:56.087313 4791 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-5t9vv" event={"ID":"465ba82f-1142-4cf1-b2ec-1f73e565ec4d","Type":"ContainerStarted","Data":"3f4a625f44c279f956531d80439783a052f7b7b159a4d69a372a952073be167e"} Dec 10 22:51:56 crc kubenswrapper[4791]: I1210 22:51:56.092924 4791 patch_prober.go:28] interesting pod/router-default-5444994796-p97fc container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Dec 10 22:51:56 crc kubenswrapper[4791]: [-]has-synced failed: reason withheld Dec 10 22:51:56 crc kubenswrapper[4791]: [+]process-running ok Dec 10 22:51:56 crc kubenswrapper[4791]: healthz check failed Dec 10 22:51:56 crc kubenswrapper[4791]: I1210 22:51:56.093108 4791 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-p97fc" podUID="d3de1ba0-b014-4f98-b597-cc89f5a70364" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Dec 10 22:51:56 crc kubenswrapper[4791]: I1210 22:51:56.102193 4791 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-bws6n" event={"ID":"06eb2b28-22d6-4116-9ce4-1abf7a21db36","Type":"ContainerStarted","Data":"101a0b0df93291fe6a85b2048d936ab30807d60a36b8c2f7c67f36f431326f31"} Dec 10 22:51:56 crc kubenswrapper[4791]: I1210 22:51:56.111958 4791 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 10 22:51:56 crc kubenswrapper[4791]: E1210 22:51:56.113688 4791 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-10 22:51:56.613663176 +0000 UTC m=+151.043280789 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 10 22:51:56 crc kubenswrapper[4791]: I1210 22:51:56.132109 4791 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-storage-version-migrator/migrator-59844c95c7-vscpm" event={"ID":"868d6095-79bd-41fc-ae14-f369fdf68fde","Type":"ContainerStarted","Data":"d63e4ca801e471f85c224e98870b018f8df17beea56d082b4ba6cf66bc714708"} Dec 10 22:51:56 crc kubenswrapper[4791]: I1210 22:51:56.135822 4791 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-srd52\" (UID: \"e93dbd5d-0ca7-4a15-9eb3-0d71595fada0\") " pod="openshift-image-registry/image-registry-697d97f7c8-srd52" Dec 10 22:51:56 crc kubenswrapper[4791]: E1210 22:51:56.136416 4791 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-10 22:51:56.636396108 +0000 UTC m=+151.066013721 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-srd52" (UID: "e93dbd5d-0ca7-4a15-9eb3-0d71595fada0") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 10 22:51:56 crc kubenswrapper[4791]: I1210 22:51:56.140423 4791 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-ms8jw"] Dec 10 22:51:56 crc kubenswrapper[4791]: I1210 22:51:56.154080 4791 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-console/downloads-7954f5f757-mlb9p" Dec 10 22:51:56 crc kubenswrapper[4791]: I1210 22:51:56.154128 4791 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-console/downloads-7954f5f757-mlb9p" event={"ID":"680d2296-dd48-4786-bea7-64c32edc94c7","Type":"ContainerStarted","Data":"72866ec37c4628747469d84c8a751d1566ed39c7a131b2a0c2ff0d1706ae5415"} Dec 10 22:51:56 crc kubenswrapper[4791]: I1210 22:51:56.154154 4791 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-console/downloads-7954f5f757-mlb9p" event={"ID":"680d2296-dd48-4786-bea7-64c32edc94c7","Type":"ContainerStarted","Data":"199fd8730cb3ee3734bc260d331857c8cb58cc7448f4bb0d1d38d47b3f19147d"} Dec 10 22:51:56 crc kubenswrapper[4791]: I1210 22:51:56.169107 4791 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-k7qsx"] Dec 10 22:51:56 crc kubenswrapper[4791]: I1210 22:51:56.169493 4791 patch_prober.go:28] interesting pod/downloads-7954f5f757-mlb9p container/download-server namespace/openshift-console: Readiness probe status=failure output="Get \"http://10.217.0.10:8080/\": dial tcp 10.217.0.10:8080: connect: connection refused" start-of-body= Dec 10 22:51:56 crc kubenswrapper[4791]: I1210 22:51:56.169576 4791 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-console/downloads-7954f5f757-mlb9p" podUID="680d2296-dd48-4786-bea7-64c32edc94c7" containerName="download-server" probeResult="failure" output="Get \"http://10.217.0.10:8080/\": dial tcp 10.217.0.10:8080: connect: connection refused" Dec 10 22:51:56 crc kubenswrapper[4791]: I1210 22:51:56.182426 4791 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/marketplace-operator-79b997595-7g828"] Dec 10 22:51:56 crc kubenswrapper[4791]: I1210 22:51:56.204518 4791 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-z6gpf"] Dec 10 22:51:56 crc kubenswrapper[4791]: I1210 22:51:56.204566 4791 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-w5dpr"] Dec 10 22:51:56 crc kubenswrapper[4791]: I1210 22:51:56.213628 4791 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-rmjd2"] Dec 10 22:51:56 crc kubenswrapper[4791]: I1210 22:51:56.228425 4791 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-dns-operator/dns-operator-744455d44c-cggs2"] Dec 10 22:51:56 crc kubenswrapper[4791]: I1210 22:51:56.229359 4791 generic.go:334] "Generic (PLEG): container finished" podID="b419e5d6-d6a4-4258-a1bd-a462bd7f7a12" containerID="7fe31745d72ad0a24ab80eefcb1028ec4a49794064d6cef25167ecd530b743d3" exitCode=0 Dec 10 22:51:56 crc kubenswrapper[4791]: I1210 22:51:56.231007 4791 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-config-operator/openshift-config-operator-7777fb866f-wwpsp" event={"ID":"b419e5d6-d6a4-4258-a1bd-a462bd7f7a12","Type":"ContainerDied","Data":"7fe31745d72ad0a24ab80eefcb1028ec4a49794064d6cef25167ecd530b743d3"} Dec 10 22:51:56 crc kubenswrapper[4791]: I1210 22:51:56.231038 4791 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-dns/dns-default-qs77n"] Dec 10 22:51:56 crc kubenswrapper[4791]: I1210 22:51:56.231051 4791 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-config-operator/openshift-config-operator-7777fb866f-wwpsp" event={"ID":"b419e5d6-d6a4-4258-a1bd-a462bd7f7a12","Type":"ContainerStarted","Data":"c3ff611a802a93d7cd48c0a382a4a4f10ace5b0955f9431c78652ae4feed1aef"} Dec 10 22:51:56 crc kubenswrapper[4791]: I1210 22:51:56.236470 4791 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-ingress-canary/ingress-canary-9dwf5"] Dec 10 22:51:56 crc kubenswrapper[4791]: I1210 22:51:56.236508 4791 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-service-ca-operator/service-ca-operator-777779d784-b8v4c"] Dec 10 22:51:56 crc kubenswrapper[4791]: I1210 22:51:56.237037 4791 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 10 22:51:56 crc kubenswrapper[4791]: E1210 22:51:56.238290 4791 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-10 22:51:56.738274437 +0000 UTC m=+151.167892050 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 10 22:51:56 crc kubenswrapper[4791]: I1210 22:51:56.243432 4791 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-service-ca/service-ca-9c57cc56f-x9pw8"] Dec 10 22:51:56 crc kubenswrapper[4791]: I1210 22:51:56.244414 4791 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-q4cpx" event={"ID":"33c30ace-bd17-4a3b-a43a-ddfeccb88349","Type":"ContainerStarted","Data":"b4683d63557333aad9f323a1cf40174f7a727a6fc3538ce5ef9dc0b034d45169"} Dec 10 22:51:56 crc kubenswrapper[4791]: I1210 22:51:56.249946 4791 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["hostpath-provisioner/csi-hostpathplugin-mkz5r"] Dec 10 22:51:56 crc kubenswrapper[4791]: I1210 22:51:56.255531 4791 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29423445-b9cxb"] Dec 10 22:51:56 crc kubenswrapper[4791]: I1210 22:51:56.312195 4791 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ingress-operator/ingress-operator-5b745b69d9-8npst" event={"ID":"e92ca7e3-d360-4e14-aeb9-76738a44c391","Type":"ContainerStarted","Data":"afc17026b0673c5d860fea9ddb1ab4b0198eae8f1bf3369f34e708498e80952b"} Dec 10 22:51:56 crc kubenswrapper[4791]: I1210 22:51:56.312538 4791 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ingress-operator/ingress-operator-5b745b69d9-8npst" event={"ID":"e92ca7e3-d360-4e14-aeb9-76738a44c391","Type":"ContainerStarted","Data":"675018a16342172a763fe858fc105c5f5bdf54b0e66377e94f08ce5647604dba"} Dec 10 22:51:56 crc kubenswrapper[4791]: I1210 22:51:56.325465 4791 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-497m8" event={"ID":"8f0772d5-db77-42b2-b712-0293c0422161","Type":"ContainerStarted","Data":"2a673925b6a9b5369c5dec28c58733217e066e08890de9acac1a4b057a70ff83"} Dec 10 22:51:56 crc kubenswrapper[4791]: I1210 22:51:56.338585 4791 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-srd52\" (UID: \"e93dbd5d-0ca7-4a15-9eb3-0d71595fada0\") " pod="openshift-image-registry/image-registry-697d97f7c8-srd52" Dec 10 22:51:56 crc kubenswrapper[4791]: E1210 22:51:56.338961 4791 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-10 22:51:56.838948482 +0000 UTC m=+151.268566095 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-srd52" (UID: "e93dbd5d-0ca7-4a15-9eb3-0d71595fada0") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 10 22:51:56 crc kubenswrapper[4791]: I1210 22:51:56.363257 4791 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-vm842"] Dec 10 22:51:56 crc kubenswrapper[4791]: I1210 22:51:56.363299 4791 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-apiserver/apiserver-76f77b778f-kqtbx" event={"ID":"4a1cf441-8a08-4677-a480-947c8cddd6a6","Type":"ContainerStarted","Data":"fc03a28215b432769fd9607896bfb42ad90c061e9c09a2ad1d235d974e2a0450"} Dec 10 22:51:56 crc kubenswrapper[4791]: I1210 22:51:56.380726 4791 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-operator-74547568cd-rcnrn" event={"ID":"2fee22d4-2e9b-4768-9603-22e11d9dad9a","Type":"ContainerStarted","Data":"f19854c3d2e7df0b083828a7ce9784ca5daf7050a817ec871f5ded2ecbdd20bf"} Dec 10 22:51:56 crc kubenswrapper[4791]: I1210 22:51:56.385743 4791 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-vsnsh" event={"ID":"5369b8df-3c36-4108-b4e7-cfb11478b3a7","Type":"ContainerStarted","Data":"99a5d5b9b8638ab2472e67f0037b48896f9464090eb279f1bfb13addf1ffd77d"} Dec 10 22:51:56 crc kubenswrapper[4791]: I1210 22:51:56.386368 4791 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-vsnsh" Dec 10 22:51:56 crc kubenswrapper[4791]: I1210 22:51:56.390583 4791 patch_prober.go:28] interesting pod/route-controller-manager-6576b87f9c-vsnsh container/route-controller-manager namespace/openshift-route-controller-manager: Readiness probe status=failure output="Get \"https://10.217.0.8:8443/healthz\": dial tcp 10.217.0.8:8443: connect: connection refused" start-of-body= Dec 10 22:51:56 crc kubenswrapper[4791]: I1210 22:51:56.390678 4791 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-vsnsh" podUID="5369b8df-3c36-4108-b4e7-cfb11478b3a7" containerName="route-controller-manager" probeResult="failure" output="Get \"https://10.217.0.8:8443/healthz\": dial tcp 10.217.0.8:8443: connect: connection refused" Dec 10 22:51:56 crc kubenswrapper[4791]: I1210 22:51:56.411318 4791 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-t7742" event={"ID":"4205f348-3276-40f3-a35b-69306b331d80","Type":"ContainerStarted","Data":"248f9b167248762d9f65bfb46ac6fbd8f9367815f2c6a30e055a2724eaa2c1ae"} Dec 10 22:51:56 crc kubenswrapper[4791]: I1210 22:51:56.415804 4791 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-console/console-f9d7485db-zkfkv" event={"ID":"1b61a973-db5e-4aa7-96e3-7fbdd9dd8cc0","Type":"ContainerStarted","Data":"8be5e83a023b172a495e79f8fe719fc8e54ba3500d683e4fce82952ba90d0523"} Dec 10 22:51:56 crc kubenswrapper[4791]: I1210 22:51:56.439415 4791 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 10 22:51:56 crc kubenswrapper[4791]: E1210 22:51:56.440265 4791 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-10 22:51:56.940227504 +0000 UTC m=+151.369845187 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 10 22:51:56 crc kubenswrapper[4791]: I1210 22:51:56.441606 4791 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-mldvd" event={"ID":"5be73604-c6c9-4ef6-a53d-6c3f6df27202","Type":"ContainerStarted","Data":"dc51f4a9c5063fb3535d141c18eed04074d235ed082c038e90fb7b048cdb353e"} Dec 10 22:51:56 crc kubenswrapper[4791]: I1210 22:51:56.441667 4791 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-mldvd" event={"ID":"5be73604-c6c9-4ef6-a53d-6c3f6df27202","Type":"ContainerStarted","Data":"fdf99ce46ea76b994f89654fee996e46a4c14b5a6a4dab95849768db51e90894"} Dec 10 22:51:56 crc kubenswrapper[4791]: I1210 22:51:56.464392 4791 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-authentication-operator/authentication-operator-69f744f599-7rhk9" event={"ID":"cff60894-8deb-4354-9ad1-3a83baa98714","Type":"ContainerStarted","Data":"8790da15304edad266227c9ea255f602bd0f6f4a39c3eaece30199b43a5c4c63"} Dec 10 22:51:56 crc kubenswrapper[4791]: I1210 22:51:56.507425 4791 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-console-operator/console-operator-58897d9998-hd6zl" Dec 10 22:51:56 crc kubenswrapper[4791]: I1210 22:51:56.559443 4791 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-srd52\" (UID: \"e93dbd5d-0ca7-4a15-9eb3-0d71595fada0\") " pod="openshift-image-registry/image-registry-697d97f7c8-srd52" Dec 10 22:51:56 crc kubenswrapper[4791]: E1210 22:51:56.562072 4791 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-10 22:51:57.062059626 +0000 UTC m=+151.491677239 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-srd52" (UID: "e93dbd5d-0ca7-4a15-9eb3-0d71595fada0") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 10 22:51:56 crc kubenswrapper[4791]: I1210 22:51:56.665765 4791 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 10 22:51:56 crc kubenswrapper[4791]: E1210 22:51:56.667389 4791 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-10 22:51:57.167367672 +0000 UTC m=+151.596985355 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 10 22:51:56 crc kubenswrapper[4791]: I1210 22:51:56.782470 4791 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-srd52\" (UID: \"e93dbd5d-0ca7-4a15-9eb3-0d71595fada0\") " pod="openshift-image-registry/image-registry-697d97f7c8-srd52" Dec 10 22:51:56 crc kubenswrapper[4791]: E1210 22:51:56.783293 4791 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-10 22:51:57.283273977 +0000 UTC m=+151.712891590 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-srd52" (UID: "e93dbd5d-0ca7-4a15-9eb3-0d71595fada0") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 10 22:51:56 crc kubenswrapper[4791]: I1210 22:51:56.813919 4791 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-authentication-operator/authentication-operator-69f744f599-7rhk9" podStartSLOduration=126.813901623 podStartE2EDuration="2m6.813901623s" podCreationTimestamp="2025-12-10 22:49:50 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-10 22:51:56.813487151 +0000 UTC m=+151.243104764" watchObservedRunningTime="2025-12-10 22:51:56.813901623 +0000 UTC m=+151.243519236" Dec 10 22:51:56 crc kubenswrapper[4791]: I1210 22:51:56.861917 4791 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-console/downloads-7954f5f757-mlb9p" podStartSLOduration=125.861899919 podStartE2EDuration="2m5.861899919s" podCreationTimestamp="2025-12-10 22:49:51 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-10 22:51:56.842004437 +0000 UTC m=+151.271622050" watchObservedRunningTime="2025-12-10 22:51:56.861899919 +0000 UTC m=+151.291517532" Dec 10 22:51:56 crc kubenswrapper[4791]: I1210 22:51:56.885506 4791 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 10 22:51:56 crc kubenswrapper[4791]: E1210 22:51:56.885948 4791 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-10 22:51:57.385933408 +0000 UTC m=+151.815551021 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 10 22:51:56 crc kubenswrapper[4791]: I1210 22:51:56.915023 4791 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-authentication/oauth-openshift-558db77b4-ht5t7" podStartSLOduration=126.91500456 podStartE2EDuration="2m6.91500456s" podCreationTimestamp="2025-12-10 22:49:50 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-10 22:51:56.895642302 +0000 UTC m=+151.325259915" watchObservedRunningTime="2025-12-10 22:51:56.91500456 +0000 UTC m=+151.344622173" Dec 10 22:51:56 crc kubenswrapper[4791]: I1210 22:51:56.979215 4791 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-machine-config-operator/machine-config-controller-84d6567774-sc74x" podStartSLOduration=125.979199114 podStartE2EDuration="2m5.979199114s" podCreationTimestamp="2025-12-10 22:49:51 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-10 22:51:56.977880576 +0000 UTC m=+151.407498189" watchObservedRunningTime="2025-12-10 22:51:56.979199114 +0000 UTC m=+151.408816727" Dec 10 22:51:56 crc kubenswrapper[4791]: I1210 22:51:56.979997 4791 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-mldvd" podStartSLOduration=125.979993006 podStartE2EDuration="2m5.979993006s" podCreationTimestamp="2025-12-10 22:49:51 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-10 22:51:56.922408269 +0000 UTC m=+151.352025882" watchObservedRunningTime="2025-12-10 22:51:56.979993006 +0000 UTC m=+151.409610619" Dec 10 22:51:56 crc kubenswrapper[4791]: I1210 22:51:56.986782 4791 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-srd52\" (UID: \"e93dbd5d-0ca7-4a15-9eb3-0d71595fada0\") " pod="openshift-image-registry/image-registry-697d97f7c8-srd52" Dec 10 22:51:56 crc kubenswrapper[4791]: E1210 22:51:56.987126 4791 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-10 22:51:57.487112887 +0000 UTC m=+151.916730500 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-srd52" (UID: "e93dbd5d-0ca7-4a15-9eb3-0d71595fada0") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 10 22:51:57 crc kubenswrapper[4791]: I1210 22:51:57.047564 4791 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-vsnsh" podStartSLOduration=126.047550275 podStartE2EDuration="2m6.047550275s" podCreationTimestamp="2025-12-10 22:49:51 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-10 22:51:57.012476884 +0000 UTC m=+151.442094497" watchObservedRunningTime="2025-12-10 22:51:57.047550275 +0000 UTC m=+151.477167888" Dec 10 22:51:57 crc kubenswrapper[4791]: I1210 22:51:57.077419 4791 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-5t9vv" podStartSLOduration=126.077406739 podStartE2EDuration="2m6.077406739s" podCreationTimestamp="2025-12-10 22:49:51 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-10 22:51:57.076370869 +0000 UTC m=+151.505988482" watchObservedRunningTime="2025-12-10 22:51:57.077406739 +0000 UTC m=+151.507024352" Dec 10 22:51:57 crc kubenswrapper[4791]: I1210 22:51:57.078229 4791 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-apiserver/apiserver-76f77b778f-kqtbx" podStartSLOduration=127.078224172 podStartE2EDuration="2m7.078224172s" podCreationTimestamp="2025-12-10 22:49:50 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-10 22:51:57.049469619 +0000 UTC m=+151.479087232" watchObservedRunningTime="2025-12-10 22:51:57.078224172 +0000 UTC m=+151.507841785" Dec 10 22:51:57 crc kubenswrapper[4791]: I1210 22:51:57.088164 4791 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 10 22:51:57 crc kubenswrapper[4791]: E1210 22:51:57.088540 4791 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-10 22:51:57.588525453 +0000 UTC m=+152.018143056 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 10 22:51:57 crc kubenswrapper[4791]: I1210 22:51:57.094740 4791 patch_prober.go:28] interesting pod/router-default-5444994796-p97fc container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Dec 10 22:51:57 crc kubenswrapper[4791]: [-]has-synced failed: reason withheld Dec 10 22:51:57 crc kubenswrapper[4791]: [+]process-running ok Dec 10 22:51:57 crc kubenswrapper[4791]: healthz check failed Dec 10 22:51:57 crc kubenswrapper[4791]: I1210 22:51:57.095090 4791 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-p97fc" podUID="d3de1ba0-b014-4f98-b597-cc89f5a70364" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Dec 10 22:51:57 crc kubenswrapper[4791]: I1210 22:51:57.189683 4791 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-srd52\" (UID: \"e93dbd5d-0ca7-4a15-9eb3-0d71595fada0\") " pod="openshift-image-registry/image-registry-697d97f7c8-srd52" Dec 10 22:51:57 crc kubenswrapper[4791]: E1210 22:51:57.190200 4791 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-10 22:51:57.690189056 +0000 UTC m=+152.119806669 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-srd52" (UID: "e93dbd5d-0ca7-4a15-9eb3-0d71595fada0") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 10 22:51:57 crc kubenswrapper[4791]: I1210 22:51:57.296905 4791 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 10 22:51:57 crc kubenswrapper[4791]: E1210 22:51:57.297242 4791 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-10 22:51:57.79722762 +0000 UTC m=+152.226845233 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 10 22:51:57 crc kubenswrapper[4791]: I1210 22:51:57.398179 4791 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-srd52\" (UID: \"e93dbd5d-0ca7-4a15-9eb3-0d71595fada0\") " pod="openshift-image-registry/image-registry-697d97f7c8-srd52" Dec 10 22:51:57 crc kubenswrapper[4791]: E1210 22:51:57.398503 4791 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-10 22:51:57.898491332 +0000 UTC m=+152.328108945 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-srd52" (UID: "e93dbd5d-0ca7-4a15-9eb3-0d71595fada0") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 10 22:51:57 crc kubenswrapper[4791]: I1210 22:51:57.494419 4791 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-apiserver/apiserver-76f77b778f-kqtbx" Dec 10 22:51:57 crc kubenswrapper[4791]: I1210 22:51:57.494858 4791 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-apiserver/apiserver-76f77b778f-kqtbx" Dec 10 22:51:57 crc kubenswrapper[4791]: I1210 22:51:57.500786 4791 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 10 22:51:57 crc kubenswrapper[4791]: E1210 22:51:57.501172 4791 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-10 22:51:58.001157603 +0000 UTC m=+152.430775216 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 10 22:51:57 crc kubenswrapper[4791]: I1210 22:51:57.507043 4791 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-k7qsx" event={"ID":"169ebb73-7b32-4f87-a8d7-9272bff04856","Type":"ContainerStarted","Data":"e5de1310f0be962e52d8e0fbe23b45061106fcd04157a47e799428e263936701"} Dec 10 22:51:57 crc kubenswrapper[4791]: I1210 22:51:57.507687 4791 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-dns/dns-default-qs77n" event={"ID":"e9978f48-c33a-4765-9b05-16e37f5cdb24","Type":"ContainerStarted","Data":"ac395d18d432dca20f411762e3e8342f22f4d8e4b6b1b5a8716ac31caaf6e31b"} Dec 10 22:51:57 crc kubenswrapper[4791]: I1210 22:51:57.508516 4791 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-z6gpf" event={"ID":"8f993103-9ee4-446f-a91a-a03dcd3ed9ea","Type":"ContainerStarted","Data":"436565b57357c782ca813e4b15823855c7f50d884e93e6a98318feb144029a3f"} Dec 10 22:51:57 crc kubenswrapper[4791]: I1210 22:51:57.508538 4791 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-z6gpf" event={"ID":"8f993103-9ee4-446f-a91a-a03dcd3ed9ea","Type":"ContainerStarted","Data":"e6e1494733bc5a9a4136d66ccfbec996fb27c7eedefeb220e126838b6a9182e5"} Dec 10 22:51:57 crc kubenswrapper[4791]: I1210 22:51:57.532312 4791 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-storage-version-migrator/migrator-59844c95c7-vscpm" event={"ID":"868d6095-79bd-41fc-ae14-f369fdf68fde","Type":"ContainerStarted","Data":"9417737d91f525a9747e4f0aa3d5255172dfd272c86c331d73a1fde9930aada3"} Dec 10 22:51:57 crc kubenswrapper[4791]: I1210 22:51:57.532376 4791 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-storage-version-migrator/migrator-59844c95c7-vscpm" event={"ID":"868d6095-79bd-41fc-ae14-f369fdf68fde","Type":"ContainerStarted","Data":"4c8441fb9f5d57642c0e7f3767c2847cb2b83bd84b9fa6eeeaa7882075577f4a"} Dec 10 22:51:57 crc kubenswrapper[4791]: I1210 22:51:57.577434 4791 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-nwwpb" Dec 10 22:51:57 crc kubenswrapper[4791]: I1210 22:51:57.580685 4791 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-nwwpb" Dec 10 22:51:57 crc kubenswrapper[4791]: I1210 22:51:57.595221 4791 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-z6gpf" podStartSLOduration=126.59520492 podStartE2EDuration="2m6.59520492s" podCreationTimestamp="2025-12-10 22:49:51 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-10 22:51:57.54671901 +0000 UTC m=+151.976336643" watchObservedRunningTime="2025-12-10 22:51:57.59520492 +0000 UTC m=+152.024822533" Dec 10 22:51:57 crc kubenswrapper[4791]: I1210 22:51:57.595554 4791 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-kube-storage-version-migrator/migrator-59844c95c7-vscpm" podStartSLOduration=126.59555103 podStartE2EDuration="2m6.59555103s" podCreationTimestamp="2025-12-10 22:49:51 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-10 22:51:57.593595105 +0000 UTC m=+152.023212718" watchObservedRunningTime="2025-12-10 22:51:57.59555103 +0000 UTC m=+152.025168643" Dec 10 22:51:57 crc kubenswrapper[4791]: I1210 22:51:57.602458 4791 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-srd52\" (UID: \"e93dbd5d-0ca7-4a15-9eb3-0d71595fada0\") " pod="openshift-image-registry/image-registry-697d97f7c8-srd52" Dec 10 22:51:57 crc kubenswrapper[4791]: E1210 22:51:57.603579 4791 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-10 22:51:58.103562806 +0000 UTC m=+152.533180419 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-srd52" (UID: "e93dbd5d-0ca7-4a15-9eb3-0d71595fada0") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 10 22:51:57 crc kubenswrapper[4791]: I1210 22:51:57.604018 4791 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-nwwpb" Dec 10 22:51:57 crc kubenswrapper[4791]: I1210 22:51:57.651806 4791 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-bws6n" event={"ID":"06eb2b28-22d6-4116-9ce4-1abf7a21db36","Type":"ContainerStarted","Data":"b9277324222bf644c860b62c30de58b11630f8e507eed433de44b74f8dd4d661"} Dec 10 22:51:57 crc kubenswrapper[4791]: I1210 22:51:57.652513 4791 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-bws6n" Dec 10 22:51:57 crc kubenswrapper[4791]: I1210 22:51:57.670772 4791 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ingress-operator/ingress-operator-5b745b69d9-8npst" event={"ID":"e92ca7e3-d360-4e14-aeb9-76738a44c391","Type":"ContainerStarted","Data":"ec5bbc341b16f333896641afd6abeda5fd4cb5d7c0051ba004702f1719934053"} Dec 10 22:51:57 crc kubenswrapper[4791]: I1210 22:51:57.705205 4791 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 10 22:51:57 crc kubenswrapper[4791]: E1210 22:51:57.705534 4791 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-10 22:51:58.205518017 +0000 UTC m=+152.635135620 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 10 22:51:57 crc kubenswrapper[4791]: I1210 22:51:57.706670 4791 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-bws6n" podStartSLOduration=126.70666213 podStartE2EDuration="2m6.70666213s" podCreationTimestamp="2025-12-10 22:49:51 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-10 22:51:57.705696842 +0000 UTC m=+152.135314455" watchObservedRunningTime="2025-12-10 22:51:57.70666213 +0000 UTC m=+152.136279743" Dec 10 22:51:57 crc kubenswrapper[4791]: I1210 22:51:57.717643 4791 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-bws6n" Dec 10 22:51:57 crc kubenswrapper[4791]: I1210 22:51:57.737516 4791 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-v445v" event={"ID":"a05f3579-4703-4c46-bb11-5de60eedbc02","Type":"ContainerStarted","Data":"ad51e868c86ca16b33537ab226913804cbf6dbfff939c5b6ecd4a60e19782832"} Dec 10 22:51:57 crc kubenswrapper[4791]: I1210 22:51:57.737560 4791 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-v445v" event={"ID":"a05f3579-4703-4c46-bb11-5de60eedbc02","Type":"ContainerStarted","Data":"fc9b9a8f70b5718715ebd65b9d71eefb675509cd8b431bfc7f804f01b456c6fd"} Dec 10 22:51:57 crc kubenswrapper[4791]: I1210 22:51:57.738354 4791 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-v445v" Dec 10 22:51:57 crc kubenswrapper[4791]: I1210 22:51:57.748520 4791 patch_prober.go:28] interesting pod/packageserver-d55dfcdfc-v445v container/packageserver namespace/openshift-operator-lifecycle-manager: Readiness probe status=failure output="Get \"https://10.217.0.36:5443/healthz\": dial tcp 10.217.0.36:5443: connect: connection refused" start-of-body= Dec 10 22:51:57 crc kubenswrapper[4791]: I1210 22:51:57.748580 4791 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-v445v" podUID="a05f3579-4703-4c46-bb11-5de60eedbc02" containerName="packageserver" probeResult="failure" output="Get \"https://10.217.0.36:5443/healthz\": dial tcp 10.217.0.36:5443: connect: connection refused" Dec 10 22:51:57 crc kubenswrapper[4791]: I1210 22:51:57.749482 4791 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-dns-operator/dns-operator-744455d44c-cggs2" event={"ID":"5a2896d3-fca8-4c64-abeb-6320f0b78d4e","Type":"ContainerStarted","Data":"167a3c86e6fe38f1602763df32ee0751a7b0598473fad4d417e5c5a28d457794"} Dec 10 22:51:57 crc kubenswrapper[4791]: I1210 22:51:57.763325 4791 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-497m8" event={"ID":"8f0772d5-db77-42b2-b712-0293c0422161","Type":"ContainerStarted","Data":"3ea8e116aabfca4b66f71ec8d3a0241864fcb5f8e8f6d7e1ee9b611af63084b2"} Dec 10 22:51:57 crc kubenswrapper[4791]: I1210 22:51:57.764458 4791 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-497m8" Dec 10 22:51:57 crc kubenswrapper[4791]: I1210 22:51:57.780397 4791 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-497m8" Dec 10 22:51:57 crc kubenswrapper[4791]: I1210 22:51:57.792186 4791 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-vsnsh" event={"ID":"5369b8df-3c36-4108-b4e7-cfb11478b3a7","Type":"ContainerStarted","Data":"4b9f827d94a26dadac2b0de34bb8c127f77e2fb7e7156bfd1f63d2ea2036e664"} Dec 10 22:51:57 crc kubenswrapper[4791]: I1210 22:51:57.804568 4791 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-vsnsh" Dec 10 22:51:57 crc kubenswrapper[4791]: I1210 22:51:57.806178 4791 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-ingress-operator/ingress-operator-5b745b69d9-8npst" podStartSLOduration=126.806168292 podStartE2EDuration="2m6.806168292s" podCreationTimestamp="2025-12-10 22:49:51 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-10 22:51:57.738563011 +0000 UTC m=+152.168180624" watchObservedRunningTime="2025-12-10 22:51:57.806168292 +0000 UTC m=+152.235785905" Dec 10 22:51:57 crc kubenswrapper[4791]: I1210 22:51:57.806404 4791 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-v445v" podStartSLOduration=126.806400378 podStartE2EDuration="2m6.806400378s" podCreationTimestamp="2025-12-10 22:49:51 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-10 22:51:57.781744921 +0000 UTC m=+152.211362534" watchObservedRunningTime="2025-12-10 22:51:57.806400378 +0000 UTC m=+152.236017991" Dec 10 22:51:57 crc kubenswrapper[4791]: I1210 22:51:57.807823 4791 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-srd52\" (UID: \"e93dbd5d-0ca7-4a15-9eb3-0d71595fada0\") " pod="openshift-image-registry/image-registry-697d97f7c8-srd52" Dec 10 22:51:57 crc kubenswrapper[4791]: E1210 22:51:57.811143 4791 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-10 22:51:58.311130362 +0000 UTC m=+152.740747975 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-srd52" (UID: "e93dbd5d-0ca7-4a15-9eb3-0d71595fada0") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 10 22:51:57 crc kubenswrapper[4791]: I1210 22:51:57.834160 4791 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-service-ca-operator/service-ca-operator-777779d784-b8v4c" event={"ID":"64199728-979b-4628-8b37-2dd792f86439","Type":"ContainerStarted","Data":"23a58577db6816237987bd8e2ec4590e747f7011bcc1e901ae7d00ce6655b098"} Dec 10 22:51:57 crc kubenswrapper[4791]: I1210 22:51:57.846982 4791 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29423445-b9cxb" event={"ID":"90640468-53d7-42bc-a835-f7acf48a486e","Type":"ContainerStarted","Data":"e2c74df205f03bdc1e7206068bf864feaf6f07d7ae52dd750d7b9bb336e3c712"} Dec 10 22:51:57 crc kubenswrapper[4791]: I1210 22:51:57.847023 4791 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29423445-b9cxb" event={"ID":"90640468-53d7-42bc-a835-f7acf48a486e","Type":"ContainerStarted","Data":"b6daf82e1d63eb3f8e9105be94dd5bfa39e380aaaa223e2aa77660eccd4f2d07"} Dec 10 22:51:57 crc kubenswrapper[4791]: I1210 22:51:57.868004 4791 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/marketplace-operator-79b997595-7g828" event={"ID":"a9185ee1-adaf-48e8-b9a9-90ef658f8212","Type":"ContainerStarted","Data":"a2010281d4a0a094eb30b9d1a4f837090035e9f671f22891a41b5288b0f507fc"} Dec 10 22:51:57 crc kubenswrapper[4791]: I1210 22:51:57.869216 4791 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/marketplace-operator-79b997595-7g828" Dec 10 22:51:57 crc kubenswrapper[4791]: I1210 22:51:57.882213 4791 patch_prober.go:28] interesting pod/marketplace-operator-79b997595-7g828 container/marketplace-operator namespace/openshift-marketplace: Readiness probe status=failure output="Get \"http://10.217.0.42:8080/healthz\": dial tcp 10.217.0.42:8080: connect: connection refused" start-of-body= Dec 10 22:51:57 crc kubenswrapper[4791]: I1210 22:51:57.882260 4791 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-marketplace/marketplace-operator-79b997595-7g828" podUID="a9185ee1-adaf-48e8-b9a9-90ef658f8212" containerName="marketplace-operator" probeResult="failure" output="Get \"http://10.217.0.42:8080/healthz\": dial tcp 10.217.0.42:8080: connect: connection refused" Dec 10 22:51:57 crc kubenswrapper[4791]: I1210 22:51:57.908428 4791 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 10 22:51:57 crc kubenswrapper[4791]: E1210 22:51:57.909616 4791 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-10 22:51:58.409600564 +0000 UTC m=+152.839218177 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 10 22:51:57 crc kubenswrapper[4791]: I1210 22:51:57.929819 4791 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-497m8" podStartSLOduration=126.929801865 podStartE2EDuration="2m6.929801865s" podCreationTimestamp="2025-12-10 22:49:51 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-10 22:51:57.861851645 +0000 UTC m=+152.291469248" watchObservedRunningTime="2025-12-10 22:51:57.929801865 +0000 UTC m=+152.359419478" Dec 10 22:51:57 crc kubenswrapper[4791]: I1210 22:51:57.960440 4791 patch_prober.go:28] interesting pod/apiserver-76f77b778f-kqtbx container/openshift-apiserver namespace/openshift-apiserver: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[+]ping ok Dec 10 22:51:57 crc kubenswrapper[4791]: [+]log ok Dec 10 22:51:57 crc kubenswrapper[4791]: [+]etcd ok Dec 10 22:51:57 crc kubenswrapper[4791]: [+]poststarthook/start-apiserver-admission-initializer ok Dec 10 22:51:57 crc kubenswrapper[4791]: [+]poststarthook/generic-apiserver-start-informers ok Dec 10 22:51:57 crc kubenswrapper[4791]: [+]poststarthook/max-in-flight-filter ok Dec 10 22:51:57 crc kubenswrapper[4791]: [+]poststarthook/storage-object-count-tracker-hook ok Dec 10 22:51:57 crc kubenswrapper[4791]: [+]poststarthook/image.openshift.io-apiserver-caches ok Dec 10 22:51:57 crc kubenswrapper[4791]: [-]poststarthook/authorization.openshift.io-bootstrapclusterroles failed: reason withheld Dec 10 22:51:57 crc kubenswrapper[4791]: [-]poststarthook/authorization.openshift.io-ensurenodebootstrap-sa failed: reason withheld Dec 10 22:51:57 crc kubenswrapper[4791]: [+]poststarthook/project.openshift.io-projectcache ok Dec 10 22:51:57 crc kubenswrapper[4791]: [+]poststarthook/project.openshift.io-projectauthorizationcache ok Dec 10 22:51:57 crc kubenswrapper[4791]: [+]poststarthook/openshift.io-startinformers ok Dec 10 22:51:57 crc kubenswrapper[4791]: [+]poststarthook/openshift.io-restmapperupdater ok Dec 10 22:51:57 crc kubenswrapper[4791]: [+]poststarthook/quota.openshift.io-clusterquotamapping ok Dec 10 22:51:57 crc kubenswrapper[4791]: livez check failed Dec 10 22:51:57 crc kubenswrapper[4791]: I1210 22:51:57.960505 4791 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-apiserver/apiserver-76f77b778f-kqtbx" podUID="4a1cf441-8a08-4677-a480-947c8cddd6a6" containerName="openshift-apiserver" probeResult="failure" output="HTTP probe failed with statuscode: 500" Dec 10 22:51:57 crc kubenswrapper[4791]: I1210 22:51:57.997837 4791 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/marketplace-operator-79b997595-7g828" podStartSLOduration=126.997821617 podStartE2EDuration="2m6.997821617s" podCreationTimestamp="2025-12-10 22:49:51 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-10 22:51:57.997532239 +0000 UTC m=+152.427149852" watchObservedRunningTime="2025-12-10 22:51:57.997821617 +0000 UTC m=+152.427439220" Dec 10 22:51:58 crc kubenswrapper[4791]: I1210 22:51:58.002393 4791 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-operator-74547568cd-rcnrn" event={"ID":"2fee22d4-2e9b-4768-9603-22e11d9dad9a","Type":"ContainerStarted","Data":"f18e70d7e5947924627ebb213cda9a0f291046589877680338349aa31b16e8f2"} Dec 10 22:51:58 crc kubenswrapper[4791]: I1210 22:51:58.005136 4791 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-service-ca/service-ca-9c57cc56f-x9pw8" event={"ID":"34da0cc5-3e59-4c25-af94-e755e772b89a","Type":"ContainerStarted","Data":"e6d13f216bf94fc630841087c8d779ac1fa125104d836f75cc6e55036929206c"} Dec 10 22:51:58 crc kubenswrapper[4791]: I1210 22:51:58.010185 4791 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-srd52\" (UID: \"e93dbd5d-0ca7-4a15-9eb3-0d71595fada0\") " pod="openshift-image-registry/image-registry-697d97f7c8-srd52" Dec 10 22:51:58 crc kubenswrapper[4791]: E1210 22:51:58.011410 4791 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-10 22:51:58.511390981 +0000 UTC m=+152.941008594 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-srd52" (UID: "e93dbd5d-0ca7-4a15-9eb3-0d71595fada0") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 10 22:51:58 crc kubenswrapper[4791]: I1210 22:51:58.037775 4791 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-rmjd2" event={"ID":"93a38a79-5c1c-4296-849e-94137de8f974","Type":"ContainerStarted","Data":"50bfa6369511bc41fca5d54c0bd3b7dfe7578522cad6921379bd498d572d077c"} Dec 10 22:51:58 crc kubenswrapper[4791]: I1210 22:51:58.065969 4791 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-operator-lifecycle-manager/collect-profiles-29423445-b9cxb" podStartSLOduration=128.065956282 podStartE2EDuration="2m8.065956282s" podCreationTimestamp="2025-12-10 22:49:50 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-10 22:51:58.038551208 +0000 UTC m=+152.468168821" watchObservedRunningTime="2025-12-10 22:51:58.065956282 +0000 UTC m=+152.495573895" Dec 10 22:51:58 crc kubenswrapper[4791]: I1210 22:51:58.066937 4791 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-rmjd2" podStartSLOduration=128.06693163 podStartE2EDuration="2m8.06693163s" podCreationTimestamp="2025-12-10 22:49:50 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-10 22:51:58.065562751 +0000 UTC m=+152.495180364" watchObservedRunningTime="2025-12-10 22:51:58.06693163 +0000 UTC m=+152.496549243" Dec 10 22:51:58 crc kubenswrapper[4791]: I1210 22:51:58.111982 4791 patch_prober.go:28] interesting pod/router-default-5444994796-p97fc container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Dec 10 22:51:58 crc kubenswrapper[4791]: [-]has-synced failed: reason withheld Dec 10 22:51:58 crc kubenswrapper[4791]: [+]process-running ok Dec 10 22:51:58 crc kubenswrapper[4791]: healthz check failed Dec 10 22:51:58 crc kubenswrapper[4791]: I1210 22:51:58.112299 4791 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-p97fc" podUID="d3de1ba0-b014-4f98-b597-cc89f5a70364" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Dec 10 22:51:58 crc kubenswrapper[4791]: I1210 22:51:58.113084 4791 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 10 22:51:58 crc kubenswrapper[4791]: E1210 22:51:58.114034 4791 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-10 22:51:58.614019221 +0000 UTC m=+153.043636834 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 10 22:51:58 crc kubenswrapper[4791]: I1210 22:51:58.143471 4791 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-w5dpr" event={"ID":"3f132df5-3de5-4455-a3da-7848b0de69cc","Type":"ContainerStarted","Data":"de7e1b00332155f842d43e93e5dc309adacb1267f13c2e788663e98527302f33"} Dec 10 22:51:58 crc kubenswrapper[4791]: I1210 22:51:58.143516 4791 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-w5dpr" event={"ID":"3f132df5-3de5-4455-a3da-7848b0de69cc","Type":"ContainerStarted","Data":"aa6dfc42a40a78838a6f4a6dc5c7287330268c734f5e6758932c45b0278d7d1e"} Dec 10 22:51:58 crc kubenswrapper[4791]: I1210 22:51:58.154611 4791 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-q4cpx" event={"ID":"33c30ace-bd17-4a3b-a43a-ddfeccb88349","Type":"ContainerStarted","Data":"73496ad773b823cb687e9cfda6ca5182a304903626fb88d639c77f14869d061f"} Dec 10 22:51:58 crc kubenswrapper[4791]: I1210 22:51:58.158267 4791 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-admission-controller-857f4d67dd-fbpcr" event={"ID":"ded98820-5308-46e3-b87b-18d0a81c0519","Type":"ContainerStarted","Data":"dba9883df5dc0dad0a392889ebc0120e0bfef2bf34f86a9e4e49a42924db9bbb"} Dec 10 22:51:58 crc kubenswrapper[4791]: I1210 22:51:58.165179 4791 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ingress-canary/ingress-canary-9dwf5" event={"ID":"0921ef8b-8842-4d66-890b-3ad15bd48e70","Type":"ContainerStarted","Data":"37441e10e6ed193dc3b3bb514570fce78d7da83fb7f1252cb2b326bdd302a444"} Dec 10 22:51:58 crc kubenswrapper[4791]: I1210 22:51:58.175462 4791 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-etcd-operator/etcd-operator-b45778765-nkt8g" event={"ID":"97c517e6-0af1-4c17-a2b9-f7a7db3368b3","Type":"ContainerStarted","Data":"57cb925eaa5c99ef1144c892c05e8682b360f450ebf759b8e02fb0e9d7f453c0"} Dec 10 22:51:58 crc kubenswrapper[4791]: I1210 22:51:58.184516 4791 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-q4cpx" podStartSLOduration=127.184497552 podStartE2EDuration="2m7.184497552s" podCreationTimestamp="2025-12-10 22:49:51 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-10 22:51:58.182763083 +0000 UTC m=+152.612380706" watchObservedRunningTime="2025-12-10 22:51:58.184497552 +0000 UTC m=+152.614115165" Dec 10 22:51:58 crc kubenswrapper[4791]: I1210 22:51:58.193789 4791 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-config-operator/openshift-config-operator-7777fb866f-wwpsp" event={"ID":"b419e5d6-d6a4-4258-a1bd-a462bd7f7a12","Type":"ContainerStarted","Data":"dac88009f5f39b1949b5334229c10e950a635375742d5aea3a880f64b35fe273"} Dec 10 22:51:58 crc kubenswrapper[4791]: I1210 22:51:58.193876 4791 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-config-operator/openshift-config-operator-7777fb866f-wwpsp" Dec 10 22:51:58 crc kubenswrapper[4791]: I1210 22:51:58.205089 4791 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-ms8jw" event={"ID":"c6c773c2-29e4-477d-91fb-f63052272f06","Type":"ContainerStarted","Data":"c188c5c3e7dfdca139d80b8c0f0eefc5803acc681b23eb9981921ab87ce5877f"} Dec 10 22:51:58 crc kubenswrapper[4791]: I1210 22:51:58.205131 4791 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-ms8jw" event={"ID":"c6c773c2-29e4-477d-91fb-f63052272f06","Type":"ContainerStarted","Data":"139682954e15ed0ffc573ed3ab9ac1bb12d8b606b0076f49655616f962e329f3"} Dec 10 22:51:58 crc kubenswrapper[4791]: I1210 22:51:58.220959 4791 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-config-operator/openshift-config-operator-7777fb866f-wwpsp" podStartSLOduration=128.220942592 podStartE2EDuration="2m8.220942592s" podCreationTimestamp="2025-12-10 22:49:50 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-10 22:51:58.21838572 +0000 UTC m=+152.648003333" watchObservedRunningTime="2025-12-10 22:51:58.220942592 +0000 UTC m=+152.650560205" Dec 10 22:51:58 crc kubenswrapper[4791]: I1210 22:51:58.221243 4791 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-srd52\" (UID: \"e93dbd5d-0ca7-4a15-9eb3-0d71595fada0\") " pod="openshift-image-registry/image-registry-697d97f7c8-srd52" Dec 10 22:51:58 crc kubenswrapper[4791]: E1210 22:51:58.221536 4791 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-10 22:51:58.721526088 +0000 UTC m=+153.151143701 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-srd52" (UID: "e93dbd5d-0ca7-4a15-9eb3-0d71595fada0") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 10 22:51:58 crc kubenswrapper[4791]: I1210 22:51:58.234021 4791 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-vm842" event={"ID":"6e58f465-ef21-4b20-9eb8-5ade4810688a","Type":"ContainerStarted","Data":"60b2b562771055381327d1a68c0a6821c16b5a81b125c14c759be4650730a4b7"} Dec 10 22:51:58 crc kubenswrapper[4791]: I1210 22:51:58.243835 4791 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-ms8jw" podStartSLOduration=127.243821228 podStartE2EDuration="2m7.243821228s" podCreationTimestamp="2025-12-10 22:49:51 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-10 22:51:58.242489281 +0000 UTC m=+152.672106894" watchObservedRunningTime="2025-12-10 22:51:58.243821228 +0000 UTC m=+152.673438841" Dec 10 22:51:58 crc kubenswrapper[4791]: I1210 22:51:58.262497 4791 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-t7742" event={"ID":"4205f348-3276-40f3-a35b-69306b331d80","Type":"ContainerStarted","Data":"a3ce51755c4a160d5897403d6c6145c478b92553d49d4cb27e04103a2490c6cc"} Dec 10 22:51:58 crc kubenswrapper[4791]: I1210 22:51:58.262560 4791 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-t7742" event={"ID":"4205f348-3276-40f3-a35b-69306b331d80","Type":"ContainerStarted","Data":"f2a7fae45cbe681b628d8a8a7346a37ab062499f742f81fa5183b3fe47c24eb6"} Dec 10 22:51:58 crc kubenswrapper[4791]: I1210 22:51:58.281035 4791 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-t7742" podStartSLOduration=128.281017508 podStartE2EDuration="2m8.281017508s" podCreationTimestamp="2025-12-10 22:49:50 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-10 22:51:58.279889317 +0000 UTC m=+152.709506930" watchObservedRunningTime="2025-12-10 22:51:58.281017508 +0000 UTC m=+152.710635121" Dec 10 22:51:58 crc kubenswrapper[4791]: I1210 22:51:58.289297 4791 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-console/console-f9d7485db-zkfkv" event={"ID":"1b61a973-db5e-4aa7-96e3-7fbdd9dd8cc0","Type":"ContainerStarted","Data":"82877649b51625497004f259c4773b2cd94796ab144537ed6ed9dc6355b4a741"} Dec 10 22:51:58 crc kubenswrapper[4791]: I1210 22:51:58.296151 4791 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="hostpath-provisioner/csi-hostpathplugin-mkz5r" event={"ID":"1666d5bb-295d-4828-868e-b1b17b5741bb","Type":"ContainerStarted","Data":"1473363eaa02d91b680620ce7240e749940eea21ce67f89d6c62dd810334b4b3"} Dec 10 22:51:58 crc kubenswrapper[4791]: I1210 22:51:58.304450 4791 patch_prober.go:28] interesting pod/downloads-7954f5f757-mlb9p container/download-server namespace/openshift-console: Readiness probe status=failure output="Get \"http://10.217.0.10:8080/\": dial tcp 10.217.0.10:8080: connect: connection refused" start-of-body= Dec 10 22:51:58 crc kubenswrapper[4791]: I1210 22:51:58.304503 4791 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-console/downloads-7954f5f757-mlb9p" podUID="680d2296-dd48-4786-bea7-64c32edc94c7" containerName="download-server" probeResult="failure" output="Get \"http://10.217.0.10:8080/\": dial tcp 10.217.0.10:8080: connect: connection refused" Dec 10 22:51:58 crc kubenswrapper[4791]: I1210 22:51:58.309302 4791 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-nwwpb" Dec 10 22:51:58 crc kubenswrapper[4791]: I1210 22:51:58.310013 4791 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-authentication/oauth-openshift-558db77b4-ht5t7" Dec 10 22:51:58 crc kubenswrapper[4791]: I1210 22:51:58.322903 4791 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 10 22:51:58 crc kubenswrapper[4791]: E1210 22:51:58.323274 4791 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-10 22:51:58.823246742 +0000 UTC m=+153.252864355 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 10 22:51:58 crc kubenswrapper[4791]: I1210 22:51:58.344169 4791 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-console/console-f9d7485db-zkfkv" podStartSLOduration=127.344146652 podStartE2EDuration="2m7.344146652s" podCreationTimestamp="2025-12-10 22:49:51 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-10 22:51:58.31432698 +0000 UTC m=+152.743944603" watchObservedRunningTime="2025-12-10 22:51:58.344146652 +0000 UTC m=+152.773764265" Dec 10 22:51:58 crc kubenswrapper[4791]: I1210 22:51:58.425780 4791 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-srd52\" (UID: \"e93dbd5d-0ca7-4a15-9eb3-0d71595fada0\") " pod="openshift-image-registry/image-registry-697d97f7c8-srd52" Dec 10 22:51:58 crc kubenswrapper[4791]: E1210 22:51:58.432251 4791 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-10 22:51:58.932236571 +0000 UTC m=+153.361854184 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-srd52" (UID: "e93dbd5d-0ca7-4a15-9eb3-0d71595fada0") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 10 22:51:58 crc kubenswrapper[4791]: I1210 22:51:58.530895 4791 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 10 22:51:58 crc kubenswrapper[4791]: E1210 22:51:58.531572 4791 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-10 22:51:59.031552628 +0000 UTC m=+153.461170241 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 10 22:51:58 crc kubenswrapper[4791]: I1210 22:51:58.632223 4791 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-srd52\" (UID: \"e93dbd5d-0ca7-4a15-9eb3-0d71595fada0\") " pod="openshift-image-registry/image-registry-697d97f7c8-srd52" Dec 10 22:51:58 crc kubenswrapper[4791]: E1210 22:51:58.632627 4791 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-10 22:51:59.132615564 +0000 UTC m=+153.562233177 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-srd52" (UID: "e93dbd5d-0ca7-4a15-9eb3-0d71595fada0") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 10 22:51:58 crc kubenswrapper[4791]: I1210 22:51:58.732928 4791 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 10 22:51:58 crc kubenswrapper[4791]: E1210 22:51:58.733221 4791 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-10 22:51:59.233206526 +0000 UTC m=+153.662824139 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 10 22:51:58 crc kubenswrapper[4791]: I1210 22:51:58.733312 4791 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-srd52\" (UID: \"e93dbd5d-0ca7-4a15-9eb3-0d71595fada0\") " pod="openshift-image-registry/image-registry-697d97f7c8-srd52" Dec 10 22:51:58 crc kubenswrapper[4791]: E1210 22:51:58.733569 4791 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-10 22:51:59.233561606 +0000 UTC m=+153.663179219 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-srd52" (UID: "e93dbd5d-0ca7-4a15-9eb3-0d71595fada0") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 10 22:51:58 crc kubenswrapper[4791]: I1210 22:51:58.834297 4791 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 10 22:51:58 crc kubenswrapper[4791]: E1210 22:51:58.834490 4791 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-10 22:51:59.334459837 +0000 UTC m=+153.764077450 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 10 22:51:58 crc kubenswrapper[4791]: I1210 22:51:58.834683 4791 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-srd52\" (UID: \"e93dbd5d-0ca7-4a15-9eb3-0d71595fada0\") " pod="openshift-image-registry/image-registry-697d97f7c8-srd52" Dec 10 22:51:58 crc kubenswrapper[4791]: E1210 22:51:58.834994 4791 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-10 22:51:59.334985532 +0000 UTC m=+153.764603145 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-srd52" (UID: "e93dbd5d-0ca7-4a15-9eb3-0d71595fada0") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 10 22:51:58 crc kubenswrapper[4791]: I1210 22:51:58.936723 4791 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 10 22:51:58 crc kubenswrapper[4791]: I1210 22:51:58.937189 4791 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cqllr\" (UniqueName: \"kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr\") pod \"network-check-target-xd92c\" (UID: \"3b6479f0-333b-4a96-9adf-2099afdc2447\") " pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 10 22:51:58 crc kubenswrapper[4791]: I1210 22:51:58.937219 4791 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"networking-console-plugin-cert\" (UniqueName: \"kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 10 22:51:58 crc kubenswrapper[4791]: I1210 22:51:58.937250 4791 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-s2dwl\" (UniqueName: \"kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl\") pod \"network-check-source-55646444c4-trplf\" (UID: \"9d751cbb-f2e2-430d-9754-c882a5e924a5\") " pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 10 22:51:58 crc kubenswrapper[4791]: E1210 22:51:58.937353 4791 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-10 22:51:59.437309234 +0000 UTC m=+153.866926847 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 10 22:51:58 crc kubenswrapper[4791]: I1210 22:51:58.937440 4791 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nginx-conf\" (UniqueName: \"kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 10 22:51:58 crc kubenswrapper[4791]: I1210 22:51:58.938319 4791 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"nginx-conf\" (UniqueName: \"kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 10 22:51:58 crc kubenswrapper[4791]: I1210 22:51:58.951381 4791 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-s2dwl\" (UniqueName: \"kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl\") pod \"network-check-source-55646444c4-trplf\" (UID: \"9d751cbb-f2e2-430d-9754-c882a5e924a5\") " pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 10 22:51:58 crc kubenswrapper[4791]: I1210 22:51:58.951959 4791 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-cqllr\" (UniqueName: \"kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr\") pod \"network-check-target-xd92c\" (UID: \"3b6479f0-333b-4a96-9adf-2099afdc2447\") " pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 10 22:51:58 crc kubenswrapper[4791]: I1210 22:51:58.953283 4791 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"networking-console-plugin-cert\" (UniqueName: \"kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 10 22:51:59 crc kubenswrapper[4791]: I1210 22:51:59.038481 4791 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-srd52\" (UID: \"e93dbd5d-0ca7-4a15-9eb3-0d71595fada0\") " pod="openshift-image-registry/image-registry-697d97f7c8-srd52" Dec 10 22:51:59 crc kubenswrapper[4791]: E1210 22:51:59.038906 4791 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-10 22:51:59.538888714 +0000 UTC m=+153.968506327 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-srd52" (UID: "e93dbd5d-0ca7-4a15-9eb3-0d71595fada0") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 10 22:51:59 crc kubenswrapper[4791]: I1210 22:51:59.089231 4791 patch_prober.go:28] interesting pod/router-default-5444994796-p97fc container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Dec 10 22:51:59 crc kubenswrapper[4791]: [-]has-synced failed: reason withheld Dec 10 22:51:59 crc kubenswrapper[4791]: [+]process-running ok Dec 10 22:51:59 crc kubenswrapper[4791]: healthz check failed Dec 10 22:51:59 crc kubenswrapper[4791]: I1210 22:51:59.089305 4791 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-p97fc" podUID="d3de1ba0-b014-4f98-b597-cc89f5a70364" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Dec 10 22:51:59 crc kubenswrapper[4791]: I1210 22:51:59.139375 4791 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 10 22:51:59 crc kubenswrapper[4791]: E1210 22:51:59.139543 4791 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-10 22:51:59.639513077 +0000 UTC m=+154.069130690 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 10 22:51:59 crc kubenswrapper[4791]: I1210 22:51:59.139615 4791 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-srd52\" (UID: \"e93dbd5d-0ca7-4a15-9eb3-0d71595fada0\") " pod="openshift-image-registry/image-registry-697d97f7c8-srd52" Dec 10 22:51:59 crc kubenswrapper[4791]: E1210 22:51:59.139927 4791 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-10 22:51:59.639914858 +0000 UTC m=+154.069532461 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-srd52" (UID: "e93dbd5d-0ca7-4a15-9eb3-0d71595fada0") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 10 22:51:59 crc kubenswrapper[4791]: I1210 22:51:59.213557 4791 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 10 22:51:59 crc kubenswrapper[4791]: I1210 22:51:59.228385 4791 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 10 22:51:59 crc kubenswrapper[4791]: I1210 22:51:59.240506 4791 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 10 22:51:59 crc kubenswrapper[4791]: I1210 22:51:59.240768 4791 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 10 22:51:59 crc kubenswrapper[4791]: E1210 22:51:59.240818 4791 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-10 22:51:59.74080106 +0000 UTC m=+154.170418673 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 10 22:51:59 crc kubenswrapper[4791]: I1210 22:51:59.240853 4791 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-srd52\" (UID: \"e93dbd5d-0ca7-4a15-9eb3-0d71595fada0\") " pod="openshift-image-registry/image-registry-697d97f7c8-srd52" Dec 10 22:51:59 crc kubenswrapper[4791]: E1210 22:51:59.241215 4791 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-10 22:51:59.741197081 +0000 UTC m=+154.170814694 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-srd52" (UID: "e93dbd5d-0ca7-4a15-9eb3-0d71595fada0") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 10 22:51:59 crc kubenswrapper[4791]: I1210 22:51:59.326588 4791 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-service-ca/service-ca-9c57cc56f-x9pw8" event={"ID":"34da0cc5-3e59-4c25-af94-e755e772b89a","Type":"ContainerStarted","Data":"c6d613633c2597af102cd4bf8c665c85c9857c9b88b19514806cba9e692c796e"} Dec 10 22:51:59 crc kubenswrapper[4791]: I1210 22:51:59.327926 4791 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/certified-operators-8x2rc"] Dec 10 22:51:59 crc kubenswrapper[4791]: I1210 22:51:59.330445 4791 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-8x2rc" Dec 10 22:51:59 crc kubenswrapper[4791]: I1210 22:51:59.334955 4791 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"certified-operators-dockercfg-4rs5g" Dec 10 22:51:59 crc kubenswrapper[4791]: I1210 22:51:59.341935 4791 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-8x2rc"] Dec 10 22:51:59 crc kubenswrapper[4791]: I1210 22:51:59.342155 4791 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 10 22:51:59 crc kubenswrapper[4791]: E1210 22:51:59.342572 4791 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-10 22:51:59.842551435 +0000 UTC m=+154.272169048 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 10 22:51:59 crc kubenswrapper[4791]: I1210 22:51:59.357286 4791 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-service-ca/service-ca-9c57cc56f-x9pw8" podStartSLOduration=128.35726521 podStartE2EDuration="2m8.35726521s" podCreationTimestamp="2025-12-10 22:49:51 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-10 22:51:59.351695213 +0000 UTC m=+153.781312826" watchObservedRunningTime="2025-12-10 22:51:59.35726521 +0000 UTC m=+153.786882823" Dec 10 22:51:59 crc kubenswrapper[4791]: I1210 22:51:59.359511 4791 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-dns-operator/dns-operator-744455d44c-cggs2" event={"ID":"5a2896d3-fca8-4c64-abeb-6320f0b78d4e","Type":"ContainerStarted","Data":"f592574125e1ac5dc83bd782b12f0f525e2e3b4305438c3b222196a6e5299e00"} Dec 10 22:51:59 crc kubenswrapper[4791]: I1210 22:51:59.359555 4791 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-dns-operator/dns-operator-744455d44c-cggs2" event={"ID":"5a2896d3-fca8-4c64-abeb-6320f0b78d4e","Type":"ContainerStarted","Data":"3380e9d4aeac11fb24ad3f51ba985368e349f7dbb7ea3983aaeaef82e2613f18"} Dec 10 22:51:59 crc kubenswrapper[4791]: I1210 22:51:59.389597 4791 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-rmjd2" event={"ID":"93a38a79-5c1c-4296-849e-94137de8f974","Type":"ContainerStarted","Data":"1f9e69fc2d53d98f9b39d5b7e76eb99795bbe38d27c472b521f608c927c21d2d"} Dec 10 22:51:59 crc kubenswrapper[4791]: I1210 22:51:59.400157 4791 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-dns-operator/dns-operator-744455d44c-cggs2" podStartSLOduration=128.400139112 podStartE2EDuration="2m8.400139112s" podCreationTimestamp="2025-12-10 22:49:51 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-10 22:51:59.398629419 +0000 UTC m=+153.828247032" watchObservedRunningTime="2025-12-10 22:51:59.400139112 +0000 UTC m=+153.829756725" Dec 10 22:51:59 crc kubenswrapper[4791]: I1210 22:51:59.406236 4791 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-admission-controller-857f4d67dd-fbpcr" event={"ID":"ded98820-5308-46e3-b87b-18d0a81c0519","Type":"ContainerStarted","Data":"e9a947076a21b63392c2022cbdc8cb62f9600278c20774bc458ad3662baa4faf"} Dec 10 22:51:59 crc kubenswrapper[4791]: I1210 22:51:59.406280 4791 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-admission-controller-857f4d67dd-fbpcr" event={"ID":"ded98820-5308-46e3-b87b-18d0a81c0519","Type":"ContainerStarted","Data":"1e7fa8a7ea01e33c016dff5cefa696377409c1f63c40de999184471fb8020f52"} Dec 10 22:51:59 crc kubenswrapper[4791]: I1210 22:51:59.425752 4791 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-dns/dns-default-qs77n" event={"ID":"e9978f48-c33a-4765-9b05-16e37f5cdb24","Type":"ContainerStarted","Data":"f50699d69b534099d54a64cbf5edf2a0aa0c4b5abc63161ee27a7341fbe8f8f4"} Dec 10 22:51:59 crc kubenswrapper[4791]: I1210 22:51:59.425803 4791 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-dns/dns-default-qs77n" event={"ID":"e9978f48-c33a-4765-9b05-16e37f5cdb24","Type":"ContainerStarted","Data":"604a40c19db023163023a8ed1538384e236fdf51c7c36953f6a0f1b881b62c35"} Dec 10 22:51:59 crc kubenswrapper[4791]: I1210 22:51:59.426485 4791 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-dns/dns-default-qs77n" Dec 10 22:51:59 crc kubenswrapper[4791]: I1210 22:51:59.434900 4791 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-multus/multus-admission-controller-857f4d67dd-fbpcr" podStartSLOduration=128.434885534 podStartE2EDuration="2m8.434885534s" podCreationTimestamp="2025-12-10 22:49:51 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-10 22:51:59.427197467 +0000 UTC m=+153.856815080" watchObservedRunningTime="2025-12-10 22:51:59.434885534 +0000 UTC m=+153.864503147" Dec 10 22:51:59 crc kubenswrapper[4791]: I1210 22:51:59.443790 4791 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-srd52\" (UID: \"e93dbd5d-0ca7-4a15-9eb3-0d71595fada0\") " pod="openshift-image-registry/image-registry-697d97f7c8-srd52" Dec 10 22:51:59 crc kubenswrapper[4791]: I1210 22:51:59.443873 4791 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/009efc35-5943-429f-98fd-3307a5f557f2-utilities\") pod \"certified-operators-8x2rc\" (UID: \"009efc35-5943-429f-98fd-3307a5f557f2\") " pod="openshift-marketplace/certified-operators-8x2rc" Dec 10 22:51:59 crc kubenswrapper[4791]: I1210 22:51:59.443898 4791 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/009efc35-5943-429f-98fd-3307a5f557f2-catalog-content\") pod \"certified-operators-8x2rc\" (UID: \"009efc35-5943-429f-98fd-3307a5f557f2\") " pod="openshift-marketplace/certified-operators-8x2rc" Dec 10 22:51:59 crc kubenswrapper[4791]: I1210 22:51:59.443981 4791 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-z2kjq\" (UniqueName: \"kubernetes.io/projected/009efc35-5943-429f-98fd-3307a5f557f2-kube-api-access-z2kjq\") pod \"certified-operators-8x2rc\" (UID: \"009efc35-5943-429f-98fd-3307a5f557f2\") " pod="openshift-marketplace/certified-operators-8x2rc" Dec 10 22:51:59 crc kubenswrapper[4791]: E1210 22:51:59.444171 4791 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-10 22:51:59.944153476 +0000 UTC m=+154.373771089 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-srd52" (UID: "e93dbd5d-0ca7-4a15-9eb3-0d71595fada0") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 10 22:51:59 crc kubenswrapper[4791]: I1210 22:51:59.457874 4791 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-w5dpr" event={"ID":"3f132df5-3de5-4455-a3da-7848b0de69cc","Type":"ContainerStarted","Data":"9b37e4b15ade5cfbeeece6b493f0a89b597a6cdeff5d2b6f871be3c091d055d1"} Dec 10 22:51:59 crc kubenswrapper[4791]: I1210 22:51:59.458400 4791 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-w5dpr" Dec 10 22:51:59 crc kubenswrapper[4791]: I1210 22:51:59.496864 4791 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-dns/dns-default-qs77n" podStartSLOduration=9.496842785 podStartE2EDuration="9.496842785s" podCreationTimestamp="2025-12-10 22:51:50 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-10 22:51:59.457686088 +0000 UTC m=+153.887303701" watchObservedRunningTime="2025-12-10 22:51:59.496842785 +0000 UTC m=+153.926460398" Dec 10 22:51:59 crc kubenswrapper[4791]: I1210 22:51:59.497093 4791 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-w5dpr" podStartSLOduration=128.497087882 podStartE2EDuration="2m8.497087882s" podCreationTimestamp="2025-12-10 22:49:51 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-10 22:51:59.495749834 +0000 UTC m=+153.925367467" watchObservedRunningTime="2025-12-10 22:51:59.497087882 +0000 UTC m=+153.926705495" Dec 10 22:51:59 crc kubenswrapper[4791]: I1210 22:51:59.499108 4791 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-operator-74547568cd-rcnrn" event={"ID":"2fee22d4-2e9b-4768-9603-22e11d9dad9a","Type":"ContainerStarted","Data":"4bfd9a77e1386b737419d35171beca34a541ba4ecb7c65a23f8c17548fb1c980"} Dec 10 22:51:59 crc kubenswrapper[4791]: I1210 22:51:59.546832 4791 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 10 22:51:59 crc kubenswrapper[4791]: I1210 22:51:59.547112 4791 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/009efc35-5943-429f-98fd-3307a5f557f2-utilities\") pod \"certified-operators-8x2rc\" (UID: \"009efc35-5943-429f-98fd-3307a5f557f2\") " pod="openshift-marketplace/certified-operators-8x2rc" Dec 10 22:51:59 crc kubenswrapper[4791]: I1210 22:51:59.547179 4791 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/009efc35-5943-429f-98fd-3307a5f557f2-catalog-content\") pod \"certified-operators-8x2rc\" (UID: \"009efc35-5943-429f-98fd-3307a5f557f2\") " pod="openshift-marketplace/certified-operators-8x2rc" Dec 10 22:51:59 crc kubenswrapper[4791]: I1210 22:51:59.547248 4791 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-z2kjq\" (UniqueName: \"kubernetes.io/projected/009efc35-5943-429f-98fd-3307a5f557f2-kube-api-access-z2kjq\") pod \"certified-operators-8x2rc\" (UID: \"009efc35-5943-429f-98fd-3307a5f557f2\") " pod="openshift-marketplace/certified-operators-8x2rc" Dec 10 22:51:59 crc kubenswrapper[4791]: E1210 22:51:59.548097 4791 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-10 22:52:00.048080342 +0000 UTC m=+154.477697955 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 10 22:51:59 crc kubenswrapper[4791]: I1210 22:51:59.548917 4791 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/009efc35-5943-429f-98fd-3307a5f557f2-utilities\") pod \"certified-operators-8x2rc\" (UID: \"009efc35-5943-429f-98fd-3307a5f557f2\") " pod="openshift-marketplace/certified-operators-8x2rc" Dec 10 22:51:59 crc kubenswrapper[4791]: I1210 22:51:59.554289 4791 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-machine-config-operator/machine-config-operator-74547568cd-rcnrn" podStartSLOduration=128.554270347 podStartE2EDuration="2m8.554270347s" podCreationTimestamp="2025-12-10 22:49:51 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-10 22:51:59.54269166 +0000 UTC m=+153.972309303" watchObservedRunningTime="2025-12-10 22:51:59.554270347 +0000 UTC m=+153.983887960" Dec 10 22:51:59 crc kubenswrapper[4791]: I1210 22:51:59.555874 4791 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="hostpath-provisioner/csi-hostpathplugin-mkz5r" event={"ID":"1666d5bb-295d-4828-868e-b1b17b5741bb","Type":"ContainerStarted","Data":"aecd7ed785852ac65f1d0efc0abac119ee90a361ae80c9b62b9361a878c1f999"} Dec 10 22:51:59 crc kubenswrapper[4791]: I1210 22:51:59.556609 4791 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/community-operators-gskzr"] Dec 10 22:51:59 crc kubenswrapper[4791]: I1210 22:51:59.557094 4791 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/009efc35-5943-429f-98fd-3307a5f557f2-catalog-content\") pod \"certified-operators-8x2rc\" (UID: \"009efc35-5943-429f-98fd-3307a5f557f2\") " pod="openshift-marketplace/certified-operators-8x2rc" Dec 10 22:51:59 crc kubenswrapper[4791]: I1210 22:51:59.569675 4791 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-gskzr" Dec 10 22:51:59 crc kubenswrapper[4791]: I1210 22:51:59.571429 4791 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"community-operators-dockercfg-dmngl" Dec 10 22:51:59 crc kubenswrapper[4791]: I1210 22:51:59.586008 4791 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-service-ca-operator/service-ca-operator-777779d784-b8v4c" event={"ID":"64199728-979b-4628-8b37-2dd792f86439","Type":"ContainerStarted","Data":"4e08cbdbd6f02afe7b14aec7a8d72349558f58172c78ef55beec6864fae77336"} Dec 10 22:51:59 crc kubenswrapper[4791]: I1210 22:51:59.601271 4791 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-gskzr"] Dec 10 22:51:59 crc kubenswrapper[4791]: I1210 22:51:59.602779 4791 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-z2kjq\" (UniqueName: \"kubernetes.io/projected/009efc35-5943-429f-98fd-3307a5f557f2-kube-api-access-z2kjq\") pod \"certified-operators-8x2rc\" (UID: \"009efc35-5943-429f-98fd-3307a5f557f2\") " pod="openshift-marketplace/certified-operators-8x2rc" Dec 10 22:51:59 crc kubenswrapper[4791]: I1210 22:51:59.608884 4791 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/marketplace-operator-79b997595-7g828" event={"ID":"a9185ee1-adaf-48e8-b9a9-90ef658f8212","Type":"ContainerStarted","Data":"b4b2f13d92bcf32cde11b26e411412c60deeba2f579aca917b1aea04f10c82e2"} Dec 10 22:51:59 crc kubenswrapper[4791]: I1210 22:51:59.609301 4791 patch_prober.go:28] interesting pod/marketplace-operator-79b997595-7g828 container/marketplace-operator namespace/openshift-marketplace: Readiness probe status=failure output="Get \"http://10.217.0.42:8080/healthz\": dial tcp 10.217.0.42:8080: connect: connection refused" start-of-body= Dec 10 22:51:59 crc kubenswrapper[4791]: I1210 22:51:59.609327 4791 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-marketplace/marketplace-operator-79b997595-7g828" podUID="a9185ee1-adaf-48e8-b9a9-90ef658f8212" containerName="marketplace-operator" probeResult="failure" output="Get \"http://10.217.0.42:8080/healthz\": dial tcp 10.217.0.42:8080: connect: connection refused" Dec 10 22:51:59 crc kubenswrapper[4791]: I1210 22:51:59.633851 4791 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-service-ca-operator/service-ca-operator-777779d784-b8v4c" podStartSLOduration=128.633832166 podStartE2EDuration="2m8.633832166s" podCreationTimestamp="2025-12-10 22:49:51 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-10 22:51:59.632816037 +0000 UTC m=+154.062433650" watchObservedRunningTime="2025-12-10 22:51:59.633832166 +0000 UTC m=+154.063449779" Dec 10 22:51:59 crc kubenswrapper[4791]: I1210 22:51:59.634013 4791 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-vm842" event={"ID":"6e58f465-ef21-4b20-9eb8-5ade4810688a","Type":"ContainerStarted","Data":"70ddd1c9e00e0f1a6cd3be3585b8fdfd2d26a36e39bfd61ffac1de6078745870"} Dec 10 22:51:59 crc kubenswrapper[4791]: I1210 22:51:59.648593 4791 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/08d50c37-7a4d-4731-beb8-fbc6d2e10bb0-utilities\") pod \"community-operators-gskzr\" (UID: \"08d50c37-7a4d-4731-beb8-fbc6d2e10bb0\") " pod="openshift-marketplace/community-operators-gskzr" Dec 10 22:51:59 crc kubenswrapper[4791]: I1210 22:51:59.648642 4791 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-srd52\" (UID: \"e93dbd5d-0ca7-4a15-9eb3-0d71595fada0\") " pod="openshift-image-registry/image-registry-697d97f7c8-srd52" Dec 10 22:51:59 crc kubenswrapper[4791]: I1210 22:51:59.648662 4791 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/08d50c37-7a4d-4731-beb8-fbc6d2e10bb0-catalog-content\") pod \"community-operators-gskzr\" (UID: \"08d50c37-7a4d-4731-beb8-fbc6d2e10bb0\") " pod="openshift-marketplace/community-operators-gskzr" Dec 10 22:51:59 crc kubenswrapper[4791]: I1210 22:51:59.648697 4791 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-qm95v\" (UniqueName: \"kubernetes.io/projected/08d50c37-7a4d-4731-beb8-fbc6d2e10bb0-kube-api-access-qm95v\") pod \"community-operators-gskzr\" (UID: \"08d50c37-7a4d-4731-beb8-fbc6d2e10bb0\") " pod="openshift-marketplace/community-operators-gskzr" Dec 10 22:51:59 crc kubenswrapper[4791]: I1210 22:51:59.659066 4791 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ingress-canary/ingress-canary-9dwf5" event={"ID":"0921ef8b-8842-4d66-890b-3ad15bd48e70","Type":"ContainerStarted","Data":"e5ff465aaa39be648b35cc470a281101d033019d9924785fba7115a1e5de8acc"} Dec 10 22:51:59 crc kubenswrapper[4791]: E1210 22:51:59.659824 4791 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-10 22:52:00.15981191 +0000 UTC m=+154.589429523 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-srd52" (UID: "e93dbd5d-0ca7-4a15-9eb3-0d71595fada0") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 10 22:51:59 crc kubenswrapper[4791]: I1210 22:51:59.663970 4791 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-8x2rc" Dec 10 22:51:59 crc kubenswrapper[4791]: I1210 22:51:59.677762 4791 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-vm842" podStartSLOduration=128.677743376 podStartE2EDuration="2m8.677743376s" podCreationTimestamp="2025-12-10 22:49:51 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-10 22:51:59.676878202 +0000 UTC m=+154.106495815" watchObservedRunningTime="2025-12-10 22:51:59.677743376 +0000 UTC m=+154.107360989" Dec 10 22:51:59 crc kubenswrapper[4791]: I1210 22:51:59.685592 4791 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-k7qsx" event={"ID":"169ebb73-7b32-4f87-a8d7-9272bff04856","Type":"ContainerStarted","Data":"ed7fee3d2079cc8c4f6c44e45a6c33a9bbd4e95b108a34cdac65f7d42d2a73f3"} Dec 10 22:51:59 crc kubenswrapper[4791]: I1210 22:51:59.708914 4791 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-etcd-operator/etcd-operator-b45778765-nkt8g" event={"ID":"97c517e6-0af1-4c17-a2b9-f7a7db3368b3","Type":"ContainerStarted","Data":"03b2205478dbf8057366d8f51436feb9c93950168574756c7170b7d204e3e01f"} Dec 10 22:51:59 crc kubenswrapper[4791]: I1210 22:51:59.723540 4791 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-ingress-canary/ingress-canary-9dwf5" podStartSLOduration=9.72350752 podStartE2EDuration="9.72350752s" podCreationTimestamp="2025-12-10 22:51:50 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-10 22:51:59.714513306 +0000 UTC m=+154.144130919" watchObservedRunningTime="2025-12-10 22:51:59.72350752 +0000 UTC m=+154.153125153" Dec 10 22:51:59 crc kubenswrapper[4791]: I1210 22:51:59.750611 4791 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 10 22:51:59 crc kubenswrapper[4791]: I1210 22:51:59.751010 4791 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/08d50c37-7a4d-4731-beb8-fbc6d2e10bb0-utilities\") pod \"community-operators-gskzr\" (UID: \"08d50c37-7a4d-4731-beb8-fbc6d2e10bb0\") " pod="openshift-marketplace/community-operators-gskzr" Dec 10 22:51:59 crc kubenswrapper[4791]: I1210 22:51:59.751112 4791 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/08d50c37-7a4d-4731-beb8-fbc6d2e10bb0-catalog-content\") pod \"community-operators-gskzr\" (UID: \"08d50c37-7a4d-4731-beb8-fbc6d2e10bb0\") " pod="openshift-marketplace/community-operators-gskzr" Dec 10 22:51:59 crc kubenswrapper[4791]: I1210 22:51:59.751189 4791 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-qm95v\" (UniqueName: \"kubernetes.io/projected/08d50c37-7a4d-4731-beb8-fbc6d2e10bb0-kube-api-access-qm95v\") pod \"community-operators-gskzr\" (UID: \"08d50c37-7a4d-4731-beb8-fbc6d2e10bb0\") " pod="openshift-marketplace/community-operators-gskzr" Dec 10 22:51:59 crc kubenswrapper[4791]: E1210 22:51:59.752970 4791 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-10 22:52:00.252931591 +0000 UTC m=+154.682549204 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 10 22:51:59 crc kubenswrapper[4791]: I1210 22:51:59.758527 4791 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/08d50c37-7a4d-4731-beb8-fbc6d2e10bb0-utilities\") pod \"community-operators-gskzr\" (UID: \"08d50c37-7a4d-4731-beb8-fbc6d2e10bb0\") " pod="openshift-marketplace/community-operators-gskzr" Dec 10 22:51:59 crc kubenswrapper[4791]: I1210 22:51:59.758803 4791 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/08d50c37-7a4d-4731-beb8-fbc6d2e10bb0-catalog-content\") pod \"community-operators-gskzr\" (UID: \"08d50c37-7a4d-4731-beb8-fbc6d2e10bb0\") " pod="openshift-marketplace/community-operators-gskzr" Dec 10 22:51:59 crc kubenswrapper[4791]: I1210 22:51:59.761880 4791 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-config-operator/openshift-config-operator-7777fb866f-wwpsp" Dec 10 22:51:59 crc kubenswrapper[4791]: I1210 22:51:59.771443 4791 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/certified-operators-9cbtj"] Dec 10 22:51:59 crc kubenswrapper[4791]: I1210 22:51:59.781785 4791 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-qm95v\" (UniqueName: \"kubernetes.io/projected/08d50c37-7a4d-4731-beb8-fbc6d2e10bb0-kube-api-access-qm95v\") pod \"community-operators-gskzr\" (UID: \"08d50c37-7a4d-4731-beb8-fbc6d2e10bb0\") " pod="openshift-marketplace/community-operators-gskzr" Dec 10 22:51:59 crc kubenswrapper[4791]: I1210 22:51:59.791757 4791 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-k7qsx" podStartSLOduration=128.791736768 podStartE2EDuration="2m8.791736768s" podCreationTimestamp="2025-12-10 22:49:51 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-10 22:51:59.772951247 +0000 UTC m=+154.202568860" watchObservedRunningTime="2025-12-10 22:51:59.791736768 +0000 UTC m=+154.221354371" Dec 10 22:51:59 crc kubenswrapper[4791]: I1210 22:51:59.806396 4791 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-v445v" Dec 10 22:51:59 crc kubenswrapper[4791]: I1210 22:51:59.806432 4791 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-9cbtj"] Dec 10 22:51:59 crc kubenswrapper[4791]: I1210 22:51:59.806528 4791 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-9cbtj" Dec 10 22:51:59 crc kubenswrapper[4791]: I1210 22:51:59.845565 4791 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-etcd-operator/etcd-operator-b45778765-nkt8g" podStartSLOduration=128.845547928 podStartE2EDuration="2m8.845547928s" podCreationTimestamp="2025-12-10 22:49:51 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-10 22:51:59.844257252 +0000 UTC m=+154.273874875" watchObservedRunningTime="2025-12-10 22:51:59.845547928 +0000 UTC m=+154.275165541" Dec 10 22:51:59 crc kubenswrapper[4791]: I1210 22:51:59.855404 4791 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-srd52\" (UID: \"e93dbd5d-0ca7-4a15-9eb3-0d71595fada0\") " pod="openshift-image-registry/image-registry-697d97f7c8-srd52" Dec 10 22:51:59 crc kubenswrapper[4791]: I1210 22:51:59.855465 4791 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/8d2ae7b5-156f-4230-b011-3b107066e18e-utilities\") pod \"certified-operators-9cbtj\" (UID: \"8d2ae7b5-156f-4230-b011-3b107066e18e\") " pod="openshift-marketplace/certified-operators-9cbtj" Dec 10 22:51:59 crc kubenswrapper[4791]: I1210 22:51:59.855551 4791 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/8d2ae7b5-156f-4230-b011-3b107066e18e-catalog-content\") pod \"certified-operators-9cbtj\" (UID: \"8d2ae7b5-156f-4230-b011-3b107066e18e\") " pod="openshift-marketplace/certified-operators-9cbtj" Dec 10 22:51:59 crc kubenswrapper[4791]: I1210 22:51:59.855653 4791 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-8t7vg\" (UniqueName: \"kubernetes.io/projected/8d2ae7b5-156f-4230-b011-3b107066e18e-kube-api-access-8t7vg\") pod \"certified-operators-9cbtj\" (UID: \"8d2ae7b5-156f-4230-b011-3b107066e18e\") " pod="openshift-marketplace/certified-operators-9cbtj" Dec 10 22:51:59 crc kubenswrapper[4791]: E1210 22:51:59.859670 4791 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-10 22:52:00.359653447 +0000 UTC m=+154.789271060 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-srd52" (UID: "e93dbd5d-0ca7-4a15-9eb3-0d71595fada0") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 10 22:51:59 crc kubenswrapper[4791]: I1210 22:51:59.909647 4791 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-gskzr" Dec 10 22:51:59 crc kubenswrapper[4791]: I1210 22:51:59.956941 4791 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 10 22:51:59 crc kubenswrapper[4791]: I1210 22:51:59.957155 4791 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/8d2ae7b5-156f-4230-b011-3b107066e18e-catalog-content\") pod \"certified-operators-9cbtj\" (UID: \"8d2ae7b5-156f-4230-b011-3b107066e18e\") " pod="openshift-marketplace/certified-operators-9cbtj" Dec 10 22:51:59 crc kubenswrapper[4791]: I1210 22:51:59.957185 4791 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-8t7vg\" (UniqueName: \"kubernetes.io/projected/8d2ae7b5-156f-4230-b011-3b107066e18e-kube-api-access-8t7vg\") pod \"certified-operators-9cbtj\" (UID: \"8d2ae7b5-156f-4230-b011-3b107066e18e\") " pod="openshift-marketplace/certified-operators-9cbtj" Dec 10 22:51:59 crc kubenswrapper[4791]: I1210 22:51:59.957261 4791 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/8d2ae7b5-156f-4230-b011-3b107066e18e-utilities\") pod \"certified-operators-9cbtj\" (UID: \"8d2ae7b5-156f-4230-b011-3b107066e18e\") " pod="openshift-marketplace/certified-operators-9cbtj" Dec 10 22:51:59 crc kubenswrapper[4791]: E1210 22:51:59.957509 4791 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-10 22:52:00.457480041 +0000 UTC m=+154.887097654 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 10 22:51:59 crc kubenswrapper[4791]: I1210 22:51:59.957935 4791 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/8d2ae7b5-156f-4230-b011-3b107066e18e-utilities\") pod \"certified-operators-9cbtj\" (UID: \"8d2ae7b5-156f-4230-b011-3b107066e18e\") " pod="openshift-marketplace/certified-operators-9cbtj" Dec 10 22:51:59 crc kubenswrapper[4791]: I1210 22:51:59.958097 4791 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/8d2ae7b5-156f-4230-b011-3b107066e18e-catalog-content\") pod \"certified-operators-9cbtj\" (UID: \"8d2ae7b5-156f-4230-b011-3b107066e18e\") " pod="openshift-marketplace/certified-operators-9cbtj" Dec 10 22:52:00 crc kubenswrapper[4791]: I1210 22:52:00.001904 4791 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-8t7vg\" (UniqueName: \"kubernetes.io/projected/8d2ae7b5-156f-4230-b011-3b107066e18e-kube-api-access-8t7vg\") pod \"certified-operators-9cbtj\" (UID: \"8d2ae7b5-156f-4230-b011-3b107066e18e\") " pod="openshift-marketplace/certified-operators-9cbtj" Dec 10 22:52:00 crc kubenswrapper[4791]: I1210 22:52:00.042595 4791 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/community-operators-q7lbx"] Dec 10 22:52:00 crc kubenswrapper[4791]: I1210 22:52:00.043997 4791 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-q7lbx"] Dec 10 22:52:00 crc kubenswrapper[4791]: I1210 22:52:00.044100 4791 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-q7lbx" Dec 10 22:52:00 crc kubenswrapper[4791]: I1210 22:52:00.058180 4791 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-srd52\" (UID: \"e93dbd5d-0ca7-4a15-9eb3-0d71595fada0\") " pod="openshift-image-registry/image-registry-697d97f7c8-srd52" Dec 10 22:52:00 crc kubenswrapper[4791]: E1210 22:52:00.058609 4791 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-10 22:52:00.558595509 +0000 UTC m=+154.988213122 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-srd52" (UID: "e93dbd5d-0ca7-4a15-9eb3-0d71595fada0") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 10 22:52:00 crc kubenswrapper[4791]: I1210 22:52:00.094892 4791 patch_prober.go:28] interesting pod/router-default-5444994796-p97fc container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Dec 10 22:52:00 crc kubenswrapper[4791]: [-]has-synced failed: reason withheld Dec 10 22:52:00 crc kubenswrapper[4791]: [+]process-running ok Dec 10 22:52:00 crc kubenswrapper[4791]: healthz check failed Dec 10 22:52:00 crc kubenswrapper[4791]: I1210 22:52:00.094953 4791 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-p97fc" podUID="d3de1ba0-b014-4f98-b597-cc89f5a70364" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Dec 10 22:52:00 crc kubenswrapper[4791]: I1210 22:52:00.158944 4791 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 10 22:52:00 crc kubenswrapper[4791]: I1210 22:52:00.159210 4791 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/776c00fb-1f2c-4ea2-ba50-62d4d24ffd72-utilities\") pod \"community-operators-q7lbx\" (UID: \"776c00fb-1f2c-4ea2-ba50-62d4d24ffd72\") " pod="openshift-marketplace/community-operators-q7lbx" Dec 10 22:52:00 crc kubenswrapper[4791]: I1210 22:52:00.159240 4791 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-xcs9j\" (UniqueName: \"kubernetes.io/projected/776c00fb-1f2c-4ea2-ba50-62d4d24ffd72-kube-api-access-xcs9j\") pod \"community-operators-q7lbx\" (UID: \"776c00fb-1f2c-4ea2-ba50-62d4d24ffd72\") " pod="openshift-marketplace/community-operators-q7lbx" Dec 10 22:52:00 crc kubenswrapper[4791]: I1210 22:52:00.159276 4791 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/776c00fb-1f2c-4ea2-ba50-62d4d24ffd72-catalog-content\") pod \"community-operators-q7lbx\" (UID: \"776c00fb-1f2c-4ea2-ba50-62d4d24ffd72\") " pod="openshift-marketplace/community-operators-q7lbx" Dec 10 22:52:00 crc kubenswrapper[4791]: E1210 22:52:00.159379 4791 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-10 22:52:00.659364466 +0000 UTC m=+155.088982079 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 10 22:52:00 crc kubenswrapper[4791]: I1210 22:52:00.170546 4791 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-8x2rc"] Dec 10 22:52:00 crc kubenswrapper[4791]: I1210 22:52:00.170765 4791 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-9cbtj" Dec 10 22:52:00 crc kubenswrapper[4791]: W1210 22:52:00.185350 4791 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod009efc35_5943_429f_98fd_3307a5f557f2.slice/crio-d76e6bdaae80e7f5f3b052f64ad503cba5daecead3682b207dbfbbbb7671bc6a WatchSource:0}: Error finding container d76e6bdaae80e7f5f3b052f64ad503cba5daecead3682b207dbfbbbb7671bc6a: Status 404 returned error can't find the container with id d76e6bdaae80e7f5f3b052f64ad503cba5daecead3682b207dbfbbbb7671bc6a Dec 10 22:52:00 crc kubenswrapper[4791]: I1210 22:52:00.260139 4791 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/776c00fb-1f2c-4ea2-ba50-62d4d24ffd72-utilities\") pod \"community-operators-q7lbx\" (UID: \"776c00fb-1f2c-4ea2-ba50-62d4d24ffd72\") " pod="openshift-marketplace/community-operators-q7lbx" Dec 10 22:52:00 crc kubenswrapper[4791]: I1210 22:52:00.260516 4791 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-xcs9j\" (UniqueName: \"kubernetes.io/projected/776c00fb-1f2c-4ea2-ba50-62d4d24ffd72-kube-api-access-xcs9j\") pod \"community-operators-q7lbx\" (UID: \"776c00fb-1f2c-4ea2-ba50-62d4d24ffd72\") " pod="openshift-marketplace/community-operators-q7lbx" Dec 10 22:52:00 crc kubenswrapper[4791]: I1210 22:52:00.260560 4791 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/776c00fb-1f2c-4ea2-ba50-62d4d24ffd72-catalog-content\") pod \"community-operators-q7lbx\" (UID: \"776c00fb-1f2c-4ea2-ba50-62d4d24ffd72\") " pod="openshift-marketplace/community-operators-q7lbx" Dec 10 22:52:00 crc kubenswrapper[4791]: I1210 22:52:00.260603 4791 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-srd52\" (UID: \"e93dbd5d-0ca7-4a15-9eb3-0d71595fada0\") " pod="openshift-image-registry/image-registry-697d97f7c8-srd52" Dec 10 22:52:00 crc kubenswrapper[4791]: E1210 22:52:00.260902 4791 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-10 22:52:00.760886835 +0000 UTC m=+155.190504448 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-srd52" (UID: "e93dbd5d-0ca7-4a15-9eb3-0d71595fada0") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 10 22:52:00 crc kubenswrapper[4791]: I1210 22:52:00.261661 4791 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/776c00fb-1f2c-4ea2-ba50-62d4d24ffd72-catalog-content\") pod \"community-operators-q7lbx\" (UID: \"776c00fb-1f2c-4ea2-ba50-62d4d24ffd72\") " pod="openshift-marketplace/community-operators-q7lbx" Dec 10 22:52:00 crc kubenswrapper[4791]: I1210 22:52:00.262176 4791 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/776c00fb-1f2c-4ea2-ba50-62d4d24ffd72-utilities\") pod \"community-operators-q7lbx\" (UID: \"776c00fb-1f2c-4ea2-ba50-62d4d24ffd72\") " pod="openshift-marketplace/community-operators-q7lbx" Dec 10 22:52:00 crc kubenswrapper[4791]: I1210 22:52:00.290278 4791 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-xcs9j\" (UniqueName: \"kubernetes.io/projected/776c00fb-1f2c-4ea2-ba50-62d4d24ffd72-kube-api-access-xcs9j\") pod \"community-operators-q7lbx\" (UID: \"776c00fb-1f2c-4ea2-ba50-62d4d24ffd72\") " pod="openshift-marketplace/community-operators-q7lbx" Dec 10 22:52:00 crc kubenswrapper[4791]: I1210 22:52:00.362844 4791 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 10 22:52:00 crc kubenswrapper[4791]: E1210 22:52:00.363207 4791 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-10 22:52:00.863178475 +0000 UTC m=+155.292796088 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 10 22:52:00 crc kubenswrapper[4791]: I1210 22:52:00.363275 4791 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-srd52\" (UID: \"e93dbd5d-0ca7-4a15-9eb3-0d71595fada0\") " pod="openshift-image-registry/image-registry-697d97f7c8-srd52" Dec 10 22:52:00 crc kubenswrapper[4791]: E1210 22:52:00.363805 4791 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-10 22:52:00.863795193 +0000 UTC m=+155.293412806 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-srd52" (UID: "e93dbd5d-0ca7-4a15-9eb3-0d71595fada0") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 10 22:52:00 crc kubenswrapper[4791]: I1210 22:52:00.465973 4791 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 10 22:52:00 crc kubenswrapper[4791]: E1210 22:52:00.466169 4791 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-10 22:52:00.966146635 +0000 UTC m=+155.395764248 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 10 22:52:00 crc kubenswrapper[4791]: I1210 22:52:00.466419 4791 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-srd52\" (UID: \"e93dbd5d-0ca7-4a15-9eb3-0d71595fada0\") " pod="openshift-image-registry/image-registry-697d97f7c8-srd52" Dec 10 22:52:00 crc kubenswrapper[4791]: E1210 22:52:00.466772 4791 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-10 22:52:00.966761782 +0000 UTC m=+155.396379395 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-srd52" (UID: "e93dbd5d-0ca7-4a15-9eb3-0d71595fada0") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 10 22:52:00 crc kubenswrapper[4791]: I1210 22:52:00.471472 4791 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-q7lbx" Dec 10 22:52:00 crc kubenswrapper[4791]: I1210 22:52:00.479306 4791 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-9cbtj"] Dec 10 22:52:00 crc kubenswrapper[4791]: I1210 22:52:00.559246 4791 plugin_watcher.go:194] "Adding socket path or updating timestamp to desired state cache" path="/var/lib/kubelet/plugins_registry/kubevirt.io.hostpath-provisioner-reg.sock" Dec 10 22:52:00 crc kubenswrapper[4791]: I1210 22:52:00.567605 4791 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 10 22:52:00 crc kubenswrapper[4791]: E1210 22:52:00.567772 4791 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-10 22:52:01.067739406 +0000 UTC m=+155.497357029 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 10 22:52:00 crc kubenswrapper[4791]: I1210 22:52:00.567920 4791 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-srd52\" (UID: \"e93dbd5d-0ca7-4a15-9eb3-0d71595fada0\") " pod="openshift-image-registry/image-registry-697d97f7c8-srd52" Dec 10 22:52:00 crc kubenswrapper[4791]: E1210 22:52:00.568242 4791 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-10 22:52:01.06823124 +0000 UTC m=+155.497848853 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-srd52" (UID: "e93dbd5d-0ca7-4a15-9eb3-0d71595fada0") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 10 22:52:00 crc kubenswrapper[4791]: I1210 22:52:00.618983 4791 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-gskzr"] Dec 10 22:52:00 crc kubenswrapper[4791]: I1210 22:52:00.671529 4791 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 10 22:52:00 crc kubenswrapper[4791]: E1210 22:52:00.671928 4791 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-10 22:52:01.171876398 +0000 UTC m=+155.601494011 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 10 22:52:00 crc kubenswrapper[4791]: I1210 22:52:00.672148 4791 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-srd52\" (UID: \"e93dbd5d-0ca7-4a15-9eb3-0d71595fada0\") " pod="openshift-image-registry/image-registry-697d97f7c8-srd52" Dec 10 22:52:00 crc kubenswrapper[4791]: E1210 22:52:00.672552 4791 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-10 22:52:01.172534487 +0000 UTC m=+155.602152100 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-srd52" (UID: "e93dbd5d-0ca7-4a15-9eb3-0d71595fada0") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 10 22:52:00 crc kubenswrapper[4791]: I1210 22:52:00.709376 4791 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-gskzr" event={"ID":"08d50c37-7a4d-4731-beb8-fbc6d2e10bb0","Type":"ContainerStarted","Data":"fb24af949a64f3aba6e54c13b2a56bcabd7e6f909d37593e4fc58035bb20bf7e"} Dec 10 22:52:00 crc kubenswrapper[4791]: I1210 22:52:00.711098 4791 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" event={"ID":"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8","Type":"ContainerStarted","Data":"b3862def867ea7a79ffb384afb1017a90b59289390e769916d7747646b558587"} Dec 10 22:52:00 crc kubenswrapper[4791]: I1210 22:52:00.711131 4791 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" event={"ID":"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8","Type":"ContainerStarted","Data":"cca41aab76be99f787faeb47a88a8701db95cedd86792376464c6bc820d89b0f"} Dec 10 22:52:00 crc kubenswrapper[4791]: I1210 22:52:00.715227 4791 generic.go:334] "Generic (PLEG): container finished" podID="8d2ae7b5-156f-4230-b011-3b107066e18e" containerID="3075d884ea3976218f0e1cf4a93febd7ec448d80696660570342ac1050864c00" exitCode=0 Dec 10 22:52:00 crc kubenswrapper[4791]: I1210 22:52:00.715276 4791 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-9cbtj" event={"ID":"8d2ae7b5-156f-4230-b011-3b107066e18e","Type":"ContainerDied","Data":"3075d884ea3976218f0e1cf4a93febd7ec448d80696660570342ac1050864c00"} Dec 10 22:52:00 crc kubenswrapper[4791]: I1210 22:52:00.715296 4791 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-9cbtj" event={"ID":"8d2ae7b5-156f-4230-b011-3b107066e18e","Type":"ContainerStarted","Data":"de343f1792da5a284c1755dbd2680fe93e360afd4d0b606f505bf48ed94e0e07"} Dec 10 22:52:00 crc kubenswrapper[4791]: I1210 22:52:00.717008 4791 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Dec 10 22:52:00 crc kubenswrapper[4791]: I1210 22:52:00.719688 4791 generic.go:334] "Generic (PLEG): container finished" podID="009efc35-5943-429f-98fd-3307a5f557f2" containerID="200d18941348a8e2c514d3071c544e59138565cf3001822a3d80bf82b72a1861" exitCode=0 Dec 10 22:52:00 crc kubenswrapper[4791]: I1210 22:52:00.719736 4791 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-8x2rc" event={"ID":"009efc35-5943-429f-98fd-3307a5f557f2","Type":"ContainerDied","Data":"200d18941348a8e2c514d3071c544e59138565cf3001822a3d80bf82b72a1861"} Dec 10 22:52:00 crc kubenswrapper[4791]: I1210 22:52:00.719759 4791 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-8x2rc" event={"ID":"009efc35-5943-429f-98fd-3307a5f557f2","Type":"ContainerStarted","Data":"d76e6bdaae80e7f5f3b052f64ad503cba5daecead3682b207dbfbbbb7671bc6a"} Dec 10 22:52:00 crc kubenswrapper[4791]: I1210 22:52:00.750866 4791 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-q7lbx"] Dec 10 22:52:00 crc kubenswrapper[4791]: I1210 22:52:00.762857 4791 generic.go:334] "Generic (PLEG): container finished" podID="90640468-53d7-42bc-a835-f7acf48a486e" containerID="e2c74df205f03bdc1e7206068bf864feaf6f07d7ae52dd750d7b9bb336e3c712" exitCode=0 Dec 10 22:52:00 crc kubenswrapper[4791]: I1210 22:52:00.762982 4791 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29423445-b9cxb" event={"ID":"90640468-53d7-42bc-a835-f7acf48a486e","Type":"ContainerDied","Data":"e2c74df205f03bdc1e7206068bf864feaf6f07d7ae52dd750d7b9bb336e3c712"} Dec 10 22:52:00 crc kubenswrapper[4791]: I1210 22:52:00.776394 4791 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 10 22:52:00 crc kubenswrapper[4791]: E1210 22:52:00.776486 4791 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-10 22:52:01.276466384 +0000 UTC m=+155.706084007 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 10 22:52:00 crc kubenswrapper[4791]: I1210 22:52:00.776828 4791 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-srd52\" (UID: \"e93dbd5d-0ca7-4a15-9eb3-0d71595fada0\") " pod="openshift-image-registry/image-registry-697d97f7c8-srd52" Dec 10 22:52:00 crc kubenswrapper[4791]: I1210 22:52:00.778623 4791 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" event={"ID":"9d751cbb-f2e2-430d-9754-c882a5e924a5","Type":"ContainerStarted","Data":"d9eedd98caf38a44db5abd9e30fea81485dd632d3c01c4fab293eb79f555df22"} Dec 10 22:52:00 crc kubenswrapper[4791]: I1210 22:52:00.778675 4791 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" event={"ID":"9d751cbb-f2e2-430d-9754-c882a5e924a5","Type":"ContainerStarted","Data":"b4a494496db2e62eae634cf56e019dfd3976572aebee30da85e6818bfe6a55ce"} Dec 10 22:52:00 crc kubenswrapper[4791]: E1210 22:52:00.778740 4791 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-10 22:52:01.278724588 +0000 UTC m=+155.708342211 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-srd52" (UID: "e93dbd5d-0ca7-4a15-9eb3-0d71595fada0") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 10 22:52:00 crc kubenswrapper[4791]: I1210 22:52:00.805296 4791 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="hostpath-provisioner/csi-hostpathplugin-mkz5r" event={"ID":"1666d5bb-295d-4828-868e-b1b17b5741bb","Type":"ContainerStarted","Data":"11ac1d087e64b27017a202f6078e696ebf1fe24c79249add54d4d602e6022e5b"} Dec 10 22:52:00 crc kubenswrapper[4791]: I1210 22:52:00.815872 4791 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" event={"ID":"3b6479f0-333b-4a96-9adf-2099afdc2447","Type":"ContainerStarted","Data":"bd28c671341677b6963f90d4ca2b34237008b6020fd131eda58c8825eb7aca8a"} Dec 10 22:52:00 crc kubenswrapper[4791]: I1210 22:52:00.815921 4791 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" event={"ID":"3b6479f0-333b-4a96-9adf-2099afdc2447","Type":"ContainerStarted","Data":"26c03113e0c807679cb752c60ef68de404a4bab361d760fe75d33db577a91870"} Dec 10 22:52:00 crc kubenswrapper[4791]: I1210 22:52:00.821244 4791 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/marketplace-operator-79b997595-7g828" Dec 10 22:52:00 crc kubenswrapper[4791]: I1210 22:52:00.877385 4791 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 10 22:52:00 crc kubenswrapper[4791]: E1210 22:52:00.877584 4791 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-10 22:52:01.377559091 +0000 UTC m=+155.807176704 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 10 22:52:00 crc kubenswrapper[4791]: I1210 22:52:00.878115 4791 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-srd52\" (UID: \"e93dbd5d-0ca7-4a15-9eb3-0d71595fada0\") " pod="openshift-image-registry/image-registry-697d97f7c8-srd52" Dec 10 22:52:00 crc kubenswrapper[4791]: E1210 22:52:00.882032 4791 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-10 22:52:01.382017167 +0000 UTC m=+155.811634780 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-srd52" (UID: "e93dbd5d-0ca7-4a15-9eb3-0d71595fada0") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 10 22:52:00 crc kubenswrapper[4791]: I1210 22:52:00.979493 4791 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 10 22:52:00 crc kubenswrapper[4791]: E1210 22:52:00.979636 4791 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-10 22:52:01.479608654 +0000 UTC m=+155.909226277 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 10 22:52:00 crc kubenswrapper[4791]: I1210 22:52:00.979731 4791 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-srd52\" (UID: \"e93dbd5d-0ca7-4a15-9eb3-0d71595fada0\") " pod="openshift-image-registry/image-registry-697d97f7c8-srd52" Dec 10 22:52:00 crc kubenswrapper[4791]: E1210 22:52:00.980055 4791 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-10 22:52:01.480034536 +0000 UTC m=+155.909652149 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-srd52" (UID: "e93dbd5d-0ca7-4a15-9eb3-0d71595fada0") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 10 22:52:01 crc kubenswrapper[4791]: I1210 22:52:01.044393 4791 reconciler.go:161] "OperationExecutor.RegisterPlugin started" plugin={"SocketPath":"/var/lib/kubelet/plugins_registry/kubevirt.io.hostpath-provisioner-reg.sock","Timestamp":"2025-12-10T22:52:00.559270707Z","Handler":null,"Name":""} Dec 10 22:52:01 crc kubenswrapper[4791]: I1210 22:52:01.055292 4791 csi_plugin.go:100] kubernetes.io/csi: Trying to validate a new CSI Driver with name: kubevirt.io.hostpath-provisioner endpoint: /var/lib/kubelet/plugins/csi-hostpath/csi.sock versions: 1.0.0 Dec 10 22:52:01 crc kubenswrapper[4791]: I1210 22:52:01.055329 4791 csi_plugin.go:113] kubernetes.io/csi: Register new plugin with name: kubevirt.io.hostpath-provisioner at endpoint: /var/lib/kubelet/plugins/csi-hostpath/csi.sock Dec 10 22:52:01 crc kubenswrapper[4791]: I1210 22:52:01.080552 4791 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 10 22:52:01 crc kubenswrapper[4791]: I1210 22:52:01.090937 4791 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (OuterVolumeSpecName: "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b"). InnerVolumeSpecName "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8". PluginName "kubernetes.io/csi", VolumeGidValue "" Dec 10 22:52:01 crc kubenswrapper[4791]: I1210 22:52:01.091177 4791 patch_prober.go:28] interesting pod/router-default-5444994796-p97fc container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Dec 10 22:52:01 crc kubenswrapper[4791]: [-]has-synced failed: reason withheld Dec 10 22:52:01 crc kubenswrapper[4791]: [+]process-running ok Dec 10 22:52:01 crc kubenswrapper[4791]: healthz check failed Dec 10 22:52:01 crc kubenswrapper[4791]: I1210 22:52:01.091247 4791 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-p97fc" podUID="d3de1ba0-b014-4f98-b597-cc89f5a70364" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Dec 10 22:52:01 crc kubenswrapper[4791]: I1210 22:52:01.181731 4791 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-srd52\" (UID: \"e93dbd5d-0ca7-4a15-9eb3-0d71595fada0\") " pod="openshift-image-registry/image-registry-697d97f7c8-srd52" Dec 10 22:52:01 crc kubenswrapper[4791]: I1210 22:52:01.228004 4791 csi_attacher.go:380] kubernetes.io/csi: attacher.MountDevice STAGE_UNSTAGE_VOLUME capability not set. Skipping MountDevice... Dec 10 22:52:01 crc kubenswrapper[4791]: I1210 22:52:01.228060 4791 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-srd52\" (UID: \"e93dbd5d-0ca7-4a15-9eb3-0d71595fada0\") device mount path \"/var/lib/kubelet/plugins/kubernetes.io/csi/kubevirt.io.hostpath-provisioner/1f4776af88835e41c12b831b4c9fed40233456d14189815a54dbe7f892fc1983/globalmount\"" pod="openshift-image-registry/image-registry-697d97f7c8-srd52" Dec 10 22:52:01 crc kubenswrapper[4791]: I1210 22:52:01.519539 4791 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-marketplace-wkwbl"] Dec 10 22:52:01 crc kubenswrapper[4791]: I1210 22:52:01.520794 4791 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-wkwbl" Dec 10 22:52:01 crc kubenswrapper[4791]: I1210 22:52:01.522307 4791 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"redhat-marketplace-dockercfg-x2ctb" Dec 10 22:52:01 crc kubenswrapper[4791]: I1210 22:52:01.530501 4791 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-wkwbl"] Dec 10 22:52:01 crc kubenswrapper[4791]: I1210 22:52:01.586563 4791 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/396f55d6-c96d-4520-8165-862b5905a8c0-catalog-content\") pod \"redhat-marketplace-wkwbl\" (UID: \"396f55d6-c96d-4520-8165-862b5905a8c0\") " pod="openshift-marketplace/redhat-marketplace-wkwbl" Dec 10 22:52:01 crc kubenswrapper[4791]: I1210 22:52:01.586643 4791 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/396f55d6-c96d-4520-8165-862b5905a8c0-utilities\") pod \"redhat-marketplace-wkwbl\" (UID: \"396f55d6-c96d-4520-8165-862b5905a8c0\") " pod="openshift-marketplace/redhat-marketplace-wkwbl" Dec 10 22:52:01 crc kubenswrapper[4791]: I1210 22:52:01.586680 4791 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-mq8d4\" (UniqueName: \"kubernetes.io/projected/396f55d6-c96d-4520-8165-862b5905a8c0-kube-api-access-mq8d4\") pod \"redhat-marketplace-wkwbl\" (UID: \"396f55d6-c96d-4520-8165-862b5905a8c0\") " pod="openshift-marketplace/redhat-marketplace-wkwbl" Dec 10 22:52:01 crc kubenswrapper[4791]: I1210 22:52:01.689019 4791 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/396f55d6-c96d-4520-8165-862b5905a8c0-catalog-content\") pod \"redhat-marketplace-wkwbl\" (UID: \"396f55d6-c96d-4520-8165-862b5905a8c0\") " pod="openshift-marketplace/redhat-marketplace-wkwbl" Dec 10 22:52:01 crc kubenswrapper[4791]: I1210 22:52:01.689163 4791 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/396f55d6-c96d-4520-8165-862b5905a8c0-utilities\") pod \"redhat-marketplace-wkwbl\" (UID: \"396f55d6-c96d-4520-8165-862b5905a8c0\") " pod="openshift-marketplace/redhat-marketplace-wkwbl" Dec 10 22:52:01 crc kubenswrapper[4791]: I1210 22:52:01.689199 4791 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-mq8d4\" (UniqueName: \"kubernetes.io/projected/396f55d6-c96d-4520-8165-862b5905a8c0-kube-api-access-mq8d4\") pod \"redhat-marketplace-wkwbl\" (UID: \"396f55d6-c96d-4520-8165-862b5905a8c0\") " pod="openshift-marketplace/redhat-marketplace-wkwbl" Dec 10 22:52:01 crc kubenswrapper[4791]: I1210 22:52:01.690151 4791 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/396f55d6-c96d-4520-8165-862b5905a8c0-catalog-content\") pod \"redhat-marketplace-wkwbl\" (UID: \"396f55d6-c96d-4520-8165-862b5905a8c0\") " pod="openshift-marketplace/redhat-marketplace-wkwbl" Dec 10 22:52:01 crc kubenswrapper[4791]: I1210 22:52:01.690454 4791 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/396f55d6-c96d-4520-8165-862b5905a8c0-utilities\") pod \"redhat-marketplace-wkwbl\" (UID: \"396f55d6-c96d-4520-8165-862b5905a8c0\") " pod="openshift-marketplace/redhat-marketplace-wkwbl" Dec 10 22:52:01 crc kubenswrapper[4791]: I1210 22:52:01.711449 4791 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-mq8d4\" (UniqueName: \"kubernetes.io/projected/396f55d6-c96d-4520-8165-862b5905a8c0-kube-api-access-mq8d4\") pod \"redhat-marketplace-wkwbl\" (UID: \"396f55d6-c96d-4520-8165-862b5905a8c0\") " pod="openshift-marketplace/redhat-marketplace-wkwbl" Dec 10 22:52:01 crc kubenswrapper[4791]: I1210 22:52:01.716536 4791 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-kube-controller-manager/revision-pruner-9-crc"] Dec 10 22:52:01 crc kubenswrapper[4791]: I1210 22:52:01.717415 4791 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-controller-manager/revision-pruner-9-crc" Dec 10 22:52:01 crc kubenswrapper[4791]: I1210 22:52:01.718931 4791 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-controller-manager"/"installer-sa-dockercfg-kjl2n" Dec 10 22:52:01 crc kubenswrapper[4791]: I1210 22:52:01.721079 4791 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-controller-manager"/"kube-root-ca.crt" Dec 10 22:52:01 crc kubenswrapper[4791]: I1210 22:52:01.724046 4791 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-controller-manager/revision-pruner-9-crc"] Dec 10 22:52:01 crc kubenswrapper[4791]: I1210 22:52:01.790716 4791 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/0cc263fd-5b37-4c41-99c0-0d00cc068806-kube-api-access\") pod \"revision-pruner-9-crc\" (UID: \"0cc263fd-5b37-4c41-99c0-0d00cc068806\") " pod="openshift-kube-controller-manager/revision-pruner-9-crc" Dec 10 22:52:01 crc kubenswrapper[4791]: I1210 22:52:01.790785 4791 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/0cc263fd-5b37-4c41-99c0-0d00cc068806-kubelet-dir\") pod \"revision-pruner-9-crc\" (UID: \"0cc263fd-5b37-4c41-99c0-0d00cc068806\") " pod="openshift-kube-controller-manager/revision-pruner-9-crc" Dec 10 22:52:01 crc kubenswrapper[4791]: I1210 22:52:01.837866 4791 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-wkwbl" Dec 10 22:52:01 crc kubenswrapper[4791]: I1210 22:52:01.891501 4791 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/0cc263fd-5b37-4c41-99c0-0d00cc068806-kube-api-access\") pod \"revision-pruner-9-crc\" (UID: \"0cc263fd-5b37-4c41-99c0-0d00cc068806\") " pod="openshift-kube-controller-manager/revision-pruner-9-crc" Dec 10 22:52:01 crc kubenswrapper[4791]: I1210 22:52:01.891588 4791 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/0cc263fd-5b37-4c41-99c0-0d00cc068806-kubelet-dir\") pod \"revision-pruner-9-crc\" (UID: \"0cc263fd-5b37-4c41-99c0-0d00cc068806\") " pod="openshift-kube-controller-manager/revision-pruner-9-crc" Dec 10 22:52:01 crc kubenswrapper[4791]: I1210 22:52:01.891804 4791 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/0cc263fd-5b37-4c41-99c0-0d00cc068806-kubelet-dir\") pod \"revision-pruner-9-crc\" (UID: \"0cc263fd-5b37-4c41-99c0-0d00cc068806\") " pod="openshift-kube-controller-manager/revision-pruner-9-crc" Dec 10 22:52:01 crc kubenswrapper[4791]: I1210 22:52:01.914937 4791 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/0cc263fd-5b37-4c41-99c0-0d00cc068806-kube-api-access\") pod \"revision-pruner-9-crc\" (UID: \"0cc263fd-5b37-4c41-99c0-0d00cc068806\") " pod="openshift-kube-controller-manager/revision-pruner-9-crc" Dec 10 22:52:02 crc kubenswrapper[4791]: I1210 22:52:02.046800 4791 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-controller-manager/revision-pruner-9-crc" Dec 10 22:52:02 crc kubenswrapper[4791]: I1210 22:52:02.090413 4791 patch_prober.go:28] interesting pod/router-default-5444994796-p97fc container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Dec 10 22:52:02 crc kubenswrapper[4791]: [-]has-synced failed: reason withheld Dec 10 22:52:02 crc kubenswrapper[4791]: [+]process-running ok Dec 10 22:52:02 crc kubenswrapper[4791]: healthz check failed Dec 10 22:52:02 crc kubenswrapper[4791]: I1210 22:52:02.090483 4791 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-p97fc" podUID="d3de1ba0-b014-4f98-b597-cc89f5a70364" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Dec 10 22:52:02 crc kubenswrapper[4791]: I1210 22:52:02.131699 4791 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="8f668bae-612b-4b75-9490-919e737c6a3b" path="/var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes" Dec 10 22:52:02 crc kubenswrapper[4791]: I1210 22:52:02.132655 4791 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-marketplace-gtcrt"] Dec 10 22:52:02 crc kubenswrapper[4791]: I1210 22:52:02.134041 4791 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-gtcrt"] Dec 10 22:52:02 crc kubenswrapper[4791]: I1210 22:52:02.134205 4791 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-gtcrt" Dec 10 22:52:02 crc kubenswrapper[4791]: I1210 22:52:02.137538 4791 generic.go:334] "Generic (PLEG): container finished" podID="776c00fb-1f2c-4ea2-ba50-62d4d24ffd72" containerID="46c4394fb64814b43f7aa1d6e09152469f011e20b50f1603a8d3dfd24956021e" exitCode=0 Dec 10 22:52:02 crc kubenswrapper[4791]: I1210 22:52:02.137656 4791 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-q7lbx" event={"ID":"776c00fb-1f2c-4ea2-ba50-62d4d24ffd72","Type":"ContainerDied","Data":"46c4394fb64814b43f7aa1d6e09152469f011e20b50f1603a8d3dfd24956021e"} Dec 10 22:52:02 crc kubenswrapper[4791]: I1210 22:52:02.137716 4791 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-q7lbx" event={"ID":"776c00fb-1f2c-4ea2-ba50-62d4d24ffd72","Type":"ContainerStarted","Data":"b366355516b15a3cad414d8d67508f83152c420f1eb999866923386ef9535bce"} Dec 10 22:52:02 crc kubenswrapper[4791]: I1210 22:52:02.141869 4791 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="hostpath-provisioner/csi-hostpathplugin-mkz5r" event={"ID":"1666d5bb-295d-4828-868e-b1b17b5741bb","Type":"ContainerStarted","Data":"370cc275838024f455f7ab2eb24d00956fe422137700b44b1882933c2caf3d75"} Dec 10 22:52:02 crc kubenswrapper[4791]: I1210 22:52:02.143958 4791 generic.go:334] "Generic (PLEG): container finished" podID="08d50c37-7a4d-4731-beb8-fbc6d2e10bb0" containerID="efaa4cb838ed737e6af746d2de57003b1499ef7bbcbd43dcd52edbb01871d25b" exitCode=0 Dec 10 22:52:02 crc kubenswrapper[4791]: I1210 22:52:02.144020 4791 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-gskzr" event={"ID":"08d50c37-7a4d-4731-beb8-fbc6d2e10bb0","Type":"ContainerDied","Data":"efaa4cb838ed737e6af746d2de57003b1499ef7bbcbd43dcd52edbb01871d25b"} Dec 10 22:52:02 crc kubenswrapper[4791]: I1210 22:52:02.154435 4791 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-srd52\" (UID: \"e93dbd5d-0ca7-4a15-9eb3-0d71595fada0\") " pod="openshift-image-registry/image-registry-697d97f7c8-srd52" Dec 10 22:52:02 crc kubenswrapper[4791]: I1210 22:52:02.199954 4791 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/cd4347cf-ed66-480c-884f-b1d1c58ce143-catalog-content\") pod \"redhat-marketplace-gtcrt\" (UID: \"cd4347cf-ed66-480c-884f-b1d1c58ce143\") " pod="openshift-marketplace/redhat-marketplace-gtcrt" Dec 10 22:52:02 crc kubenswrapper[4791]: I1210 22:52:02.200113 4791 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-x4n6r\" (UniqueName: \"kubernetes.io/projected/cd4347cf-ed66-480c-884f-b1d1c58ce143-kube-api-access-x4n6r\") pod \"redhat-marketplace-gtcrt\" (UID: \"cd4347cf-ed66-480c-884f-b1d1c58ce143\") " pod="openshift-marketplace/redhat-marketplace-gtcrt" Dec 10 22:52:02 crc kubenswrapper[4791]: I1210 22:52:02.200140 4791 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/cd4347cf-ed66-480c-884f-b1d1c58ce143-utilities\") pod \"redhat-marketplace-gtcrt\" (UID: \"cd4347cf-ed66-480c-884f-b1d1c58ce143\") " pod="openshift-marketplace/redhat-marketplace-gtcrt" Dec 10 22:52:02 crc kubenswrapper[4791]: I1210 22:52:02.298289 4791 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-image-registry/image-registry-697d97f7c8-srd52" Dec 10 22:52:02 crc kubenswrapper[4791]: I1210 22:52:02.302234 4791 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-x4n6r\" (UniqueName: \"kubernetes.io/projected/cd4347cf-ed66-480c-884f-b1d1c58ce143-kube-api-access-x4n6r\") pod \"redhat-marketplace-gtcrt\" (UID: \"cd4347cf-ed66-480c-884f-b1d1c58ce143\") " pod="openshift-marketplace/redhat-marketplace-gtcrt" Dec 10 22:52:02 crc kubenswrapper[4791]: I1210 22:52:02.302288 4791 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/cd4347cf-ed66-480c-884f-b1d1c58ce143-utilities\") pod \"redhat-marketplace-gtcrt\" (UID: \"cd4347cf-ed66-480c-884f-b1d1c58ce143\") " pod="openshift-marketplace/redhat-marketplace-gtcrt" Dec 10 22:52:02 crc kubenswrapper[4791]: I1210 22:52:02.302378 4791 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/cd4347cf-ed66-480c-884f-b1d1c58ce143-catalog-content\") pod \"redhat-marketplace-gtcrt\" (UID: \"cd4347cf-ed66-480c-884f-b1d1c58ce143\") " pod="openshift-marketplace/redhat-marketplace-gtcrt" Dec 10 22:52:02 crc kubenswrapper[4791]: I1210 22:52:02.302934 4791 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/cd4347cf-ed66-480c-884f-b1d1c58ce143-catalog-content\") pod \"redhat-marketplace-gtcrt\" (UID: \"cd4347cf-ed66-480c-884f-b1d1c58ce143\") " pod="openshift-marketplace/redhat-marketplace-gtcrt" Dec 10 22:52:02 crc kubenswrapper[4791]: I1210 22:52:02.303654 4791 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/cd4347cf-ed66-480c-884f-b1d1c58ce143-utilities\") pod \"redhat-marketplace-gtcrt\" (UID: \"cd4347cf-ed66-480c-884f-b1d1c58ce143\") " pod="openshift-marketplace/redhat-marketplace-gtcrt" Dec 10 22:52:02 crc kubenswrapper[4791]: I1210 22:52:02.325822 4791 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-x4n6r\" (UniqueName: \"kubernetes.io/projected/cd4347cf-ed66-480c-884f-b1d1c58ce143-kube-api-access-x4n6r\") pod \"redhat-marketplace-gtcrt\" (UID: \"cd4347cf-ed66-480c-884f-b1d1c58ce143\") " pod="openshift-marketplace/redhat-marketplace-gtcrt" Dec 10 22:52:02 crc kubenswrapper[4791]: I1210 22:52:02.436919 4791 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29423445-b9cxb" Dec 10 22:52:02 crc kubenswrapper[4791]: I1210 22:52:02.492405 4791 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-apiserver/apiserver-76f77b778f-kqtbx" Dec 10 22:52:02 crc kubenswrapper[4791]: I1210 22:52:02.497171 4791 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-apiserver/apiserver-76f77b778f-kqtbx" Dec 10 22:52:02 crc kubenswrapper[4791]: I1210 22:52:02.510746 4791 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-gnhrf\" (UniqueName: \"kubernetes.io/projected/90640468-53d7-42bc-a835-f7acf48a486e-kube-api-access-gnhrf\") pod \"90640468-53d7-42bc-a835-f7acf48a486e\" (UID: \"90640468-53d7-42bc-a835-f7acf48a486e\") " Dec 10 22:52:02 crc kubenswrapper[4791]: I1210 22:52:02.510847 4791 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/90640468-53d7-42bc-a835-f7acf48a486e-config-volume\") pod \"90640468-53d7-42bc-a835-f7acf48a486e\" (UID: \"90640468-53d7-42bc-a835-f7acf48a486e\") " Dec 10 22:52:02 crc kubenswrapper[4791]: I1210 22:52:02.510889 4791 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/90640468-53d7-42bc-a835-f7acf48a486e-secret-volume\") pod \"90640468-53d7-42bc-a835-f7acf48a486e\" (UID: \"90640468-53d7-42bc-a835-f7acf48a486e\") " Dec 10 22:52:02 crc kubenswrapper[4791]: I1210 22:52:02.512748 4791 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/90640468-53d7-42bc-a835-f7acf48a486e-config-volume" (OuterVolumeSpecName: "config-volume") pod "90640468-53d7-42bc-a835-f7acf48a486e" (UID: "90640468-53d7-42bc-a835-f7acf48a486e"). InnerVolumeSpecName "config-volume". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 10 22:52:02 crc kubenswrapper[4791]: I1210 22:52:02.523569 4791 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/90640468-53d7-42bc-a835-f7acf48a486e-kube-api-access-gnhrf" (OuterVolumeSpecName: "kube-api-access-gnhrf") pod "90640468-53d7-42bc-a835-f7acf48a486e" (UID: "90640468-53d7-42bc-a835-f7acf48a486e"). InnerVolumeSpecName "kube-api-access-gnhrf". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 10 22:52:02 crc kubenswrapper[4791]: I1210 22:52:02.530725 4791 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/90640468-53d7-42bc-a835-f7acf48a486e-secret-volume" (OuterVolumeSpecName: "secret-volume") pod "90640468-53d7-42bc-a835-f7acf48a486e" (UID: "90640468-53d7-42bc-a835-f7acf48a486e"). InnerVolumeSpecName "secret-volume". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 10 22:52:02 crc kubenswrapper[4791]: I1210 22:52:02.534057 4791 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-operators-tj8hn"] Dec 10 22:52:02 crc kubenswrapper[4791]: E1210 22:52:02.534376 4791 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="90640468-53d7-42bc-a835-f7acf48a486e" containerName="collect-profiles" Dec 10 22:52:02 crc kubenswrapper[4791]: I1210 22:52:02.534390 4791 state_mem.go:107] "Deleted CPUSet assignment" podUID="90640468-53d7-42bc-a835-f7acf48a486e" containerName="collect-profiles" Dec 10 22:52:02 crc kubenswrapper[4791]: I1210 22:52:02.534521 4791 memory_manager.go:354] "RemoveStaleState removing state" podUID="90640468-53d7-42bc-a835-f7acf48a486e" containerName="collect-profiles" Dec 10 22:52:02 crc kubenswrapper[4791]: I1210 22:52:02.537666 4791 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-gtcrt" Dec 10 22:52:02 crc kubenswrapper[4791]: I1210 22:52:02.543485 4791 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-tj8hn" Dec 10 22:52:02 crc kubenswrapper[4791]: I1210 22:52:02.575847 4791 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"redhat-operators-dockercfg-ct8rh" Dec 10 22:52:02 crc kubenswrapper[4791]: I1210 22:52:02.602960 4791 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-tj8hn"] Dec 10 22:52:02 crc kubenswrapper[4791]: I1210 22:52:02.612421 4791 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/ab4c6e9a-3413-49c1-8f49-fb22c8366cb6-utilities\") pod \"redhat-operators-tj8hn\" (UID: \"ab4c6e9a-3413-49c1-8f49-fb22c8366cb6\") " pod="openshift-marketplace/redhat-operators-tj8hn" Dec 10 22:52:02 crc kubenswrapper[4791]: I1210 22:52:02.612612 4791 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/ab4c6e9a-3413-49c1-8f49-fb22c8366cb6-catalog-content\") pod \"redhat-operators-tj8hn\" (UID: \"ab4c6e9a-3413-49c1-8f49-fb22c8366cb6\") " pod="openshift-marketplace/redhat-operators-tj8hn" Dec 10 22:52:02 crc kubenswrapper[4791]: I1210 22:52:02.612638 4791 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-ck7dp\" (UniqueName: \"kubernetes.io/projected/ab4c6e9a-3413-49c1-8f49-fb22c8366cb6-kube-api-access-ck7dp\") pod \"redhat-operators-tj8hn\" (UID: \"ab4c6e9a-3413-49c1-8f49-fb22c8366cb6\") " pod="openshift-marketplace/redhat-operators-tj8hn" Dec 10 22:52:02 crc kubenswrapper[4791]: I1210 22:52:02.612690 4791 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-gnhrf\" (UniqueName: \"kubernetes.io/projected/90640468-53d7-42bc-a835-f7acf48a486e-kube-api-access-gnhrf\") on node \"crc\" DevicePath \"\"" Dec 10 22:52:02 crc kubenswrapper[4791]: I1210 22:52:02.612703 4791 reconciler_common.go:293] "Volume detached for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/90640468-53d7-42bc-a835-f7acf48a486e-config-volume\") on node \"crc\" DevicePath \"\"" Dec 10 22:52:02 crc kubenswrapper[4791]: I1210 22:52:02.612712 4791 reconciler_common.go:293] "Volume detached for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/90640468-53d7-42bc-a835-f7acf48a486e-secret-volume\") on node \"crc\" DevicePath \"\"" Dec 10 22:52:02 crc kubenswrapper[4791]: W1210 22:52:02.656522 4791 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pode93dbd5d_0ca7_4a15_9eb3_0d71595fada0.slice/crio-188e6bd81aba3d08f6a97b4a555211ed5d1b7ae3905df653258fe88d8a31fc5e WatchSource:0}: Error finding container 188e6bd81aba3d08f6a97b4a555211ed5d1b7ae3905df653258fe88d8a31fc5e: Status 404 returned error can't find the container with id 188e6bd81aba3d08f6a97b4a555211ed5d1b7ae3905df653258fe88d8a31fc5e Dec 10 22:52:02 crc kubenswrapper[4791]: I1210 22:52:02.665232 4791 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-image-registry/image-registry-697d97f7c8-srd52"] Dec 10 22:52:02 crc kubenswrapper[4791]: I1210 22:52:02.714241 4791 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/ab4c6e9a-3413-49c1-8f49-fb22c8366cb6-catalog-content\") pod \"redhat-operators-tj8hn\" (UID: \"ab4c6e9a-3413-49c1-8f49-fb22c8366cb6\") " pod="openshift-marketplace/redhat-operators-tj8hn" Dec 10 22:52:02 crc kubenswrapper[4791]: I1210 22:52:02.714287 4791 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-ck7dp\" (UniqueName: \"kubernetes.io/projected/ab4c6e9a-3413-49c1-8f49-fb22c8366cb6-kube-api-access-ck7dp\") pod \"redhat-operators-tj8hn\" (UID: \"ab4c6e9a-3413-49c1-8f49-fb22c8366cb6\") " pod="openshift-marketplace/redhat-operators-tj8hn" Dec 10 22:52:02 crc kubenswrapper[4791]: I1210 22:52:02.714320 4791 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/ab4c6e9a-3413-49c1-8f49-fb22c8366cb6-utilities\") pod \"redhat-operators-tj8hn\" (UID: \"ab4c6e9a-3413-49c1-8f49-fb22c8366cb6\") " pod="openshift-marketplace/redhat-operators-tj8hn" Dec 10 22:52:02 crc kubenswrapper[4791]: I1210 22:52:02.715150 4791 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/ab4c6e9a-3413-49c1-8f49-fb22c8366cb6-utilities\") pod \"redhat-operators-tj8hn\" (UID: \"ab4c6e9a-3413-49c1-8f49-fb22c8366cb6\") " pod="openshift-marketplace/redhat-operators-tj8hn" Dec 10 22:52:02 crc kubenswrapper[4791]: I1210 22:52:02.715366 4791 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/ab4c6e9a-3413-49c1-8f49-fb22c8366cb6-catalog-content\") pod \"redhat-operators-tj8hn\" (UID: \"ab4c6e9a-3413-49c1-8f49-fb22c8366cb6\") " pod="openshift-marketplace/redhat-operators-tj8hn" Dec 10 22:52:02 crc kubenswrapper[4791]: I1210 22:52:02.720946 4791 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-wkwbl"] Dec 10 22:52:02 crc kubenswrapper[4791]: I1210 22:52:02.732397 4791 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-controller-manager/revision-pruner-9-crc"] Dec 10 22:52:02 crc kubenswrapper[4791]: I1210 22:52:02.761865 4791 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-ck7dp\" (UniqueName: \"kubernetes.io/projected/ab4c6e9a-3413-49c1-8f49-fb22c8366cb6-kube-api-access-ck7dp\") pod \"redhat-operators-tj8hn\" (UID: \"ab4c6e9a-3413-49c1-8f49-fb22c8366cb6\") " pod="openshift-marketplace/redhat-operators-tj8hn" Dec 10 22:52:02 crc kubenswrapper[4791]: I1210 22:52:02.893280 4791 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-tj8hn" Dec 10 22:52:02 crc kubenswrapper[4791]: I1210 22:52:02.922530 4791 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-operators-p75kv"] Dec 10 22:52:02 crc kubenswrapper[4791]: I1210 22:52:02.923883 4791 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-p75kv" Dec 10 22:52:02 crc kubenswrapper[4791]: I1210 22:52:02.925884 4791 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-gtcrt"] Dec 10 22:52:02 crc kubenswrapper[4791]: I1210 22:52:02.929566 4791 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-p75kv"] Dec 10 22:52:03 crc kubenswrapper[4791]: I1210 22:52:03.024783 4791 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/abaf8cb8-8e36-4995-a926-8c4dadf4f56a-utilities\") pod \"redhat-operators-p75kv\" (UID: \"abaf8cb8-8e36-4995-a926-8c4dadf4f56a\") " pod="openshift-marketplace/redhat-operators-p75kv" Dec 10 22:52:03 crc kubenswrapper[4791]: I1210 22:52:03.024845 4791 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/abaf8cb8-8e36-4995-a926-8c4dadf4f56a-catalog-content\") pod \"redhat-operators-p75kv\" (UID: \"abaf8cb8-8e36-4995-a926-8c4dadf4f56a\") " pod="openshift-marketplace/redhat-operators-p75kv" Dec 10 22:52:03 crc kubenswrapper[4791]: I1210 22:52:03.024983 4791 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-m5w5q\" (UniqueName: \"kubernetes.io/projected/abaf8cb8-8e36-4995-a926-8c4dadf4f56a-kube-api-access-m5w5q\") pod \"redhat-operators-p75kv\" (UID: \"abaf8cb8-8e36-4995-a926-8c4dadf4f56a\") " pod="openshift-marketplace/redhat-operators-p75kv" Dec 10 22:52:03 crc kubenswrapper[4791]: I1210 22:52:03.085795 4791 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-ingress/router-default-5444994796-p97fc" Dec 10 22:52:03 crc kubenswrapper[4791]: I1210 22:52:03.089829 4791 patch_prober.go:28] interesting pod/router-default-5444994796-p97fc container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Dec 10 22:52:03 crc kubenswrapper[4791]: [-]has-synced failed: reason withheld Dec 10 22:52:03 crc kubenswrapper[4791]: [+]process-running ok Dec 10 22:52:03 crc kubenswrapper[4791]: healthz check failed Dec 10 22:52:03 crc kubenswrapper[4791]: I1210 22:52:03.089897 4791 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-p97fc" podUID="d3de1ba0-b014-4f98-b597-cc89f5a70364" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Dec 10 22:52:03 crc kubenswrapper[4791]: I1210 22:52:03.128279 4791 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/abaf8cb8-8e36-4995-a926-8c4dadf4f56a-utilities\") pod \"redhat-operators-p75kv\" (UID: \"abaf8cb8-8e36-4995-a926-8c4dadf4f56a\") " pod="openshift-marketplace/redhat-operators-p75kv" Dec 10 22:52:03 crc kubenswrapper[4791]: I1210 22:52:03.128307 4791 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/abaf8cb8-8e36-4995-a926-8c4dadf4f56a-catalog-content\") pod \"redhat-operators-p75kv\" (UID: \"abaf8cb8-8e36-4995-a926-8c4dadf4f56a\") " pod="openshift-marketplace/redhat-operators-p75kv" Dec 10 22:52:03 crc kubenswrapper[4791]: I1210 22:52:03.128378 4791 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-m5w5q\" (UniqueName: \"kubernetes.io/projected/abaf8cb8-8e36-4995-a926-8c4dadf4f56a-kube-api-access-m5w5q\") pod \"redhat-operators-p75kv\" (UID: \"abaf8cb8-8e36-4995-a926-8c4dadf4f56a\") " pod="openshift-marketplace/redhat-operators-p75kv" Dec 10 22:52:03 crc kubenswrapper[4791]: I1210 22:52:03.129542 4791 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/abaf8cb8-8e36-4995-a926-8c4dadf4f56a-utilities\") pod \"redhat-operators-p75kv\" (UID: \"abaf8cb8-8e36-4995-a926-8c4dadf4f56a\") " pod="openshift-marketplace/redhat-operators-p75kv" Dec 10 22:52:03 crc kubenswrapper[4791]: I1210 22:52:03.129835 4791 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/abaf8cb8-8e36-4995-a926-8c4dadf4f56a-catalog-content\") pod \"redhat-operators-p75kv\" (UID: \"abaf8cb8-8e36-4995-a926-8c4dadf4f56a\") " pod="openshift-marketplace/redhat-operators-p75kv" Dec 10 22:52:03 crc kubenswrapper[4791]: I1210 22:52:03.168492 4791 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-m5w5q\" (UniqueName: \"kubernetes.io/projected/abaf8cb8-8e36-4995-a926-8c4dadf4f56a-kube-api-access-m5w5q\") pod \"redhat-operators-p75kv\" (UID: \"abaf8cb8-8e36-4995-a926-8c4dadf4f56a\") " pod="openshift-marketplace/redhat-operators-p75kv" Dec 10 22:52:03 crc kubenswrapper[4791]: I1210 22:52:03.169396 4791 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-wkwbl" event={"ID":"396f55d6-c96d-4520-8165-862b5905a8c0","Type":"ContainerStarted","Data":"867bd37f5f9510139aaf03f18fd25afbac0100646d83d88e2ae1c3b62c657cff"} Dec 10 22:52:03 crc kubenswrapper[4791]: I1210 22:52:03.174200 4791 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29423445-b9cxb" event={"ID":"90640468-53d7-42bc-a835-f7acf48a486e","Type":"ContainerDied","Data":"b6daf82e1d63eb3f8e9105be94dd5bfa39e380aaaa223e2aa77660eccd4f2d07"} Dec 10 22:52:03 crc kubenswrapper[4791]: I1210 22:52:03.174234 4791 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="b6daf82e1d63eb3f8e9105be94dd5bfa39e380aaaa223e2aa77660eccd4f2d07" Dec 10 22:52:03 crc kubenswrapper[4791]: I1210 22:52:03.174356 4791 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29423445-b9cxb" Dec 10 22:52:03 crc kubenswrapper[4791]: I1210 22:52:03.177835 4791 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-gtcrt" event={"ID":"cd4347cf-ed66-480c-884f-b1d1c58ce143","Type":"ContainerStarted","Data":"138c31a05fe2a71cecac04b09a5f86e0a18c4dab110562b6c6278d5f8e4bb8fb"} Dec 10 22:52:03 crc kubenswrapper[4791]: I1210 22:52:03.187451 4791 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-tj8hn"] Dec 10 22:52:03 crc kubenswrapper[4791]: I1210 22:52:03.189541 4791 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-image-registry/image-registry-697d97f7c8-srd52" event={"ID":"e93dbd5d-0ca7-4a15-9eb3-0d71595fada0","Type":"ContainerStarted","Data":"e6eab22df9424a9028c3071cbd5cae655435fbbba257f037a06dfc92facc867d"} Dec 10 22:52:03 crc kubenswrapper[4791]: I1210 22:52:03.189572 4791 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-image-registry/image-registry-697d97f7c8-srd52" event={"ID":"e93dbd5d-0ca7-4a15-9eb3-0d71595fada0","Type":"ContainerStarted","Data":"188e6bd81aba3d08f6a97b4a555211ed5d1b7ae3905df653258fe88d8a31fc5e"} Dec 10 22:52:03 crc kubenswrapper[4791]: I1210 22:52:03.190063 4791 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-image-registry/image-registry-697d97f7c8-srd52" Dec 10 22:52:03 crc kubenswrapper[4791]: I1210 22:52:03.199831 4791 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="hostpath-provisioner/csi-hostpathplugin-mkz5r" event={"ID":"1666d5bb-295d-4828-868e-b1b17b5741bb","Type":"ContainerStarted","Data":"beb3533c43f007ff1ef9a4f568926862ed1c993de01282cec3a05fc2ddcd1ad6"} Dec 10 22:52:03 crc kubenswrapper[4791]: I1210 22:52:03.202768 4791 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-controller-manager/revision-pruner-9-crc" event={"ID":"0cc263fd-5b37-4c41-99c0-0d00cc068806","Type":"ContainerStarted","Data":"e2861200d9d84a41f34512e7bf44a393680974f5a7410553f507695100e1d7fe"} Dec 10 22:52:03 crc kubenswrapper[4791]: I1210 22:52:03.213564 4791 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-image-registry/image-registry-697d97f7c8-srd52" podStartSLOduration=132.213547649 podStartE2EDuration="2m12.213547649s" podCreationTimestamp="2025-12-10 22:49:51 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-10 22:52:03.210227765 +0000 UTC m=+157.639845368" watchObservedRunningTime="2025-12-10 22:52:03.213547649 +0000 UTC m=+157.643165262" Dec 10 22:52:03 crc kubenswrapper[4791]: W1210 22:52:03.233373 4791 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podab4c6e9a_3413_49c1_8f49_fb22c8366cb6.slice/crio-44af7aee753f8d061171bf3f5302bff67aec21202f33f7425d7c967c105223c0 WatchSource:0}: Error finding container 44af7aee753f8d061171bf3f5302bff67aec21202f33f7425d7c967c105223c0: Status 404 returned error can't find the container with id 44af7aee753f8d061171bf3f5302bff67aec21202f33f7425d7c967c105223c0 Dec 10 22:52:03 crc kubenswrapper[4791]: I1210 22:52:03.251395 4791 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-p75kv" Dec 10 22:52:03 crc kubenswrapper[4791]: I1210 22:52:03.463303 4791 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="hostpath-provisioner/csi-hostpathplugin-mkz5r" podStartSLOduration=13.463287236 podStartE2EDuration="13.463287236s" podCreationTimestamp="2025-12-10 22:51:50 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-10 22:52:03.233376509 +0000 UTC m=+157.662994122" watchObservedRunningTime="2025-12-10 22:52:03.463287236 +0000 UTC m=+157.892904849" Dec 10 22:52:03 crc kubenswrapper[4791]: I1210 22:52:03.466798 4791 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-p75kv"] Dec 10 22:52:03 crc kubenswrapper[4791]: W1210 22:52:03.476167 4791 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podabaf8cb8_8e36_4995_a926_8c4dadf4f56a.slice/crio-94c90e6557b125e09be0d59c19374ae4cd27dc735ba804210c0856e9055229fc WatchSource:0}: Error finding container 94c90e6557b125e09be0d59c19374ae4cd27dc735ba804210c0856e9055229fc: Status 404 returned error can't find the container with id 94c90e6557b125e09be0d59c19374ae4cd27dc735ba804210c0856e9055229fc Dec 10 22:52:03 crc kubenswrapper[4791]: I1210 22:52:03.518552 4791 patch_prober.go:28] interesting pod/downloads-7954f5f757-mlb9p container/download-server namespace/openshift-console: Readiness probe status=failure output="Get \"http://10.217.0.10:8080/\": dial tcp 10.217.0.10:8080: connect: connection refused" start-of-body= Dec 10 22:52:03 crc kubenswrapper[4791]: I1210 22:52:03.518612 4791 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-console/downloads-7954f5f757-mlb9p" podUID="680d2296-dd48-4786-bea7-64c32edc94c7" containerName="download-server" probeResult="failure" output="Get \"http://10.217.0.10:8080/\": dial tcp 10.217.0.10:8080: connect: connection refused" Dec 10 22:52:03 crc kubenswrapper[4791]: I1210 22:52:03.518754 4791 patch_prober.go:28] interesting pod/downloads-7954f5f757-mlb9p container/download-server namespace/openshift-console: Liveness probe status=failure output="Get \"http://10.217.0.10:8080/\": dial tcp 10.217.0.10:8080: connect: connection refused" start-of-body= Dec 10 22:52:03 crc kubenswrapper[4791]: I1210 22:52:03.518801 4791 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-console/downloads-7954f5f757-mlb9p" podUID="680d2296-dd48-4786-bea7-64c32edc94c7" containerName="download-server" probeResult="failure" output="Get \"http://10.217.0.10:8080/\": dial tcp 10.217.0.10:8080: connect: connection refused" Dec 10 22:52:03 crc kubenswrapper[4791]: I1210 22:52:03.518872 4791 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-console/console-f9d7485db-zkfkv" Dec 10 22:52:03 crc kubenswrapper[4791]: I1210 22:52:03.519147 4791 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-console/console-f9d7485db-zkfkv" Dec 10 22:52:03 crc kubenswrapper[4791]: I1210 22:52:03.520332 4791 patch_prober.go:28] interesting pod/console-f9d7485db-zkfkv container/console namespace/openshift-console: Startup probe status=failure output="Get \"https://10.217.0.31:8443/health\": dial tcp 10.217.0.31:8443: connect: connection refused" start-of-body= Dec 10 22:52:03 crc kubenswrapper[4791]: I1210 22:52:03.520375 4791 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-console/console-f9d7485db-zkfkv" podUID="1b61a973-db5e-4aa7-96e3-7fbdd9dd8cc0" containerName="console" probeResult="failure" output="Get \"https://10.217.0.31:8443/health\": dial tcp 10.217.0.31:8443: connect: connection refused" Dec 10 22:52:04 crc kubenswrapper[4791]: I1210 22:52:04.091863 4791 patch_prober.go:28] interesting pod/router-default-5444994796-p97fc container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Dec 10 22:52:04 crc kubenswrapper[4791]: [-]has-synced failed: reason withheld Dec 10 22:52:04 crc kubenswrapper[4791]: [+]process-running ok Dec 10 22:52:04 crc kubenswrapper[4791]: healthz check failed Dec 10 22:52:04 crc kubenswrapper[4791]: I1210 22:52:04.091938 4791 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-p97fc" podUID="d3de1ba0-b014-4f98-b597-cc89f5a70364" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Dec 10 22:52:04 crc kubenswrapper[4791]: I1210 22:52:04.224907 4791 generic.go:334] "Generic (PLEG): container finished" podID="cd4347cf-ed66-480c-884f-b1d1c58ce143" containerID="2da96993499b7ba23ccd6ddc9e5aa1fb62c1c9aa98a3d4089b3afa5e90f291d0" exitCode=0 Dec 10 22:52:04 crc kubenswrapper[4791]: I1210 22:52:04.225023 4791 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-gtcrt" event={"ID":"cd4347cf-ed66-480c-884f-b1d1c58ce143","Type":"ContainerDied","Data":"2da96993499b7ba23ccd6ddc9e5aa1fb62c1c9aa98a3d4089b3afa5e90f291d0"} Dec 10 22:52:04 crc kubenswrapper[4791]: I1210 22:52:04.228712 4791 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-controller-manager/revision-pruner-9-crc" event={"ID":"0cc263fd-5b37-4c41-99c0-0d00cc068806","Type":"ContainerStarted","Data":"a8316124339720856fa137ed274f67af1deaabe96c2f52fb0ba8cc8c66d078d5"} Dec 10 22:52:04 crc kubenswrapper[4791]: I1210 22:52:04.234108 4791 generic.go:334] "Generic (PLEG): container finished" podID="ab4c6e9a-3413-49c1-8f49-fb22c8366cb6" containerID="8cb71106b36181af379ea2e38b1fd99f7c46ecb2e6edd181c06339af142843e1" exitCode=0 Dec 10 22:52:04 crc kubenswrapper[4791]: I1210 22:52:04.234169 4791 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-tj8hn" event={"ID":"ab4c6e9a-3413-49c1-8f49-fb22c8366cb6","Type":"ContainerDied","Data":"8cb71106b36181af379ea2e38b1fd99f7c46ecb2e6edd181c06339af142843e1"} Dec 10 22:52:04 crc kubenswrapper[4791]: I1210 22:52:04.234204 4791 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-tj8hn" event={"ID":"ab4c6e9a-3413-49c1-8f49-fb22c8366cb6","Type":"ContainerStarted","Data":"44af7aee753f8d061171bf3f5302bff67aec21202f33f7425d7c967c105223c0"} Dec 10 22:52:04 crc kubenswrapper[4791]: I1210 22:52:04.236905 4791 generic.go:334] "Generic (PLEG): container finished" podID="abaf8cb8-8e36-4995-a926-8c4dadf4f56a" containerID="35438f70b7ffc3d38c8e35729e9f71d94d2ca6cbfdf2d54bbc57add550ee4eb7" exitCode=0 Dec 10 22:52:04 crc kubenswrapper[4791]: I1210 22:52:04.236989 4791 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-p75kv" event={"ID":"abaf8cb8-8e36-4995-a926-8c4dadf4f56a","Type":"ContainerDied","Data":"35438f70b7ffc3d38c8e35729e9f71d94d2ca6cbfdf2d54bbc57add550ee4eb7"} Dec 10 22:52:04 crc kubenswrapper[4791]: I1210 22:52:04.237063 4791 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-p75kv" event={"ID":"abaf8cb8-8e36-4995-a926-8c4dadf4f56a","Type":"ContainerStarted","Data":"94c90e6557b125e09be0d59c19374ae4cd27dc735ba804210c0856e9055229fc"} Dec 10 22:52:04 crc kubenswrapper[4791]: I1210 22:52:04.240613 4791 generic.go:334] "Generic (PLEG): container finished" podID="396f55d6-c96d-4520-8165-862b5905a8c0" containerID="2168264d99228a03c59b448c46e5492deb2f9d8b5c2dc0f6b1d727fe35dd6679" exitCode=0 Dec 10 22:52:04 crc kubenswrapper[4791]: I1210 22:52:04.240790 4791 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-wkwbl" event={"ID":"396f55d6-c96d-4520-8165-862b5905a8c0","Type":"ContainerDied","Data":"2168264d99228a03c59b448c46e5492deb2f9d8b5c2dc0f6b1d727fe35dd6679"} Dec 10 22:52:04 crc kubenswrapper[4791]: I1210 22:52:04.309285 4791 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-kube-controller-manager/revision-pruner-9-crc" podStartSLOduration=3.309267491 podStartE2EDuration="3.309267491s" podCreationTimestamp="2025-12-10 22:52:01 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-10 22:52:04.306801601 +0000 UTC m=+158.736419224" watchObservedRunningTime="2025-12-10 22:52:04.309267491 +0000 UTC m=+158.738885104" Dec 10 22:52:05 crc kubenswrapper[4791]: I1210 22:52:05.089148 4791 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-ingress/router-default-5444994796-p97fc" Dec 10 22:52:05 crc kubenswrapper[4791]: I1210 22:52:05.091740 4791 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-ingress/router-default-5444994796-p97fc" Dec 10 22:52:05 crc kubenswrapper[4791]: I1210 22:52:05.248580 4791 generic.go:334] "Generic (PLEG): container finished" podID="0cc263fd-5b37-4c41-99c0-0d00cc068806" containerID="a8316124339720856fa137ed274f67af1deaabe96c2f52fb0ba8cc8c66d078d5" exitCode=0 Dec 10 22:52:05 crc kubenswrapper[4791]: I1210 22:52:05.249247 4791 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-controller-manager/revision-pruner-9-crc" event={"ID":"0cc263fd-5b37-4c41-99c0-0d00cc068806","Type":"ContainerDied","Data":"a8316124339720856fa137ed274f67af1deaabe96c2f52fb0ba8cc8c66d078d5"} Dec 10 22:52:07 crc kubenswrapper[4791]: I1210 22:52:07.387946 4791 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-kube-apiserver/revision-pruner-8-crc"] Dec 10 22:52:07 crc kubenswrapper[4791]: I1210 22:52:07.388812 4791 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/revision-pruner-8-crc" Dec 10 22:52:07 crc kubenswrapper[4791]: I1210 22:52:07.391734 4791 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-apiserver"/"kube-root-ca.crt" Dec 10 22:52:07 crc kubenswrapper[4791]: I1210 22:52:07.393512 4791 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-apiserver"/"installer-sa-dockercfg-5pr6n" Dec 10 22:52:07 crc kubenswrapper[4791]: I1210 22:52:07.401268 4791 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-apiserver/revision-pruner-8-crc"] Dec 10 22:52:07 crc kubenswrapper[4791]: I1210 22:52:07.408332 4791 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/76668a03-41c5-4892-8a62-ba917064ed0a-kubelet-dir\") pod \"revision-pruner-8-crc\" (UID: \"76668a03-41c5-4892-8a62-ba917064ed0a\") " pod="openshift-kube-apiserver/revision-pruner-8-crc" Dec 10 22:52:07 crc kubenswrapper[4791]: I1210 22:52:07.408410 4791 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/76668a03-41c5-4892-8a62-ba917064ed0a-kube-api-access\") pod \"revision-pruner-8-crc\" (UID: \"76668a03-41c5-4892-8a62-ba917064ed0a\") " pod="openshift-kube-apiserver/revision-pruner-8-crc" Dec 10 22:52:07 crc kubenswrapper[4791]: I1210 22:52:07.509888 4791 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/76668a03-41c5-4892-8a62-ba917064ed0a-kubelet-dir\") pod \"revision-pruner-8-crc\" (UID: \"76668a03-41c5-4892-8a62-ba917064ed0a\") " pod="openshift-kube-apiserver/revision-pruner-8-crc" Dec 10 22:52:07 crc kubenswrapper[4791]: I1210 22:52:07.509936 4791 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/76668a03-41c5-4892-8a62-ba917064ed0a-kube-api-access\") pod \"revision-pruner-8-crc\" (UID: \"76668a03-41c5-4892-8a62-ba917064ed0a\") " pod="openshift-kube-apiserver/revision-pruner-8-crc" Dec 10 22:52:07 crc kubenswrapper[4791]: I1210 22:52:07.510207 4791 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/76668a03-41c5-4892-8a62-ba917064ed0a-kubelet-dir\") pod \"revision-pruner-8-crc\" (UID: \"76668a03-41c5-4892-8a62-ba917064ed0a\") " pod="openshift-kube-apiserver/revision-pruner-8-crc" Dec 10 22:52:07 crc kubenswrapper[4791]: I1210 22:52:07.528900 4791 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/76668a03-41c5-4892-8a62-ba917064ed0a-kube-api-access\") pod \"revision-pruner-8-crc\" (UID: \"76668a03-41c5-4892-8a62-ba917064ed0a\") " pod="openshift-kube-apiserver/revision-pruner-8-crc" Dec 10 22:52:07 crc kubenswrapper[4791]: I1210 22:52:07.718837 4791 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/revision-pruner-8-crc" Dec 10 22:52:08 crc kubenswrapper[4791]: I1210 22:52:08.105264 4791 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-ovn-kubernetes/ovnkube-node-zhq64" Dec 10 22:52:09 crc kubenswrapper[4791]: I1210 22:52:09.229485 4791 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 10 22:52:09 crc kubenswrapper[4791]: I1210 22:52:09.440743 4791 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-dns/dns-default-qs77n" Dec 10 22:52:12 crc kubenswrapper[4791]: I1210 22:52:12.569968 4791 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-kube-controller-manager/revision-pruner-9-crc" Dec 10 22:52:12 crc kubenswrapper[4791]: I1210 22:52:12.694461 4791 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/0cc263fd-5b37-4c41-99c0-0d00cc068806-kubelet-dir\") pod \"0cc263fd-5b37-4c41-99c0-0d00cc068806\" (UID: \"0cc263fd-5b37-4c41-99c0-0d00cc068806\") " Dec 10 22:52:12 crc kubenswrapper[4791]: I1210 22:52:12.694598 4791 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/0cc263fd-5b37-4c41-99c0-0d00cc068806-kube-api-access\") pod \"0cc263fd-5b37-4c41-99c0-0d00cc068806\" (UID: \"0cc263fd-5b37-4c41-99c0-0d00cc068806\") " Dec 10 22:52:12 crc kubenswrapper[4791]: I1210 22:52:12.694617 4791 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/0cc263fd-5b37-4c41-99c0-0d00cc068806-kubelet-dir" (OuterVolumeSpecName: "kubelet-dir") pod "0cc263fd-5b37-4c41-99c0-0d00cc068806" (UID: "0cc263fd-5b37-4c41-99c0-0d00cc068806"). InnerVolumeSpecName "kubelet-dir". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 10 22:52:12 crc kubenswrapper[4791]: I1210 22:52:12.694950 4791 reconciler_common.go:293] "Volume detached for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/0cc263fd-5b37-4c41-99c0-0d00cc068806-kubelet-dir\") on node \"crc\" DevicePath \"\"" Dec 10 22:52:12 crc kubenswrapper[4791]: I1210 22:52:12.715090 4791 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/0cc263fd-5b37-4c41-99c0-0d00cc068806-kube-api-access" (OuterVolumeSpecName: "kube-api-access") pod "0cc263fd-5b37-4c41-99c0-0d00cc068806" (UID: "0cc263fd-5b37-4c41-99c0-0d00cc068806"). InnerVolumeSpecName "kube-api-access". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 10 22:52:12 crc kubenswrapper[4791]: I1210 22:52:12.803554 4791 reconciler_common.go:293] "Volume detached for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/0cc263fd-5b37-4c41-99c0-0d00cc068806-kube-api-access\") on node \"crc\" DevicePath \"\"" Dec 10 22:52:12 crc kubenswrapper[4791]: I1210 22:52:12.905085 4791 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/b4527e71-bb58-4a39-a99c-fe614aa4c5f9-metrics-certs\") pod \"network-metrics-daemon-b6kf6\" (UID: \"b4527e71-bb58-4a39-a99c-fe614aa4c5f9\") " pod="openshift-multus/network-metrics-daemon-b6kf6" Dec 10 22:52:12 crc kubenswrapper[4791]: I1210 22:52:12.908853 4791 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/b4527e71-bb58-4a39-a99c-fe614aa4c5f9-metrics-certs\") pod \"network-metrics-daemon-b6kf6\" (UID: \"b4527e71-bb58-4a39-a99c-fe614aa4c5f9\") " pod="openshift-multus/network-metrics-daemon-b6kf6" Dec 10 22:52:13 crc kubenswrapper[4791]: I1210 22:52:13.107503 4791 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-b6kf6" Dec 10 22:52:13 crc kubenswrapper[4791]: I1210 22:52:13.313382 4791 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-controller-manager/revision-pruner-9-crc" event={"ID":"0cc263fd-5b37-4c41-99c0-0d00cc068806","Type":"ContainerDied","Data":"e2861200d9d84a41f34512e7bf44a393680974f5a7410553f507695100e1d7fe"} Dec 10 22:52:13 crc kubenswrapper[4791]: I1210 22:52:13.313452 4791 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="e2861200d9d84a41f34512e7bf44a393680974f5a7410553f507695100e1d7fe" Dec 10 22:52:13 crc kubenswrapper[4791]: I1210 22:52:13.313449 4791 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-kube-controller-manager/revision-pruner-9-crc" Dec 10 22:52:13 crc kubenswrapper[4791]: I1210 22:52:13.522180 4791 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-console/console-f9d7485db-zkfkv" Dec 10 22:52:13 crc kubenswrapper[4791]: I1210 22:52:13.525611 4791 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-console/console-f9d7485db-zkfkv" Dec 10 22:52:13 crc kubenswrapper[4791]: I1210 22:52:13.531126 4791 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-console/downloads-7954f5f757-mlb9p" Dec 10 22:52:22 crc kubenswrapper[4791]: I1210 22:52:22.307128 4791 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-image-registry/image-registry-697d97f7c8-srd52" Dec 10 22:52:24 crc kubenswrapper[4791]: I1210 22:52:24.127542 4791 patch_prober.go:28] interesting pod/router-default-5444994796-p97fc container/router namespace/openshift-ingress: Readiness probe status=failure output="Get \"http://localhost:1936/healthz/ready\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" start-of-body= Dec 10 22:52:24 crc kubenswrapper[4791]: I1210 22:52:24.127813 4791 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-ingress/router-default-5444994796-p97fc" podUID="d3de1ba0-b014-4f98-b597-cc89f5a70364" containerName="router" probeResult="failure" output="Get \"http://localhost:1936/healthz/ready\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" Dec 10 22:52:25 crc kubenswrapper[4791]: I1210 22:52:25.038110 4791 patch_prober.go:28] interesting pod/machine-config-daemon-5rb5l container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 10 22:52:25 crc kubenswrapper[4791]: I1210 22:52:25.038275 4791 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-5rb5l" podUID="ba35653c-6e06-4cee-a4d6-137764090d18" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 10 22:52:34 crc kubenswrapper[4791]: I1210 22:52:34.363312 4791 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-w5dpr" Dec 10 22:52:38 crc kubenswrapper[4791]: I1210 22:52:38.584158 4791 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-kube-apiserver/revision-pruner-9-crc"] Dec 10 22:52:38 crc kubenswrapper[4791]: E1210 22:52:38.584481 4791 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="0cc263fd-5b37-4c41-99c0-0d00cc068806" containerName="pruner" Dec 10 22:52:38 crc kubenswrapper[4791]: I1210 22:52:38.584497 4791 state_mem.go:107] "Deleted CPUSet assignment" podUID="0cc263fd-5b37-4c41-99c0-0d00cc068806" containerName="pruner" Dec 10 22:52:38 crc kubenswrapper[4791]: I1210 22:52:38.584638 4791 memory_manager.go:354] "RemoveStaleState removing state" podUID="0cc263fd-5b37-4c41-99c0-0d00cc068806" containerName="pruner" Dec 10 22:52:38 crc kubenswrapper[4791]: I1210 22:52:38.585069 4791 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/revision-pruner-9-crc" Dec 10 22:52:38 crc kubenswrapper[4791]: I1210 22:52:38.598309 4791 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-apiserver/revision-pruner-9-crc"] Dec 10 22:52:38 crc kubenswrapper[4791]: I1210 22:52:38.652789 4791 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/602d7666-cce8-4158-a369-6921a5f6605b-kube-api-access\") pod \"revision-pruner-9-crc\" (UID: \"602d7666-cce8-4158-a369-6921a5f6605b\") " pod="openshift-kube-apiserver/revision-pruner-9-crc" Dec 10 22:52:38 crc kubenswrapper[4791]: I1210 22:52:38.652870 4791 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/602d7666-cce8-4158-a369-6921a5f6605b-kubelet-dir\") pod \"revision-pruner-9-crc\" (UID: \"602d7666-cce8-4158-a369-6921a5f6605b\") " pod="openshift-kube-apiserver/revision-pruner-9-crc" Dec 10 22:52:38 crc kubenswrapper[4791]: I1210 22:52:38.754073 4791 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/602d7666-cce8-4158-a369-6921a5f6605b-kubelet-dir\") pod \"revision-pruner-9-crc\" (UID: \"602d7666-cce8-4158-a369-6921a5f6605b\") " pod="openshift-kube-apiserver/revision-pruner-9-crc" Dec 10 22:52:38 crc kubenswrapper[4791]: I1210 22:52:38.754145 4791 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/602d7666-cce8-4158-a369-6921a5f6605b-kube-api-access\") pod \"revision-pruner-9-crc\" (UID: \"602d7666-cce8-4158-a369-6921a5f6605b\") " pod="openshift-kube-apiserver/revision-pruner-9-crc" Dec 10 22:52:38 crc kubenswrapper[4791]: I1210 22:52:38.754204 4791 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/602d7666-cce8-4158-a369-6921a5f6605b-kubelet-dir\") pod \"revision-pruner-9-crc\" (UID: \"602d7666-cce8-4158-a369-6921a5f6605b\") " pod="openshift-kube-apiserver/revision-pruner-9-crc" Dec 10 22:52:38 crc kubenswrapper[4791]: I1210 22:52:38.776879 4791 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/602d7666-cce8-4158-a369-6921a5f6605b-kube-api-access\") pod \"revision-pruner-9-crc\" (UID: \"602d7666-cce8-4158-a369-6921a5f6605b\") " pod="openshift-kube-apiserver/revision-pruner-9-crc" Dec 10 22:52:38 crc kubenswrapper[4791]: I1210 22:52:38.910941 4791 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/revision-pruner-9-crc" Dec 10 22:52:40 crc kubenswrapper[4791]: E1210 22:52:40.680826 4791 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" image="registry.redhat.io/redhat/community-operator-index:v4.18" Dec 10 22:52:40 crc kubenswrapper[4791]: E1210 22:52:40.681328 4791 kuberuntime_manager.go:1274] "Unhandled Error" err="init container &Container{Name:extract-content,Image:registry.redhat.io/redhat/community-operator-index:v4.18,Command:[/utilities/copy-content],Args:[--catalog.from=/configs --catalog.to=/extracted-catalog/catalog --cache.from=/tmp/cache --cache.to=/extracted-catalog/cache],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:utilities,ReadOnly:false,MountPath:/utilities,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:catalog-content,ReadOnly:false,MountPath:/extracted-catalog,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-xcs9j,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:Always,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000170000,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:FallbackToLogsOnError,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod community-operators-q7lbx_openshift-marketplace(776c00fb-1f2c-4ea2-ba50-62d4d24ffd72): ErrImagePull: rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" logger="UnhandledError" Dec 10 22:52:40 crc kubenswrapper[4791]: E1210 22:52:40.682601 4791 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ErrImagePull: \"rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled\"" pod="openshift-marketplace/community-operators-q7lbx" podUID="776c00fb-1f2c-4ea2-ba50-62d4d24ffd72" Dec 10 22:52:40 crc kubenswrapper[4791]: I1210 22:52:40.743033 4791 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 10 22:52:42 crc kubenswrapper[4791]: E1210 22:52:42.891928 4791 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ImagePullBackOff: \"Back-off pulling image \\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"\"" pod="openshift-marketplace/community-operators-q7lbx" podUID="776c00fb-1f2c-4ea2-ba50-62d4d24ffd72" Dec 10 22:52:43 crc kubenswrapper[4791]: I1210 22:52:43.311390 4791 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-apiserver/revision-pruner-8-crc"] Dec 10 22:52:43 crc kubenswrapper[4791]: I1210 22:52:43.577726 4791 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-kube-apiserver/installer-9-crc"] Dec 10 22:52:43 crc kubenswrapper[4791]: I1210 22:52:43.579010 4791 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/installer-9-crc" Dec 10 22:52:43 crc kubenswrapper[4791]: I1210 22:52:43.598689 4791 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-apiserver/installer-9-crc"] Dec 10 22:52:43 crc kubenswrapper[4791]: I1210 22:52:43.641250 4791 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/31bd0efc-25ef-4a1f-b4b9-e4c5b14768ea-kube-api-access\") pod \"installer-9-crc\" (UID: \"31bd0efc-25ef-4a1f-b4b9-e4c5b14768ea\") " pod="openshift-kube-apiserver/installer-9-crc" Dec 10 22:52:43 crc kubenswrapper[4791]: I1210 22:52:43.641319 4791 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-lock\" (UniqueName: \"kubernetes.io/host-path/31bd0efc-25ef-4a1f-b4b9-e4c5b14768ea-var-lock\") pod \"installer-9-crc\" (UID: \"31bd0efc-25ef-4a1f-b4b9-e4c5b14768ea\") " pod="openshift-kube-apiserver/installer-9-crc" Dec 10 22:52:43 crc kubenswrapper[4791]: I1210 22:52:43.641373 4791 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/31bd0efc-25ef-4a1f-b4b9-e4c5b14768ea-kubelet-dir\") pod \"installer-9-crc\" (UID: \"31bd0efc-25ef-4a1f-b4b9-e4c5b14768ea\") " pod="openshift-kube-apiserver/installer-9-crc" Dec 10 22:52:43 crc kubenswrapper[4791]: E1210 22:52:43.723334 4791 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" image="registry.redhat.io/redhat/certified-operator-index:v4.18" Dec 10 22:52:43 crc kubenswrapper[4791]: E1210 22:52:43.723699 4791 kuberuntime_manager.go:1274] "Unhandled Error" err="init container &Container{Name:extract-content,Image:registry.redhat.io/redhat/certified-operator-index:v4.18,Command:[/utilities/copy-content],Args:[--catalog.from=/configs --catalog.to=/extracted-catalog/catalog --cache.from=/tmp/cache --cache.to=/extracted-catalog/cache],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:utilities,ReadOnly:false,MountPath:/utilities,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:catalog-content,ReadOnly:false,MountPath:/extracted-catalog,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-z2kjq,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:Always,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000170000,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:FallbackToLogsOnError,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod certified-operators-8x2rc_openshift-marketplace(009efc35-5943-429f-98fd-3307a5f557f2): ErrImagePull: rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" logger="UnhandledError" Dec 10 22:52:43 crc kubenswrapper[4791]: E1210 22:52:43.724912 4791 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ErrImagePull: \"rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled\"" pod="openshift-marketplace/certified-operators-8x2rc" podUID="009efc35-5943-429f-98fd-3307a5f557f2" Dec 10 22:52:43 crc kubenswrapper[4791]: I1210 22:52:43.742905 4791 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/31bd0efc-25ef-4a1f-b4b9-e4c5b14768ea-kube-api-access\") pod \"installer-9-crc\" (UID: \"31bd0efc-25ef-4a1f-b4b9-e4c5b14768ea\") " pod="openshift-kube-apiserver/installer-9-crc" Dec 10 22:52:43 crc kubenswrapper[4791]: I1210 22:52:43.742982 4791 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-lock\" (UniqueName: \"kubernetes.io/host-path/31bd0efc-25ef-4a1f-b4b9-e4c5b14768ea-var-lock\") pod \"installer-9-crc\" (UID: \"31bd0efc-25ef-4a1f-b4b9-e4c5b14768ea\") " pod="openshift-kube-apiserver/installer-9-crc" Dec 10 22:52:43 crc kubenswrapper[4791]: I1210 22:52:43.743015 4791 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/31bd0efc-25ef-4a1f-b4b9-e4c5b14768ea-kubelet-dir\") pod \"installer-9-crc\" (UID: \"31bd0efc-25ef-4a1f-b4b9-e4c5b14768ea\") " pod="openshift-kube-apiserver/installer-9-crc" Dec 10 22:52:43 crc kubenswrapper[4791]: I1210 22:52:43.743135 4791 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/31bd0efc-25ef-4a1f-b4b9-e4c5b14768ea-kubelet-dir\") pod \"installer-9-crc\" (UID: \"31bd0efc-25ef-4a1f-b4b9-e4c5b14768ea\") " pod="openshift-kube-apiserver/installer-9-crc" Dec 10 22:52:43 crc kubenswrapper[4791]: I1210 22:52:43.743177 4791 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-lock\" (UniqueName: \"kubernetes.io/host-path/31bd0efc-25ef-4a1f-b4b9-e4c5b14768ea-var-lock\") pod \"installer-9-crc\" (UID: \"31bd0efc-25ef-4a1f-b4b9-e4c5b14768ea\") " pod="openshift-kube-apiserver/installer-9-crc" Dec 10 22:52:43 crc kubenswrapper[4791]: I1210 22:52:43.762267 4791 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/31bd0efc-25ef-4a1f-b4b9-e4c5b14768ea-kube-api-access\") pod \"installer-9-crc\" (UID: \"31bd0efc-25ef-4a1f-b4b9-e4c5b14768ea\") " pod="openshift-kube-apiserver/installer-9-crc" Dec 10 22:52:43 crc kubenswrapper[4791]: I1210 22:52:43.917950 4791 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/installer-9-crc" Dec 10 22:52:44 crc kubenswrapper[4791]: E1210 22:52:44.727267 4791 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" image="registry.redhat.io/redhat/certified-operator-index:v4.18" Dec 10 22:52:44 crc kubenswrapper[4791]: E1210 22:52:44.727444 4791 kuberuntime_manager.go:1274] "Unhandled Error" err="init container &Container{Name:extract-content,Image:registry.redhat.io/redhat/certified-operator-index:v4.18,Command:[/utilities/copy-content],Args:[--catalog.from=/configs --catalog.to=/extracted-catalog/catalog --cache.from=/tmp/cache --cache.to=/extracted-catalog/cache],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:utilities,ReadOnly:false,MountPath:/utilities,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:catalog-content,ReadOnly:false,MountPath:/extracted-catalog,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-8t7vg,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:Always,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000170000,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:FallbackToLogsOnError,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod certified-operators-9cbtj_openshift-marketplace(8d2ae7b5-156f-4230-b011-3b107066e18e): ErrImagePull: rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" logger="UnhandledError" Dec 10 22:52:44 crc kubenswrapper[4791]: E1210 22:52:44.728584 4791 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ErrImagePull: \"rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled\"" pod="openshift-marketplace/certified-operators-9cbtj" podUID="8d2ae7b5-156f-4230-b011-3b107066e18e" Dec 10 22:52:45 crc kubenswrapper[4791]: E1210 22:52:45.502703 4791 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ImagePullBackOff: \"Back-off pulling image \\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"\"" pod="openshift-marketplace/certified-operators-8x2rc" podUID="009efc35-5943-429f-98fd-3307a5f557f2" Dec 10 22:52:50 crc kubenswrapper[4791]: E1210 22:52:50.519918 4791 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" image="registry.redhat.io/redhat/redhat-marketplace-index:v4.18" Dec 10 22:52:50 crc kubenswrapper[4791]: E1210 22:52:50.520586 4791 kuberuntime_manager.go:1274] "Unhandled Error" err="init container &Container{Name:extract-content,Image:registry.redhat.io/redhat/redhat-marketplace-index:v4.18,Command:[/utilities/copy-content],Args:[--catalog.from=/configs --catalog.to=/extracted-catalog/catalog --cache.from=/tmp/cache --cache.to=/extracted-catalog/cache],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:utilities,ReadOnly:false,MountPath:/utilities,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:catalog-content,ReadOnly:false,MountPath:/extracted-catalog,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-mq8d4,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:Always,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000170000,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:FallbackToLogsOnError,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod redhat-marketplace-wkwbl_openshift-marketplace(396f55d6-c96d-4520-8165-862b5905a8c0): ErrImagePull: rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" logger="UnhandledError" Dec 10 22:52:50 crc kubenswrapper[4791]: E1210 22:52:50.522007 4791 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ErrImagePull: \"rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled\"" pod="openshift-marketplace/redhat-marketplace-wkwbl" podUID="396f55d6-c96d-4520-8165-862b5905a8c0" Dec 10 22:52:51 crc kubenswrapper[4791]: E1210 22:52:51.567135 4791 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" image="registry.redhat.io/redhat/community-operator-index:v4.18" Dec 10 22:52:51 crc kubenswrapper[4791]: E1210 22:52:51.567283 4791 kuberuntime_manager.go:1274] "Unhandled Error" err="init container &Container{Name:extract-content,Image:registry.redhat.io/redhat/community-operator-index:v4.18,Command:[/utilities/copy-content],Args:[--catalog.from=/configs --catalog.to=/extracted-catalog/catalog --cache.from=/tmp/cache --cache.to=/extracted-catalog/cache],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:utilities,ReadOnly:false,MountPath:/utilities,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:catalog-content,ReadOnly:false,MountPath:/extracted-catalog,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-qm95v,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:Always,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000170000,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:FallbackToLogsOnError,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod community-operators-gskzr_openshift-marketplace(08d50c37-7a4d-4731-beb8-fbc6d2e10bb0): ErrImagePull: rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" logger="UnhandledError" Dec 10 22:52:51 crc kubenswrapper[4791]: E1210 22:52:51.568435 4791 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ErrImagePull: \"rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled\"" pod="openshift-marketplace/community-operators-gskzr" podUID="08d50c37-7a4d-4731-beb8-fbc6d2e10bb0" Dec 10 22:52:51 crc kubenswrapper[4791]: E1210 22:52:51.574249 4791 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ImagePullBackOff: \"Back-off pulling image \\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"\"" pod="openshift-marketplace/redhat-marketplace-wkwbl" podUID="396f55d6-c96d-4520-8165-862b5905a8c0" Dec 10 22:52:51 crc kubenswrapper[4791]: E1210 22:52:51.574375 4791 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ImagePullBackOff: \"Back-off pulling image \\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"\"" pod="openshift-marketplace/certified-operators-9cbtj" podUID="8d2ae7b5-156f-4230-b011-3b107066e18e" Dec 10 22:52:51 crc kubenswrapper[4791]: E1210 22:52:51.575607 4791 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" image="registry.redhat.io/redhat/redhat-marketplace-index:v4.18" Dec 10 22:52:51 crc kubenswrapper[4791]: E1210 22:52:51.576303 4791 kuberuntime_manager.go:1274] "Unhandled Error" err="init container &Container{Name:extract-content,Image:registry.redhat.io/redhat/redhat-marketplace-index:v4.18,Command:[/utilities/copy-content],Args:[--catalog.from=/configs --catalog.to=/extracted-catalog/catalog --cache.from=/tmp/cache --cache.to=/extracted-catalog/cache],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:utilities,ReadOnly:false,MountPath:/utilities,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:catalog-content,ReadOnly:false,MountPath:/extracted-catalog,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-x4n6r,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:Always,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000170000,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:FallbackToLogsOnError,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod redhat-marketplace-gtcrt_openshift-marketplace(cd4347cf-ed66-480c-884f-b1d1c58ce143): ErrImagePull: rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" logger="UnhandledError" Dec 10 22:52:51 crc kubenswrapper[4791]: E1210 22:52:51.577532 4791 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ErrImagePull: \"rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled\"" pod="openshift-marketplace/redhat-marketplace-gtcrt" podUID="cd4347cf-ed66-480c-884f-b1d1c58ce143" Dec 10 22:52:51 crc kubenswrapper[4791]: E1210 22:52:51.603560 4791 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" image="registry.redhat.io/redhat/redhat-operator-index:v4.18" Dec 10 22:52:51 crc kubenswrapper[4791]: E1210 22:52:51.603750 4791 kuberuntime_manager.go:1274] "Unhandled Error" err="init container &Container{Name:extract-content,Image:registry.redhat.io/redhat/redhat-operator-index:v4.18,Command:[/utilities/copy-content],Args:[--catalog.from=/configs --catalog.to=/extracted-catalog/catalog --cache.from=/tmp/cache --cache.to=/extracted-catalog/cache],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:utilities,ReadOnly:false,MountPath:/utilities,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:catalog-content,ReadOnly:false,MountPath:/extracted-catalog,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-m5w5q,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:Always,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000170000,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:FallbackToLogsOnError,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod redhat-operators-p75kv_openshift-marketplace(abaf8cb8-8e36-4995-a926-8c4dadf4f56a): ErrImagePull: rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" logger="UnhandledError" Dec 10 22:52:51 crc kubenswrapper[4791]: E1210 22:52:51.604928 4791 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ErrImagePull: \"rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled\"" pod="openshift-marketplace/redhat-operators-p75kv" podUID="abaf8cb8-8e36-4995-a926-8c4dadf4f56a" Dec 10 22:52:51 crc kubenswrapper[4791]: E1210 22:52:51.616717 4791 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" image="registry.redhat.io/redhat/redhat-operator-index:v4.18" Dec 10 22:52:51 crc kubenswrapper[4791]: E1210 22:52:51.616893 4791 kuberuntime_manager.go:1274] "Unhandled Error" err="init container &Container{Name:extract-content,Image:registry.redhat.io/redhat/redhat-operator-index:v4.18,Command:[/utilities/copy-content],Args:[--catalog.from=/configs --catalog.to=/extracted-catalog/catalog --cache.from=/tmp/cache --cache.to=/extracted-catalog/cache],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:utilities,ReadOnly:false,MountPath:/utilities,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:catalog-content,ReadOnly:false,MountPath:/extracted-catalog,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-ck7dp,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:Always,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000170000,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:FallbackToLogsOnError,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod redhat-operators-tj8hn_openshift-marketplace(ab4c6e9a-3413-49c1-8f49-fb22c8366cb6): ErrImagePull: rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" logger="UnhandledError" Dec 10 22:52:51 crc kubenswrapper[4791]: E1210 22:52:51.618469 4791 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ErrImagePull: \"rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled\"" pod="openshift-marketplace/redhat-operators-tj8hn" podUID="ab4c6e9a-3413-49c1-8f49-fb22c8366cb6" Dec 10 22:52:52 crc kubenswrapper[4791]: I1210 22:52:52.002464 4791 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-apiserver/installer-9-crc"] Dec 10 22:52:52 crc kubenswrapper[4791]: I1210 22:52:52.009537 4791 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-apiserver/revision-pruner-9-crc"] Dec 10 22:52:52 crc kubenswrapper[4791]: W1210 22:52:52.018250 4791 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-pod602d7666_cce8_4158_a369_6921a5f6605b.slice/crio-0c2040d34c8278741356c2a2d1b4ae854b42d828abd18ee7d71a0d93ad7f311c WatchSource:0}: Error finding container 0c2040d34c8278741356c2a2d1b4ae854b42d828abd18ee7d71a0d93ad7f311c: Status 404 returned error can't find the container with id 0c2040d34c8278741356c2a2d1b4ae854b42d828abd18ee7d71a0d93ad7f311c Dec 10 22:52:52 crc kubenswrapper[4791]: W1210 22:52:52.020100 4791 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-pod31bd0efc_25ef_4a1f_b4b9_e4c5b14768ea.slice/crio-2fb0253edf06e6a7393d013814126be22811166fe5f712405de81225ba67c3a1 WatchSource:0}: Error finding container 2fb0253edf06e6a7393d013814126be22811166fe5f712405de81225ba67c3a1: Status 404 returned error can't find the container with id 2fb0253edf06e6a7393d013814126be22811166fe5f712405de81225ba67c3a1 Dec 10 22:52:52 crc kubenswrapper[4791]: I1210 22:52:52.076791 4791 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-multus/network-metrics-daemon-b6kf6"] Dec 10 22:52:52 crc kubenswrapper[4791]: W1210 22:52:52.086779 4791 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podb4527e71_bb58_4a39_a99c_fe614aa4c5f9.slice/crio-b39ec7dbc59b2875165c7c4a3890a2fd7fd5e3e95314a623e3ef1c995b4aee40 WatchSource:0}: Error finding container b39ec7dbc59b2875165c7c4a3890a2fd7fd5e3e95314a623e3ef1c995b4aee40: Status 404 returned error can't find the container with id b39ec7dbc59b2875165c7c4a3890a2fd7fd5e3e95314a623e3ef1c995b4aee40 Dec 10 22:52:52 crc kubenswrapper[4791]: I1210 22:52:52.545711 4791 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/network-metrics-daemon-b6kf6" event={"ID":"b4527e71-bb58-4a39-a99c-fe614aa4c5f9","Type":"ContainerStarted","Data":"e93a967842476573b5e570c3918f541c223b54f7e75fd38bb6c77771809bb3ed"} Dec 10 22:52:52 crc kubenswrapper[4791]: I1210 22:52:52.545761 4791 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/network-metrics-daemon-b6kf6" event={"ID":"b4527e71-bb58-4a39-a99c-fe614aa4c5f9","Type":"ContainerStarted","Data":"b39ec7dbc59b2875165c7c4a3890a2fd7fd5e3e95314a623e3ef1c995b4aee40"} Dec 10 22:52:52 crc kubenswrapper[4791]: I1210 22:52:52.547179 4791 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/revision-pruner-9-crc" event={"ID":"602d7666-cce8-4158-a369-6921a5f6605b","Type":"ContainerStarted","Data":"5a4a8ae2213f28f65702daaaaa6d7b15893ecc868a63f30b992354ea525cb33a"} Dec 10 22:52:52 crc kubenswrapper[4791]: I1210 22:52:52.547204 4791 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/revision-pruner-9-crc" event={"ID":"602d7666-cce8-4158-a369-6921a5f6605b","Type":"ContainerStarted","Data":"0c2040d34c8278741356c2a2d1b4ae854b42d828abd18ee7d71a0d93ad7f311c"} Dec 10 22:52:52 crc kubenswrapper[4791]: I1210 22:52:52.549088 4791 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/revision-pruner-8-crc" event={"ID":"76668a03-41c5-4892-8a62-ba917064ed0a","Type":"ContainerStarted","Data":"2d296a58c0980cd1c28e8e15131eccc29c6047b545463e0edfd389875bf6e1cd"} Dec 10 22:52:52 crc kubenswrapper[4791]: I1210 22:52:52.549137 4791 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/revision-pruner-8-crc" event={"ID":"76668a03-41c5-4892-8a62-ba917064ed0a","Type":"ContainerStarted","Data":"e3eaba4aa912ef24b545fba25d0bff214d8d9af851ec9df25d800be5bad0aeea"} Dec 10 22:52:52 crc kubenswrapper[4791]: I1210 22:52:52.552157 4791 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/installer-9-crc" event={"ID":"31bd0efc-25ef-4a1f-b4b9-e4c5b14768ea","Type":"ContainerStarted","Data":"b26dcf3a5303fefc9276862018a0f98295928a8d297d12f058c57e37ea06d831"} Dec 10 22:52:52 crc kubenswrapper[4791]: I1210 22:52:52.552194 4791 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/installer-9-crc" event={"ID":"31bd0efc-25ef-4a1f-b4b9-e4c5b14768ea","Type":"ContainerStarted","Data":"2fb0253edf06e6a7393d013814126be22811166fe5f712405de81225ba67c3a1"} Dec 10 22:52:52 crc kubenswrapper[4791]: E1210 22:52:52.558130 4791 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ImagePullBackOff: \"Back-off pulling image \\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"\"" pod="openshift-marketplace/redhat-marketplace-gtcrt" podUID="cd4347cf-ed66-480c-884f-b1d1c58ce143" Dec 10 22:52:52 crc kubenswrapper[4791]: E1210 22:52:52.559111 4791 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ImagePullBackOff: \"Back-off pulling image \\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"\"" pod="openshift-marketplace/community-operators-gskzr" podUID="08d50c37-7a4d-4731-beb8-fbc6d2e10bb0" Dec 10 22:52:52 crc kubenswrapper[4791]: E1210 22:52:52.560177 4791 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ImagePullBackOff: \"Back-off pulling image \\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"\"" pod="openshift-marketplace/redhat-operators-p75kv" podUID="abaf8cb8-8e36-4995-a926-8c4dadf4f56a" Dec 10 22:52:52 crc kubenswrapper[4791]: E1210 22:52:52.560224 4791 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ImagePullBackOff: \"Back-off pulling image \\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"\"" pod="openshift-marketplace/redhat-operators-tj8hn" podUID="ab4c6e9a-3413-49c1-8f49-fb22c8366cb6" Dec 10 22:52:52 crc kubenswrapper[4791]: I1210 22:52:52.568730 4791 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-kube-apiserver/revision-pruner-9-crc" podStartSLOduration=14.568710171 podStartE2EDuration="14.568710171s" podCreationTimestamp="2025-12-10 22:52:38 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-10 22:52:52.565663313 +0000 UTC m=+206.995280926" watchObservedRunningTime="2025-12-10 22:52:52.568710171 +0000 UTC m=+206.998327784" Dec 10 22:52:52 crc kubenswrapper[4791]: I1210 22:52:52.604167 4791 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-kube-apiserver/installer-9-crc" podStartSLOduration=9.604146204 podStartE2EDuration="9.604146204s" podCreationTimestamp="2025-12-10 22:52:43 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-10 22:52:52.601886014 +0000 UTC m=+207.031503627" watchObservedRunningTime="2025-12-10 22:52:52.604146204 +0000 UTC m=+207.033763817" Dec 10 22:52:52 crc kubenswrapper[4791]: I1210 22:52:52.620372 4791 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-kube-apiserver/revision-pruner-8-crc" podStartSLOduration=45.620352878 podStartE2EDuration="45.620352878s" podCreationTimestamp="2025-12-10 22:52:07 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-10 22:52:52.617704174 +0000 UTC m=+207.047321787" watchObservedRunningTime="2025-12-10 22:52:52.620352878 +0000 UTC m=+207.049970491" Dec 10 22:52:53 crc kubenswrapper[4791]: I1210 22:52:53.558845 4791 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/network-metrics-daemon-b6kf6" event={"ID":"b4527e71-bb58-4a39-a99c-fe614aa4c5f9","Type":"ContainerStarted","Data":"9bc3b27f33c0ba438bc06c35457b99d9e8e8e449e9278d11fb4b45408f0294f4"} Dec 10 22:52:53 crc kubenswrapper[4791]: I1210 22:52:53.560547 4791 generic.go:334] "Generic (PLEG): container finished" podID="602d7666-cce8-4158-a369-6921a5f6605b" containerID="5a4a8ae2213f28f65702daaaaa6d7b15893ecc868a63f30b992354ea525cb33a" exitCode=0 Dec 10 22:52:53 crc kubenswrapper[4791]: I1210 22:52:53.560615 4791 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/revision-pruner-9-crc" event={"ID":"602d7666-cce8-4158-a369-6921a5f6605b","Type":"ContainerDied","Data":"5a4a8ae2213f28f65702daaaaa6d7b15893ecc868a63f30b992354ea525cb33a"} Dec 10 22:52:53 crc kubenswrapper[4791]: I1210 22:52:53.562576 4791 generic.go:334] "Generic (PLEG): container finished" podID="76668a03-41c5-4892-8a62-ba917064ed0a" containerID="2d296a58c0980cd1c28e8e15131eccc29c6047b545463e0edfd389875bf6e1cd" exitCode=0 Dec 10 22:52:53 crc kubenswrapper[4791]: I1210 22:52:53.563127 4791 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/revision-pruner-8-crc" event={"ID":"76668a03-41c5-4892-8a62-ba917064ed0a","Type":"ContainerDied","Data":"2d296a58c0980cd1c28e8e15131eccc29c6047b545463e0edfd389875bf6e1cd"} Dec 10 22:52:53 crc kubenswrapper[4791]: I1210 22:52:53.574594 4791 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-multus/network-metrics-daemon-b6kf6" podStartSLOduration=182.574573984 podStartE2EDuration="3m2.574573984s" podCreationTimestamp="2025-12-10 22:49:51 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-10 22:52:53.574509502 +0000 UTC m=+208.004127135" watchObservedRunningTime="2025-12-10 22:52:53.574573984 +0000 UTC m=+208.004191607" Dec 10 22:52:54 crc kubenswrapper[4791]: I1210 22:52:54.842036 4791 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/revision-pruner-9-crc" Dec 10 22:52:54 crc kubenswrapper[4791]: I1210 22:52:54.847372 4791 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/revision-pruner-8-crc" Dec 10 22:52:55 crc kubenswrapper[4791]: I1210 22:52:55.001654 4791 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/602d7666-cce8-4158-a369-6921a5f6605b-kubelet-dir\") pod \"602d7666-cce8-4158-a369-6921a5f6605b\" (UID: \"602d7666-cce8-4158-a369-6921a5f6605b\") " Dec 10 22:52:55 crc kubenswrapper[4791]: I1210 22:52:55.001814 4791 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/602d7666-cce8-4158-a369-6921a5f6605b-kubelet-dir" (OuterVolumeSpecName: "kubelet-dir") pod "602d7666-cce8-4158-a369-6921a5f6605b" (UID: "602d7666-cce8-4158-a369-6921a5f6605b"). InnerVolumeSpecName "kubelet-dir". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 10 22:52:55 crc kubenswrapper[4791]: I1210 22:52:55.001869 4791 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/76668a03-41c5-4892-8a62-ba917064ed0a-kube-api-access\") pod \"76668a03-41c5-4892-8a62-ba917064ed0a\" (UID: \"76668a03-41c5-4892-8a62-ba917064ed0a\") " Dec 10 22:52:55 crc kubenswrapper[4791]: I1210 22:52:55.001914 4791 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/76668a03-41c5-4892-8a62-ba917064ed0a-kubelet-dir\") pod \"76668a03-41c5-4892-8a62-ba917064ed0a\" (UID: \"76668a03-41c5-4892-8a62-ba917064ed0a\") " Dec 10 22:52:55 crc kubenswrapper[4791]: I1210 22:52:55.001997 4791 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/602d7666-cce8-4158-a369-6921a5f6605b-kube-api-access\") pod \"602d7666-cce8-4158-a369-6921a5f6605b\" (UID: \"602d7666-cce8-4158-a369-6921a5f6605b\") " Dec 10 22:52:55 crc kubenswrapper[4791]: I1210 22:52:55.002220 4791 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/76668a03-41c5-4892-8a62-ba917064ed0a-kubelet-dir" (OuterVolumeSpecName: "kubelet-dir") pod "76668a03-41c5-4892-8a62-ba917064ed0a" (UID: "76668a03-41c5-4892-8a62-ba917064ed0a"). InnerVolumeSpecName "kubelet-dir". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 10 22:52:55 crc kubenswrapper[4791]: I1210 22:52:55.002613 4791 reconciler_common.go:293] "Volume detached for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/602d7666-cce8-4158-a369-6921a5f6605b-kubelet-dir\") on node \"crc\" DevicePath \"\"" Dec 10 22:52:55 crc kubenswrapper[4791]: I1210 22:52:55.002664 4791 reconciler_common.go:293] "Volume detached for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/76668a03-41c5-4892-8a62-ba917064ed0a-kubelet-dir\") on node \"crc\" DevicePath \"\"" Dec 10 22:52:55 crc kubenswrapper[4791]: I1210 22:52:55.008653 4791 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/76668a03-41c5-4892-8a62-ba917064ed0a-kube-api-access" (OuterVolumeSpecName: "kube-api-access") pod "76668a03-41c5-4892-8a62-ba917064ed0a" (UID: "76668a03-41c5-4892-8a62-ba917064ed0a"). InnerVolumeSpecName "kube-api-access". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 10 22:52:55 crc kubenswrapper[4791]: I1210 22:52:55.008954 4791 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/602d7666-cce8-4158-a369-6921a5f6605b-kube-api-access" (OuterVolumeSpecName: "kube-api-access") pod "602d7666-cce8-4158-a369-6921a5f6605b" (UID: "602d7666-cce8-4158-a369-6921a5f6605b"). InnerVolumeSpecName "kube-api-access". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 10 22:52:55 crc kubenswrapper[4791]: I1210 22:52:55.038038 4791 patch_prober.go:28] interesting pod/machine-config-daemon-5rb5l container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 10 22:52:55 crc kubenswrapper[4791]: I1210 22:52:55.038106 4791 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-5rb5l" podUID="ba35653c-6e06-4cee-a4d6-137764090d18" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 10 22:52:55 crc kubenswrapper[4791]: I1210 22:52:55.038175 4791 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-5rb5l" Dec 10 22:52:55 crc kubenswrapper[4791]: I1210 22:52:55.038986 4791 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"185eb2ff44221a514878283ec873773a6c9a9073dc64e296f982e60dee1338e4"} pod="openshift-machine-config-operator/machine-config-daemon-5rb5l" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Dec 10 22:52:55 crc kubenswrapper[4791]: I1210 22:52:55.039157 4791 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-5rb5l" podUID="ba35653c-6e06-4cee-a4d6-137764090d18" containerName="machine-config-daemon" containerID="cri-o://185eb2ff44221a514878283ec873773a6c9a9073dc64e296f982e60dee1338e4" gracePeriod=600 Dec 10 22:52:55 crc kubenswrapper[4791]: I1210 22:52:55.104082 4791 reconciler_common.go:293] "Volume detached for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/76668a03-41c5-4892-8a62-ba917064ed0a-kube-api-access\") on node \"crc\" DevicePath \"\"" Dec 10 22:52:55 crc kubenswrapper[4791]: I1210 22:52:55.104143 4791 reconciler_common.go:293] "Volume detached for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/602d7666-cce8-4158-a369-6921a5f6605b-kube-api-access\") on node \"crc\" DevicePath \"\"" Dec 10 22:52:55 crc kubenswrapper[4791]: I1210 22:52:55.578047 4791 generic.go:334] "Generic (PLEG): container finished" podID="ba35653c-6e06-4cee-a4d6-137764090d18" containerID="185eb2ff44221a514878283ec873773a6c9a9073dc64e296f982e60dee1338e4" exitCode=0 Dec 10 22:52:55 crc kubenswrapper[4791]: I1210 22:52:55.578627 4791 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-5rb5l" event={"ID":"ba35653c-6e06-4cee-a4d6-137764090d18","Type":"ContainerDied","Data":"185eb2ff44221a514878283ec873773a6c9a9073dc64e296f982e60dee1338e4"} Dec 10 22:52:55 crc kubenswrapper[4791]: I1210 22:52:55.578680 4791 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-5rb5l" event={"ID":"ba35653c-6e06-4cee-a4d6-137764090d18","Type":"ContainerStarted","Data":"05fce03d9a939befc9dac71cbc9750ba888a8f61825ef9c21bf847dd73aa10dc"} Dec 10 22:52:55 crc kubenswrapper[4791]: I1210 22:52:55.586454 4791 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/revision-pruner-9-crc" event={"ID":"602d7666-cce8-4158-a369-6921a5f6605b","Type":"ContainerDied","Data":"0c2040d34c8278741356c2a2d1b4ae854b42d828abd18ee7d71a0d93ad7f311c"} Dec 10 22:52:55 crc kubenswrapper[4791]: I1210 22:52:55.586523 4791 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="0c2040d34c8278741356c2a2d1b4ae854b42d828abd18ee7d71a0d93ad7f311c" Dec 10 22:52:55 crc kubenswrapper[4791]: I1210 22:52:55.586472 4791 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/revision-pruner-9-crc" Dec 10 22:52:55 crc kubenswrapper[4791]: I1210 22:52:55.592791 4791 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/revision-pruner-8-crc" event={"ID":"76668a03-41c5-4892-8a62-ba917064ed0a","Type":"ContainerDied","Data":"e3eaba4aa912ef24b545fba25d0bff214d8d9af851ec9df25d800be5bad0aeea"} Dec 10 22:52:55 crc kubenswrapper[4791]: I1210 22:52:55.592861 4791 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="e3eaba4aa912ef24b545fba25d0bff214d8d9af851ec9df25d800be5bad0aeea" Dec 10 22:52:55 crc kubenswrapper[4791]: I1210 22:52:55.592954 4791 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/revision-pruner-8-crc" Dec 10 22:52:57 crc kubenswrapper[4791]: E1210 22:52:57.521419 4791 cadvisor_stats_provider.go:516] "Partial failure issuing cadvisor.ContainerInfoV2" err="partial failures: [\"/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod776c00fb_1f2c_4ea2_ba50_62d4d24ffd72.slice/crio-fce20cd2c354361b367fb6b564db76f6644816a8c9b8b41ffa751b26fc372e91.scope\": RecentStats: unable to find data in memory cache]" Dec 10 22:52:57 crc kubenswrapper[4791]: I1210 22:52:57.608079 4791 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-q7lbx" event={"ID":"776c00fb-1f2c-4ea2-ba50-62d4d24ffd72","Type":"ContainerStarted","Data":"fce20cd2c354361b367fb6b564db76f6644816a8c9b8b41ffa751b26fc372e91"} Dec 10 22:52:58 crc kubenswrapper[4791]: I1210 22:52:58.615865 4791 generic.go:334] "Generic (PLEG): container finished" podID="776c00fb-1f2c-4ea2-ba50-62d4d24ffd72" containerID="fce20cd2c354361b367fb6b564db76f6644816a8c9b8b41ffa751b26fc372e91" exitCode=0 Dec 10 22:52:58 crc kubenswrapper[4791]: I1210 22:52:58.616134 4791 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-q7lbx" event={"ID":"776c00fb-1f2c-4ea2-ba50-62d4d24ffd72","Type":"ContainerDied","Data":"fce20cd2c354361b367fb6b564db76f6644816a8c9b8b41ffa751b26fc372e91"} Dec 10 22:52:58 crc kubenswrapper[4791]: I1210 22:52:58.620666 4791 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-8x2rc" event={"ID":"009efc35-5943-429f-98fd-3307a5f557f2","Type":"ContainerStarted","Data":"d82a5b0dd8195ec88a52c56bfdb70f420c467627cf24a4ed0e5c7aa05f60366b"} Dec 10 22:52:59 crc kubenswrapper[4791]: I1210 22:52:59.628575 4791 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-q7lbx" event={"ID":"776c00fb-1f2c-4ea2-ba50-62d4d24ffd72","Type":"ContainerStarted","Data":"de6a32dd963ea3e0aa9686ac2942243e283fc9082c53e3a945fe1c89f377aaa3"} Dec 10 22:52:59 crc kubenswrapper[4791]: I1210 22:52:59.631014 4791 generic.go:334] "Generic (PLEG): container finished" podID="009efc35-5943-429f-98fd-3307a5f557f2" containerID="d82a5b0dd8195ec88a52c56bfdb70f420c467627cf24a4ed0e5c7aa05f60366b" exitCode=0 Dec 10 22:52:59 crc kubenswrapper[4791]: I1210 22:52:59.631051 4791 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-8x2rc" event={"ID":"009efc35-5943-429f-98fd-3307a5f557f2","Type":"ContainerDied","Data":"d82a5b0dd8195ec88a52c56bfdb70f420c467627cf24a4ed0e5c7aa05f60366b"} Dec 10 22:52:59 crc kubenswrapper[4791]: I1210 22:52:59.649002 4791 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/community-operators-q7lbx" podStartSLOduration=3.740692649 podStartE2EDuration="1m0.648978982s" podCreationTimestamp="2025-12-10 22:51:59 +0000 UTC" firstStartedPulling="2025-12-10 22:52:02.154464602 +0000 UTC m=+156.584087745" lastFinishedPulling="2025-12-10 22:52:59.062756425 +0000 UTC m=+213.492374078" observedRunningTime="2025-12-10 22:52:59.648483064 +0000 UTC m=+214.078100677" watchObservedRunningTime="2025-12-10 22:52:59.648978982 +0000 UTC m=+214.078596605" Dec 10 22:53:00 crc kubenswrapper[4791]: I1210 22:53:00.472502 4791 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/community-operators-q7lbx" Dec 10 22:53:00 crc kubenswrapper[4791]: I1210 22:53:00.472819 4791 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/community-operators-q7lbx" Dec 10 22:53:00 crc kubenswrapper[4791]: I1210 22:53:00.637202 4791 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-8x2rc" event={"ID":"009efc35-5943-429f-98fd-3307a5f557f2","Type":"ContainerStarted","Data":"716a7c482ddb51c0626fcc775f7f3b405fbe182e7fd48c9aa126321e8e6267d9"} Dec 10 22:53:00 crc kubenswrapper[4791]: I1210 22:53:00.662967 4791 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/certified-operators-8x2rc" podStartSLOduration=2.189945016 podStartE2EDuration="1m1.662951172s" podCreationTimestamp="2025-12-10 22:51:59 +0000 UTC" firstStartedPulling="2025-12-10 22:52:00.723838937 +0000 UTC m=+155.153456540" lastFinishedPulling="2025-12-10 22:53:00.196845083 +0000 UTC m=+214.626462696" observedRunningTime="2025-12-10 22:53:00.660848878 +0000 UTC m=+215.090466491" watchObservedRunningTime="2025-12-10 22:53:00.662951172 +0000 UTC m=+215.092568785" Dec 10 22:53:01 crc kubenswrapper[4791]: I1210 22:53:01.570526 4791 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-marketplace/community-operators-q7lbx" podUID="776c00fb-1f2c-4ea2-ba50-62d4d24ffd72" containerName="registry-server" probeResult="failure" output=< Dec 10 22:53:01 crc kubenswrapper[4791]: timeout: failed to connect service ":50051" within 1s Dec 10 22:53:01 crc kubenswrapper[4791]: > Dec 10 22:53:03 crc kubenswrapper[4791]: I1210 22:53:03.656299 4791 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-9cbtj" event={"ID":"8d2ae7b5-156f-4230-b011-3b107066e18e","Type":"ContainerStarted","Data":"c37c40f11f486df939f9f98ab808fdd7a80801a8c9193c5e8b5a05b992e95f32"} Dec 10 22:53:04 crc kubenswrapper[4791]: I1210 22:53:04.675294 4791 generic.go:334] "Generic (PLEG): container finished" podID="8d2ae7b5-156f-4230-b011-3b107066e18e" containerID="c37c40f11f486df939f9f98ab808fdd7a80801a8c9193c5e8b5a05b992e95f32" exitCode=0 Dec 10 22:53:04 crc kubenswrapper[4791]: I1210 22:53:04.675679 4791 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-9cbtj" event={"ID":"8d2ae7b5-156f-4230-b011-3b107066e18e","Type":"ContainerDied","Data":"c37c40f11f486df939f9f98ab808fdd7a80801a8c9193c5e8b5a05b992e95f32"} Dec 10 22:53:04 crc kubenswrapper[4791]: I1210 22:53:04.679844 4791 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-p75kv" event={"ID":"abaf8cb8-8e36-4995-a926-8c4dadf4f56a","Type":"ContainerStarted","Data":"c980b3f6b2a7ba27e18ee03dd4ffba17d68607ffa6a2993b5bb3253c1ae6db81"} Dec 10 22:53:05 crc kubenswrapper[4791]: I1210 22:53:05.686211 4791 generic.go:334] "Generic (PLEG): container finished" podID="abaf8cb8-8e36-4995-a926-8c4dadf4f56a" containerID="c980b3f6b2a7ba27e18ee03dd4ffba17d68607ffa6a2993b5bb3253c1ae6db81" exitCode=0 Dec 10 22:53:05 crc kubenswrapper[4791]: I1210 22:53:05.686301 4791 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-p75kv" event={"ID":"abaf8cb8-8e36-4995-a926-8c4dadf4f56a","Type":"ContainerDied","Data":"c980b3f6b2a7ba27e18ee03dd4ffba17d68607ffa6a2993b5bb3253c1ae6db81"} Dec 10 22:53:05 crc kubenswrapper[4791]: I1210 22:53:05.690848 4791 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-9cbtj" event={"ID":"8d2ae7b5-156f-4230-b011-3b107066e18e","Type":"ContainerStarted","Data":"f5fc77d5681c70e5bb0dca3b478f21a6ceb3ab51d43c62eb5d0065b2d8dfd0ff"} Dec 10 22:53:05 crc kubenswrapper[4791]: I1210 22:53:05.694144 4791 generic.go:334] "Generic (PLEG): container finished" podID="cd4347cf-ed66-480c-884f-b1d1c58ce143" containerID="911d6f105637e8d675e67750a3e098ef643375687beb2a462c1cc2f6e78cf672" exitCode=0 Dec 10 22:53:05 crc kubenswrapper[4791]: I1210 22:53:05.694194 4791 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-gtcrt" event={"ID":"cd4347cf-ed66-480c-884f-b1d1c58ce143","Type":"ContainerDied","Data":"911d6f105637e8d675e67750a3e098ef643375687beb2a462c1cc2f6e78cf672"} Dec 10 22:53:05 crc kubenswrapper[4791]: I1210 22:53:05.733193 4791 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/certified-operators-9cbtj" podStartSLOduration=1.972330018 podStartE2EDuration="1m6.733172636s" podCreationTimestamp="2025-12-10 22:51:59 +0000 UTC" firstStartedPulling="2025-12-10 22:52:00.716762087 +0000 UTC m=+155.146379700" lastFinishedPulling="2025-12-10 22:53:05.477604705 +0000 UTC m=+219.907222318" observedRunningTime="2025-12-10 22:53:05.732679349 +0000 UTC m=+220.162296982" watchObservedRunningTime="2025-12-10 22:53:05.733172636 +0000 UTC m=+220.162790249" Dec 10 22:53:06 crc kubenswrapper[4791]: I1210 22:53:06.700978 4791 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-gtcrt" event={"ID":"cd4347cf-ed66-480c-884f-b1d1c58ce143","Type":"ContainerStarted","Data":"232a0ff4b1edef9a71ed32c1d896f15475de5455664168e1ab5ec35133102bdf"} Dec 10 22:53:06 crc kubenswrapper[4791]: I1210 22:53:06.703619 4791 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-p75kv" event={"ID":"abaf8cb8-8e36-4995-a926-8c4dadf4f56a","Type":"ContainerStarted","Data":"131d98c26ad370bb3bc579b99f8c655d394abc9d9c901b8292ab99ae1b55a558"} Dec 10 22:53:06 crc kubenswrapper[4791]: I1210 22:53:06.721848 4791 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-marketplace-gtcrt" podStartSLOduration=3.808866229 podStartE2EDuration="1m5.72183421s" podCreationTimestamp="2025-12-10 22:52:01 +0000 UTC" firstStartedPulling="2025-12-10 22:52:04.22712154 +0000 UTC m=+158.656739153" lastFinishedPulling="2025-12-10 22:53:06.140089521 +0000 UTC m=+220.569707134" observedRunningTime="2025-12-10 22:53:06.717742645 +0000 UTC m=+221.147360258" watchObservedRunningTime="2025-12-10 22:53:06.72183421 +0000 UTC m=+221.151451823" Dec 10 22:53:06 crc kubenswrapper[4791]: I1210 22:53:06.740904 4791 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-operators-p75kv" podStartSLOduration=2.808943408 podStartE2EDuration="1m4.740876484s" podCreationTimestamp="2025-12-10 22:52:02 +0000 UTC" firstStartedPulling="2025-12-10 22:52:04.238901293 +0000 UTC m=+158.668518906" lastFinishedPulling="2025-12-10 22:53:06.170834369 +0000 UTC m=+220.600451982" observedRunningTime="2025-12-10 22:53:06.738069604 +0000 UTC m=+221.167687227" watchObservedRunningTime="2025-12-10 22:53:06.740876484 +0000 UTC m=+221.170494097" Dec 10 22:53:07 crc kubenswrapper[4791]: I1210 22:53:07.710776 4791 generic.go:334] "Generic (PLEG): container finished" podID="08d50c37-7a4d-4731-beb8-fbc6d2e10bb0" containerID="58d691ba8d536c11c04f63e8e4cf0ff5bafc842c07ce006053f9244a17a81b0d" exitCode=0 Dec 10 22:53:07 crc kubenswrapper[4791]: I1210 22:53:07.710850 4791 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-gskzr" event={"ID":"08d50c37-7a4d-4731-beb8-fbc6d2e10bb0","Type":"ContainerDied","Data":"58d691ba8d536c11c04f63e8e4cf0ff5bafc842c07ce006053f9244a17a81b0d"} Dec 10 22:53:07 crc kubenswrapper[4791]: I1210 22:53:07.723789 4791 generic.go:334] "Generic (PLEG): container finished" podID="ab4c6e9a-3413-49c1-8f49-fb22c8366cb6" containerID="f62e716ba96d2eeb3fa3d37a632ba2bf20f383b5d039007cb3bb88040fa60ac5" exitCode=0 Dec 10 22:53:07 crc kubenswrapper[4791]: I1210 22:53:07.723833 4791 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-tj8hn" event={"ID":"ab4c6e9a-3413-49c1-8f49-fb22c8366cb6","Type":"ContainerDied","Data":"f62e716ba96d2eeb3fa3d37a632ba2bf20f383b5d039007cb3bb88040fa60ac5"} Dec 10 22:53:08 crc kubenswrapper[4791]: I1210 22:53:08.730643 4791 generic.go:334] "Generic (PLEG): container finished" podID="396f55d6-c96d-4520-8165-862b5905a8c0" containerID="5f188d26e1f99d267391af0a2204832707a2d80f1b5747cb124469c3048ea6d8" exitCode=0 Dec 10 22:53:08 crc kubenswrapper[4791]: I1210 22:53:08.730731 4791 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-wkwbl" event={"ID":"396f55d6-c96d-4520-8165-862b5905a8c0","Type":"ContainerDied","Data":"5f188d26e1f99d267391af0a2204832707a2d80f1b5747cb124469c3048ea6d8"} Dec 10 22:53:09 crc kubenswrapper[4791]: I1210 22:53:09.664925 4791 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/certified-operators-8x2rc" Dec 10 22:53:09 crc kubenswrapper[4791]: I1210 22:53:09.664990 4791 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/certified-operators-8x2rc" Dec 10 22:53:09 crc kubenswrapper[4791]: I1210 22:53:09.890103 4791 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/certified-operators-8x2rc" Dec 10 22:53:09 crc kubenswrapper[4791]: I1210 22:53:09.928913 4791 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/certified-operators-8x2rc" Dec 10 22:53:10 crc kubenswrapper[4791]: I1210 22:53:10.171391 4791 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/certified-operators-9cbtj" Dec 10 22:53:10 crc kubenswrapper[4791]: I1210 22:53:10.171453 4791 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/certified-operators-9cbtj" Dec 10 22:53:10 crc kubenswrapper[4791]: I1210 22:53:10.215840 4791 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/certified-operators-9cbtj" Dec 10 22:53:10 crc kubenswrapper[4791]: I1210 22:53:10.516476 4791 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/community-operators-q7lbx" Dec 10 22:53:10 crc kubenswrapper[4791]: I1210 22:53:10.552624 4791 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/community-operators-q7lbx" Dec 10 22:53:10 crc kubenswrapper[4791]: I1210 22:53:10.743378 4791 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-tj8hn" event={"ID":"ab4c6e9a-3413-49c1-8f49-fb22c8366cb6","Type":"ContainerStarted","Data":"ec98715ef0499713aacf43ea6bdd8b414c5f4931f812cc72b9d0f6235bd510cf"} Dec 10 22:53:10 crc kubenswrapper[4791]: I1210 22:53:10.747289 4791 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-gskzr" event={"ID":"08d50c37-7a4d-4731-beb8-fbc6d2e10bb0","Type":"ContainerStarted","Data":"b95dc6273b4eb68303d46abd4b7a567379de20b6c589ae921a68bc99c91fd48e"} Dec 10 22:53:10 crc kubenswrapper[4791]: I1210 22:53:10.749308 4791 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-wkwbl" event={"ID":"396f55d6-c96d-4520-8165-862b5905a8c0","Type":"ContainerStarted","Data":"bc477107547cc777dce54e21ac00704217e22d0bd8695c6a9bac537e5daa72f1"} Dec 10 22:53:10 crc kubenswrapper[4791]: I1210 22:53:10.767405 4791 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-operators-tj8hn" podStartSLOduration=2.762139565 podStartE2EDuration="1m8.767383706s" podCreationTimestamp="2025-12-10 22:52:02 +0000 UTC" firstStartedPulling="2025-12-10 22:52:04.235889298 +0000 UTC m=+158.665506911" lastFinishedPulling="2025-12-10 22:53:10.241133449 +0000 UTC m=+224.670751052" observedRunningTime="2025-12-10 22:53:10.763657714 +0000 UTC m=+225.193275327" watchObservedRunningTime="2025-12-10 22:53:10.767383706 +0000 UTC m=+225.197001319" Dec 10 22:53:10 crc kubenswrapper[4791]: I1210 22:53:10.786650 4791 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-marketplace-wkwbl" podStartSLOduration=3.684370748 podStartE2EDuration="1m9.786632516s" podCreationTimestamp="2025-12-10 22:52:01 +0000 UTC" firstStartedPulling="2025-12-10 22:52:04.246388294 +0000 UTC m=+158.676005907" lastFinishedPulling="2025-12-10 22:53:10.348650062 +0000 UTC m=+224.778267675" observedRunningTime="2025-12-10 22:53:10.78559438 +0000 UTC m=+225.215212003" watchObservedRunningTime="2025-12-10 22:53:10.786632516 +0000 UTC m=+225.216250129" Dec 10 22:53:10 crc kubenswrapper[4791]: I1210 22:53:10.797419 4791 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/certified-operators-9cbtj" Dec 10 22:53:10 crc kubenswrapper[4791]: I1210 22:53:10.805895 4791 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/community-operators-gskzr" podStartSLOduration=3.7048374109999997 podStartE2EDuration="1m11.805872737s" podCreationTimestamp="2025-12-10 22:51:59 +0000 UTC" firstStartedPulling="2025-12-10 22:52:02.153186096 +0000 UTC m=+156.582803709" lastFinishedPulling="2025-12-10 22:53:10.254221432 +0000 UTC m=+224.683839035" observedRunningTime="2025-12-10 22:53:10.800499987 +0000 UTC m=+225.230117600" watchObservedRunningTime="2025-12-10 22:53:10.805872737 +0000 UTC m=+225.235490350" Dec 10 22:53:11 crc kubenswrapper[4791]: I1210 22:53:11.797045 4791 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-9cbtj"] Dec 10 22:53:11 crc kubenswrapper[4791]: I1210 22:53:11.838221 4791 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-marketplace-wkwbl" Dec 10 22:53:11 crc kubenswrapper[4791]: I1210 22:53:11.838288 4791 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-marketplace-wkwbl" Dec 10 22:53:12 crc kubenswrapper[4791]: I1210 22:53:12.539104 4791 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-marketplace-gtcrt" Dec 10 22:53:12 crc kubenswrapper[4791]: I1210 22:53:12.539152 4791 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-marketplace-gtcrt" Dec 10 22:53:12 crc kubenswrapper[4791]: I1210 22:53:12.583208 4791 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-marketplace-gtcrt" Dec 10 22:53:12 crc kubenswrapper[4791]: I1210 22:53:12.760883 4791 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/certified-operators-9cbtj" podUID="8d2ae7b5-156f-4230-b011-3b107066e18e" containerName="registry-server" containerID="cri-o://f5fc77d5681c70e5bb0dca3b478f21a6ceb3ab51d43c62eb5d0065b2d8dfd0ff" gracePeriod=2 Dec 10 22:53:12 crc kubenswrapper[4791]: I1210 22:53:12.796662 4791 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-marketplace-gtcrt" Dec 10 22:53:12 crc kubenswrapper[4791]: I1210 22:53:12.879642 4791 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-marketplace/redhat-marketplace-wkwbl" podUID="396f55d6-c96d-4520-8165-862b5905a8c0" containerName="registry-server" probeResult="failure" output=< Dec 10 22:53:12 crc kubenswrapper[4791]: timeout: failed to connect service ":50051" within 1s Dec 10 22:53:12 crc kubenswrapper[4791]: > Dec 10 22:53:12 crc kubenswrapper[4791]: I1210 22:53:12.893758 4791 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-operators-tj8hn" Dec 10 22:53:12 crc kubenswrapper[4791]: I1210 22:53:12.893814 4791 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-operators-tj8hn" Dec 10 22:53:13 crc kubenswrapper[4791]: I1210 22:53:13.252434 4791 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-operators-p75kv" Dec 10 22:53:13 crc kubenswrapper[4791]: I1210 22:53:13.252486 4791 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-operators-p75kv" Dec 10 22:53:13 crc kubenswrapper[4791]: I1210 22:53:13.292142 4791 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-operators-p75kv" Dec 10 22:53:13 crc kubenswrapper[4791]: I1210 22:53:13.808837 4791 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-operators-p75kv" Dec 10 22:53:13 crc kubenswrapper[4791]: I1210 22:53:13.932154 4791 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-marketplace/redhat-operators-tj8hn" podUID="ab4c6e9a-3413-49c1-8f49-fb22c8366cb6" containerName="registry-server" probeResult="failure" output=< Dec 10 22:53:13 crc kubenswrapper[4791]: timeout: failed to connect service ":50051" within 1s Dec 10 22:53:13 crc kubenswrapper[4791]: > Dec 10 22:53:14 crc kubenswrapper[4791]: I1210 22:53:14.196120 4791 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-q7lbx"] Dec 10 22:53:14 crc kubenswrapper[4791]: I1210 22:53:14.196332 4791 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/community-operators-q7lbx" podUID="776c00fb-1f2c-4ea2-ba50-62d4d24ffd72" containerName="registry-server" containerID="cri-o://de6a32dd963ea3e0aa9686ac2942243e283fc9082c53e3a945fe1c89f377aaa3" gracePeriod=2 Dec 10 22:53:14 crc kubenswrapper[4791]: I1210 22:53:14.772276 4791 generic.go:334] "Generic (PLEG): container finished" podID="8d2ae7b5-156f-4230-b011-3b107066e18e" containerID="f5fc77d5681c70e5bb0dca3b478f21a6ceb3ab51d43c62eb5d0065b2d8dfd0ff" exitCode=0 Dec 10 22:53:14 crc kubenswrapper[4791]: I1210 22:53:14.772368 4791 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-9cbtj" event={"ID":"8d2ae7b5-156f-4230-b011-3b107066e18e","Type":"ContainerDied","Data":"f5fc77d5681c70e5bb0dca3b478f21a6ceb3ab51d43c62eb5d0065b2d8dfd0ff"} Dec 10 22:53:15 crc kubenswrapper[4791]: I1210 22:53:15.516806 4791 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-9cbtj" Dec 10 22:53:15 crc kubenswrapper[4791]: I1210 22:53:15.682791 4791 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-8t7vg\" (UniqueName: \"kubernetes.io/projected/8d2ae7b5-156f-4230-b011-3b107066e18e-kube-api-access-8t7vg\") pod \"8d2ae7b5-156f-4230-b011-3b107066e18e\" (UID: \"8d2ae7b5-156f-4230-b011-3b107066e18e\") " Dec 10 22:53:15 crc kubenswrapper[4791]: I1210 22:53:15.682904 4791 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/8d2ae7b5-156f-4230-b011-3b107066e18e-catalog-content\") pod \"8d2ae7b5-156f-4230-b011-3b107066e18e\" (UID: \"8d2ae7b5-156f-4230-b011-3b107066e18e\") " Dec 10 22:53:15 crc kubenswrapper[4791]: I1210 22:53:15.684452 4791 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/8d2ae7b5-156f-4230-b011-3b107066e18e-utilities\") pod \"8d2ae7b5-156f-4230-b011-3b107066e18e\" (UID: \"8d2ae7b5-156f-4230-b011-3b107066e18e\") " Dec 10 22:53:15 crc kubenswrapper[4791]: I1210 22:53:15.685402 4791 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/8d2ae7b5-156f-4230-b011-3b107066e18e-utilities" (OuterVolumeSpecName: "utilities") pod "8d2ae7b5-156f-4230-b011-3b107066e18e" (UID: "8d2ae7b5-156f-4230-b011-3b107066e18e"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 10 22:53:15 crc kubenswrapper[4791]: I1210 22:53:15.688265 4791 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/8d2ae7b5-156f-4230-b011-3b107066e18e-kube-api-access-8t7vg" (OuterVolumeSpecName: "kube-api-access-8t7vg") pod "8d2ae7b5-156f-4230-b011-3b107066e18e" (UID: "8d2ae7b5-156f-4230-b011-3b107066e18e"). InnerVolumeSpecName "kube-api-access-8t7vg". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 10 22:53:15 crc kubenswrapper[4791]: I1210 22:53:15.736307 4791 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/8d2ae7b5-156f-4230-b011-3b107066e18e-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "8d2ae7b5-156f-4230-b011-3b107066e18e" (UID: "8d2ae7b5-156f-4230-b011-3b107066e18e"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 10 22:53:15 crc kubenswrapper[4791]: I1210 22:53:15.779274 4791 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-9cbtj" event={"ID":"8d2ae7b5-156f-4230-b011-3b107066e18e","Type":"ContainerDied","Data":"de343f1792da5a284c1755dbd2680fe93e360afd4d0b606f505bf48ed94e0e07"} Dec 10 22:53:15 crc kubenswrapper[4791]: I1210 22:53:15.779329 4791 scope.go:117] "RemoveContainer" containerID="f5fc77d5681c70e5bb0dca3b478f21a6ceb3ab51d43c62eb5d0065b2d8dfd0ff" Dec 10 22:53:15 crc kubenswrapper[4791]: I1210 22:53:15.779508 4791 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-9cbtj" Dec 10 22:53:15 crc kubenswrapper[4791]: I1210 22:53:15.786140 4791 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/8d2ae7b5-156f-4230-b011-3b107066e18e-utilities\") on node \"crc\" DevicePath \"\"" Dec 10 22:53:15 crc kubenswrapper[4791]: I1210 22:53:15.786174 4791 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-8t7vg\" (UniqueName: \"kubernetes.io/projected/8d2ae7b5-156f-4230-b011-3b107066e18e-kube-api-access-8t7vg\") on node \"crc\" DevicePath \"\"" Dec 10 22:53:15 crc kubenswrapper[4791]: I1210 22:53:15.786186 4791 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/8d2ae7b5-156f-4230-b011-3b107066e18e-catalog-content\") on node \"crc\" DevicePath \"\"" Dec 10 22:53:15 crc kubenswrapper[4791]: I1210 22:53:15.797778 4791 scope.go:117] "RemoveContainer" containerID="c37c40f11f486df939f9f98ab808fdd7a80801a8c9193c5e8b5a05b992e95f32" Dec 10 22:53:15 crc kubenswrapper[4791]: I1210 22:53:15.813950 4791 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-9cbtj"] Dec 10 22:53:15 crc kubenswrapper[4791]: I1210 22:53:15.818545 4791 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/certified-operators-9cbtj"] Dec 10 22:53:15 crc kubenswrapper[4791]: I1210 22:53:15.820230 4791 scope.go:117] "RemoveContainer" containerID="3075d884ea3976218f0e1cf4a93febd7ec448d80696660570342ac1050864c00" Dec 10 22:53:15 crc kubenswrapper[4791]: I1210 22:53:15.892441 4791 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="8d2ae7b5-156f-4230-b011-3b107066e18e" path="/var/lib/kubelet/pods/8d2ae7b5-156f-4230-b011-3b107066e18e/volumes" Dec 10 22:53:15 crc kubenswrapper[4791]: I1210 22:53:15.996020 4791 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-gtcrt"] Dec 10 22:53:15 crc kubenswrapper[4791]: I1210 22:53:15.996281 4791 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-marketplace-gtcrt" podUID="cd4347cf-ed66-480c-884f-b1d1c58ce143" containerName="registry-server" containerID="cri-o://232a0ff4b1edef9a71ed32c1d896f15475de5455664168e1ab5ec35133102bdf" gracePeriod=2 Dec 10 22:53:16 crc kubenswrapper[4791]: I1210 22:53:16.788836 4791 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-q7lbx" event={"ID":"776c00fb-1f2c-4ea2-ba50-62d4d24ffd72","Type":"ContainerDied","Data":"de6a32dd963ea3e0aa9686ac2942243e283fc9082c53e3a945fe1c89f377aaa3"} Dec 10 22:53:16 crc kubenswrapper[4791]: I1210 22:53:16.788740 4791 generic.go:334] "Generic (PLEG): container finished" podID="776c00fb-1f2c-4ea2-ba50-62d4d24ffd72" containerID="de6a32dd963ea3e0aa9686ac2942243e283fc9082c53e3a945fe1c89f377aaa3" exitCode=0 Dec 10 22:53:16 crc kubenswrapper[4791]: I1210 22:53:16.792546 4791 generic.go:334] "Generic (PLEG): container finished" podID="cd4347cf-ed66-480c-884f-b1d1c58ce143" containerID="232a0ff4b1edef9a71ed32c1d896f15475de5455664168e1ab5ec35133102bdf" exitCode=0 Dec 10 22:53:16 crc kubenswrapper[4791]: I1210 22:53:16.792583 4791 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-gtcrt" event={"ID":"cd4347cf-ed66-480c-884f-b1d1c58ce143","Type":"ContainerDied","Data":"232a0ff4b1edef9a71ed32c1d896f15475de5455664168e1ab5ec35133102bdf"} Dec 10 22:53:17 crc kubenswrapper[4791]: I1210 22:53:17.146585 4791 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-q7lbx" Dec 10 22:53:17 crc kubenswrapper[4791]: I1210 22:53:17.203089 4791 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/776c00fb-1f2c-4ea2-ba50-62d4d24ffd72-catalog-content\") pod \"776c00fb-1f2c-4ea2-ba50-62d4d24ffd72\" (UID: \"776c00fb-1f2c-4ea2-ba50-62d4d24ffd72\") " Dec 10 22:53:17 crc kubenswrapper[4791]: I1210 22:53:17.203178 4791 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/776c00fb-1f2c-4ea2-ba50-62d4d24ffd72-utilities\") pod \"776c00fb-1f2c-4ea2-ba50-62d4d24ffd72\" (UID: \"776c00fb-1f2c-4ea2-ba50-62d4d24ffd72\") " Dec 10 22:53:17 crc kubenswrapper[4791]: I1210 22:53:17.203239 4791 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-xcs9j\" (UniqueName: \"kubernetes.io/projected/776c00fb-1f2c-4ea2-ba50-62d4d24ffd72-kube-api-access-xcs9j\") pod \"776c00fb-1f2c-4ea2-ba50-62d4d24ffd72\" (UID: \"776c00fb-1f2c-4ea2-ba50-62d4d24ffd72\") " Dec 10 22:53:17 crc kubenswrapper[4791]: I1210 22:53:17.204191 4791 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/776c00fb-1f2c-4ea2-ba50-62d4d24ffd72-utilities" (OuterVolumeSpecName: "utilities") pod "776c00fb-1f2c-4ea2-ba50-62d4d24ffd72" (UID: "776c00fb-1f2c-4ea2-ba50-62d4d24ffd72"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 10 22:53:17 crc kubenswrapper[4791]: I1210 22:53:17.207966 4791 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/776c00fb-1f2c-4ea2-ba50-62d4d24ffd72-kube-api-access-xcs9j" (OuterVolumeSpecName: "kube-api-access-xcs9j") pod "776c00fb-1f2c-4ea2-ba50-62d4d24ffd72" (UID: "776c00fb-1f2c-4ea2-ba50-62d4d24ffd72"). InnerVolumeSpecName "kube-api-access-xcs9j". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 10 22:53:17 crc kubenswrapper[4791]: I1210 22:53:17.252612 4791 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/776c00fb-1f2c-4ea2-ba50-62d4d24ffd72-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "776c00fb-1f2c-4ea2-ba50-62d4d24ffd72" (UID: "776c00fb-1f2c-4ea2-ba50-62d4d24ffd72"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 10 22:53:17 crc kubenswrapper[4791]: I1210 22:53:17.304638 4791 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/776c00fb-1f2c-4ea2-ba50-62d4d24ffd72-utilities\") on node \"crc\" DevicePath \"\"" Dec 10 22:53:17 crc kubenswrapper[4791]: I1210 22:53:17.304671 4791 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-xcs9j\" (UniqueName: \"kubernetes.io/projected/776c00fb-1f2c-4ea2-ba50-62d4d24ffd72-kube-api-access-xcs9j\") on node \"crc\" DevicePath \"\"" Dec 10 22:53:17 crc kubenswrapper[4791]: I1210 22:53:17.304682 4791 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/776c00fb-1f2c-4ea2-ba50-62d4d24ffd72-catalog-content\") on node \"crc\" DevicePath \"\"" Dec 10 22:53:17 crc kubenswrapper[4791]: I1210 22:53:17.800058 4791 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-q7lbx" event={"ID":"776c00fb-1f2c-4ea2-ba50-62d4d24ffd72","Type":"ContainerDied","Data":"b366355516b15a3cad414d8d67508f83152c420f1eb999866923386ef9535bce"} Dec 10 22:53:17 crc kubenswrapper[4791]: I1210 22:53:17.800115 4791 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-q7lbx" Dec 10 22:53:17 crc kubenswrapper[4791]: I1210 22:53:17.800124 4791 scope.go:117] "RemoveContainer" containerID="de6a32dd963ea3e0aa9686ac2942243e283fc9082c53e3a945fe1c89f377aaa3" Dec 10 22:53:17 crc kubenswrapper[4791]: I1210 22:53:17.818719 4791 scope.go:117] "RemoveContainer" containerID="fce20cd2c354361b367fb6b564db76f6644816a8c9b8b41ffa751b26fc372e91" Dec 10 22:53:17 crc kubenswrapper[4791]: I1210 22:53:17.826749 4791 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-q7lbx"] Dec 10 22:53:17 crc kubenswrapper[4791]: I1210 22:53:17.831101 4791 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/community-operators-q7lbx"] Dec 10 22:53:17 crc kubenswrapper[4791]: I1210 22:53:17.863258 4791 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-gtcrt" Dec 10 22:53:17 crc kubenswrapper[4791]: I1210 22:53:17.874503 4791 scope.go:117] "RemoveContainer" containerID="46c4394fb64814b43f7aa1d6e09152469f011e20b50f1603a8d3dfd24956021e" Dec 10 22:53:17 crc kubenswrapper[4791]: I1210 22:53:17.893571 4791 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="776c00fb-1f2c-4ea2-ba50-62d4d24ffd72" path="/var/lib/kubelet/pods/776c00fb-1f2c-4ea2-ba50-62d4d24ffd72/volumes" Dec 10 22:53:17 crc kubenswrapper[4791]: I1210 22:53:17.912571 4791 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/cd4347cf-ed66-480c-884f-b1d1c58ce143-utilities\") pod \"cd4347cf-ed66-480c-884f-b1d1c58ce143\" (UID: \"cd4347cf-ed66-480c-884f-b1d1c58ce143\") " Dec 10 22:53:17 crc kubenswrapper[4791]: I1210 22:53:17.912622 4791 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/cd4347cf-ed66-480c-884f-b1d1c58ce143-catalog-content\") pod \"cd4347cf-ed66-480c-884f-b1d1c58ce143\" (UID: \"cd4347cf-ed66-480c-884f-b1d1c58ce143\") " Dec 10 22:53:17 crc kubenswrapper[4791]: I1210 22:53:17.912701 4791 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-x4n6r\" (UniqueName: \"kubernetes.io/projected/cd4347cf-ed66-480c-884f-b1d1c58ce143-kube-api-access-x4n6r\") pod \"cd4347cf-ed66-480c-884f-b1d1c58ce143\" (UID: \"cd4347cf-ed66-480c-884f-b1d1c58ce143\") " Dec 10 22:53:17 crc kubenswrapper[4791]: I1210 22:53:17.913804 4791 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/cd4347cf-ed66-480c-884f-b1d1c58ce143-utilities" (OuterVolumeSpecName: "utilities") pod "cd4347cf-ed66-480c-884f-b1d1c58ce143" (UID: "cd4347cf-ed66-480c-884f-b1d1c58ce143"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 10 22:53:17 crc kubenswrapper[4791]: I1210 22:53:17.914587 4791 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/cd4347cf-ed66-480c-884f-b1d1c58ce143-utilities\") on node \"crc\" DevicePath \"\"" Dec 10 22:53:17 crc kubenswrapper[4791]: I1210 22:53:17.915977 4791 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/cd4347cf-ed66-480c-884f-b1d1c58ce143-kube-api-access-x4n6r" (OuterVolumeSpecName: "kube-api-access-x4n6r") pod "cd4347cf-ed66-480c-884f-b1d1c58ce143" (UID: "cd4347cf-ed66-480c-884f-b1d1c58ce143"). InnerVolumeSpecName "kube-api-access-x4n6r". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 10 22:53:17 crc kubenswrapper[4791]: I1210 22:53:17.941077 4791 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/cd4347cf-ed66-480c-884f-b1d1c58ce143-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "cd4347cf-ed66-480c-884f-b1d1c58ce143" (UID: "cd4347cf-ed66-480c-884f-b1d1c58ce143"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 10 22:53:18 crc kubenswrapper[4791]: I1210 22:53:18.015432 4791 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-x4n6r\" (UniqueName: \"kubernetes.io/projected/cd4347cf-ed66-480c-884f-b1d1c58ce143-kube-api-access-x4n6r\") on node \"crc\" DevicePath \"\"" Dec 10 22:53:18 crc kubenswrapper[4791]: I1210 22:53:18.015461 4791 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/cd4347cf-ed66-480c-884f-b1d1c58ce143-catalog-content\") on node \"crc\" DevicePath \"\"" Dec 10 22:53:18 crc kubenswrapper[4791]: I1210 22:53:18.399532 4791 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-p75kv"] Dec 10 22:53:18 crc kubenswrapper[4791]: I1210 22:53:18.399760 4791 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-operators-p75kv" podUID="abaf8cb8-8e36-4995-a926-8c4dadf4f56a" containerName="registry-server" containerID="cri-o://131d98c26ad370bb3bc579b99f8c655d394abc9d9c901b8292ab99ae1b55a558" gracePeriod=2 Dec 10 22:53:18 crc kubenswrapper[4791]: I1210 22:53:18.807917 4791 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-gtcrt" event={"ID":"cd4347cf-ed66-480c-884f-b1d1c58ce143","Type":"ContainerDied","Data":"138c31a05fe2a71cecac04b09a5f86e0a18c4dab110562b6c6278d5f8e4bb8fb"} Dec 10 22:53:18 crc kubenswrapper[4791]: I1210 22:53:18.808211 4791 scope.go:117] "RemoveContainer" containerID="232a0ff4b1edef9a71ed32c1d896f15475de5455664168e1ab5ec35133102bdf" Dec 10 22:53:18 crc kubenswrapper[4791]: I1210 22:53:18.807962 4791 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-gtcrt" Dec 10 22:53:18 crc kubenswrapper[4791]: I1210 22:53:18.825289 4791 scope.go:117] "RemoveContainer" containerID="911d6f105637e8d675e67750a3e098ef643375687beb2a462c1cc2f6e78cf672" Dec 10 22:53:18 crc kubenswrapper[4791]: I1210 22:53:18.844772 4791 scope.go:117] "RemoveContainer" containerID="2da96993499b7ba23ccd6ddc9e5aa1fb62c1c9aa98a3d4089b3afa5e90f291d0" Dec 10 22:53:18 crc kubenswrapper[4791]: I1210 22:53:18.849179 4791 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-gtcrt"] Dec 10 22:53:18 crc kubenswrapper[4791]: I1210 22:53:18.861989 4791 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-marketplace-gtcrt"] Dec 10 22:53:19 crc kubenswrapper[4791]: I1210 22:53:19.818949 4791 generic.go:334] "Generic (PLEG): container finished" podID="abaf8cb8-8e36-4995-a926-8c4dadf4f56a" containerID="131d98c26ad370bb3bc579b99f8c655d394abc9d9c901b8292ab99ae1b55a558" exitCode=0 Dec 10 22:53:19 crc kubenswrapper[4791]: I1210 22:53:19.818995 4791 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-p75kv" event={"ID":"abaf8cb8-8e36-4995-a926-8c4dadf4f56a","Type":"ContainerDied","Data":"131d98c26ad370bb3bc579b99f8c655d394abc9d9c901b8292ab99ae1b55a558"} Dec 10 22:53:19 crc kubenswrapper[4791]: I1210 22:53:19.896688 4791 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="cd4347cf-ed66-480c-884f-b1d1c58ce143" path="/var/lib/kubelet/pods/cd4347cf-ed66-480c-884f-b1d1c58ce143/volumes" Dec 10 22:53:19 crc kubenswrapper[4791]: I1210 22:53:19.910074 4791 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/community-operators-gskzr" Dec 10 22:53:19 crc kubenswrapper[4791]: I1210 22:53:19.910154 4791 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/community-operators-gskzr" Dec 10 22:53:19 crc kubenswrapper[4791]: I1210 22:53:19.958085 4791 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/community-operators-gskzr" Dec 10 22:53:20 crc kubenswrapper[4791]: I1210 22:53:20.896297 4791 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/community-operators-gskzr" Dec 10 22:53:21 crc kubenswrapper[4791]: I1210 22:53:21.324229 4791 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-p75kv" Dec 10 22:53:21 crc kubenswrapper[4791]: I1210 22:53:21.453666 4791 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-m5w5q\" (UniqueName: \"kubernetes.io/projected/abaf8cb8-8e36-4995-a926-8c4dadf4f56a-kube-api-access-m5w5q\") pod \"abaf8cb8-8e36-4995-a926-8c4dadf4f56a\" (UID: \"abaf8cb8-8e36-4995-a926-8c4dadf4f56a\") " Dec 10 22:53:21 crc kubenswrapper[4791]: I1210 22:53:21.453805 4791 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/abaf8cb8-8e36-4995-a926-8c4dadf4f56a-utilities\") pod \"abaf8cb8-8e36-4995-a926-8c4dadf4f56a\" (UID: \"abaf8cb8-8e36-4995-a926-8c4dadf4f56a\") " Dec 10 22:53:21 crc kubenswrapper[4791]: I1210 22:53:21.453965 4791 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/abaf8cb8-8e36-4995-a926-8c4dadf4f56a-catalog-content\") pod \"abaf8cb8-8e36-4995-a926-8c4dadf4f56a\" (UID: \"abaf8cb8-8e36-4995-a926-8c4dadf4f56a\") " Dec 10 22:53:21 crc kubenswrapper[4791]: I1210 22:53:21.454723 4791 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/abaf8cb8-8e36-4995-a926-8c4dadf4f56a-utilities" (OuterVolumeSpecName: "utilities") pod "abaf8cb8-8e36-4995-a926-8c4dadf4f56a" (UID: "abaf8cb8-8e36-4995-a926-8c4dadf4f56a"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 10 22:53:21 crc kubenswrapper[4791]: I1210 22:53:21.471178 4791 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/abaf8cb8-8e36-4995-a926-8c4dadf4f56a-kube-api-access-m5w5q" (OuterVolumeSpecName: "kube-api-access-m5w5q") pod "abaf8cb8-8e36-4995-a926-8c4dadf4f56a" (UID: "abaf8cb8-8e36-4995-a926-8c4dadf4f56a"). InnerVolumeSpecName "kube-api-access-m5w5q". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 10 22:53:21 crc kubenswrapper[4791]: I1210 22:53:21.556573 4791 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-m5w5q\" (UniqueName: \"kubernetes.io/projected/abaf8cb8-8e36-4995-a926-8c4dadf4f56a-kube-api-access-m5w5q\") on node \"crc\" DevicePath \"\"" Dec 10 22:53:21 crc kubenswrapper[4791]: I1210 22:53:21.556615 4791 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/abaf8cb8-8e36-4995-a926-8c4dadf4f56a-utilities\") on node \"crc\" DevicePath \"\"" Dec 10 22:53:21 crc kubenswrapper[4791]: I1210 22:53:21.557119 4791 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/abaf8cb8-8e36-4995-a926-8c4dadf4f56a-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "abaf8cb8-8e36-4995-a926-8c4dadf4f56a" (UID: "abaf8cb8-8e36-4995-a926-8c4dadf4f56a"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 10 22:53:21 crc kubenswrapper[4791]: I1210 22:53:21.657516 4791 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/abaf8cb8-8e36-4995-a926-8c4dadf4f56a-catalog-content\") on node \"crc\" DevicePath \"\"" Dec 10 22:53:21 crc kubenswrapper[4791]: I1210 22:53:21.830730 4791 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-p75kv" event={"ID":"abaf8cb8-8e36-4995-a926-8c4dadf4f56a","Type":"ContainerDied","Data":"94c90e6557b125e09be0d59c19374ae4cd27dc735ba804210c0856e9055229fc"} Dec 10 22:53:21 crc kubenswrapper[4791]: I1210 22:53:21.830792 4791 scope.go:117] "RemoveContainer" containerID="131d98c26ad370bb3bc579b99f8c655d394abc9d9c901b8292ab99ae1b55a558" Dec 10 22:53:21 crc kubenswrapper[4791]: I1210 22:53:21.830836 4791 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-p75kv" Dec 10 22:53:21 crc kubenswrapper[4791]: I1210 22:53:21.843727 4791 scope.go:117] "RemoveContainer" containerID="c980b3f6b2a7ba27e18ee03dd4ffba17d68607ffa6a2993b5bb3253c1ae6db81" Dec 10 22:53:21 crc kubenswrapper[4791]: I1210 22:53:21.870328 4791 scope.go:117] "RemoveContainer" containerID="35438f70b7ffc3d38c8e35729e9f71d94d2ca6cbfdf2d54bbc57add550ee4eb7" Dec 10 22:53:21 crc kubenswrapper[4791]: I1210 22:53:21.875078 4791 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-p75kv"] Dec 10 22:53:21 crc kubenswrapper[4791]: I1210 22:53:21.878726 4791 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-operators-p75kv"] Dec 10 22:53:21 crc kubenswrapper[4791]: I1210 22:53:21.892715 4791 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="abaf8cb8-8e36-4995-a926-8c4dadf4f56a" path="/var/lib/kubelet/pods/abaf8cb8-8e36-4995-a926-8c4dadf4f56a/volumes" Dec 10 22:53:21 crc kubenswrapper[4791]: I1210 22:53:21.895028 4791 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-marketplace-wkwbl" Dec 10 22:53:21 crc kubenswrapper[4791]: I1210 22:53:21.942995 4791 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-marketplace-wkwbl" Dec 10 22:53:22 crc kubenswrapper[4791]: I1210 22:53:22.858867 4791 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-authentication/oauth-openshift-558db77b4-ht5t7"] Dec 10 22:53:22 crc kubenswrapper[4791]: I1210 22:53:22.962189 4791 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-operators-tj8hn" Dec 10 22:53:23 crc kubenswrapper[4791]: I1210 22:53:23.016870 4791 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-operators-tj8hn" Dec 10 22:53:30 crc kubenswrapper[4791]: I1210 22:53:30.002177 4791 kubelet.go:2421] "SyncLoop ADD" source="file" pods=["openshift-kube-apiserver/kube-apiserver-startup-monitor-crc"] Dec 10 22:53:30 crc kubenswrapper[4791]: E1210 22:53:30.002883 4791 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="8d2ae7b5-156f-4230-b011-3b107066e18e" containerName="extract-content" Dec 10 22:53:30 crc kubenswrapper[4791]: I1210 22:53:30.002896 4791 state_mem.go:107] "Deleted CPUSet assignment" podUID="8d2ae7b5-156f-4230-b011-3b107066e18e" containerName="extract-content" Dec 10 22:53:30 crc kubenswrapper[4791]: E1210 22:53:30.002904 4791 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="8d2ae7b5-156f-4230-b011-3b107066e18e" containerName="registry-server" Dec 10 22:53:30 crc kubenswrapper[4791]: I1210 22:53:30.002910 4791 state_mem.go:107] "Deleted CPUSet assignment" podUID="8d2ae7b5-156f-4230-b011-3b107066e18e" containerName="registry-server" Dec 10 22:53:30 crc kubenswrapper[4791]: E1210 22:53:30.002918 4791 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="abaf8cb8-8e36-4995-a926-8c4dadf4f56a" containerName="extract-utilities" Dec 10 22:53:30 crc kubenswrapper[4791]: I1210 22:53:30.002923 4791 state_mem.go:107] "Deleted CPUSet assignment" podUID="abaf8cb8-8e36-4995-a926-8c4dadf4f56a" containerName="extract-utilities" Dec 10 22:53:30 crc kubenswrapper[4791]: E1210 22:53:30.002936 4791 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="8d2ae7b5-156f-4230-b011-3b107066e18e" containerName="extract-utilities" Dec 10 22:53:30 crc kubenswrapper[4791]: I1210 22:53:30.002941 4791 state_mem.go:107] "Deleted CPUSet assignment" podUID="8d2ae7b5-156f-4230-b011-3b107066e18e" containerName="extract-utilities" Dec 10 22:53:30 crc kubenswrapper[4791]: E1210 22:53:30.002950 4791 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="cd4347cf-ed66-480c-884f-b1d1c58ce143" containerName="extract-content" Dec 10 22:53:30 crc kubenswrapper[4791]: I1210 22:53:30.002956 4791 state_mem.go:107] "Deleted CPUSet assignment" podUID="cd4347cf-ed66-480c-884f-b1d1c58ce143" containerName="extract-content" Dec 10 22:53:30 crc kubenswrapper[4791]: E1210 22:53:30.002964 4791 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="cd4347cf-ed66-480c-884f-b1d1c58ce143" containerName="extract-utilities" Dec 10 22:53:30 crc kubenswrapper[4791]: I1210 22:53:30.002970 4791 state_mem.go:107] "Deleted CPUSet assignment" podUID="cd4347cf-ed66-480c-884f-b1d1c58ce143" containerName="extract-utilities" Dec 10 22:53:30 crc kubenswrapper[4791]: E1210 22:53:30.002978 4791 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="776c00fb-1f2c-4ea2-ba50-62d4d24ffd72" containerName="extract-utilities" Dec 10 22:53:30 crc kubenswrapper[4791]: I1210 22:53:30.002983 4791 state_mem.go:107] "Deleted CPUSet assignment" podUID="776c00fb-1f2c-4ea2-ba50-62d4d24ffd72" containerName="extract-utilities" Dec 10 22:53:30 crc kubenswrapper[4791]: E1210 22:53:30.002990 4791 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="76668a03-41c5-4892-8a62-ba917064ed0a" containerName="pruner" Dec 10 22:53:30 crc kubenswrapper[4791]: I1210 22:53:30.002996 4791 state_mem.go:107] "Deleted CPUSet assignment" podUID="76668a03-41c5-4892-8a62-ba917064ed0a" containerName="pruner" Dec 10 22:53:30 crc kubenswrapper[4791]: E1210 22:53:30.003004 4791 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="776c00fb-1f2c-4ea2-ba50-62d4d24ffd72" containerName="extract-content" Dec 10 22:53:30 crc kubenswrapper[4791]: I1210 22:53:30.003010 4791 state_mem.go:107] "Deleted CPUSet assignment" podUID="776c00fb-1f2c-4ea2-ba50-62d4d24ffd72" containerName="extract-content" Dec 10 22:53:30 crc kubenswrapper[4791]: E1210 22:53:30.003018 4791 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="602d7666-cce8-4158-a369-6921a5f6605b" containerName="pruner" Dec 10 22:53:30 crc kubenswrapper[4791]: I1210 22:53:30.003023 4791 state_mem.go:107] "Deleted CPUSet assignment" podUID="602d7666-cce8-4158-a369-6921a5f6605b" containerName="pruner" Dec 10 22:53:30 crc kubenswrapper[4791]: E1210 22:53:30.003030 4791 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="abaf8cb8-8e36-4995-a926-8c4dadf4f56a" containerName="extract-content" Dec 10 22:53:30 crc kubenswrapper[4791]: I1210 22:53:30.003035 4791 state_mem.go:107] "Deleted CPUSet assignment" podUID="abaf8cb8-8e36-4995-a926-8c4dadf4f56a" containerName="extract-content" Dec 10 22:53:30 crc kubenswrapper[4791]: E1210 22:53:30.003043 4791 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="cd4347cf-ed66-480c-884f-b1d1c58ce143" containerName="registry-server" Dec 10 22:53:30 crc kubenswrapper[4791]: I1210 22:53:30.003049 4791 state_mem.go:107] "Deleted CPUSet assignment" podUID="cd4347cf-ed66-480c-884f-b1d1c58ce143" containerName="registry-server" Dec 10 22:53:30 crc kubenswrapper[4791]: E1210 22:53:30.003059 4791 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="776c00fb-1f2c-4ea2-ba50-62d4d24ffd72" containerName="registry-server" Dec 10 22:53:30 crc kubenswrapper[4791]: I1210 22:53:30.003065 4791 state_mem.go:107] "Deleted CPUSet assignment" podUID="776c00fb-1f2c-4ea2-ba50-62d4d24ffd72" containerName="registry-server" Dec 10 22:53:30 crc kubenswrapper[4791]: E1210 22:53:30.003074 4791 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="abaf8cb8-8e36-4995-a926-8c4dadf4f56a" containerName="registry-server" Dec 10 22:53:30 crc kubenswrapper[4791]: I1210 22:53:30.003079 4791 state_mem.go:107] "Deleted CPUSet assignment" podUID="abaf8cb8-8e36-4995-a926-8c4dadf4f56a" containerName="registry-server" Dec 10 22:53:30 crc kubenswrapper[4791]: I1210 22:53:30.003168 4791 memory_manager.go:354] "RemoveStaleState removing state" podUID="cd4347cf-ed66-480c-884f-b1d1c58ce143" containerName="registry-server" Dec 10 22:53:30 crc kubenswrapper[4791]: I1210 22:53:30.003182 4791 memory_manager.go:354] "RemoveStaleState removing state" podUID="602d7666-cce8-4158-a369-6921a5f6605b" containerName="pruner" Dec 10 22:53:30 crc kubenswrapper[4791]: I1210 22:53:30.003189 4791 memory_manager.go:354] "RemoveStaleState removing state" podUID="76668a03-41c5-4892-8a62-ba917064ed0a" containerName="pruner" Dec 10 22:53:30 crc kubenswrapper[4791]: I1210 22:53:30.003198 4791 memory_manager.go:354] "RemoveStaleState removing state" podUID="abaf8cb8-8e36-4995-a926-8c4dadf4f56a" containerName="registry-server" Dec 10 22:53:30 crc kubenswrapper[4791]: I1210 22:53:30.003208 4791 memory_manager.go:354] "RemoveStaleState removing state" podUID="776c00fb-1f2c-4ea2-ba50-62d4d24ffd72" containerName="registry-server" Dec 10 22:53:30 crc kubenswrapper[4791]: I1210 22:53:30.003217 4791 memory_manager.go:354] "RemoveStaleState removing state" podUID="8d2ae7b5-156f-4230-b011-3b107066e18e" containerName="registry-server" Dec 10 22:53:30 crc kubenswrapper[4791]: I1210 22:53:30.003576 4791 kubelet.go:2431] "SyncLoop REMOVE" source="file" pods=["openshift-kube-apiserver/kube-apiserver-crc"] Dec 10 22:53:30 crc kubenswrapper[4791]: I1210 22:53:30.003802 4791 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Dec 10 22:53:30 crc kubenswrapper[4791]: I1210 22:53:30.003847 4791 kubelet.go:2421] "SyncLoop ADD" source="file" pods=["openshift-kube-apiserver/kube-apiserver-crc"] Dec 10 22:53:30 crc kubenswrapper[4791]: I1210 22:53:30.003954 4791 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver" containerID="cri-o://acafe12186df99d1059a31260de0eafce35ff7771bcb8f5c6753b72665d6f60b" gracePeriod=15 Dec 10 22:53:30 crc kubenswrapper[4791]: I1210 22:53:30.004037 4791 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-check-endpoints" containerID="cri-o://1b2e11ff2bf8789fa1d0d23ae648b3e6ef653106e5bc4bbb5dd6447f7cd6f05d" gracePeriod=15 Dec 10 22:53:30 crc kubenswrapper[4791]: I1210 22:53:30.004037 4791 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-cert-regeneration-controller" containerID="cri-o://ea4c346507862a8c12756a3061d6e25e788a8d79f4621b62f81f4322185c82a8" gracePeriod=15 Dec 10 22:53:30 crc kubenswrapper[4791]: I1210 22:53:30.004073 4791 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-cert-syncer" containerID="cri-o://2e6e324ec6f4f7bb69fdfab4ea3ac1aa55e480dd8e2a8984b0fe5b4c9f66636e" gracePeriod=15 Dec 10 22:53:30 crc kubenswrapper[4791]: I1210 22:53:30.004037 4791 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-insecure-readyz" containerID="cri-o://0c4d036ab269f62561fe7ef1586012cb3a16e3583a08b9488a0616ec00c8e0ae" gracePeriod=15 Dec 10 22:53:30 crc kubenswrapper[4791]: E1210 22:53:30.004606 4791 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="setup" Dec 10 22:53:30 crc kubenswrapper[4791]: I1210 22:53:30.004622 4791 state_mem.go:107] "Deleted CPUSet assignment" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="setup" Dec 10 22:53:30 crc kubenswrapper[4791]: E1210 22:53:30.004633 4791 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-insecure-readyz" Dec 10 22:53:30 crc kubenswrapper[4791]: I1210 22:53:30.004639 4791 state_mem.go:107] "Deleted CPUSet assignment" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-insecure-readyz" Dec 10 22:53:30 crc kubenswrapper[4791]: E1210 22:53:30.004647 4791 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-cert-syncer" Dec 10 22:53:30 crc kubenswrapper[4791]: I1210 22:53:30.004653 4791 state_mem.go:107] "Deleted CPUSet assignment" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-cert-syncer" Dec 10 22:53:30 crc kubenswrapper[4791]: E1210 22:53:30.004666 4791 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-check-endpoints" Dec 10 22:53:30 crc kubenswrapper[4791]: I1210 22:53:30.004691 4791 state_mem.go:107] "Deleted CPUSet assignment" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-check-endpoints" Dec 10 22:53:30 crc kubenswrapper[4791]: E1210 22:53:30.004700 4791 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-cert-regeneration-controller" Dec 10 22:53:30 crc kubenswrapper[4791]: I1210 22:53:30.004705 4791 state_mem.go:107] "Deleted CPUSet assignment" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-cert-regeneration-controller" Dec 10 22:53:30 crc kubenswrapper[4791]: E1210 22:53:30.004713 4791 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-check-endpoints" Dec 10 22:53:30 crc kubenswrapper[4791]: I1210 22:53:30.004718 4791 state_mem.go:107] "Deleted CPUSet assignment" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-check-endpoints" Dec 10 22:53:30 crc kubenswrapper[4791]: E1210 22:53:30.005220 4791 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver" Dec 10 22:53:30 crc kubenswrapper[4791]: I1210 22:53:30.005236 4791 state_mem.go:107] "Deleted CPUSet assignment" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver" Dec 10 22:53:30 crc kubenswrapper[4791]: I1210 22:53:30.005368 4791 memory_manager.go:354] "RemoveStaleState removing state" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver" Dec 10 22:53:30 crc kubenswrapper[4791]: I1210 22:53:30.005381 4791 memory_manager.go:354] "RemoveStaleState removing state" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-check-endpoints" Dec 10 22:53:30 crc kubenswrapper[4791]: I1210 22:53:30.005389 4791 memory_manager.go:354] "RemoveStaleState removing state" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-cert-regeneration-controller" Dec 10 22:53:30 crc kubenswrapper[4791]: I1210 22:53:30.005395 4791 memory_manager.go:354] "RemoveStaleState removing state" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-cert-syncer" Dec 10 22:53:30 crc kubenswrapper[4791]: I1210 22:53:30.005401 4791 memory_manager.go:354] "RemoveStaleState removing state" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-insecure-readyz" Dec 10 22:53:30 crc kubenswrapper[4791]: I1210 22:53:30.005571 4791 memory_manager.go:354] "RemoveStaleState removing state" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-check-endpoints" Dec 10 22:53:30 crc kubenswrapper[4791]: I1210 22:53:30.007257 4791 status_manager.go:861] "Pod was deleted and then recreated, skipping status update" pod="openshift-kube-apiserver/kube-apiserver-crc" oldPodUID="f4b27818a5e8e43d0dc095d08835c792" podUID="71bb4a3aecc4ba5b26c4b7318770ce13" Dec 10 22:53:30 crc kubenswrapper[4791]: I1210 22:53:30.171428 4791 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/71bb4a3aecc4ba5b26c4b7318770ce13-resource-dir\") pod \"kube-apiserver-crc\" (UID: \"71bb4a3aecc4ba5b26c4b7318770ce13\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Dec 10 22:53:30 crc kubenswrapper[4791]: I1210 22:53:30.171838 4791 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/71bb4a3aecc4ba5b26c4b7318770ce13-cert-dir\") pod \"kube-apiserver-crc\" (UID: \"71bb4a3aecc4ba5b26c4b7318770ce13\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Dec 10 22:53:30 crc kubenswrapper[4791]: I1210 22:53:30.171870 4791 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"manifests\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-manifests\") pod \"kube-apiserver-startup-monitor-crc\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Dec 10 22:53:30 crc kubenswrapper[4791]: I1210 22:53:30.171895 4791 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-lock\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-var-lock\") pod \"kube-apiserver-startup-monitor-crc\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Dec 10 22:53:30 crc kubenswrapper[4791]: I1210 22:53:30.171921 4791 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-log\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-var-log\") pod \"kube-apiserver-startup-monitor-crc\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Dec 10 22:53:30 crc kubenswrapper[4791]: I1210 22:53:30.171972 4791 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-resource-dir\") pod \"kube-apiserver-startup-monitor-crc\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Dec 10 22:53:30 crc kubenswrapper[4791]: I1210 22:53:30.172027 4791 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"pod-resource-dir\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-pod-resource-dir\") pod \"kube-apiserver-startup-monitor-crc\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Dec 10 22:53:30 crc kubenswrapper[4791]: I1210 22:53:30.172082 4791 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/71bb4a3aecc4ba5b26c4b7318770ce13-audit-dir\") pod \"kube-apiserver-crc\" (UID: \"71bb4a3aecc4ba5b26c4b7318770ce13\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Dec 10 22:53:30 crc kubenswrapper[4791]: I1210 22:53:30.273135 4791 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/71bb4a3aecc4ba5b26c4b7318770ce13-cert-dir\") pod \"kube-apiserver-crc\" (UID: \"71bb4a3aecc4ba5b26c4b7318770ce13\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Dec 10 22:53:30 crc kubenswrapper[4791]: I1210 22:53:30.273186 4791 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"manifests\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-manifests\") pod \"kube-apiserver-startup-monitor-crc\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Dec 10 22:53:30 crc kubenswrapper[4791]: I1210 22:53:30.273204 4791 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-lock\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-var-lock\") pod \"kube-apiserver-startup-monitor-crc\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Dec 10 22:53:30 crc kubenswrapper[4791]: I1210 22:53:30.273222 4791 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-log\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-var-log\") pod \"kube-apiserver-startup-monitor-crc\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Dec 10 22:53:30 crc kubenswrapper[4791]: I1210 22:53:30.273262 4791 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-resource-dir\") pod \"kube-apiserver-startup-monitor-crc\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Dec 10 22:53:30 crc kubenswrapper[4791]: I1210 22:53:30.273280 4791 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pod-resource-dir\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-pod-resource-dir\") pod \"kube-apiserver-startup-monitor-crc\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Dec 10 22:53:30 crc kubenswrapper[4791]: I1210 22:53:30.273307 4791 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/71bb4a3aecc4ba5b26c4b7318770ce13-audit-dir\") pod \"kube-apiserver-crc\" (UID: \"71bb4a3aecc4ba5b26c4b7318770ce13\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Dec 10 22:53:30 crc kubenswrapper[4791]: I1210 22:53:30.273285 4791 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/71bb4a3aecc4ba5b26c4b7318770ce13-cert-dir\") pod \"kube-apiserver-crc\" (UID: \"71bb4a3aecc4ba5b26c4b7318770ce13\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Dec 10 22:53:30 crc kubenswrapper[4791]: I1210 22:53:30.273301 4791 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-lock\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-var-lock\") pod \"kube-apiserver-startup-monitor-crc\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Dec 10 22:53:30 crc kubenswrapper[4791]: I1210 22:53:30.273347 4791 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/71bb4a3aecc4ba5b26c4b7318770ce13-resource-dir\") pod \"kube-apiserver-crc\" (UID: \"71bb4a3aecc4ba5b26c4b7318770ce13\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Dec 10 22:53:30 crc kubenswrapper[4791]: I1210 22:53:30.273372 4791 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/71bb4a3aecc4ba5b26c4b7318770ce13-resource-dir\") pod \"kube-apiserver-crc\" (UID: \"71bb4a3aecc4ba5b26c4b7318770ce13\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Dec 10 22:53:30 crc kubenswrapper[4791]: I1210 22:53:30.273315 4791 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"manifests\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-manifests\") pod \"kube-apiserver-startup-monitor-crc\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Dec 10 22:53:30 crc kubenswrapper[4791]: I1210 22:53:30.273371 4791 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"pod-resource-dir\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-pod-resource-dir\") pod \"kube-apiserver-startup-monitor-crc\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Dec 10 22:53:30 crc kubenswrapper[4791]: I1210 22:53:30.273399 4791 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/71bb4a3aecc4ba5b26c4b7318770ce13-audit-dir\") pod \"kube-apiserver-crc\" (UID: \"71bb4a3aecc4ba5b26c4b7318770ce13\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Dec 10 22:53:30 crc kubenswrapper[4791]: I1210 22:53:30.273426 4791 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-resource-dir\") pod \"kube-apiserver-startup-monitor-crc\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Dec 10 22:53:30 crc kubenswrapper[4791]: I1210 22:53:30.273395 4791 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-log\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-var-log\") pod \"kube-apiserver-startup-monitor-crc\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Dec 10 22:53:30 crc kubenswrapper[4791]: E1210 22:53:30.814942 4791 controller.go:195] "Failed to update lease" err="Put \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": dial tcp 38.102.83.30:6443: connect: connection refused" Dec 10 22:53:30 crc kubenswrapper[4791]: E1210 22:53:30.815483 4791 controller.go:195] "Failed to update lease" err="Put \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": dial tcp 38.102.83.30:6443: connect: connection refused" Dec 10 22:53:30 crc kubenswrapper[4791]: E1210 22:53:30.815721 4791 controller.go:195] "Failed to update lease" err="Put \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": dial tcp 38.102.83.30:6443: connect: connection refused" Dec 10 22:53:30 crc kubenswrapper[4791]: E1210 22:53:30.816083 4791 controller.go:195] "Failed to update lease" err="Put \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": dial tcp 38.102.83.30:6443: connect: connection refused" Dec 10 22:53:30 crc kubenswrapper[4791]: E1210 22:53:30.816504 4791 controller.go:195] "Failed to update lease" err="Put \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": dial tcp 38.102.83.30:6443: connect: connection refused" Dec 10 22:53:30 crc kubenswrapper[4791]: I1210 22:53:30.816534 4791 controller.go:115] "failed to update lease using latest lease, fallback to ensure lease" err="failed 5 attempts to update lease" Dec 10 22:53:30 crc kubenswrapper[4791]: E1210 22:53:30.816756 4791 controller.go:145] "Failed to ensure lease exists, will retry" err="Get \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": dial tcp 38.102.83.30:6443: connect: connection refused" interval="200ms" Dec 10 22:53:30 crc kubenswrapper[4791]: I1210 22:53:30.896228 4791 generic.go:334] "Generic (PLEG): container finished" podID="31bd0efc-25ef-4a1f-b4b9-e4c5b14768ea" containerID="b26dcf3a5303fefc9276862018a0f98295928a8d297d12f058c57e37ea06d831" exitCode=0 Dec 10 22:53:30 crc kubenswrapper[4791]: I1210 22:53:30.896331 4791 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/installer-9-crc" event={"ID":"31bd0efc-25ef-4a1f-b4b9-e4c5b14768ea","Type":"ContainerDied","Data":"b26dcf3a5303fefc9276862018a0f98295928a8d297d12f058c57e37ea06d831"} Dec 10 22:53:30 crc kubenswrapper[4791]: I1210 22:53:30.897232 4791 status_manager.go:851] "Failed to get status for pod" podUID="31bd0efc-25ef-4a1f-b4b9-e4c5b14768ea" pod="openshift-kube-apiserver/installer-9-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/installer-9-crc\": dial tcp 38.102.83.30:6443: connect: connection refused" Dec 10 22:53:30 crc kubenswrapper[4791]: I1210 22:53:30.898742 4791 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-kube-apiserver_kube-apiserver-crc_f4b27818a5e8e43d0dc095d08835c792/kube-apiserver-check-endpoints/0.log" Dec 10 22:53:30 crc kubenswrapper[4791]: I1210 22:53:30.900115 4791 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-kube-apiserver_kube-apiserver-crc_f4b27818a5e8e43d0dc095d08835c792/kube-apiserver-cert-syncer/0.log" Dec 10 22:53:30 crc kubenswrapper[4791]: I1210 22:53:30.900850 4791 generic.go:334] "Generic (PLEG): container finished" podID="f4b27818a5e8e43d0dc095d08835c792" containerID="1b2e11ff2bf8789fa1d0d23ae648b3e6ef653106e5bc4bbb5dd6447f7cd6f05d" exitCode=0 Dec 10 22:53:30 crc kubenswrapper[4791]: I1210 22:53:30.900884 4791 generic.go:334] "Generic (PLEG): container finished" podID="f4b27818a5e8e43d0dc095d08835c792" containerID="0c4d036ab269f62561fe7ef1586012cb3a16e3583a08b9488a0616ec00c8e0ae" exitCode=0 Dec 10 22:53:30 crc kubenswrapper[4791]: I1210 22:53:30.900900 4791 generic.go:334] "Generic (PLEG): container finished" podID="f4b27818a5e8e43d0dc095d08835c792" containerID="ea4c346507862a8c12756a3061d6e25e788a8d79f4621b62f81f4322185c82a8" exitCode=0 Dec 10 22:53:30 crc kubenswrapper[4791]: I1210 22:53:30.900910 4791 generic.go:334] "Generic (PLEG): container finished" podID="f4b27818a5e8e43d0dc095d08835c792" containerID="2e6e324ec6f4f7bb69fdfab4ea3ac1aa55e480dd8e2a8984b0fe5b4c9f66636e" exitCode=2 Dec 10 22:53:30 crc kubenswrapper[4791]: I1210 22:53:30.900947 4791 scope.go:117] "RemoveContainer" containerID="d6aa7f8c5d908d90008f2df49860ac8a3480136da567e3d85f86135d779204f9" Dec 10 22:53:31 crc kubenswrapper[4791]: E1210 22:53:31.018400 4791 controller.go:145] "Failed to ensure lease exists, will retry" err="Get \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": dial tcp 38.102.83.30:6443: connect: connection refused" interval="400ms" Dec 10 22:53:31 crc kubenswrapper[4791]: E1210 22:53:31.419322 4791 controller.go:145] "Failed to ensure lease exists, will retry" err="Get \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": dial tcp 38.102.83.30:6443: connect: connection refused" interval="800ms" Dec 10 22:53:31 crc kubenswrapper[4791]: I1210 22:53:31.909039 4791 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-kube-apiserver_kube-apiserver-crc_f4b27818a5e8e43d0dc095d08835c792/kube-apiserver-cert-syncer/0.log" Dec 10 22:53:32 crc kubenswrapper[4791]: I1210 22:53:32.217377 4791 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/installer-9-crc" Dec 10 22:53:32 crc kubenswrapper[4791]: I1210 22:53:32.218116 4791 status_manager.go:851] "Failed to get status for pod" podUID="31bd0efc-25ef-4a1f-b4b9-e4c5b14768ea" pod="openshift-kube-apiserver/installer-9-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/installer-9-crc\": dial tcp 38.102.83.30:6443: connect: connection refused" Dec 10 22:53:32 crc kubenswrapper[4791]: E1210 22:53:32.219976 4791 controller.go:145] "Failed to ensure lease exists, will retry" err="Get \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": dial tcp 38.102.83.30:6443: connect: connection refused" interval="1.6s" Dec 10 22:53:32 crc kubenswrapper[4791]: I1210 22:53:32.364186 4791 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-kube-apiserver_kube-apiserver-crc_f4b27818a5e8e43d0dc095d08835c792/kube-apiserver-cert-syncer/0.log" Dec 10 22:53:32 crc kubenswrapper[4791]: I1210 22:53:32.365225 4791 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/kube-apiserver-crc" Dec 10 22:53:32 crc kubenswrapper[4791]: I1210 22:53:32.365820 4791 status_manager.go:851] "Failed to get status for pod" podUID="31bd0efc-25ef-4a1f-b4b9-e4c5b14768ea" pod="openshift-kube-apiserver/installer-9-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/installer-9-crc\": dial tcp 38.102.83.30:6443: connect: connection refused" Dec 10 22:53:32 crc kubenswrapper[4791]: I1210 22:53:32.366330 4791 status_manager.go:851] "Failed to get status for pod" podUID="f4b27818a5e8e43d0dc095d08835c792" pod="openshift-kube-apiserver/kube-apiserver-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\": dial tcp 38.102.83.30:6443: connect: connection refused" Dec 10 22:53:32 crc kubenswrapper[4791]: I1210 22:53:32.402213 4791 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"var-lock\" (UniqueName: \"kubernetes.io/host-path/31bd0efc-25ef-4a1f-b4b9-e4c5b14768ea-var-lock\") pod \"31bd0efc-25ef-4a1f-b4b9-e4c5b14768ea\" (UID: \"31bd0efc-25ef-4a1f-b4b9-e4c5b14768ea\") " Dec 10 22:53:32 crc kubenswrapper[4791]: I1210 22:53:32.402322 4791 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/31bd0efc-25ef-4a1f-b4b9-e4c5b14768ea-var-lock" (OuterVolumeSpecName: "var-lock") pod "31bd0efc-25ef-4a1f-b4b9-e4c5b14768ea" (UID: "31bd0efc-25ef-4a1f-b4b9-e4c5b14768ea"). InnerVolumeSpecName "var-lock". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 10 22:53:32 crc kubenswrapper[4791]: I1210 22:53:32.402403 4791 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/31bd0efc-25ef-4a1f-b4b9-e4c5b14768ea-kubelet-dir\") pod \"31bd0efc-25ef-4a1f-b4b9-e4c5b14768ea\" (UID: \"31bd0efc-25ef-4a1f-b4b9-e4c5b14768ea\") " Dec 10 22:53:32 crc kubenswrapper[4791]: I1210 22:53:32.402442 4791 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/31bd0efc-25ef-4a1f-b4b9-e4c5b14768ea-kube-api-access\") pod \"31bd0efc-25ef-4a1f-b4b9-e4c5b14768ea\" (UID: \"31bd0efc-25ef-4a1f-b4b9-e4c5b14768ea\") " Dec 10 22:53:32 crc kubenswrapper[4791]: I1210 22:53:32.402482 4791 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/31bd0efc-25ef-4a1f-b4b9-e4c5b14768ea-kubelet-dir" (OuterVolumeSpecName: "kubelet-dir") pod "31bd0efc-25ef-4a1f-b4b9-e4c5b14768ea" (UID: "31bd0efc-25ef-4a1f-b4b9-e4c5b14768ea"). InnerVolumeSpecName "kubelet-dir". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 10 22:53:32 crc kubenswrapper[4791]: I1210 22:53:32.402683 4791 reconciler_common.go:293] "Volume detached for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/31bd0efc-25ef-4a1f-b4b9-e4c5b14768ea-kubelet-dir\") on node \"crc\" DevicePath \"\"" Dec 10 22:53:32 crc kubenswrapper[4791]: I1210 22:53:32.402706 4791 reconciler_common.go:293] "Volume detached for volume \"var-lock\" (UniqueName: \"kubernetes.io/host-path/31bd0efc-25ef-4a1f-b4b9-e4c5b14768ea-var-lock\") on node \"crc\" DevicePath \"\"" Dec 10 22:53:32 crc kubenswrapper[4791]: I1210 22:53:32.408963 4791 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/31bd0efc-25ef-4a1f-b4b9-e4c5b14768ea-kube-api-access" (OuterVolumeSpecName: "kube-api-access") pod "31bd0efc-25ef-4a1f-b4b9-e4c5b14768ea" (UID: "31bd0efc-25ef-4a1f-b4b9-e4c5b14768ea"). InnerVolumeSpecName "kube-api-access". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 10 22:53:32 crc kubenswrapper[4791]: I1210 22:53:32.503457 4791 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-cert-dir\") pod \"f4b27818a5e8e43d0dc095d08835c792\" (UID: \"f4b27818a5e8e43d0dc095d08835c792\") " Dec 10 22:53:32 crc kubenswrapper[4791]: I1210 22:53:32.503540 4791 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-audit-dir\") pod \"f4b27818a5e8e43d0dc095d08835c792\" (UID: \"f4b27818a5e8e43d0dc095d08835c792\") " Dec 10 22:53:32 crc kubenswrapper[4791]: I1210 22:53:32.503630 4791 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-resource-dir\") pod \"f4b27818a5e8e43d0dc095d08835c792\" (UID: \"f4b27818a5e8e43d0dc095d08835c792\") " Dec 10 22:53:32 crc kubenswrapper[4791]: I1210 22:53:32.504074 4791 reconciler_common.go:293] "Volume detached for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/31bd0efc-25ef-4a1f-b4b9-e4c5b14768ea-kube-api-access\") on node \"crc\" DevicePath \"\"" Dec 10 22:53:32 crc kubenswrapper[4791]: I1210 22:53:32.504163 4791 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-resource-dir" (OuterVolumeSpecName: "resource-dir") pod "f4b27818a5e8e43d0dc095d08835c792" (UID: "f4b27818a5e8e43d0dc095d08835c792"). InnerVolumeSpecName "resource-dir". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 10 22:53:32 crc kubenswrapper[4791]: I1210 22:53:32.504212 4791 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-cert-dir" (OuterVolumeSpecName: "cert-dir") pod "f4b27818a5e8e43d0dc095d08835c792" (UID: "f4b27818a5e8e43d0dc095d08835c792"). InnerVolumeSpecName "cert-dir". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 10 22:53:32 crc kubenswrapper[4791]: I1210 22:53:32.504244 4791 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-audit-dir" (OuterVolumeSpecName: "audit-dir") pod "f4b27818a5e8e43d0dc095d08835c792" (UID: "f4b27818a5e8e43d0dc095d08835c792"). InnerVolumeSpecName "audit-dir". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 10 22:53:32 crc kubenswrapper[4791]: I1210 22:53:32.605070 4791 reconciler_common.go:293] "Volume detached for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-resource-dir\") on node \"crc\" DevicePath \"\"" Dec 10 22:53:32 crc kubenswrapper[4791]: I1210 22:53:32.605100 4791 reconciler_common.go:293] "Volume detached for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-cert-dir\") on node \"crc\" DevicePath \"\"" Dec 10 22:53:32 crc kubenswrapper[4791]: I1210 22:53:32.605108 4791 reconciler_common.go:293] "Volume detached for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-audit-dir\") on node \"crc\" DevicePath \"\"" Dec 10 22:53:32 crc kubenswrapper[4791]: I1210 22:53:32.917496 4791 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-kube-apiserver_kube-apiserver-crc_f4b27818a5e8e43d0dc095d08835c792/kube-apiserver-cert-syncer/0.log" Dec 10 22:53:32 crc kubenswrapper[4791]: I1210 22:53:32.918371 4791 generic.go:334] "Generic (PLEG): container finished" podID="f4b27818a5e8e43d0dc095d08835c792" containerID="acafe12186df99d1059a31260de0eafce35ff7771bcb8f5c6753b72665d6f60b" exitCode=0 Dec 10 22:53:32 crc kubenswrapper[4791]: I1210 22:53:32.918452 4791 scope.go:117] "RemoveContainer" containerID="1b2e11ff2bf8789fa1d0d23ae648b3e6ef653106e5bc4bbb5dd6447f7cd6f05d" Dec 10 22:53:32 crc kubenswrapper[4791]: I1210 22:53:32.918457 4791 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/kube-apiserver-crc" Dec 10 22:53:32 crc kubenswrapper[4791]: I1210 22:53:32.922952 4791 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/installer-9-crc" event={"ID":"31bd0efc-25ef-4a1f-b4b9-e4c5b14768ea","Type":"ContainerDied","Data":"2fb0253edf06e6a7393d013814126be22811166fe5f712405de81225ba67c3a1"} Dec 10 22:53:32 crc kubenswrapper[4791]: I1210 22:53:32.922991 4791 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="2fb0253edf06e6a7393d013814126be22811166fe5f712405de81225ba67c3a1" Dec 10 22:53:32 crc kubenswrapper[4791]: I1210 22:53:32.923384 4791 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/installer-9-crc" Dec 10 22:53:32 crc kubenswrapper[4791]: I1210 22:53:32.940551 4791 status_manager.go:851] "Failed to get status for pod" podUID="f4b27818a5e8e43d0dc095d08835c792" pod="openshift-kube-apiserver/kube-apiserver-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\": dial tcp 38.102.83.30:6443: connect: connection refused" Dec 10 22:53:32 crc kubenswrapper[4791]: I1210 22:53:32.941728 4791 status_manager.go:851] "Failed to get status for pod" podUID="31bd0efc-25ef-4a1f-b4b9-e4c5b14768ea" pod="openshift-kube-apiserver/installer-9-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/installer-9-crc\": dial tcp 38.102.83.30:6443: connect: connection refused" Dec 10 22:53:32 crc kubenswrapper[4791]: I1210 22:53:32.948548 4791 scope.go:117] "RemoveContainer" containerID="0c4d036ab269f62561fe7ef1586012cb3a16e3583a08b9488a0616ec00c8e0ae" Dec 10 22:53:32 crc kubenswrapper[4791]: I1210 22:53:32.952842 4791 status_manager.go:851] "Failed to get status for pod" podUID="31bd0efc-25ef-4a1f-b4b9-e4c5b14768ea" pod="openshift-kube-apiserver/installer-9-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/installer-9-crc\": dial tcp 38.102.83.30:6443: connect: connection refused" Dec 10 22:53:32 crc kubenswrapper[4791]: I1210 22:53:32.953176 4791 status_manager.go:851] "Failed to get status for pod" podUID="f4b27818a5e8e43d0dc095d08835c792" pod="openshift-kube-apiserver/kube-apiserver-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\": dial tcp 38.102.83.30:6443: connect: connection refused" Dec 10 22:53:32 crc kubenswrapper[4791]: I1210 22:53:32.959032 4791 scope.go:117] "RemoveContainer" containerID="ea4c346507862a8c12756a3061d6e25e788a8d79f4621b62f81f4322185c82a8" Dec 10 22:53:32 crc kubenswrapper[4791]: I1210 22:53:32.970676 4791 scope.go:117] "RemoveContainer" containerID="2e6e324ec6f4f7bb69fdfab4ea3ac1aa55e480dd8e2a8984b0fe5b4c9f66636e" Dec 10 22:53:32 crc kubenswrapper[4791]: I1210 22:53:32.981086 4791 scope.go:117] "RemoveContainer" containerID="acafe12186df99d1059a31260de0eafce35ff7771bcb8f5c6753b72665d6f60b" Dec 10 22:53:32 crc kubenswrapper[4791]: I1210 22:53:32.996587 4791 scope.go:117] "RemoveContainer" containerID="c5856d6b26c19fda13b5c9ac2892a7e2f9c6f79feb8173ecfd9b33b673327d67" Dec 10 22:53:33 crc kubenswrapper[4791]: I1210 22:53:33.015192 4791 scope.go:117] "RemoveContainer" containerID="1b2e11ff2bf8789fa1d0d23ae648b3e6ef653106e5bc4bbb5dd6447f7cd6f05d" Dec 10 22:53:33 crc kubenswrapper[4791]: E1210 22:53:33.015585 4791 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"1b2e11ff2bf8789fa1d0d23ae648b3e6ef653106e5bc4bbb5dd6447f7cd6f05d\": container with ID starting with 1b2e11ff2bf8789fa1d0d23ae648b3e6ef653106e5bc4bbb5dd6447f7cd6f05d not found: ID does not exist" containerID="1b2e11ff2bf8789fa1d0d23ae648b3e6ef653106e5bc4bbb5dd6447f7cd6f05d" Dec 10 22:53:33 crc kubenswrapper[4791]: I1210 22:53:33.015638 4791 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"1b2e11ff2bf8789fa1d0d23ae648b3e6ef653106e5bc4bbb5dd6447f7cd6f05d"} err="failed to get container status \"1b2e11ff2bf8789fa1d0d23ae648b3e6ef653106e5bc4bbb5dd6447f7cd6f05d\": rpc error: code = NotFound desc = could not find container \"1b2e11ff2bf8789fa1d0d23ae648b3e6ef653106e5bc4bbb5dd6447f7cd6f05d\": container with ID starting with 1b2e11ff2bf8789fa1d0d23ae648b3e6ef653106e5bc4bbb5dd6447f7cd6f05d not found: ID does not exist" Dec 10 22:53:33 crc kubenswrapper[4791]: I1210 22:53:33.015670 4791 scope.go:117] "RemoveContainer" containerID="0c4d036ab269f62561fe7ef1586012cb3a16e3583a08b9488a0616ec00c8e0ae" Dec 10 22:53:33 crc kubenswrapper[4791]: E1210 22:53:33.015952 4791 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"0c4d036ab269f62561fe7ef1586012cb3a16e3583a08b9488a0616ec00c8e0ae\": container with ID starting with 0c4d036ab269f62561fe7ef1586012cb3a16e3583a08b9488a0616ec00c8e0ae not found: ID does not exist" containerID="0c4d036ab269f62561fe7ef1586012cb3a16e3583a08b9488a0616ec00c8e0ae" Dec 10 22:53:33 crc kubenswrapper[4791]: I1210 22:53:33.015981 4791 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"0c4d036ab269f62561fe7ef1586012cb3a16e3583a08b9488a0616ec00c8e0ae"} err="failed to get container status \"0c4d036ab269f62561fe7ef1586012cb3a16e3583a08b9488a0616ec00c8e0ae\": rpc error: code = NotFound desc = could not find container \"0c4d036ab269f62561fe7ef1586012cb3a16e3583a08b9488a0616ec00c8e0ae\": container with ID starting with 0c4d036ab269f62561fe7ef1586012cb3a16e3583a08b9488a0616ec00c8e0ae not found: ID does not exist" Dec 10 22:53:33 crc kubenswrapper[4791]: I1210 22:53:33.015999 4791 scope.go:117] "RemoveContainer" containerID="ea4c346507862a8c12756a3061d6e25e788a8d79f4621b62f81f4322185c82a8" Dec 10 22:53:33 crc kubenswrapper[4791]: E1210 22:53:33.016261 4791 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"ea4c346507862a8c12756a3061d6e25e788a8d79f4621b62f81f4322185c82a8\": container with ID starting with ea4c346507862a8c12756a3061d6e25e788a8d79f4621b62f81f4322185c82a8 not found: ID does not exist" containerID="ea4c346507862a8c12756a3061d6e25e788a8d79f4621b62f81f4322185c82a8" Dec 10 22:53:33 crc kubenswrapper[4791]: I1210 22:53:33.016292 4791 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"ea4c346507862a8c12756a3061d6e25e788a8d79f4621b62f81f4322185c82a8"} err="failed to get container status \"ea4c346507862a8c12756a3061d6e25e788a8d79f4621b62f81f4322185c82a8\": rpc error: code = NotFound desc = could not find container \"ea4c346507862a8c12756a3061d6e25e788a8d79f4621b62f81f4322185c82a8\": container with ID starting with ea4c346507862a8c12756a3061d6e25e788a8d79f4621b62f81f4322185c82a8 not found: ID does not exist" Dec 10 22:53:33 crc kubenswrapper[4791]: I1210 22:53:33.016314 4791 scope.go:117] "RemoveContainer" containerID="2e6e324ec6f4f7bb69fdfab4ea3ac1aa55e480dd8e2a8984b0fe5b4c9f66636e" Dec 10 22:53:33 crc kubenswrapper[4791]: E1210 22:53:33.016622 4791 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"2e6e324ec6f4f7bb69fdfab4ea3ac1aa55e480dd8e2a8984b0fe5b4c9f66636e\": container with ID starting with 2e6e324ec6f4f7bb69fdfab4ea3ac1aa55e480dd8e2a8984b0fe5b4c9f66636e not found: ID does not exist" containerID="2e6e324ec6f4f7bb69fdfab4ea3ac1aa55e480dd8e2a8984b0fe5b4c9f66636e" Dec 10 22:53:33 crc kubenswrapper[4791]: I1210 22:53:33.016651 4791 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"2e6e324ec6f4f7bb69fdfab4ea3ac1aa55e480dd8e2a8984b0fe5b4c9f66636e"} err="failed to get container status \"2e6e324ec6f4f7bb69fdfab4ea3ac1aa55e480dd8e2a8984b0fe5b4c9f66636e\": rpc error: code = NotFound desc = could not find container \"2e6e324ec6f4f7bb69fdfab4ea3ac1aa55e480dd8e2a8984b0fe5b4c9f66636e\": container with ID starting with 2e6e324ec6f4f7bb69fdfab4ea3ac1aa55e480dd8e2a8984b0fe5b4c9f66636e not found: ID does not exist" Dec 10 22:53:33 crc kubenswrapper[4791]: I1210 22:53:33.016672 4791 scope.go:117] "RemoveContainer" containerID="acafe12186df99d1059a31260de0eafce35ff7771bcb8f5c6753b72665d6f60b" Dec 10 22:53:33 crc kubenswrapper[4791]: E1210 22:53:33.017577 4791 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"acafe12186df99d1059a31260de0eafce35ff7771bcb8f5c6753b72665d6f60b\": container with ID starting with acafe12186df99d1059a31260de0eafce35ff7771bcb8f5c6753b72665d6f60b not found: ID does not exist" containerID="acafe12186df99d1059a31260de0eafce35ff7771bcb8f5c6753b72665d6f60b" Dec 10 22:53:33 crc kubenswrapper[4791]: I1210 22:53:33.017612 4791 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"acafe12186df99d1059a31260de0eafce35ff7771bcb8f5c6753b72665d6f60b"} err="failed to get container status \"acafe12186df99d1059a31260de0eafce35ff7771bcb8f5c6753b72665d6f60b\": rpc error: code = NotFound desc = could not find container \"acafe12186df99d1059a31260de0eafce35ff7771bcb8f5c6753b72665d6f60b\": container with ID starting with acafe12186df99d1059a31260de0eafce35ff7771bcb8f5c6753b72665d6f60b not found: ID does not exist" Dec 10 22:53:33 crc kubenswrapper[4791]: I1210 22:53:33.017634 4791 scope.go:117] "RemoveContainer" containerID="c5856d6b26c19fda13b5c9ac2892a7e2f9c6f79feb8173ecfd9b33b673327d67" Dec 10 22:53:33 crc kubenswrapper[4791]: E1210 22:53:33.018056 4791 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"c5856d6b26c19fda13b5c9ac2892a7e2f9c6f79feb8173ecfd9b33b673327d67\": container with ID starting with c5856d6b26c19fda13b5c9ac2892a7e2f9c6f79feb8173ecfd9b33b673327d67 not found: ID does not exist" containerID="c5856d6b26c19fda13b5c9ac2892a7e2f9c6f79feb8173ecfd9b33b673327d67" Dec 10 22:53:33 crc kubenswrapper[4791]: I1210 22:53:33.018090 4791 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"c5856d6b26c19fda13b5c9ac2892a7e2f9c6f79feb8173ecfd9b33b673327d67"} err="failed to get container status \"c5856d6b26c19fda13b5c9ac2892a7e2f9c6f79feb8173ecfd9b33b673327d67\": rpc error: code = NotFound desc = could not find container \"c5856d6b26c19fda13b5c9ac2892a7e2f9c6f79feb8173ecfd9b33b673327d67\": container with ID starting with c5856d6b26c19fda13b5c9ac2892a7e2f9c6f79feb8173ecfd9b33b673327d67 not found: ID does not exist" Dec 10 22:53:33 crc kubenswrapper[4791]: E1210 22:53:33.115965 4791 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-12-10T22:53:33Z\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-10T22:53:33Z\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-10T22:53:33Z\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-10T22:53:33Z\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"runtimeHandlers\\\":[{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":true},\\\"name\\\":\\\"crun\\\"},{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":true},\\\"name\\\":\\\"\\\"},{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":false},\\\"name\\\":\\\"runc\\\"}]}}\" for node \"crc\": Patch \"https://api-int.crc.testing:6443/api/v1/nodes/crc/status?timeout=10s\": dial tcp 38.102.83.30:6443: connect: connection refused" Dec 10 22:53:33 crc kubenswrapper[4791]: E1210 22:53:33.116298 4791 kubelet_node_status.go:585] "Error updating node status, will retry" err="error getting node \"crc\": Get \"https://api-int.crc.testing:6443/api/v1/nodes/crc?timeout=10s\": dial tcp 38.102.83.30:6443: connect: connection refused" Dec 10 22:53:33 crc kubenswrapper[4791]: E1210 22:53:33.116619 4791 kubelet_node_status.go:585] "Error updating node status, will retry" err="error getting node \"crc\": Get \"https://api-int.crc.testing:6443/api/v1/nodes/crc?timeout=10s\": dial tcp 38.102.83.30:6443: connect: connection refused" Dec 10 22:53:33 crc kubenswrapper[4791]: E1210 22:53:33.116822 4791 kubelet_node_status.go:585] "Error updating node status, will retry" err="error getting node \"crc\": Get \"https://api-int.crc.testing:6443/api/v1/nodes/crc?timeout=10s\": dial tcp 38.102.83.30:6443: connect: connection refused" Dec 10 22:53:33 crc kubenswrapper[4791]: E1210 22:53:33.117088 4791 kubelet_node_status.go:585] "Error updating node status, will retry" err="error getting node \"crc\": Get \"https://api-int.crc.testing:6443/api/v1/nodes/crc?timeout=10s\": dial tcp 38.102.83.30:6443: connect: connection refused" Dec 10 22:53:33 crc kubenswrapper[4791]: E1210 22:53:33.117107 4791 kubelet_node_status.go:572] "Unable to update node status" err="update node status exceeds retry count" Dec 10 22:53:33 crc kubenswrapper[4791]: E1210 22:53:33.821226 4791 controller.go:145] "Failed to ensure lease exists, will retry" err="Get \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": dial tcp 38.102.83.30:6443: connect: connection refused" interval="3.2s" Dec 10 22:53:33 crc kubenswrapper[4791]: I1210 22:53:33.897901 4791 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="f4b27818a5e8e43d0dc095d08835c792" path="/var/lib/kubelet/pods/f4b27818a5e8e43d0dc095d08835c792/volumes" Dec 10 22:53:35 crc kubenswrapper[4791]: E1210 22:53:35.035990 4791 kubelet.go:1929] "Failed creating a mirror pod for" err="Post \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods\": dial tcp 38.102.83.30:6443: connect: connection refused" pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Dec 10 22:53:35 crc kubenswrapper[4791]: I1210 22:53:35.036408 4791 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Dec 10 22:53:35 crc kubenswrapper[4791]: E1210 22:53:35.061587 4791 event.go:368] "Unable to write event (may retry after sleeping)" err="Post \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/events\": dial tcp 38.102.83.30:6443: connect: connection refused" event="&Event{ObjectMeta:{kube-apiserver-startup-monitor-crc.187ffc82319f29cb openshift-kube-apiserver 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-kube-apiserver,Name:kube-apiserver-startup-monitor-crc,UID:f85e55b1a89d02b0cb034b1ea31ed45a,APIVersion:v1,ResourceVersion:,FieldPath:spec.containers{startup-monitor},},Reason:Pulled,Message:Container image \"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\" already present on machine,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2025-12-10 22:53:35.060986315 +0000 UTC m=+249.490603938,LastTimestamp:2025-12-10 22:53:35.060986315 +0000 UTC m=+249.490603938,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Dec 10 22:53:35 crc kubenswrapper[4791]: I1210 22:53:35.887331 4791 status_manager.go:851] "Failed to get status for pod" podUID="31bd0efc-25ef-4a1f-b4b9-e4c5b14768ea" pod="openshift-kube-apiserver/installer-9-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/installer-9-crc\": dial tcp 38.102.83.30:6443: connect: connection refused" Dec 10 22:53:35 crc kubenswrapper[4791]: I1210 22:53:35.948239 4791 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" event={"ID":"f85e55b1a89d02b0cb034b1ea31ed45a","Type":"ContainerStarted","Data":"3c36eae76bf3819198f38fa3cb82e87295ea35971222122789121c928c1c6f03"} Dec 10 22:53:35 crc kubenswrapper[4791]: I1210 22:53:35.948314 4791 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" event={"ID":"f85e55b1a89d02b0cb034b1ea31ed45a","Type":"ContainerStarted","Data":"d9d10aa09803e2399fb129d6bf10ec58ad694c6e06351c61191a82d8b6447b0b"} Dec 10 22:53:35 crc kubenswrapper[4791]: I1210 22:53:35.948905 4791 status_manager.go:851] "Failed to get status for pod" podUID="31bd0efc-25ef-4a1f-b4b9-e4c5b14768ea" pod="openshift-kube-apiserver/installer-9-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/installer-9-crc\": dial tcp 38.102.83.30:6443: connect: connection refused" Dec 10 22:53:35 crc kubenswrapper[4791]: E1210 22:53:35.949298 4791 kubelet.go:1929] "Failed creating a mirror pod for" err="Post \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods\": dial tcp 38.102.83.30:6443: connect: connection refused" pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Dec 10 22:53:37 crc kubenswrapper[4791]: E1210 22:53:37.022973 4791 controller.go:145] "Failed to ensure lease exists, will retry" err="Get \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": dial tcp 38.102.83.30:6443: connect: connection refused" interval="6.4s" Dec 10 22:53:37 crc kubenswrapper[4791]: E1210 22:53:37.831001 4791 event.go:368] "Unable to write event (may retry after sleeping)" err="Post \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/events\": dial tcp 38.102.83.30:6443: connect: connection refused" event="&Event{ObjectMeta:{kube-apiserver-startup-monitor-crc.187ffc82319f29cb openshift-kube-apiserver 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-kube-apiserver,Name:kube-apiserver-startup-monitor-crc,UID:f85e55b1a89d02b0cb034b1ea31ed45a,APIVersion:v1,ResourceVersion:,FieldPath:spec.containers{startup-monitor},},Reason:Pulled,Message:Container image \"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\" already present on machine,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2025-12-10 22:53:35.060986315 +0000 UTC m=+249.490603938,LastTimestamp:2025-12-10 22:53:35.060986315 +0000 UTC m=+249.490603938,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Dec 10 22:53:42 crc kubenswrapper[4791]: I1210 22:53:42.985578 4791 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-kube-controller-manager_kube-controller-manager-crc_f614b9022728cf315e60c057852e563e/kube-controller-manager/0.log" Dec 10 22:53:42 crc kubenswrapper[4791]: I1210 22:53:42.986068 4791 generic.go:334] "Generic (PLEG): container finished" podID="f614b9022728cf315e60c057852e563e" containerID="372eeaffd40992c6f06efc4df6dfcb44b31de64ee8256c21508cdd7c74ecb96e" exitCode=1 Dec 10 22:53:42 crc kubenswrapper[4791]: I1210 22:53:42.986097 4791 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" event={"ID":"f614b9022728cf315e60c057852e563e","Type":"ContainerDied","Data":"372eeaffd40992c6f06efc4df6dfcb44b31de64ee8256c21508cdd7c74ecb96e"} Dec 10 22:53:42 crc kubenswrapper[4791]: I1210 22:53:42.986514 4791 scope.go:117] "RemoveContainer" containerID="372eeaffd40992c6f06efc4df6dfcb44b31de64ee8256c21508cdd7c74ecb96e" Dec 10 22:53:42 crc kubenswrapper[4791]: I1210 22:53:42.987307 4791 status_manager.go:851] "Failed to get status for pod" podUID="31bd0efc-25ef-4a1f-b4b9-e4c5b14768ea" pod="openshift-kube-apiserver/installer-9-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/installer-9-crc\": dial tcp 38.102.83.30:6443: connect: connection refused" Dec 10 22:53:42 crc kubenswrapper[4791]: I1210 22:53:42.987594 4791 status_manager.go:851] "Failed to get status for pod" podUID="f614b9022728cf315e60c057852e563e" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-controller-manager/pods/kube-controller-manager-crc\": dial tcp 38.102.83.30:6443: connect: connection refused" Dec 10 22:53:43 crc kubenswrapper[4791]: E1210 22:53:43.423878 4791 controller.go:145] "Failed to ensure lease exists, will retry" err="Get \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": dial tcp 38.102.83.30:6443: connect: connection refused" interval="7s" Dec 10 22:53:43 crc kubenswrapper[4791]: E1210 22:53:43.518507 4791 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-12-10T22:53:43Z\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-10T22:53:43Z\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-10T22:53:43Z\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-10T22:53:43Z\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"runtimeHandlers\\\":[{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":true},\\\"name\\\":\\\"crun\\\"},{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":true},\\\"name\\\":\\\"\\\"},{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":false},\\\"name\\\":\\\"runc\\\"}]}}\" for node \"crc\": Patch \"https://api-int.crc.testing:6443/api/v1/nodes/crc/status?timeout=10s\": dial tcp 38.102.83.30:6443: connect: connection refused" Dec 10 22:53:43 crc kubenswrapper[4791]: E1210 22:53:43.523685 4791 kubelet_node_status.go:585] "Error updating node status, will retry" err="error getting node \"crc\": Get \"https://api-int.crc.testing:6443/api/v1/nodes/crc?timeout=10s\": dial tcp 38.102.83.30:6443: connect: connection refused" Dec 10 22:53:43 crc kubenswrapper[4791]: E1210 22:53:43.524189 4791 kubelet_node_status.go:585] "Error updating node status, will retry" err="error getting node \"crc\": Get \"https://api-int.crc.testing:6443/api/v1/nodes/crc?timeout=10s\": dial tcp 38.102.83.30:6443: connect: connection refused" Dec 10 22:53:43 crc kubenswrapper[4791]: E1210 22:53:43.524652 4791 kubelet_node_status.go:585] "Error updating node status, will retry" err="error getting node \"crc\": Get \"https://api-int.crc.testing:6443/api/v1/nodes/crc?timeout=10s\": dial tcp 38.102.83.30:6443: connect: connection refused" Dec 10 22:53:43 crc kubenswrapper[4791]: E1210 22:53:43.525086 4791 kubelet_node_status.go:585] "Error updating node status, will retry" err="error getting node \"crc\": Get \"https://api-int.crc.testing:6443/api/v1/nodes/crc?timeout=10s\": dial tcp 38.102.83.30:6443: connect: connection refused" Dec 10 22:53:43 crc kubenswrapper[4791]: E1210 22:53:43.525129 4791 kubelet_node_status.go:572] "Unable to update node status" err="update node status exceeds retry count" Dec 10 22:53:43 crc kubenswrapper[4791]: I1210 22:53:43.995478 4791 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-kube-controller-manager_kube-controller-manager-crc_f614b9022728cf315e60c057852e563e/kube-controller-manager/0.log" Dec 10 22:53:43 crc kubenswrapper[4791]: I1210 22:53:43.995594 4791 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" event={"ID":"f614b9022728cf315e60c057852e563e","Type":"ContainerStarted","Data":"0f6062f1c6e38a55c5af87181c61a412a0ecc8ef3346869f578805c7362f4f53"} Dec 10 22:53:43 crc kubenswrapper[4791]: I1210 22:53:43.996683 4791 status_manager.go:851] "Failed to get status for pod" podUID="f614b9022728cf315e60c057852e563e" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-controller-manager/pods/kube-controller-manager-crc\": dial tcp 38.102.83.30:6443: connect: connection refused" Dec 10 22:53:43 crc kubenswrapper[4791]: I1210 22:53:43.997327 4791 status_manager.go:851] "Failed to get status for pod" podUID="31bd0efc-25ef-4a1f-b4b9-e4c5b14768ea" pod="openshift-kube-apiserver/installer-9-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/installer-9-crc\": dial tcp 38.102.83.30:6443: connect: connection refused" Dec 10 22:53:44 crc kubenswrapper[4791]: I1210 22:53:44.884096 4791 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/kube-apiserver-crc" Dec 10 22:53:44 crc kubenswrapper[4791]: I1210 22:53:44.885384 4791 status_manager.go:851] "Failed to get status for pod" podUID="f614b9022728cf315e60c057852e563e" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-controller-manager/pods/kube-controller-manager-crc\": dial tcp 38.102.83.30:6443: connect: connection refused" Dec 10 22:53:44 crc kubenswrapper[4791]: I1210 22:53:44.885812 4791 status_manager.go:851] "Failed to get status for pod" podUID="31bd0efc-25ef-4a1f-b4b9-e4c5b14768ea" pod="openshift-kube-apiserver/installer-9-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/installer-9-crc\": dial tcp 38.102.83.30:6443: connect: connection refused" Dec 10 22:53:44 crc kubenswrapper[4791]: I1210 22:53:44.897961 4791 kubelet.go:1909] "Trying to delete pod" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="f8349fea-c49b-49a1-b5ee-032c5dffe021" Dec 10 22:53:44 crc kubenswrapper[4791]: I1210 22:53:44.898001 4791 mirror_client.go:130] "Deleting a mirror pod" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="f8349fea-c49b-49a1-b5ee-032c5dffe021" Dec 10 22:53:44 crc kubenswrapper[4791]: E1210 22:53:44.898637 4791 mirror_client.go:138] "Failed deleting a mirror pod" err="Delete \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\": dial tcp 38.102.83.30:6443: connect: connection refused" pod="openshift-kube-apiserver/kube-apiserver-crc" Dec 10 22:53:44 crc kubenswrapper[4791]: I1210 22:53:44.899413 4791 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/kube-apiserver-crc" Dec 10 22:53:44 crc kubenswrapper[4791]: W1210 22:53:44.925672 4791 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod71bb4a3aecc4ba5b26c4b7318770ce13.slice/crio-de470e7b6b6edcad509453c874135919ec7ea63a0607afb42a9e833cf7b93c54 WatchSource:0}: Error finding container de470e7b6b6edcad509453c874135919ec7ea63a0607afb42a9e833cf7b93c54: Status 404 returned error can't find the container with id de470e7b6b6edcad509453c874135919ec7ea63a0607afb42a9e833cf7b93c54 Dec 10 22:53:45 crc kubenswrapper[4791]: I1210 22:53:45.000750 4791 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"71bb4a3aecc4ba5b26c4b7318770ce13","Type":"ContainerStarted","Data":"de470e7b6b6edcad509453c874135919ec7ea63a0607afb42a9e833cf7b93c54"} Dec 10 22:53:45 crc kubenswrapper[4791]: I1210 22:53:45.888964 4791 status_manager.go:851] "Failed to get status for pod" podUID="71bb4a3aecc4ba5b26c4b7318770ce13" pod="openshift-kube-apiserver/kube-apiserver-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\": dial tcp 38.102.83.30:6443: connect: connection refused" Dec 10 22:53:45 crc kubenswrapper[4791]: I1210 22:53:45.889653 4791 status_manager.go:851] "Failed to get status for pod" podUID="f614b9022728cf315e60c057852e563e" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-controller-manager/pods/kube-controller-manager-crc\": dial tcp 38.102.83.30:6443: connect: connection refused" Dec 10 22:53:45 crc kubenswrapper[4791]: I1210 22:53:45.889966 4791 status_manager.go:851] "Failed to get status for pod" podUID="31bd0efc-25ef-4a1f-b4b9-e4c5b14768ea" pod="openshift-kube-apiserver/installer-9-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/installer-9-crc\": dial tcp 38.102.83.30:6443: connect: connection refused" Dec 10 22:53:46 crc kubenswrapper[4791]: I1210 22:53:46.008889 4791 generic.go:334] "Generic (PLEG): container finished" podID="71bb4a3aecc4ba5b26c4b7318770ce13" containerID="01560e75495ee72b8a7a409b50be098d5f94e6a396c20242a9ec0a17ab5706c8" exitCode=0 Dec 10 22:53:46 crc kubenswrapper[4791]: I1210 22:53:46.008963 4791 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"71bb4a3aecc4ba5b26c4b7318770ce13","Type":"ContainerDied","Data":"01560e75495ee72b8a7a409b50be098d5f94e6a396c20242a9ec0a17ab5706c8"} Dec 10 22:53:46 crc kubenswrapper[4791]: I1210 22:53:46.009209 4791 kubelet.go:1909] "Trying to delete pod" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="f8349fea-c49b-49a1-b5ee-032c5dffe021" Dec 10 22:53:46 crc kubenswrapper[4791]: I1210 22:53:46.009238 4791 mirror_client.go:130] "Deleting a mirror pod" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="f8349fea-c49b-49a1-b5ee-032c5dffe021" Dec 10 22:53:46 crc kubenswrapper[4791]: E1210 22:53:46.009816 4791 mirror_client.go:138] "Failed deleting a mirror pod" err="Delete \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\": dial tcp 38.102.83.30:6443: connect: connection refused" pod="openshift-kube-apiserver/kube-apiserver-crc" Dec 10 22:53:46 crc kubenswrapper[4791]: I1210 22:53:46.009822 4791 status_manager.go:851] "Failed to get status for pod" podUID="f614b9022728cf315e60c057852e563e" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-controller-manager/pods/kube-controller-manager-crc\": dial tcp 38.102.83.30:6443: connect: connection refused" Dec 10 22:53:46 crc kubenswrapper[4791]: I1210 22:53:46.010570 4791 status_manager.go:851] "Failed to get status for pod" podUID="71bb4a3aecc4ba5b26c4b7318770ce13" pod="openshift-kube-apiserver/kube-apiserver-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\": dial tcp 38.102.83.30:6443: connect: connection refused" Dec 10 22:53:46 crc kubenswrapper[4791]: I1210 22:53:46.010923 4791 status_manager.go:851] "Failed to get status for pod" podUID="31bd0efc-25ef-4a1f-b4b9-e4c5b14768ea" pod="openshift-kube-apiserver/installer-9-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/installer-9-crc\": dial tcp 38.102.83.30:6443: connect: connection refused" Dec 10 22:53:46 crc kubenswrapper[4791]: I1210 22:53:46.726093 4791 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Dec 10 22:53:47 crc kubenswrapper[4791]: I1210 22:53:47.019893 4791 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"71bb4a3aecc4ba5b26c4b7318770ce13","Type":"ContainerStarted","Data":"a8f3aec18d3482e4072f26b1df575a4c525967713866b125b51c0d7e09a3d2e2"} Dec 10 22:53:47 crc kubenswrapper[4791]: I1210 22:53:47.019935 4791 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"71bb4a3aecc4ba5b26c4b7318770ce13","Type":"ContainerStarted","Data":"f259cc9e117cb804944f3ced1cfc2c7e3d77652e09009c746a1f2eca580472d6"} Dec 10 22:53:47 crc kubenswrapper[4791]: I1210 22:53:47.019946 4791 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"71bb4a3aecc4ba5b26c4b7318770ce13","Type":"ContainerStarted","Data":"462f1b7be677e154664eecc3e7117dc48e21ac34eb53309362efd03aab7c38d0"} Dec 10 22:53:47 crc kubenswrapper[4791]: I1210 22:53:47.019957 4791 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"71bb4a3aecc4ba5b26c4b7318770ce13","Type":"ContainerStarted","Data":"51a811d1a50a68c3e634c0b5beed4f0d5f707cd9674e3f34c53837a09ced31e4"} Dec 10 22:53:47 crc kubenswrapper[4791]: I1210 22:53:47.890676 4791 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-authentication/oauth-openshift-558db77b4-ht5t7" podUID="35033f1a-b91b-4e02-ba77-3841a2e0fdab" containerName="oauth-openshift" containerID="cri-o://74fcc1d511780523d69f3aea9165e4405a91db746d4444a4393c970da15aa078" gracePeriod=15 Dec 10 22:53:48 crc kubenswrapper[4791]: I1210 22:53:48.026842 4791 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"71bb4a3aecc4ba5b26c4b7318770ce13","Type":"ContainerStarted","Data":"ef117272bf3c442e67c604a535d2700330237723942e294b3eb501db591ee293"} Dec 10 22:53:48 crc kubenswrapper[4791]: I1210 22:53:48.027099 4791 kubelet.go:1909] "Trying to delete pod" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="f8349fea-c49b-49a1-b5ee-032c5dffe021" Dec 10 22:53:48 crc kubenswrapper[4791]: I1210 22:53:48.027114 4791 mirror_client.go:130] "Deleting a mirror pod" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="f8349fea-c49b-49a1-b5ee-032c5dffe021" Dec 10 22:53:48 crc kubenswrapper[4791]: I1210 22:53:48.027332 4791 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-kube-apiserver/kube-apiserver-crc" Dec 10 22:53:48 crc kubenswrapper[4791]: I1210 22:53:48.029011 4791 generic.go:334] "Generic (PLEG): container finished" podID="35033f1a-b91b-4e02-ba77-3841a2e0fdab" containerID="74fcc1d511780523d69f3aea9165e4405a91db746d4444a4393c970da15aa078" exitCode=0 Dec 10 22:53:48 crc kubenswrapper[4791]: I1210 22:53:48.029037 4791 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-authentication/oauth-openshift-558db77b4-ht5t7" event={"ID":"35033f1a-b91b-4e02-ba77-3841a2e0fdab","Type":"ContainerDied","Data":"74fcc1d511780523d69f3aea9165e4405a91db746d4444a4393c970da15aa078"} Dec 10 22:53:48 crc kubenswrapper[4791]: I1210 22:53:48.271946 4791 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-authentication/oauth-openshift-558db77b4-ht5t7" Dec 10 22:53:48 crc kubenswrapper[4791]: I1210 22:53:48.412079 4791 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/35033f1a-b91b-4e02-ba77-3841a2e0fdab-v4-0-config-system-trusted-ca-bundle\") pod \"35033f1a-b91b-4e02-ba77-3841a2e0fdab\" (UID: \"35033f1a-b91b-4e02-ba77-3841a2e0fdab\") " Dec 10 22:53:48 crc kubenswrapper[4791]: I1210 22:53:48.412188 4791 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-dlncw\" (UniqueName: \"kubernetes.io/projected/35033f1a-b91b-4e02-ba77-3841a2e0fdab-kube-api-access-dlncw\") pod \"35033f1a-b91b-4e02-ba77-3841a2e0fdab\" (UID: \"35033f1a-b91b-4e02-ba77-3841a2e0fdab\") " Dec 10 22:53:48 crc kubenswrapper[4791]: I1210 22:53:48.412236 4791 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-ocp-branding-template\" (UniqueName: \"kubernetes.io/secret/35033f1a-b91b-4e02-ba77-3841a2e0fdab-v4-0-config-system-ocp-branding-template\") pod \"35033f1a-b91b-4e02-ba77-3841a2e0fdab\" (UID: \"35033f1a-b91b-4e02-ba77-3841a2e0fdab\") " Dec 10 22:53:48 crc kubenswrapper[4791]: I1210 22:53:48.412260 4791 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-user-template-error\" (UniqueName: \"kubernetes.io/secret/35033f1a-b91b-4e02-ba77-3841a2e0fdab-v4-0-config-user-template-error\") pod \"35033f1a-b91b-4e02-ba77-3841a2e0fdab\" (UID: \"35033f1a-b91b-4e02-ba77-3841a2e0fdab\") " Dec 10 22:53:48 crc kubenswrapper[4791]: I1210 22:53:48.412309 4791 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-cliconfig\" (UniqueName: \"kubernetes.io/configmap/35033f1a-b91b-4e02-ba77-3841a2e0fdab-v4-0-config-system-cliconfig\") pod \"35033f1a-b91b-4e02-ba77-3841a2e0fdab\" (UID: \"35033f1a-b91b-4e02-ba77-3841a2e0fdab\") " Dec 10 22:53:48 crc kubenswrapper[4791]: I1210 22:53:48.412331 4791 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-session\" (UniqueName: \"kubernetes.io/secret/35033f1a-b91b-4e02-ba77-3841a2e0fdab-v4-0-config-system-session\") pod \"35033f1a-b91b-4e02-ba77-3841a2e0fdab\" (UID: \"35033f1a-b91b-4e02-ba77-3841a2e0fdab\") " Dec 10 22:53:48 crc kubenswrapper[4791]: I1210 22:53:48.412362 4791 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-router-certs\" (UniqueName: \"kubernetes.io/secret/35033f1a-b91b-4e02-ba77-3841a2e0fdab-v4-0-config-system-router-certs\") pod \"35033f1a-b91b-4e02-ba77-3841a2e0fdab\" (UID: \"35033f1a-b91b-4e02-ba77-3841a2e0fdab\") " Dec 10 22:53:48 crc kubenswrapper[4791]: I1210 22:53:48.412399 4791 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-user-template-login\" (UniqueName: \"kubernetes.io/secret/35033f1a-b91b-4e02-ba77-3841a2e0fdab-v4-0-config-user-template-login\") pod \"35033f1a-b91b-4e02-ba77-3841a2e0fdab\" (UID: \"35033f1a-b91b-4e02-ba77-3841a2e0fdab\") " Dec 10 22:53:48 crc kubenswrapper[4791]: I1210 22:53:48.412419 4791 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-service-ca\" (UniqueName: \"kubernetes.io/configmap/35033f1a-b91b-4e02-ba77-3841a2e0fdab-v4-0-config-system-service-ca\") pod \"35033f1a-b91b-4e02-ba77-3841a2e0fdab\" (UID: \"35033f1a-b91b-4e02-ba77-3841a2e0fdab\") " Dec 10 22:53:48 crc kubenswrapper[4791]: I1210 22:53:48.412439 4791 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-user-template-provider-selection\" (UniqueName: \"kubernetes.io/secret/35033f1a-b91b-4e02-ba77-3841a2e0fdab-v4-0-config-user-template-provider-selection\") pod \"35033f1a-b91b-4e02-ba77-3841a2e0fdab\" (UID: \"35033f1a-b91b-4e02-ba77-3841a2e0fdab\") " Dec 10 22:53:48 crc kubenswrapper[4791]: I1210 22:53:48.412460 4791 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/35033f1a-b91b-4e02-ba77-3841a2e0fdab-audit-policies\") pod \"35033f1a-b91b-4e02-ba77-3841a2e0fdab\" (UID: \"35033f1a-b91b-4e02-ba77-3841a2e0fdab\") " Dec 10 22:53:48 crc kubenswrapper[4791]: I1210 22:53:48.412534 4791 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-serving-cert\" (UniqueName: \"kubernetes.io/secret/35033f1a-b91b-4e02-ba77-3841a2e0fdab-v4-0-config-system-serving-cert\") pod \"35033f1a-b91b-4e02-ba77-3841a2e0fdab\" (UID: \"35033f1a-b91b-4e02-ba77-3841a2e0fdab\") " Dec 10 22:53:48 crc kubenswrapper[4791]: I1210 22:53:48.412564 4791 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-user-idp-0-file-data\" (UniqueName: \"kubernetes.io/secret/35033f1a-b91b-4e02-ba77-3841a2e0fdab-v4-0-config-user-idp-0-file-data\") pod \"35033f1a-b91b-4e02-ba77-3841a2e0fdab\" (UID: \"35033f1a-b91b-4e02-ba77-3841a2e0fdab\") " Dec 10 22:53:48 crc kubenswrapper[4791]: I1210 22:53:48.412604 4791 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/35033f1a-b91b-4e02-ba77-3841a2e0fdab-audit-dir\") pod \"35033f1a-b91b-4e02-ba77-3841a2e0fdab\" (UID: \"35033f1a-b91b-4e02-ba77-3841a2e0fdab\") " Dec 10 22:53:48 crc kubenswrapper[4791]: I1210 22:53:48.412910 4791 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/35033f1a-b91b-4e02-ba77-3841a2e0fdab-audit-dir" (OuterVolumeSpecName: "audit-dir") pod "35033f1a-b91b-4e02-ba77-3841a2e0fdab" (UID: "35033f1a-b91b-4e02-ba77-3841a2e0fdab"). InnerVolumeSpecName "audit-dir". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 10 22:53:48 crc kubenswrapper[4791]: I1210 22:53:48.414104 4791 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/35033f1a-b91b-4e02-ba77-3841a2e0fdab-v4-0-config-system-trusted-ca-bundle" (OuterVolumeSpecName: "v4-0-config-system-trusted-ca-bundle") pod "35033f1a-b91b-4e02-ba77-3841a2e0fdab" (UID: "35033f1a-b91b-4e02-ba77-3841a2e0fdab"). InnerVolumeSpecName "v4-0-config-system-trusted-ca-bundle". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 10 22:53:48 crc kubenswrapper[4791]: I1210 22:53:48.415614 4791 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/35033f1a-b91b-4e02-ba77-3841a2e0fdab-v4-0-config-system-cliconfig" (OuterVolumeSpecName: "v4-0-config-system-cliconfig") pod "35033f1a-b91b-4e02-ba77-3841a2e0fdab" (UID: "35033f1a-b91b-4e02-ba77-3841a2e0fdab"). InnerVolumeSpecName "v4-0-config-system-cliconfig". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 10 22:53:48 crc kubenswrapper[4791]: I1210 22:53:48.415869 4791 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/35033f1a-b91b-4e02-ba77-3841a2e0fdab-audit-policies" (OuterVolumeSpecName: "audit-policies") pod "35033f1a-b91b-4e02-ba77-3841a2e0fdab" (UID: "35033f1a-b91b-4e02-ba77-3841a2e0fdab"). InnerVolumeSpecName "audit-policies". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 10 22:53:48 crc kubenswrapper[4791]: I1210 22:53:48.417078 4791 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/35033f1a-b91b-4e02-ba77-3841a2e0fdab-v4-0-config-system-service-ca" (OuterVolumeSpecName: "v4-0-config-system-service-ca") pod "35033f1a-b91b-4e02-ba77-3841a2e0fdab" (UID: "35033f1a-b91b-4e02-ba77-3841a2e0fdab"). InnerVolumeSpecName "v4-0-config-system-service-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 10 22:53:48 crc kubenswrapper[4791]: I1210 22:53:48.422070 4791 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/35033f1a-b91b-4e02-ba77-3841a2e0fdab-kube-api-access-dlncw" (OuterVolumeSpecName: "kube-api-access-dlncw") pod "35033f1a-b91b-4e02-ba77-3841a2e0fdab" (UID: "35033f1a-b91b-4e02-ba77-3841a2e0fdab"). InnerVolumeSpecName "kube-api-access-dlncw". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 10 22:53:48 crc kubenswrapper[4791]: I1210 22:53:48.422456 4791 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/35033f1a-b91b-4e02-ba77-3841a2e0fdab-v4-0-config-system-session" (OuterVolumeSpecName: "v4-0-config-system-session") pod "35033f1a-b91b-4e02-ba77-3841a2e0fdab" (UID: "35033f1a-b91b-4e02-ba77-3841a2e0fdab"). InnerVolumeSpecName "v4-0-config-system-session". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 10 22:53:48 crc kubenswrapper[4791]: I1210 22:53:48.422743 4791 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/35033f1a-b91b-4e02-ba77-3841a2e0fdab-v4-0-config-user-template-error" (OuterVolumeSpecName: "v4-0-config-user-template-error") pod "35033f1a-b91b-4e02-ba77-3841a2e0fdab" (UID: "35033f1a-b91b-4e02-ba77-3841a2e0fdab"). InnerVolumeSpecName "v4-0-config-user-template-error". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 10 22:53:48 crc kubenswrapper[4791]: I1210 22:53:48.424150 4791 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/35033f1a-b91b-4e02-ba77-3841a2e0fdab-v4-0-config-system-router-certs" (OuterVolumeSpecName: "v4-0-config-system-router-certs") pod "35033f1a-b91b-4e02-ba77-3841a2e0fdab" (UID: "35033f1a-b91b-4e02-ba77-3841a2e0fdab"). InnerVolumeSpecName "v4-0-config-system-router-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 10 22:53:48 crc kubenswrapper[4791]: I1210 22:53:48.424734 4791 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/35033f1a-b91b-4e02-ba77-3841a2e0fdab-v4-0-config-user-template-login" (OuterVolumeSpecName: "v4-0-config-user-template-login") pod "35033f1a-b91b-4e02-ba77-3841a2e0fdab" (UID: "35033f1a-b91b-4e02-ba77-3841a2e0fdab"). InnerVolumeSpecName "v4-0-config-user-template-login". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 10 22:53:48 crc kubenswrapper[4791]: I1210 22:53:48.424401 4791 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/35033f1a-b91b-4e02-ba77-3841a2e0fdab-v4-0-config-system-serving-cert" (OuterVolumeSpecName: "v4-0-config-system-serving-cert") pod "35033f1a-b91b-4e02-ba77-3841a2e0fdab" (UID: "35033f1a-b91b-4e02-ba77-3841a2e0fdab"). InnerVolumeSpecName "v4-0-config-system-serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 10 22:53:48 crc kubenswrapper[4791]: I1210 22:53:48.425039 4791 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/35033f1a-b91b-4e02-ba77-3841a2e0fdab-v4-0-config-system-ocp-branding-template" (OuterVolumeSpecName: "v4-0-config-system-ocp-branding-template") pod "35033f1a-b91b-4e02-ba77-3841a2e0fdab" (UID: "35033f1a-b91b-4e02-ba77-3841a2e0fdab"). InnerVolumeSpecName "v4-0-config-system-ocp-branding-template". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 10 22:53:48 crc kubenswrapper[4791]: I1210 22:53:48.425185 4791 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/35033f1a-b91b-4e02-ba77-3841a2e0fdab-v4-0-config-user-template-provider-selection" (OuterVolumeSpecName: "v4-0-config-user-template-provider-selection") pod "35033f1a-b91b-4e02-ba77-3841a2e0fdab" (UID: "35033f1a-b91b-4e02-ba77-3841a2e0fdab"). InnerVolumeSpecName "v4-0-config-user-template-provider-selection". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 10 22:53:48 crc kubenswrapper[4791]: I1210 22:53:48.425106 4791 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/35033f1a-b91b-4e02-ba77-3841a2e0fdab-v4-0-config-user-idp-0-file-data" (OuterVolumeSpecName: "v4-0-config-user-idp-0-file-data") pod "35033f1a-b91b-4e02-ba77-3841a2e0fdab" (UID: "35033f1a-b91b-4e02-ba77-3841a2e0fdab"). InnerVolumeSpecName "v4-0-config-user-idp-0-file-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 10 22:53:48 crc kubenswrapper[4791]: I1210 22:53:48.513677 4791 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-user-template-login\" (UniqueName: \"kubernetes.io/secret/35033f1a-b91b-4e02-ba77-3841a2e0fdab-v4-0-config-user-template-login\") on node \"crc\" DevicePath \"\"" Dec 10 22:53:48 crc kubenswrapper[4791]: I1210 22:53:48.513726 4791 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-service-ca\" (UniqueName: \"kubernetes.io/configmap/35033f1a-b91b-4e02-ba77-3841a2e0fdab-v4-0-config-system-service-ca\") on node \"crc\" DevicePath \"\"" Dec 10 22:53:48 crc kubenswrapper[4791]: I1210 22:53:48.513738 4791 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-user-template-provider-selection\" (UniqueName: \"kubernetes.io/secret/35033f1a-b91b-4e02-ba77-3841a2e0fdab-v4-0-config-user-template-provider-selection\") on node \"crc\" DevicePath \"\"" Dec 10 22:53:48 crc kubenswrapper[4791]: I1210 22:53:48.513750 4791 reconciler_common.go:293] "Volume detached for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/35033f1a-b91b-4e02-ba77-3841a2e0fdab-audit-policies\") on node \"crc\" DevicePath \"\"" Dec 10 22:53:48 crc kubenswrapper[4791]: I1210 22:53:48.513759 4791 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-serving-cert\" (UniqueName: \"kubernetes.io/secret/35033f1a-b91b-4e02-ba77-3841a2e0fdab-v4-0-config-system-serving-cert\") on node \"crc\" DevicePath \"\"" Dec 10 22:53:48 crc kubenswrapper[4791]: I1210 22:53:48.513769 4791 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-user-idp-0-file-data\" (UniqueName: \"kubernetes.io/secret/35033f1a-b91b-4e02-ba77-3841a2e0fdab-v4-0-config-user-idp-0-file-data\") on node \"crc\" DevicePath \"\"" Dec 10 22:53:48 crc kubenswrapper[4791]: I1210 22:53:48.513777 4791 reconciler_common.go:293] "Volume detached for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/35033f1a-b91b-4e02-ba77-3841a2e0fdab-audit-dir\") on node \"crc\" DevicePath \"\"" Dec 10 22:53:48 crc kubenswrapper[4791]: I1210 22:53:48.513786 4791 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/35033f1a-b91b-4e02-ba77-3841a2e0fdab-v4-0-config-system-trusted-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 10 22:53:48 crc kubenswrapper[4791]: I1210 22:53:48.513796 4791 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-dlncw\" (UniqueName: \"kubernetes.io/projected/35033f1a-b91b-4e02-ba77-3841a2e0fdab-kube-api-access-dlncw\") on node \"crc\" DevicePath \"\"" Dec 10 22:53:48 crc kubenswrapper[4791]: I1210 22:53:48.513804 4791 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-user-template-error\" (UniqueName: \"kubernetes.io/secret/35033f1a-b91b-4e02-ba77-3841a2e0fdab-v4-0-config-user-template-error\") on node \"crc\" DevicePath \"\"" Dec 10 22:53:48 crc kubenswrapper[4791]: I1210 22:53:48.513813 4791 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-ocp-branding-template\" (UniqueName: \"kubernetes.io/secret/35033f1a-b91b-4e02-ba77-3841a2e0fdab-v4-0-config-system-ocp-branding-template\") on node \"crc\" DevicePath \"\"" Dec 10 22:53:48 crc kubenswrapper[4791]: I1210 22:53:48.513821 4791 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-cliconfig\" (UniqueName: \"kubernetes.io/configmap/35033f1a-b91b-4e02-ba77-3841a2e0fdab-v4-0-config-system-cliconfig\") on node \"crc\" DevicePath \"\"" Dec 10 22:53:48 crc kubenswrapper[4791]: I1210 22:53:48.513830 4791 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-session\" (UniqueName: \"kubernetes.io/secret/35033f1a-b91b-4e02-ba77-3841a2e0fdab-v4-0-config-system-session\") on node \"crc\" DevicePath \"\"" Dec 10 22:53:48 crc kubenswrapper[4791]: I1210 22:53:48.513838 4791 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-router-certs\" (UniqueName: \"kubernetes.io/secret/35033f1a-b91b-4e02-ba77-3841a2e0fdab-v4-0-config-system-router-certs\") on node \"crc\" DevicePath \"\"" Dec 10 22:53:49 crc kubenswrapper[4791]: I1210 22:53:49.040515 4791 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-authentication/oauth-openshift-558db77b4-ht5t7" event={"ID":"35033f1a-b91b-4e02-ba77-3841a2e0fdab","Type":"ContainerDied","Data":"6100b3fe0d820fdba5310d4b39a19699ff6808a46c982f7731462ffd7f68c89c"} Dec 10 22:53:49 crc kubenswrapper[4791]: I1210 22:53:49.041151 4791 scope.go:117] "RemoveContainer" containerID="74fcc1d511780523d69f3aea9165e4405a91db746d4444a4393c970da15aa078" Dec 10 22:53:49 crc kubenswrapper[4791]: I1210 22:53:49.041083 4791 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-authentication/oauth-openshift-558db77b4-ht5t7" Dec 10 22:53:49 crc kubenswrapper[4791]: I1210 22:53:49.900127 4791 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-kube-apiserver/kube-apiserver-crc" Dec 10 22:53:49 crc kubenswrapper[4791]: I1210 22:53:49.900600 4791 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-kube-apiserver/kube-apiserver-crc" Dec 10 22:53:49 crc kubenswrapper[4791]: I1210 22:53:49.905467 4791 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-kube-apiserver/kube-apiserver-crc" Dec 10 22:53:51 crc kubenswrapper[4791]: I1210 22:53:51.563386 4791 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Dec 10 22:53:51 crc kubenswrapper[4791]: I1210 22:53:51.563629 4791 patch_prober.go:28] interesting pod/kube-controller-manager-crc container/kube-controller-manager namespace/openshift-kube-controller-manager: Startup probe status=failure output="Get \"https://192.168.126.11:10257/healthz\": dial tcp 192.168.126.11:10257: connect: connection refused" start-of-body= Dec 10 22:53:51 crc kubenswrapper[4791]: I1210 22:53:51.563682 4791 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-kube-controller-manager/kube-controller-manager-crc" podUID="f614b9022728cf315e60c057852e563e" containerName="kube-controller-manager" probeResult="failure" output="Get \"https://192.168.126.11:10257/healthz\": dial tcp 192.168.126.11:10257: connect: connection refused" Dec 10 22:53:53 crc kubenswrapper[4791]: I1210 22:53:53.042641 4791 kubelet.go:1914] "Deleted mirror pod because it is outdated" pod="openshift-kube-apiserver/kube-apiserver-crc" Dec 10 22:53:53 crc kubenswrapper[4791]: I1210 22:53:53.101697 4791 status_manager.go:861] "Pod was deleted and then recreated, skipping status update" pod="openshift-kube-apiserver/kube-apiserver-crc" oldPodUID="71bb4a3aecc4ba5b26c4b7318770ce13" podUID="29eafd3a-8e53-420e-b50e-8189d8b5f9f6" Dec 10 22:53:53 crc kubenswrapper[4791]: E1210 22:53:53.314643 4791 reflector.go:158] "Unhandled Error" err="object-\"openshift-authentication\"/\"v4-0-config-system-router-certs\": Failed to watch *v1.Secret: unknown (get secrets)" logger="UnhandledError" Dec 10 22:53:53 crc kubenswrapper[4791]: E1210 22:53:53.347237 4791 reflector.go:158] "Unhandled Error" err="object-\"openshift-authentication\"/\"oauth-openshift-dockercfg-znhcc\": Failed to watch *v1.Secret: unknown (get secrets)" logger="UnhandledError" Dec 10 22:53:53 crc kubenswrapper[4791]: E1210 22:53:53.352831 4791 reflector.go:158] "Unhandled Error" err="object-\"openshift-authentication\"/\"v4-0-config-system-serving-cert\": Failed to watch *v1.Secret: unknown (get secrets)" logger="UnhandledError" Dec 10 22:53:53 crc kubenswrapper[4791]: E1210 22:53:53.441044 4791 reflector.go:158] "Unhandled Error" err="object-\"openshift-authentication\"/\"v4-0-config-system-trusted-ca-bundle\": Failed to watch *v1.ConfigMap: unknown (get configmaps)" logger="UnhandledError" Dec 10 22:53:53 crc kubenswrapper[4791]: E1210 22:53:53.551521 4791 reflector.go:158] "Unhandled Error" err="object-\"openshift-authentication\"/\"openshift-service-ca.crt\": Failed to watch *v1.ConfigMap: unknown (get configmaps)" logger="UnhandledError" Dec 10 22:53:54 crc kubenswrapper[4791]: I1210 22:53:54.071621 4791 kubelet.go:1909] "Trying to delete pod" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="f8349fea-c49b-49a1-b5ee-032c5dffe021" Dec 10 22:53:54 crc kubenswrapper[4791]: I1210 22:53:54.071916 4791 mirror_client.go:130] "Deleting a mirror pod" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="f8349fea-c49b-49a1-b5ee-032c5dffe021" Dec 10 22:53:54 crc kubenswrapper[4791]: I1210 22:53:54.076542 4791 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-kube-apiserver/kube-apiserver-crc" Dec 10 22:53:54 crc kubenswrapper[4791]: I1210 22:53:54.076681 4791 status_manager.go:861] "Pod was deleted and then recreated, skipping status update" pod="openshift-kube-apiserver/kube-apiserver-crc" oldPodUID="71bb4a3aecc4ba5b26c4b7318770ce13" podUID="29eafd3a-8e53-420e-b50e-8189d8b5f9f6" Dec 10 22:53:55 crc kubenswrapper[4791]: I1210 22:53:55.079861 4791 kubelet.go:1909] "Trying to delete pod" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="f8349fea-c49b-49a1-b5ee-032c5dffe021" Dec 10 22:53:55 crc kubenswrapper[4791]: I1210 22:53:55.079893 4791 mirror_client.go:130] "Deleting a mirror pod" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="f8349fea-c49b-49a1-b5ee-032c5dffe021" Dec 10 22:53:55 crc kubenswrapper[4791]: I1210 22:53:55.083860 4791 status_manager.go:861] "Pod was deleted and then recreated, skipping status update" pod="openshift-kube-apiserver/kube-apiserver-crc" oldPodUID="71bb4a3aecc4ba5b26c4b7318770ce13" podUID="29eafd3a-8e53-420e-b50e-8189d8b5f9f6" Dec 10 22:54:01 crc kubenswrapper[4791]: I1210 22:54:01.563832 4791 patch_prober.go:28] interesting pod/kube-controller-manager-crc container/kube-controller-manager namespace/openshift-kube-controller-manager: Startup probe status=failure output="Get \"https://192.168.126.11:10257/healthz\": dial tcp 192.168.126.11:10257: connect: connection refused" start-of-body= Dec 10 22:54:01 crc kubenswrapper[4791]: I1210 22:54:01.564937 4791 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-kube-controller-manager/kube-controller-manager-crc" podUID="f614b9022728cf315e60c057852e563e" containerName="kube-controller-manager" probeResult="failure" output="Get \"https://192.168.126.11:10257/healthz\": dial tcp 192.168.126.11:10257: connect: connection refused" Dec 10 22:54:03 crc kubenswrapper[4791]: I1210 22:54:03.410397 4791 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-scheduler-operator"/"kube-root-ca.crt" Dec 10 22:54:03 crc kubenswrapper[4791]: I1210 22:54:03.659483 4791 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-multus"/"default-dockercfg-2q5b6" Dec 10 22:54:03 crc kubenswrapper[4791]: I1210 22:54:03.745885 4791 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress-operator"/"kube-root-ca.crt" Dec 10 22:54:03 crc kubenswrapper[4791]: I1210 22:54:03.794062 4791 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-cluster-samples-operator"/"cluster-samples-operator-dockercfg-xpp9w" Dec 10 22:54:04 crc kubenswrapper[4791]: I1210 22:54:04.027576 4791 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"kube-root-ca.crt" Dec 10 22:54:04 crc kubenswrapper[4791]: I1210 22:54:04.205238 4791 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-multus"/"kube-root-ca.crt" Dec 10 22:54:04 crc kubenswrapper[4791]: I1210 22:54:04.495505 4791 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver"/"audit-1" Dec 10 22:54:04 crc kubenswrapper[4791]: I1210 22:54:04.702262 4791 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-etcd-operator"/"etcd-operator-serving-cert" Dec 10 22:54:04 crc kubenswrapper[4791]: I1210 22:54:04.750150 4791 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-apiserver-operator"/"kube-apiserver-operator-serving-cert" Dec 10 22:54:04 crc kubenswrapper[4791]: I1210 22:54:04.786386 4791 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-image-registry"/"registry-dockercfg-kzzsd" Dec 10 22:54:04 crc kubenswrapper[4791]: I1210 22:54:04.830269 4791 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console"/"kube-root-ca.crt" Dec 10 22:54:04 crc kubenswrapper[4791]: I1210 22:54:04.955938 4791 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-image-registry"/"installation-pull-secrets" Dec 10 22:54:05 crc kubenswrapper[4791]: I1210 22:54:05.078299 4791 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-controller-manager"/"serving-cert" Dec 10 22:54:05 crc kubenswrapper[4791]: I1210 22:54:05.216118 4791 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-controller-manager-operator"/"openshift-controller-manager-operator-dockercfg-vw8fw" Dec 10 22:54:05 crc kubenswrapper[4791]: I1210 22:54:05.264043 4791 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress"/"router-stats-default" Dec 10 22:54:05 crc kubenswrapper[4791]: I1210 22:54:05.360086 4791 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-dns"/"node-resolver-dockercfg-kz9s7" Dec 10 22:54:05 crc kubenswrapper[4791]: I1210 22:54:05.417054 4791 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ovn-kubernetes"/"ovn-kubernetes-node-dockercfg-pwtwl" Dec 10 22:54:05 crc kubenswrapper[4791]: I1210 22:54:05.450450 4791 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-image-registry"/"image-registry-tls" Dec 10 22:54:05 crc kubenswrapper[4791]: I1210 22:54:05.682973 4791 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-operator-lifecycle-manager"/"kube-root-ca.crt" Dec 10 22:54:05 crc kubenswrapper[4791]: I1210 22:54:05.766150 4791 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress-canary"/"default-dockercfg-2llfx" Dec 10 22:54:05 crc kubenswrapper[4791]: I1210 22:54:05.766378 4791 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver"/"trusted-ca-bundle" Dec 10 22:54:06 crc kubenswrapper[4791]: I1210 22:54:06.039947 4791 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-network-operator"/"metrics-tls" Dec 10 22:54:06 crc kubenswrapper[4791]: I1210 22:54:06.184651 4791 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ovn-kubernetes"/"openshift-service-ca.crt" Dec 10 22:54:06 crc kubenswrapper[4791]: I1210 22:54:06.191500 4791 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-oauth-apiserver"/"etcd-serving-ca" Dec 10 22:54:06 crc kubenswrapper[4791]: I1210 22:54:06.241234 4791 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console"/"openshift-service-ca.crt" Dec 10 22:54:06 crc kubenswrapper[4791]: I1210 22:54:06.247897 4791 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-etcd-operator"/"etcd-client" Dec 10 22:54:06 crc kubenswrapper[4791]: I1210 22:54:06.254454 4791 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-route-controller-manager"/"serving-cert" Dec 10 22:54:06 crc kubenswrapper[4791]: I1210 22:54:06.268045 4791 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-samples-operator"/"openshift-service-ca.crt" Dec 10 22:54:06 crc kubenswrapper[4791]: I1210 22:54:06.312807 4791 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress"/"router-certs-default" Dec 10 22:54:06 crc kubenswrapper[4791]: I1210 22:54:06.330120 4791 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-oauth-apiserver"/"openshift-service-ca.crt" Dec 10 22:54:06 crc kubenswrapper[4791]: I1210 22:54:06.373535 4791 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-config-operator"/"machine-config-operator-images" Dec 10 22:54:06 crc kubenswrapper[4791]: I1210 22:54:06.444497 4791 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"redhat-operators-dockercfg-ct8rh" Dec 10 22:54:06 crc kubenswrapper[4791]: I1210 22:54:06.471238 4791 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-storage-version-migrator-operator"/"openshift-service-ca.crt" Dec 10 22:54:06 crc kubenswrapper[4791]: I1210 22:54:06.492365 4791 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-scheduler-operator"/"kube-scheduler-operator-serving-cert" Dec 10 22:54:06 crc kubenswrapper[4791]: I1210 22:54:06.544308 4791 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-api"/"machine-api-operator-dockercfg-mfbb7" Dec 10 22:54:06 crc kubenswrapper[4791]: I1210 22:54:06.581463 4791 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-config-operator"/"kube-rbac-proxy" Dec 10 22:54:06 crc kubenswrapper[4791]: I1210 22:54:06.623506 4791 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress"/"router-dockercfg-zdk86" Dec 10 22:54:06 crc kubenswrapper[4791]: I1210 22:54:06.631087 4791 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-node-identity"/"env-overrides" Dec 10 22:54:06 crc kubenswrapper[4791]: I1210 22:54:06.635076 4791 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-controller-manager-operator"/"kube-controller-manager-operator-serving-cert" Dec 10 22:54:06 crc kubenswrapper[4791]: I1210 22:54:06.674903 4791 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"openshift-service-ca.crt" Dec 10 22:54:06 crc kubenswrapper[4791]: I1210 22:54:06.691638 4791 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-image-registry"/"openshift-service-ca.crt" Dec 10 22:54:06 crc kubenswrapper[4791]: I1210 22:54:06.788512 4791 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-marketplace"/"marketplace-trusted-ca" Dec 10 22:54:06 crc kubenswrapper[4791]: I1210 22:54:06.821366 4791 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-service-ca-operator"/"service-ca-operator-config" Dec 10 22:54:06 crc kubenswrapper[4791]: I1210 22:54:06.863382 4791 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"openshift-service-ca.crt" Dec 10 22:54:06 crc kubenswrapper[4791]: I1210 22:54:06.893345 4791 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-cluster-samples-operator"/"samples-operator-tls" Dec 10 22:54:06 crc kubenswrapper[4791]: I1210 22:54:06.946789 4791 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-oauth-apiserver"/"oauth-apiserver-sa-dockercfg-6r2bq" Dec 10 22:54:07 crc kubenswrapper[4791]: I1210 22:54:07.031591 4791 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-service-ca-operator"/"serving-cert" Dec 10 22:54:07 crc kubenswrapper[4791]: I1210 22:54:07.036263 4791 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-machine-approver"/"machine-approver-config" Dec 10 22:54:07 crc kubenswrapper[4791]: I1210 22:54:07.117022 4791 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-oauth-apiserver"/"kube-root-ca.crt" Dec 10 22:54:07 crc kubenswrapper[4791]: I1210 22:54:07.310073 4791 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress-canary"/"canary-serving-cert" Dec 10 22:54:07 crc kubenswrapper[4791]: I1210 22:54:07.352500 4791 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"marketplace-operator-dockercfg-5nsgg" Dec 10 22:54:07 crc kubenswrapper[4791]: I1210 22:54:07.378577 4791 reflector.go:368] Caches populated for *v1.RuntimeClass from k8s.io/client-go/informers/factory.go:160 Dec 10 22:54:07 crc kubenswrapper[4791]: I1210 22:54:07.479231 4791 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress-operator"/"openshift-service-ca.crt" Dec 10 22:54:07 crc kubenswrapper[4791]: I1210 22:54:07.619380 4791 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console"/"console-config" Dec 10 22:54:07 crc kubenswrapper[4791]: I1210 22:54:07.650970 4791 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-storage-version-migrator-operator"/"kube-root-ca.crt" Dec 10 22:54:07 crc kubenswrapper[4791]: I1210 22:54:07.771445 4791 reflector.go:368] Caches populated for *v1.ConfigMap from object-"hostpath-provisioner"/"kube-root-ca.crt" Dec 10 22:54:07 crc kubenswrapper[4791]: I1210 22:54:07.833199 4791 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-scheduler-operator"/"openshift-kube-scheduler-operator-dockercfg-qt55r" Dec 10 22:54:07 crc kubenswrapper[4791]: I1210 22:54:07.869732 4791 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-image-registry"/"image-registry-operator-tls" Dec 10 22:54:07 crc kubenswrapper[4791]: I1210 22:54:07.990706 4791 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-operator"/"kube-root-ca.crt" Dec 10 22:54:08 crc kubenswrapper[4791]: I1210 22:54:08.090500 4791 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console-operator"/"kube-root-ca.crt" Dec 10 22:54:08 crc kubenswrapper[4791]: I1210 22:54:08.264514 4791 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-storage-version-migrator"/"kube-root-ca.crt" Dec 10 22:54:08 crc kubenswrapper[4791]: I1210 22:54:08.375110 4791 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-controller-manager-operator"/"kube-root-ca.crt" Dec 10 22:54:08 crc kubenswrapper[4791]: I1210 22:54:08.390745 4791 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-storage-version-migrator"/"kube-storage-version-migrator-sa-dockercfg-5xfcg" Dec 10 22:54:08 crc kubenswrapper[4791]: I1210 22:54:08.410171 4791 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-oauth-apiserver"/"audit-1" Dec 10 22:54:08 crc kubenswrapper[4791]: I1210 22:54:08.495555 4791 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-api"/"kube-root-ca.crt" Dec 10 22:54:08 crc kubenswrapper[4791]: I1210 22:54:08.517975 4791 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console"/"service-ca" Dec 10 22:54:08 crc kubenswrapper[4791]: I1210 22:54:08.553765 4791 reflector.go:368] Caches populated for *v1.Secret from object-"hostpath-provisioner"/"csi-hostpath-provisioner-sa-dockercfg-qd74k" Dec 10 22:54:08 crc kubenswrapper[4791]: I1210 22:54:08.559143 4791 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ovn-kubernetes"/"ovn-node-metrics-cert" Dec 10 22:54:08 crc kubenswrapper[4791]: I1210 22:54:08.582971 4791 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ovn-kubernetes"/"ovnkube-config" Dec 10 22:54:08 crc kubenswrapper[4791]: I1210 22:54:08.634676 4791 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver"/"etcd-serving-ca" Dec 10 22:54:08 crc kubenswrapper[4791]: I1210 22:54:08.758036 4791 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress-operator"/"ingress-operator-dockercfg-7lnqk" Dec 10 22:54:09 crc kubenswrapper[4791]: I1210 22:54:09.006760 4791 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-etcd-operator"/"etcd-operator-dockercfg-r9srn" Dec 10 22:54:09 crc kubenswrapper[4791]: I1210 22:54:09.047837 4791 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"packageserver-service-cert" Dec 10 22:54:09 crc kubenswrapper[4791]: I1210 22:54:09.060661 4791 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-apiserver-operator"/"openshift-apiserver-operator-dockercfg-xtcjv" Dec 10 22:54:09 crc kubenswrapper[4791]: I1210 22:54:09.073659 4791 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-api"/"control-plane-machine-set-operator-tls" Dec 10 22:54:09 crc kubenswrapper[4791]: I1210 22:54:09.084295 4791 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-controller-manager-operator"/"openshift-controller-manager-operator-serving-cert" Dec 10 22:54:09 crc kubenswrapper[4791]: I1210 22:54:09.196440 4791 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-apiserver-operator"/"kube-apiserver-operator-config" Dec 10 22:54:09 crc kubenswrapper[4791]: I1210 22:54:09.198054 4791 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"olm-operator-serviceaccount-dockercfg-rq7zk" Dec 10 22:54:09 crc kubenswrapper[4791]: I1210 22:54:09.214081 4791 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-api"/"machine-api-operator-images" Dec 10 22:54:09 crc kubenswrapper[4791]: I1210 22:54:09.473186 4791 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console-operator"/"trusted-ca" Dec 10 22:54:09 crc kubenswrapper[4791]: I1210 22:54:09.584469 4791 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-node-identity"/"openshift-service-ca.crt" Dec 10 22:54:09 crc kubenswrapper[4791]: I1210 22:54:09.674683 4791 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-storage-version-migrator-operator"/"kube-storage-version-migrator-operator-dockercfg-2bh8d" Dec 10 22:54:09 crc kubenswrapper[4791]: I1210 22:54:09.702366 4791 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"kube-root-ca.crt" Dec 10 22:54:09 crc kubenswrapper[4791]: I1210 22:54:09.725615 4791 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"client-ca" Dec 10 22:54:09 crc kubenswrapper[4791]: I1210 22:54:09.782313 4791 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"machine-config-daemon-dockercfg-r5tcq" Dec 10 22:54:09 crc kubenswrapper[4791]: I1210 22:54:09.792890 4791 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-etcd-operator"/"etcd-service-ca-bundle" Dec 10 22:54:09 crc kubenswrapper[4791]: I1210 22:54:09.800294 4791 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"proxy-tls" Dec 10 22:54:10 crc kubenswrapper[4791]: I1210 22:54:10.010771 4791 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-api"/"machine-api-operator-tls" Dec 10 22:54:10 crc kubenswrapper[4791]: I1210 22:54:10.015780 4791 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console-operator"/"console-operator-config" Dec 10 22:54:10 crc kubenswrapper[4791]: I1210 22:54:10.019183 4791 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-operator"/"openshift-service-ca.crt" Dec 10 22:54:10 crc kubenswrapper[4791]: I1210 22:54:10.103592 4791 reflector.go:368] Caches populated for *v1.ConfigMap from object-"hostpath-provisioner"/"openshift-service-ca.crt" Dec 10 22:54:10 crc kubenswrapper[4791]: I1210 22:54:10.118932 4791 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-multus"/"metrics-daemon-sa-dockercfg-d427c" Dec 10 22:54:10 crc kubenswrapper[4791]: I1210 22:54:10.173622 4791 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-apiserver"/"encryption-config-1" Dec 10 22:54:10 crc kubenswrapper[4791]: I1210 22:54:10.190126 4791 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console-operator"/"openshift-service-ca.crt" Dec 10 22:54:10 crc kubenswrapper[4791]: I1210 22:54:10.191869 4791 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-service-ca-operator"/"openshift-service-ca.crt" Dec 10 22:54:10 crc kubenswrapper[4791]: I1210 22:54:10.235732 4791 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"machine-config-server-dockercfg-qx5rd" Dec 10 22:54:10 crc kubenswrapper[4791]: I1210 22:54:10.396598 4791 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-cluster-machine-approver"/"machine-approver-tls" Dec 10 22:54:10 crc kubenswrapper[4791]: I1210 22:54:10.400244 4791 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-etcd-operator"/"etcd-ca-bundle" Dec 10 22:54:10 crc kubenswrapper[4791]: I1210 22:54:10.413643 4791 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver"/"config" Dec 10 22:54:10 crc kubenswrapper[4791]: I1210 22:54:10.427070 4791 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-api"/"control-plane-machine-set-operator-dockercfg-k9rxt" Dec 10 22:54:10 crc kubenswrapper[4791]: I1210 22:54:10.492169 4791 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication-operator"/"serving-cert" Dec 10 22:54:10 crc kubenswrapper[4791]: I1210 22:54:10.558845 4791 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication-operator"/"kube-root-ca.crt" Dec 10 22:54:10 crc kubenswrapper[4791]: I1210 22:54:10.579857 4791 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-console"/"console-oauth-config" Dec 10 22:54:10 crc kubenswrapper[4791]: I1210 22:54:10.687866 4791 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"machine-config-operator-dockercfg-98p87" Dec 10 22:54:10 crc kubenswrapper[4791]: I1210 22:54:10.701739 4791 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-apiserver-operator"/"openshift-apiserver-operator-serving-cert" Dec 10 22:54:10 crc kubenswrapper[4791]: I1210 22:54:10.750144 4791 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"olm-operator-serving-cert" Dec 10 22:54:10 crc kubenswrapper[4791]: I1210 22:54:10.846277 4791 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-dns-operator"/"dns-operator-dockercfg-9mqw5" Dec 10 22:54:10 crc kubenswrapper[4791]: I1210 22:54:10.849002 4791 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication-operator"/"service-ca-bundle" Dec 10 22:54:10 crc kubenswrapper[4791]: I1210 22:54:10.867918 4791 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-apiserver-operator"/"kube-apiserver-operator-dockercfg-x57mr" Dec 10 22:54:10 crc kubenswrapper[4791]: I1210 22:54:10.879830 4791 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-config-operator"/"openshift-service-ca.crt" Dec 10 22:54:11 crc kubenswrapper[4791]: I1210 22:54:11.044847 4791 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager-operator"/"openshift-service-ca.crt" Dec 10 22:54:11 crc kubenswrapper[4791]: I1210 22:54:11.064215 4791 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"certified-operators-dockercfg-4rs5g" Dec 10 22:54:11 crc kubenswrapper[4791]: I1210 22:54:11.102197 4791 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ovn-kubernetes"/"ovn-control-plane-metrics-cert" Dec 10 22:54:11 crc kubenswrapper[4791]: I1210 22:54:11.145739 4791 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver-operator"/"openshift-service-ca.crt" Dec 10 22:54:11 crc kubenswrapper[4791]: I1210 22:54:11.246612 4791 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-controller-manager-operator"/"kube-controller-manager-operator-config" Dec 10 22:54:11 crc kubenswrapper[4791]: I1210 22:54:11.250293 4791 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"mco-proxy-tls" Dec 10 22:54:11 crc kubenswrapper[4791]: I1210 22:54:11.375212 4791 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-config-operator"/"kube-root-ca.crt" Dec 10 22:54:11 crc kubenswrapper[4791]: I1210 22:54:11.403858 4791 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-machine-approver"/"openshift-service-ca.crt" Dec 10 22:54:11 crc kubenswrapper[4791]: I1210 22:54:11.418271 4791 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"node-bootstrapper-token" Dec 10 22:54:11 crc kubenswrapper[4791]: I1210 22:54:11.485949 4791 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress"/"router-metrics-certs-default" Dec 10 22:54:11 crc kubenswrapper[4791]: I1210 22:54:11.568715 4791 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Dec 10 22:54:11 crc kubenswrapper[4791]: I1210 22:54:11.573271 4791 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Dec 10 22:54:11 crc kubenswrapper[4791]: I1210 22:54:11.590551 4791 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-version"/"openshift-service-ca.crt" Dec 10 22:54:11 crc kubenswrapper[4791]: I1210 22:54:11.628820 4791 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver"/"kube-root-ca.crt" Dec 10 22:54:11 crc kubenswrapper[4791]: I1210 22:54:11.752587 4791 reflector.go:368] Caches populated for *v1.Service from k8s.io/client-go/informers/factory.go:160 Dec 10 22:54:11 crc kubenswrapper[4791]: I1210 22:54:11.762890 4791 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-operator-lifecycle-manager"/"openshift-service-ca.crt" Dec 10 22:54:11 crc kubenswrapper[4791]: I1210 22:54:11.779886 4791 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver"/"image-import-ca" Dec 10 22:54:11 crc kubenswrapper[4791]: I1210 22:54:11.793812 4791 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-etcd-operator"/"kube-root-ca.crt" Dec 10 22:54:11 crc kubenswrapper[4791]: I1210 22:54:11.915889 4791 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-service-ca-operator"/"service-ca-operator-dockercfg-rg9jl" Dec 10 22:54:11 crc kubenswrapper[4791]: I1210 22:54:11.975154 4791 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"catalog-operator-serving-cert" Dec 10 22:54:12 crc kubenswrapper[4791]: I1210 22:54:12.071617 4791 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-multus"/"multus-daemon-config" Dec 10 22:54:12 crc kubenswrapper[4791]: I1210 22:54:12.155968 4791 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-diagnostics"/"kube-root-ca.crt" Dec 10 22:54:12 crc kubenswrapper[4791]: I1210 22:54:12.235694 4791 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-config-operator"/"config-operator-serving-cert" Dec 10 22:54:12 crc kubenswrapper[4791]: I1210 22:54:12.281166 4791 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"redhat-marketplace-dockercfg-x2ctb" Dec 10 22:54:12 crc kubenswrapper[4791]: I1210 22:54:12.308162 4791 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-multus"/"multus-ancillary-tools-dockercfg-vnmsz" Dec 10 22:54:12 crc kubenswrapper[4791]: I1210 22:54:12.358667 4791 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-cluster-machine-approver"/"machine-approver-sa-dockercfg-nl2j4" Dec 10 22:54:12 crc kubenswrapper[4791]: I1210 22:54:12.463457 4791 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-cluster-version"/"cluster-version-operator-serving-cert" Dec 10 22:54:12 crc kubenswrapper[4791]: I1210 22:54:12.469239 4791 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication-operator"/"authentication-operator-config" Dec 10 22:54:12 crc kubenswrapper[4791]: I1210 22:54:12.502364 4791 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-scheduler-operator"/"openshift-kube-scheduler-operator-config" Dec 10 22:54:12 crc kubenswrapper[4791]: I1210 22:54:12.503439 4791 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication-operator"/"trusted-ca-bundle" Dec 10 22:54:12 crc kubenswrapper[4791]: I1210 22:54:12.581801 4791 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress"/"service-ca-bundle" Dec 10 22:54:12 crc kubenswrapper[4791]: I1210 22:54:12.631352 4791 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-image-registry"/"trusted-ca" Dec 10 22:54:12 crc kubenswrapper[4791]: I1210 22:54:12.633059 4791 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"package-server-manager-serving-cert" Dec 10 22:54:12 crc kubenswrapper[4791]: I1210 22:54:12.708478 4791 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-config-operator"/"openshift-config-operator-dockercfg-7pc5z" Dec 10 22:54:12 crc kubenswrapper[4791]: I1210 22:54:12.847096 4791 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-apiserver"/"openshift-apiserver-sa-dockercfg-djjff" Dec 10 22:54:12 crc kubenswrapper[4791]: I1210 22:54:12.889623 4791 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-node-identity"/"ovnkube-identity-cm" Dec 10 22:54:12 crc kubenswrapper[4791]: I1210 22:54:12.891127 4791 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-service-ca"/"signing-key" Dec 10 22:54:12 crc kubenswrapper[4791]: I1210 22:54:12.906207 4791 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-multus"/"metrics-daemon-secret" Dec 10 22:54:13 crc kubenswrapper[4791]: I1210 22:54:13.011777 4791 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"openshift-global-ca" Dec 10 22:54:13 crc kubenswrapper[4791]: I1210 22:54:13.101926 4791 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"client-ca" Dec 10 22:54:13 crc kubenswrapper[4791]: I1210 22:54:13.203371 4791 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-machine-approver"/"kube-rbac-proxy" Dec 10 22:54:13 crc kubenswrapper[4791]: I1210 22:54:13.267311 4791 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-route-controller-manager"/"route-controller-manager-sa-dockercfg-h2zr2" Dec 10 22:54:13 crc kubenswrapper[4791]: I1210 22:54:13.273507 4791 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console"/"trusted-ca-bundle" Dec 10 22:54:13 crc kubenswrapper[4791]: I1210 22:54:13.333317 4791 reflector.go:368] Caches populated for *v1.Pod from pkg/kubelet/config/apiserver.go:66 Dec 10 22:54:13 crc kubenswrapper[4791]: I1210 22:54:13.337757 4791 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-kube-apiserver/kube-apiserver-crc","openshift-authentication/oauth-openshift-558db77b4-ht5t7"] Dec 10 22:54:13 crc kubenswrapper[4791]: I1210 22:54:13.337836 4791 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-kube-apiserver/kube-apiserver-crc"] Dec 10 22:54:13 crc kubenswrapper[4791]: I1210 22:54:13.342842 4791 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-kube-apiserver/kube-apiserver-crc" Dec 10 22:54:13 crc kubenswrapper[4791]: I1210 22:54:13.352880 4791 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication-operator"/"openshift-service-ca.crt" Dec 10 22:54:13 crc kubenswrapper[4791]: I1210 22:54:13.361391 4791 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-kube-apiserver/kube-apiserver-crc" podStartSLOduration=20.361372671 podStartE2EDuration="20.361372671s" podCreationTimestamp="2025-12-10 22:53:53 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-10 22:54:13.359520876 +0000 UTC m=+287.789138489" watchObservedRunningTime="2025-12-10 22:54:13.361372671 +0000 UTC m=+287.790990284" Dec 10 22:54:13 crc kubenswrapper[4791]: I1210 22:54:13.447076 4791 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-network-console"/"networking-console-plugin-cert" Dec 10 22:54:13 crc kubenswrapper[4791]: I1210 22:54:13.461069 4791 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-multus"/"multus-admission-controller-secret" Dec 10 22:54:13 crc kubenswrapper[4791]: I1210 22:54:13.473303 4791 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-service-ca"/"kube-root-ca.crt" Dec 10 22:54:13 crc kubenswrapper[4791]: I1210 22:54:13.501088 4791 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-api"/"kube-rbac-proxy" Dec 10 22:54:13 crc kubenswrapper[4791]: I1210 22:54:13.548070 4791 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-multus"/"cni-copy-resources" Dec 10 22:54:13 crc kubenswrapper[4791]: I1210 22:54:13.556249 4791 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-marketplace"/"openshift-service-ca.crt" Dec 10 22:54:13 crc kubenswrapper[4791]: I1210 22:54:13.584124 4791 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-service-ca"/"signing-cabundle" Dec 10 22:54:13 crc kubenswrapper[4791]: I1210 22:54:13.594373 4791 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-dns"/"dns-default" Dec 10 22:54:13 crc kubenswrapper[4791]: I1210 22:54:13.610100 4791 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-console"/"console-serving-cert" Dec 10 22:54:13 crc kubenswrapper[4791]: I1210 22:54:13.682424 4791 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-console-operator"/"serving-cert" Dec 10 22:54:13 crc kubenswrapper[4791]: I1210 22:54:13.731478 4791 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-controller-manager-operator"/"kube-controller-manager-operator-dockercfg-gkqpw" Dec 10 22:54:13 crc kubenswrapper[4791]: I1210 22:54:13.776133 4791 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"mcc-proxy-tls" Dec 10 22:54:13 crc kubenswrapper[4791]: I1210 22:54:13.795251 4791 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"machine-config-server-tls" Dec 10 22:54:13 crc kubenswrapper[4791]: I1210 22:54:13.795447 4791 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"marketplace-operator-metrics" Dec 10 22:54:13 crc kubenswrapper[4791]: I1210 22:54:13.819651 4791 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-multus"/"default-cni-sysctl-allowlist" Dec 10 22:54:13 crc kubenswrapper[4791]: I1210 22:54:13.833603 4791 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"config" Dec 10 22:54:13 crc kubenswrapper[4791]: I1210 22:54:13.890927 4791 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="35033f1a-b91b-4e02-ba77-3841a2e0fdab" path="/var/lib/kubelet/pods/35033f1a-b91b-4e02-ba77-3841a2e0fdab/volumes" Dec 10 22:54:13 crc kubenswrapper[4791]: I1210 22:54:13.990881 4791 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-apiserver"/"serving-cert" Dec 10 22:54:14 crc kubenswrapper[4791]: I1210 22:54:14.052299 4791 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-controller-manager"/"openshift-controller-manager-sa-dockercfg-msq4c" Dec 10 22:54:14 crc kubenswrapper[4791]: I1210 22:54:14.145453 4791 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"machine-config-controller-dockercfg-c2lfx" Dec 10 22:54:14 crc kubenswrapper[4791]: I1210 22:54:14.170922 4791 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-oauth-apiserver"/"trusted-ca-bundle" Dec 10 22:54:14 crc kubenswrapper[4791]: I1210 22:54:14.220723 4791 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-console"/"default-dockercfg-chnjx" Dec 10 22:54:14 crc kubenswrapper[4791]: I1210 22:54:14.286581 4791 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager-operator"/"openshift-controller-manager-operator-config" Dec 10 22:54:14 crc kubenswrapper[4791]: I1210 22:54:14.300570 4791 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ovn-kubernetes"/"ovn-kubernetes-control-plane-dockercfg-gs7dd" Dec 10 22:54:14 crc kubenswrapper[4791]: I1210 22:54:14.380388 4791 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console"/"oauth-serving-cert" Dec 10 22:54:14 crc kubenswrapper[4791]: I1210 22:54:14.407027 4791 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-authentication/oauth-openshift-76f6bd77b6-2dlw6"] Dec 10 22:54:14 crc kubenswrapper[4791]: E1210 22:54:14.407311 4791 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="31bd0efc-25ef-4a1f-b4b9-e4c5b14768ea" containerName="installer" Dec 10 22:54:14 crc kubenswrapper[4791]: I1210 22:54:14.407356 4791 state_mem.go:107] "Deleted CPUSet assignment" podUID="31bd0efc-25ef-4a1f-b4b9-e4c5b14768ea" containerName="installer" Dec 10 22:54:14 crc kubenswrapper[4791]: E1210 22:54:14.407384 4791 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="35033f1a-b91b-4e02-ba77-3841a2e0fdab" containerName="oauth-openshift" Dec 10 22:54:14 crc kubenswrapper[4791]: I1210 22:54:14.407395 4791 state_mem.go:107] "Deleted CPUSet assignment" podUID="35033f1a-b91b-4e02-ba77-3841a2e0fdab" containerName="oauth-openshift" Dec 10 22:54:14 crc kubenswrapper[4791]: I1210 22:54:14.407543 4791 memory_manager.go:354] "RemoveStaleState removing state" podUID="31bd0efc-25ef-4a1f-b4b9-e4c5b14768ea" containerName="installer" Dec 10 22:54:14 crc kubenswrapper[4791]: I1210 22:54:14.407565 4791 memory_manager.go:354] "RemoveStaleState removing state" podUID="35033f1a-b91b-4e02-ba77-3841a2e0fdab" containerName="oauth-openshift" Dec 10 22:54:14 crc kubenswrapper[4791]: I1210 22:54:14.408103 4791 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-authentication/oauth-openshift-76f6bd77b6-2dlw6" Dec 10 22:54:14 crc kubenswrapper[4791]: I1210 22:54:14.410863 4791 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication"/"audit" Dec 10 22:54:14 crc kubenswrapper[4791]: I1210 22:54:14.411565 4791 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-system-router-certs" Dec 10 22:54:14 crc kubenswrapper[4791]: I1210 22:54:14.414068 4791 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication"/"openshift-service-ca.crt" Dec 10 22:54:14 crc kubenswrapper[4791]: I1210 22:54:14.414400 4791 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-system-serving-cert" Dec 10 22:54:14 crc kubenswrapper[4791]: I1210 22:54:14.414573 4791 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-system-session" Dec 10 22:54:14 crc kubenswrapper[4791]: I1210 22:54:14.415255 4791 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-user-template-error" Dec 10 22:54:14 crc kubenswrapper[4791]: I1210 22:54:14.415611 4791 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication"/"kube-root-ca.crt" Dec 10 22:54:14 crc kubenswrapper[4791]: I1210 22:54:14.415681 4791 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication"/"v4-0-config-system-cliconfig" Dec 10 22:54:14 crc kubenswrapper[4791]: I1210 22:54:14.415891 4791 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication"/"v4-0-config-system-service-ca" Dec 10 22:54:14 crc kubenswrapper[4791]: I1210 22:54:14.416303 4791 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-user-template-provider-selection" Dec 10 22:54:14 crc kubenswrapper[4791]: I1210 22:54:14.417431 4791 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"oauth-openshift-dockercfg-znhcc" Dec 10 22:54:14 crc kubenswrapper[4791]: I1210 22:54:14.420360 4791 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-user-idp-0-file-data" Dec 10 22:54:14 crc kubenswrapper[4791]: I1210 22:54:14.427903 4791 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-user-template-login" Dec 10 22:54:14 crc kubenswrapper[4791]: I1210 22:54:14.428311 4791 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication"/"v4-0-config-system-trusted-ca-bundle" Dec 10 22:54:14 crc kubenswrapper[4791]: I1210 22:54:14.429070 4791 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-authentication/oauth-openshift-76f6bd77b6-2dlw6"] Dec 10 22:54:14 crc kubenswrapper[4791]: I1210 22:54:14.435375 4791 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-system-ocp-branding-template" Dec 10 22:54:14 crc kubenswrapper[4791]: I1210 22:54:14.437736 4791 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-service-ca-operator"/"kube-root-ca.crt" Dec 10 22:54:14 crc kubenswrapper[4791]: I1210 22:54:14.462896 4791 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-image-registry"/"image-registry-certificates" Dec 10 22:54:14 crc kubenswrapper[4791]: I1210 22:54:14.469722 4791 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-image-registry"/"node-ca-dockercfg-4777p" Dec 10 22:54:14 crc kubenswrapper[4791]: I1210 22:54:14.553922 4791 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-user-template-error\" (UniqueName: \"kubernetes.io/secret/d529d75e-2d97-4138-8f18-6c52ad10e7f5-v4-0-config-user-template-error\") pod \"oauth-openshift-76f6bd77b6-2dlw6\" (UID: \"d529d75e-2d97-4138-8f18-6c52ad10e7f5\") " pod="openshift-authentication/oauth-openshift-76f6bd77b6-2dlw6" Dec 10 22:54:14 crc kubenswrapper[4791]: I1210 22:54:14.554008 4791 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/d529d75e-2d97-4138-8f18-6c52ad10e7f5-audit-dir\") pod \"oauth-openshift-76f6bd77b6-2dlw6\" (UID: \"d529d75e-2d97-4138-8f18-6c52ad10e7f5\") " pod="openshift-authentication/oauth-openshift-76f6bd77b6-2dlw6" Dec 10 22:54:14 crc kubenswrapper[4791]: I1210 22:54:14.554145 4791 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-ocp-branding-template\" (UniqueName: \"kubernetes.io/secret/d529d75e-2d97-4138-8f18-6c52ad10e7f5-v4-0-config-system-ocp-branding-template\") pod \"oauth-openshift-76f6bd77b6-2dlw6\" (UID: \"d529d75e-2d97-4138-8f18-6c52ad10e7f5\") " pod="openshift-authentication/oauth-openshift-76f6bd77b6-2dlw6" Dec 10 22:54:14 crc kubenswrapper[4791]: I1210 22:54:14.554217 4791 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/d529d75e-2d97-4138-8f18-6c52ad10e7f5-v4-0-config-system-trusted-ca-bundle\") pod \"oauth-openshift-76f6bd77b6-2dlw6\" (UID: \"d529d75e-2d97-4138-8f18-6c52ad10e7f5\") " pod="openshift-authentication/oauth-openshift-76f6bd77b6-2dlw6" Dec 10 22:54:14 crc kubenswrapper[4791]: I1210 22:54:14.554276 4791 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-session\" (UniqueName: \"kubernetes.io/secret/d529d75e-2d97-4138-8f18-6c52ad10e7f5-v4-0-config-system-session\") pod \"oauth-openshift-76f6bd77b6-2dlw6\" (UID: \"d529d75e-2d97-4138-8f18-6c52ad10e7f5\") " pod="openshift-authentication/oauth-openshift-76f6bd77b6-2dlw6" Dec 10 22:54:14 crc kubenswrapper[4791]: I1210 22:54:14.554359 4791 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-serving-cert\" (UniqueName: \"kubernetes.io/secret/d529d75e-2d97-4138-8f18-6c52ad10e7f5-v4-0-config-system-serving-cert\") pod \"oauth-openshift-76f6bd77b6-2dlw6\" (UID: \"d529d75e-2d97-4138-8f18-6c52ad10e7f5\") " pod="openshift-authentication/oauth-openshift-76f6bd77b6-2dlw6" Dec 10 22:54:14 crc kubenswrapper[4791]: I1210 22:54:14.554392 4791 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-service-ca\" (UniqueName: \"kubernetes.io/configmap/d529d75e-2d97-4138-8f18-6c52ad10e7f5-v4-0-config-system-service-ca\") pod \"oauth-openshift-76f6bd77b6-2dlw6\" (UID: \"d529d75e-2d97-4138-8f18-6c52ad10e7f5\") " pod="openshift-authentication/oauth-openshift-76f6bd77b6-2dlw6" Dec 10 22:54:14 crc kubenswrapper[4791]: I1210 22:54:14.554433 4791 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-cliconfig\" (UniqueName: \"kubernetes.io/configmap/d529d75e-2d97-4138-8f18-6c52ad10e7f5-v4-0-config-system-cliconfig\") pod \"oauth-openshift-76f6bd77b6-2dlw6\" (UID: \"d529d75e-2d97-4138-8f18-6c52ad10e7f5\") " pod="openshift-authentication/oauth-openshift-76f6bd77b6-2dlw6" Dec 10 22:54:14 crc kubenswrapper[4791]: I1210 22:54:14.554526 4791 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/d529d75e-2d97-4138-8f18-6c52ad10e7f5-audit-policies\") pod \"oauth-openshift-76f6bd77b6-2dlw6\" (UID: \"d529d75e-2d97-4138-8f18-6c52ad10e7f5\") " pod="openshift-authentication/oauth-openshift-76f6bd77b6-2dlw6" Dec 10 22:54:14 crc kubenswrapper[4791]: I1210 22:54:14.554588 4791 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-router-certs\" (UniqueName: \"kubernetes.io/secret/d529d75e-2d97-4138-8f18-6c52ad10e7f5-v4-0-config-system-router-certs\") pod \"oauth-openshift-76f6bd77b6-2dlw6\" (UID: \"d529d75e-2d97-4138-8f18-6c52ad10e7f5\") " pod="openshift-authentication/oauth-openshift-76f6bd77b6-2dlw6" Dec 10 22:54:14 crc kubenswrapper[4791]: I1210 22:54:14.554624 4791 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-user-template-provider-selection\" (UniqueName: \"kubernetes.io/secret/d529d75e-2d97-4138-8f18-6c52ad10e7f5-v4-0-config-user-template-provider-selection\") pod \"oauth-openshift-76f6bd77b6-2dlw6\" (UID: \"d529d75e-2d97-4138-8f18-6c52ad10e7f5\") " pod="openshift-authentication/oauth-openshift-76f6bd77b6-2dlw6" Dec 10 22:54:14 crc kubenswrapper[4791]: I1210 22:54:14.554689 4791 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-vmb7z\" (UniqueName: \"kubernetes.io/projected/d529d75e-2d97-4138-8f18-6c52ad10e7f5-kube-api-access-vmb7z\") pod \"oauth-openshift-76f6bd77b6-2dlw6\" (UID: \"d529d75e-2d97-4138-8f18-6c52ad10e7f5\") " pod="openshift-authentication/oauth-openshift-76f6bd77b6-2dlw6" Dec 10 22:54:14 crc kubenswrapper[4791]: I1210 22:54:14.554721 4791 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-user-idp-0-file-data\" (UniqueName: \"kubernetes.io/secret/d529d75e-2d97-4138-8f18-6c52ad10e7f5-v4-0-config-user-idp-0-file-data\") pod \"oauth-openshift-76f6bd77b6-2dlw6\" (UID: \"d529d75e-2d97-4138-8f18-6c52ad10e7f5\") " pod="openshift-authentication/oauth-openshift-76f6bd77b6-2dlw6" Dec 10 22:54:14 crc kubenswrapper[4791]: I1210 22:54:14.554754 4791 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-user-template-login\" (UniqueName: \"kubernetes.io/secret/d529d75e-2d97-4138-8f18-6c52ad10e7f5-v4-0-config-user-template-login\") pod \"oauth-openshift-76f6bd77b6-2dlw6\" (UID: \"d529d75e-2d97-4138-8f18-6c52ad10e7f5\") " pod="openshift-authentication/oauth-openshift-76f6bd77b6-2dlw6" Dec 10 22:54:15 crc kubenswrapper[4791]: I1210 22:54:14.656260 4791 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-vmb7z\" (UniqueName: \"kubernetes.io/projected/d529d75e-2d97-4138-8f18-6c52ad10e7f5-kube-api-access-vmb7z\") pod \"oauth-openshift-76f6bd77b6-2dlw6\" (UID: \"d529d75e-2d97-4138-8f18-6c52ad10e7f5\") " pod="openshift-authentication/oauth-openshift-76f6bd77b6-2dlw6" Dec 10 22:54:15 crc kubenswrapper[4791]: I1210 22:54:14.656316 4791 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-user-idp-0-file-data\" (UniqueName: \"kubernetes.io/secret/d529d75e-2d97-4138-8f18-6c52ad10e7f5-v4-0-config-user-idp-0-file-data\") pod \"oauth-openshift-76f6bd77b6-2dlw6\" (UID: \"d529d75e-2d97-4138-8f18-6c52ad10e7f5\") " pod="openshift-authentication/oauth-openshift-76f6bd77b6-2dlw6" Dec 10 22:54:15 crc kubenswrapper[4791]: I1210 22:54:14.656361 4791 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-user-template-login\" (UniqueName: \"kubernetes.io/secret/d529d75e-2d97-4138-8f18-6c52ad10e7f5-v4-0-config-user-template-login\") pod \"oauth-openshift-76f6bd77b6-2dlw6\" (UID: \"d529d75e-2d97-4138-8f18-6c52ad10e7f5\") " pod="openshift-authentication/oauth-openshift-76f6bd77b6-2dlw6" Dec 10 22:54:15 crc kubenswrapper[4791]: I1210 22:54:14.656388 4791 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-user-template-error\" (UniqueName: \"kubernetes.io/secret/d529d75e-2d97-4138-8f18-6c52ad10e7f5-v4-0-config-user-template-error\") pod \"oauth-openshift-76f6bd77b6-2dlw6\" (UID: \"d529d75e-2d97-4138-8f18-6c52ad10e7f5\") " pod="openshift-authentication/oauth-openshift-76f6bd77b6-2dlw6" Dec 10 22:54:15 crc kubenswrapper[4791]: I1210 22:54:14.656418 4791 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/d529d75e-2d97-4138-8f18-6c52ad10e7f5-audit-dir\") pod \"oauth-openshift-76f6bd77b6-2dlw6\" (UID: \"d529d75e-2d97-4138-8f18-6c52ad10e7f5\") " pod="openshift-authentication/oauth-openshift-76f6bd77b6-2dlw6" Dec 10 22:54:15 crc kubenswrapper[4791]: I1210 22:54:14.656447 4791 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-ocp-branding-template\" (UniqueName: \"kubernetes.io/secret/d529d75e-2d97-4138-8f18-6c52ad10e7f5-v4-0-config-system-ocp-branding-template\") pod \"oauth-openshift-76f6bd77b6-2dlw6\" (UID: \"d529d75e-2d97-4138-8f18-6c52ad10e7f5\") " pod="openshift-authentication/oauth-openshift-76f6bd77b6-2dlw6" Dec 10 22:54:15 crc kubenswrapper[4791]: I1210 22:54:14.656468 4791 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/d529d75e-2d97-4138-8f18-6c52ad10e7f5-v4-0-config-system-trusted-ca-bundle\") pod \"oauth-openshift-76f6bd77b6-2dlw6\" (UID: \"d529d75e-2d97-4138-8f18-6c52ad10e7f5\") " pod="openshift-authentication/oauth-openshift-76f6bd77b6-2dlw6" Dec 10 22:54:15 crc kubenswrapper[4791]: I1210 22:54:14.656492 4791 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-session\" (UniqueName: \"kubernetes.io/secret/d529d75e-2d97-4138-8f18-6c52ad10e7f5-v4-0-config-system-session\") pod \"oauth-openshift-76f6bd77b6-2dlw6\" (UID: \"d529d75e-2d97-4138-8f18-6c52ad10e7f5\") " pod="openshift-authentication/oauth-openshift-76f6bd77b6-2dlw6" Dec 10 22:54:15 crc kubenswrapper[4791]: I1210 22:54:14.656540 4791 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-serving-cert\" (UniqueName: \"kubernetes.io/secret/d529d75e-2d97-4138-8f18-6c52ad10e7f5-v4-0-config-system-serving-cert\") pod \"oauth-openshift-76f6bd77b6-2dlw6\" (UID: \"d529d75e-2d97-4138-8f18-6c52ad10e7f5\") " pod="openshift-authentication/oauth-openshift-76f6bd77b6-2dlw6" Dec 10 22:54:15 crc kubenswrapper[4791]: I1210 22:54:14.656561 4791 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-service-ca\" (UniqueName: \"kubernetes.io/configmap/d529d75e-2d97-4138-8f18-6c52ad10e7f5-v4-0-config-system-service-ca\") pod \"oauth-openshift-76f6bd77b6-2dlw6\" (UID: \"d529d75e-2d97-4138-8f18-6c52ad10e7f5\") " pod="openshift-authentication/oauth-openshift-76f6bd77b6-2dlw6" Dec 10 22:54:15 crc kubenswrapper[4791]: I1210 22:54:14.656581 4791 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-cliconfig\" (UniqueName: \"kubernetes.io/configmap/d529d75e-2d97-4138-8f18-6c52ad10e7f5-v4-0-config-system-cliconfig\") pod \"oauth-openshift-76f6bd77b6-2dlw6\" (UID: \"d529d75e-2d97-4138-8f18-6c52ad10e7f5\") " pod="openshift-authentication/oauth-openshift-76f6bd77b6-2dlw6" Dec 10 22:54:15 crc kubenswrapper[4791]: I1210 22:54:14.656621 4791 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/d529d75e-2d97-4138-8f18-6c52ad10e7f5-audit-policies\") pod \"oauth-openshift-76f6bd77b6-2dlw6\" (UID: \"d529d75e-2d97-4138-8f18-6c52ad10e7f5\") " pod="openshift-authentication/oauth-openshift-76f6bd77b6-2dlw6" Dec 10 22:54:15 crc kubenswrapper[4791]: I1210 22:54:14.656647 4791 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-router-certs\" (UniqueName: \"kubernetes.io/secret/d529d75e-2d97-4138-8f18-6c52ad10e7f5-v4-0-config-system-router-certs\") pod \"oauth-openshift-76f6bd77b6-2dlw6\" (UID: \"d529d75e-2d97-4138-8f18-6c52ad10e7f5\") " pod="openshift-authentication/oauth-openshift-76f6bd77b6-2dlw6" Dec 10 22:54:15 crc kubenswrapper[4791]: I1210 22:54:14.656670 4791 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-user-template-provider-selection\" (UniqueName: \"kubernetes.io/secret/d529d75e-2d97-4138-8f18-6c52ad10e7f5-v4-0-config-user-template-provider-selection\") pod \"oauth-openshift-76f6bd77b6-2dlw6\" (UID: \"d529d75e-2d97-4138-8f18-6c52ad10e7f5\") " pod="openshift-authentication/oauth-openshift-76f6bd77b6-2dlw6" Dec 10 22:54:15 crc kubenswrapper[4791]: I1210 22:54:14.657155 4791 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/d529d75e-2d97-4138-8f18-6c52ad10e7f5-audit-dir\") pod \"oauth-openshift-76f6bd77b6-2dlw6\" (UID: \"d529d75e-2d97-4138-8f18-6c52ad10e7f5\") " pod="openshift-authentication/oauth-openshift-76f6bd77b6-2dlw6" Dec 10 22:54:15 crc kubenswrapper[4791]: I1210 22:54:14.657983 4791 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-service-ca\" (UniqueName: \"kubernetes.io/configmap/d529d75e-2d97-4138-8f18-6c52ad10e7f5-v4-0-config-system-service-ca\") pod \"oauth-openshift-76f6bd77b6-2dlw6\" (UID: \"d529d75e-2d97-4138-8f18-6c52ad10e7f5\") " pod="openshift-authentication/oauth-openshift-76f6bd77b6-2dlw6" Dec 10 22:54:15 crc kubenswrapper[4791]: I1210 22:54:14.658189 4791 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/d529d75e-2d97-4138-8f18-6c52ad10e7f5-v4-0-config-system-trusted-ca-bundle\") pod \"oauth-openshift-76f6bd77b6-2dlw6\" (UID: \"d529d75e-2d97-4138-8f18-6c52ad10e7f5\") " pod="openshift-authentication/oauth-openshift-76f6bd77b6-2dlw6" Dec 10 22:54:15 crc kubenswrapper[4791]: I1210 22:54:14.658527 4791 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-cliconfig\" (UniqueName: \"kubernetes.io/configmap/d529d75e-2d97-4138-8f18-6c52ad10e7f5-v4-0-config-system-cliconfig\") pod \"oauth-openshift-76f6bd77b6-2dlw6\" (UID: \"d529d75e-2d97-4138-8f18-6c52ad10e7f5\") " pod="openshift-authentication/oauth-openshift-76f6bd77b6-2dlw6" Dec 10 22:54:15 crc kubenswrapper[4791]: I1210 22:54:14.658671 4791 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/d529d75e-2d97-4138-8f18-6c52ad10e7f5-audit-policies\") pod \"oauth-openshift-76f6bd77b6-2dlw6\" (UID: \"d529d75e-2d97-4138-8f18-6c52ad10e7f5\") " pod="openshift-authentication/oauth-openshift-76f6bd77b6-2dlw6" Dec 10 22:54:15 crc kubenswrapper[4791]: I1210 22:54:14.663630 4791 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-user-template-provider-selection\" (UniqueName: \"kubernetes.io/secret/d529d75e-2d97-4138-8f18-6c52ad10e7f5-v4-0-config-user-template-provider-selection\") pod \"oauth-openshift-76f6bd77b6-2dlw6\" (UID: \"d529d75e-2d97-4138-8f18-6c52ad10e7f5\") " pod="openshift-authentication/oauth-openshift-76f6bd77b6-2dlw6" Dec 10 22:54:15 crc kubenswrapper[4791]: I1210 22:54:14.663801 4791 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-session\" (UniqueName: \"kubernetes.io/secret/d529d75e-2d97-4138-8f18-6c52ad10e7f5-v4-0-config-system-session\") pod \"oauth-openshift-76f6bd77b6-2dlw6\" (UID: \"d529d75e-2d97-4138-8f18-6c52ad10e7f5\") " pod="openshift-authentication/oauth-openshift-76f6bd77b6-2dlw6" Dec 10 22:54:15 crc kubenswrapper[4791]: I1210 22:54:14.664254 4791 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-user-idp-0-file-data\" (UniqueName: \"kubernetes.io/secret/d529d75e-2d97-4138-8f18-6c52ad10e7f5-v4-0-config-user-idp-0-file-data\") pod \"oauth-openshift-76f6bd77b6-2dlw6\" (UID: \"d529d75e-2d97-4138-8f18-6c52ad10e7f5\") " pod="openshift-authentication/oauth-openshift-76f6bd77b6-2dlw6" Dec 10 22:54:15 crc kubenswrapper[4791]: I1210 22:54:14.664590 4791 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-ocp-branding-template\" (UniqueName: \"kubernetes.io/secret/d529d75e-2d97-4138-8f18-6c52ad10e7f5-v4-0-config-system-ocp-branding-template\") pod \"oauth-openshift-76f6bd77b6-2dlw6\" (UID: \"d529d75e-2d97-4138-8f18-6c52ad10e7f5\") " pod="openshift-authentication/oauth-openshift-76f6bd77b6-2dlw6" Dec 10 22:54:15 crc kubenswrapper[4791]: I1210 22:54:14.664849 4791 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-router-certs\" (UniqueName: \"kubernetes.io/secret/d529d75e-2d97-4138-8f18-6c52ad10e7f5-v4-0-config-system-router-certs\") pod \"oauth-openshift-76f6bd77b6-2dlw6\" (UID: \"d529d75e-2d97-4138-8f18-6c52ad10e7f5\") " pod="openshift-authentication/oauth-openshift-76f6bd77b6-2dlw6" Dec 10 22:54:15 crc kubenswrapper[4791]: I1210 22:54:14.665762 4791 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-user-template-login\" (UniqueName: \"kubernetes.io/secret/d529d75e-2d97-4138-8f18-6c52ad10e7f5-v4-0-config-user-template-login\") pod \"oauth-openshift-76f6bd77b6-2dlw6\" (UID: \"d529d75e-2d97-4138-8f18-6c52ad10e7f5\") " pod="openshift-authentication/oauth-openshift-76f6bd77b6-2dlw6" Dec 10 22:54:15 crc kubenswrapper[4791]: I1210 22:54:14.666627 4791 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-user-template-error\" (UniqueName: \"kubernetes.io/secret/d529d75e-2d97-4138-8f18-6c52ad10e7f5-v4-0-config-user-template-error\") pod \"oauth-openshift-76f6bd77b6-2dlw6\" (UID: \"d529d75e-2d97-4138-8f18-6c52ad10e7f5\") " pod="openshift-authentication/oauth-openshift-76f6bd77b6-2dlw6" Dec 10 22:54:15 crc kubenswrapper[4791]: I1210 22:54:14.669551 4791 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-serving-cert\" (UniqueName: \"kubernetes.io/secret/d529d75e-2d97-4138-8f18-6c52ad10e7f5-v4-0-config-system-serving-cert\") pod \"oauth-openshift-76f6bd77b6-2dlw6\" (UID: \"d529d75e-2d97-4138-8f18-6c52ad10e7f5\") " pod="openshift-authentication/oauth-openshift-76f6bd77b6-2dlw6" Dec 10 22:54:15 crc kubenswrapper[4791]: I1210 22:54:14.675014 4791 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-vmb7z\" (UniqueName: \"kubernetes.io/projected/d529d75e-2d97-4138-8f18-6c52ad10e7f5-kube-api-access-vmb7z\") pod \"oauth-openshift-76f6bd77b6-2dlw6\" (UID: \"d529d75e-2d97-4138-8f18-6c52ad10e7f5\") " pod="openshift-authentication/oauth-openshift-76f6bd77b6-2dlw6" Dec 10 22:54:15 crc kubenswrapper[4791]: I1210 22:54:14.734260 4791 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-authentication/oauth-openshift-76f6bd77b6-2dlw6" Dec 10 22:54:15 crc kubenswrapper[4791]: I1210 22:54:14.867017 4791 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-multus"/"multus-ac-dockercfg-9lkdf" Dec 10 22:54:15 crc kubenswrapper[4791]: I1210 22:54:14.871667 4791 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-config-operator"/"kube-root-ca.crt" Dec 10 22:54:15 crc kubenswrapper[4791]: I1210 22:54:14.884794 4791 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress-canary"/"openshift-service-ca.crt" Dec 10 22:54:15 crc kubenswrapper[4791]: I1210 22:54:14.926928 4791 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-dns"/"dns-dockercfg-jwfmh" Dec 10 22:54:15 crc kubenswrapper[4791]: I1210 22:54:14.935091 4791 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-operator"/"iptables-alerter-script" Dec 10 22:54:15 crc kubenswrapper[4791]: I1210 22:54:15.048092 4791 reflector.go:368] Caches populated for *v1.CSIDriver from k8s.io/client-go/informers/factory.go:160 Dec 10 22:54:15 crc kubenswrapper[4791]: I1210 22:54:15.140269 4791 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-node-identity"/"kube-root-ca.crt" Dec 10 22:54:15 crc kubenswrapper[4791]: I1210 22:54:15.149456 4791 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress-operator"/"metrics-tls" Dec 10 22:54:15 crc kubenswrapper[4791]: I1210 22:54:15.154719 4791 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-apiserver-operator"/"kube-root-ca.crt" Dec 10 22:54:15 crc kubenswrapper[4791]: I1210 22:54:15.241706 4791 kubelet.go:2431] "SyncLoop REMOVE" source="file" pods=["openshift-kube-apiserver/kube-apiserver-startup-monitor-crc"] Dec 10 22:54:15 crc kubenswrapper[4791]: I1210 22:54:15.242200 4791 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" podUID="f85e55b1a89d02b0cb034b1ea31ed45a" containerName="startup-monitor" containerID="cri-o://3c36eae76bf3819198f38fa3cb82e87295ea35971222122789121c928c1c6f03" gracePeriod=5 Dec 10 22:54:15 crc kubenswrapper[4791]: I1210 22:54:15.308144 4791 reflector.go:368] Caches populated for *v1.Node from k8s.io/client-go/informers/factory.go:160 Dec 10 22:54:15 crc kubenswrapper[4791]: I1210 22:54:15.338172 4791 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-storage-version-migrator"/"openshift-service-ca.crt" Dec 10 22:54:15 crc kubenswrapper[4791]: I1210 22:54:15.411627 4791 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ovn-kubernetes"/"env-overrides" Dec 10 22:54:15 crc kubenswrapper[4791]: I1210 22:54:15.483719 4791 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-service-ca"/"openshift-service-ca.crt" Dec 10 22:54:15 crc kubenswrapper[4791]: I1210 22:54:15.486203 4791 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-oauth-apiserver"/"encryption-config-1" Dec 10 22:54:15 crc kubenswrapper[4791]: I1210 22:54:15.590448 4791 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-dns-operator"/"metrics-tls" Dec 10 22:54:15 crc kubenswrapper[4791]: I1210 22:54:15.598540 4791 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager-operator"/"kube-root-ca.crt" Dec 10 22:54:15 crc kubenswrapper[4791]: I1210 22:54:15.647003 4791 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-authentication/oauth-openshift-76f6bd77b6-2dlw6"] Dec 10 22:54:15 crc kubenswrapper[4791]: I1210 22:54:15.736887 4791 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-dns"/"dns-default-metrics-tls" Dec 10 22:54:15 crc kubenswrapper[4791]: I1210 22:54:15.765355 4791 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"pprof-cert" Dec 10 22:54:15 crc kubenswrapper[4791]: I1210 22:54:15.842330 4791 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-etcd-operator"/"openshift-service-ca.crt" Dec 10 22:54:15 crc kubenswrapper[4791]: I1210 22:54:15.845142 4791 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-storage-version-migrator-operator"/"config" Dec 10 22:54:15 crc kubenswrapper[4791]: I1210 22:54:15.883181 4791 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-apiserver"/"etcd-client" Dec 10 22:54:15 crc kubenswrapper[4791]: I1210 22:54:15.890179 4791 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver"/"openshift-service-ca.crt" Dec 10 22:54:15 crc kubenswrapper[4791]: I1210 22:54:15.949955 4791 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-machine-approver"/"kube-root-ca.crt" Dec 10 22:54:16 crc kubenswrapper[4791]: I1210 22:54:16.070585 4791 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-cluster-version"/"default-dockercfg-gxtc4" Dec 10 22:54:16 crc kubenswrapper[4791]: I1210 22:54:16.197300 4791 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-diagnostics"/"openshift-service-ca.crt" Dec 10 22:54:16 crc kubenswrapper[4791]: I1210 22:54:16.209410 4791 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-authentication/oauth-openshift-76f6bd77b6-2dlw6" event={"ID":"d529d75e-2d97-4138-8f18-6c52ad10e7f5","Type":"ContainerStarted","Data":"521332a6343ee4e130225846688f40e0893421f08126f677fbc50aabf869fb4d"} Dec 10 22:54:16 crc kubenswrapper[4791]: I1210 22:54:16.209473 4791 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-authentication/oauth-openshift-76f6bd77b6-2dlw6" event={"ID":"d529d75e-2d97-4138-8f18-6c52ad10e7f5","Type":"ContainerStarted","Data":"dafd177915488ca012742a903877451e990d6c526988f4231cc6bada6dde732e"} Dec 10 22:54:16 crc kubenswrapper[4791]: I1210 22:54:16.209809 4791 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-authentication/oauth-openshift-76f6bd77b6-2dlw6" Dec 10 22:54:16 crc kubenswrapper[4791]: I1210 22:54:16.245387 4791 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-authentication/oauth-openshift-76f6bd77b6-2dlw6" podStartSLOduration=54.245362536 podStartE2EDuration="54.245362536s" podCreationTimestamp="2025-12-10 22:53:22 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-10 22:54:16.239480774 +0000 UTC m=+290.669098397" watchObservedRunningTime="2025-12-10 22:54:16.245362536 +0000 UTC m=+290.674980189" Dec 10 22:54:16 crc kubenswrapper[4791]: I1210 22:54:16.294432 4791 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-image-registry"/"kube-root-ca.crt" Dec 10 22:54:16 crc kubenswrapper[4791]: I1210 22:54:16.370526 4791 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-multus"/"openshift-service-ca.crt" Dec 10 22:54:16 crc kubenswrapper[4791]: I1210 22:54:16.452125 4791 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-network-node-identity"/"network-node-identity-cert" Dec 10 22:54:16 crc kubenswrapper[4791]: I1210 22:54:16.550830 4791 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-authentication/oauth-openshift-76f6bd77b6-2dlw6" Dec 10 22:54:16 crc kubenswrapper[4791]: I1210 22:54:16.664600 4791 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-console"/"console-dockercfg-f62pw" Dec 10 22:54:16 crc kubenswrapper[4791]: I1210 22:54:16.678738 4791 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress"/"kube-root-ca.crt" Dec 10 22:54:16 crc kubenswrapper[4791]: I1210 22:54:16.774895 4791 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress"/"openshift-service-ca.crt" Dec 10 22:54:16 crc kubenswrapper[4791]: I1210 22:54:16.801731 4791 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"config" Dec 10 22:54:16 crc kubenswrapper[4791]: I1210 22:54:16.939678 4791 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-service-ca"/"service-ca-dockercfg-pn86c" Dec 10 22:54:17 crc kubenswrapper[4791]: I1210 22:54:17.213698 4791 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-samples-operator"/"kube-root-ca.crt" Dec 10 22:54:17 crc kubenswrapper[4791]: I1210 22:54:17.227757 4791 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ovn-kubernetes"/"kube-root-ca.crt" Dec 10 22:54:17 crc kubenswrapper[4791]: I1210 22:54:17.303189 4791 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-dns-operator"/"kube-root-ca.crt" Dec 10 22:54:17 crc kubenswrapper[4791]: I1210 22:54:17.343990 4791 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-dns"/"kube-root-ca.crt" Dec 10 22:54:17 crc kubenswrapper[4791]: I1210 22:54:17.453734 4791 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-console-operator"/"console-operator-dockercfg-4xjcr" Dec 10 22:54:17 crc kubenswrapper[4791]: I1210 22:54:17.485572 4791 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-etcd-operator"/"etcd-operator-config" Dec 10 22:54:17 crc kubenswrapper[4791]: I1210 22:54:17.571091 4791 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-image-registry"/"cluster-image-registry-operator-dockercfg-m4qtx" Dec 10 22:54:18 crc kubenswrapper[4791]: I1210 22:54:18.024023 4791 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-oauth-apiserver"/"serving-cert" Dec 10 22:54:18 crc kubenswrapper[4791]: I1210 22:54:18.135877 4791 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-version"/"kube-root-ca.crt" Dec 10 22:54:18 crc kubenswrapper[4791]: I1210 22:54:18.147093 4791 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-console"/"networking-console-plugin" Dec 10 22:54:18 crc kubenswrapper[4791]: I1210 22:54:18.229767 4791 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-config-operator"/"openshift-service-ca.crt" Dec 10 22:54:18 crc kubenswrapper[4791]: I1210 22:54:18.472140 4791 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress-operator"/"trusted-ca" Dec 10 22:54:18 crc kubenswrapper[4791]: I1210 22:54:18.570858 4791 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-storage-version-migrator-operator"/"serving-cert" Dec 10 22:54:18 crc kubenswrapper[4791]: I1210 22:54:18.753685 4791 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"community-operators-dockercfg-dmngl" Dec 10 22:54:18 crc kubenswrapper[4791]: I1210 22:54:18.964524 4791 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress-canary"/"kube-root-ca.crt" Dec 10 22:54:21 crc kubenswrapper[4791]: I1210 22:54:21.237201 4791 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-kube-apiserver_kube-apiserver-startup-monitor-crc_f85e55b1a89d02b0cb034b1ea31ed45a/startup-monitor/0.log" Dec 10 22:54:21 crc kubenswrapper[4791]: I1210 22:54:21.237436 4791 generic.go:334] "Generic (PLEG): container finished" podID="f85e55b1a89d02b0cb034b1ea31ed45a" containerID="3c36eae76bf3819198f38fa3cb82e87295ea35971222122789121c928c1c6f03" exitCode=137 Dec 10 22:54:21 crc kubenswrapper[4791]: I1210 22:54:21.693279 4791 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-kube-apiserver_kube-apiserver-startup-monitor-crc_f85e55b1a89d02b0cb034b1ea31ed45a/startup-monitor/0.log" Dec 10 22:54:21 crc kubenswrapper[4791]: I1210 22:54:21.693373 4791 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Dec 10 22:54:21 crc kubenswrapper[4791]: I1210 22:54:21.853578 4791 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"var-lock\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-var-lock\") pod \"f85e55b1a89d02b0cb034b1ea31ed45a\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " Dec 10 22:54:21 crc kubenswrapper[4791]: I1210 22:54:21.853653 4791 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pod-resource-dir\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-pod-resource-dir\") pod \"f85e55b1a89d02b0cb034b1ea31ed45a\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " Dec 10 22:54:21 crc kubenswrapper[4791]: I1210 22:54:21.853675 4791 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"manifests\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-manifests\") pod \"f85e55b1a89d02b0cb034b1ea31ed45a\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " Dec 10 22:54:21 crc kubenswrapper[4791]: I1210 22:54:21.853728 4791 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-var-lock" (OuterVolumeSpecName: "var-lock") pod "f85e55b1a89d02b0cb034b1ea31ed45a" (UID: "f85e55b1a89d02b0cb034b1ea31ed45a"). InnerVolumeSpecName "var-lock". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 10 22:54:21 crc kubenswrapper[4791]: I1210 22:54:21.853768 4791 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-resource-dir\") pod \"f85e55b1a89d02b0cb034b1ea31ed45a\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " Dec 10 22:54:21 crc kubenswrapper[4791]: I1210 22:54:21.853794 4791 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-manifests" (OuterVolumeSpecName: "manifests") pod "f85e55b1a89d02b0cb034b1ea31ed45a" (UID: "f85e55b1a89d02b0cb034b1ea31ed45a"). InnerVolumeSpecName "manifests". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 10 22:54:21 crc kubenswrapper[4791]: I1210 22:54:21.853798 4791 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"var-log\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-var-log\") pod \"f85e55b1a89d02b0cb034b1ea31ed45a\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " Dec 10 22:54:21 crc kubenswrapper[4791]: I1210 22:54:21.853836 4791 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-var-log" (OuterVolumeSpecName: "var-log") pod "f85e55b1a89d02b0cb034b1ea31ed45a" (UID: "f85e55b1a89d02b0cb034b1ea31ed45a"). InnerVolumeSpecName "var-log". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 10 22:54:21 crc kubenswrapper[4791]: I1210 22:54:21.853864 4791 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-resource-dir" (OuterVolumeSpecName: "resource-dir") pod "f85e55b1a89d02b0cb034b1ea31ed45a" (UID: "f85e55b1a89d02b0cb034b1ea31ed45a"). InnerVolumeSpecName "resource-dir". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 10 22:54:21 crc kubenswrapper[4791]: I1210 22:54:21.854246 4791 reconciler_common.go:293] "Volume detached for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-resource-dir\") on node \"crc\" DevicePath \"\"" Dec 10 22:54:21 crc kubenswrapper[4791]: I1210 22:54:21.854267 4791 reconciler_common.go:293] "Volume detached for volume \"var-log\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-var-log\") on node \"crc\" DevicePath \"\"" Dec 10 22:54:21 crc kubenswrapper[4791]: I1210 22:54:21.854278 4791 reconciler_common.go:293] "Volume detached for volume \"var-lock\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-var-lock\") on node \"crc\" DevicePath \"\"" Dec 10 22:54:21 crc kubenswrapper[4791]: I1210 22:54:21.854289 4791 reconciler_common.go:293] "Volume detached for volume \"manifests\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-manifests\") on node \"crc\" DevicePath \"\"" Dec 10 22:54:21 crc kubenswrapper[4791]: I1210 22:54:21.861263 4791 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-pod-resource-dir" (OuterVolumeSpecName: "pod-resource-dir") pod "f85e55b1a89d02b0cb034b1ea31ed45a" (UID: "f85e55b1a89d02b0cb034b1ea31ed45a"). InnerVolumeSpecName "pod-resource-dir". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 10 22:54:21 crc kubenswrapper[4791]: I1210 22:54:21.892999 4791 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="f85e55b1a89d02b0cb034b1ea31ed45a" path="/var/lib/kubelet/pods/f85e55b1a89d02b0cb034b1ea31ed45a/volumes" Dec 10 22:54:21 crc kubenswrapper[4791]: I1210 22:54:21.955874 4791 reconciler_common.go:293] "Volume detached for volume \"pod-resource-dir\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-pod-resource-dir\") on node \"crc\" DevicePath \"\"" Dec 10 22:54:22 crc kubenswrapper[4791]: I1210 22:54:22.244307 4791 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-kube-apiserver_kube-apiserver-startup-monitor-crc_f85e55b1a89d02b0cb034b1ea31ed45a/startup-monitor/0.log" Dec 10 22:54:22 crc kubenswrapper[4791]: I1210 22:54:22.244408 4791 scope.go:117] "RemoveContainer" containerID="3c36eae76bf3819198f38fa3cb82e87295ea35971222122789121c928c1c6f03" Dec 10 22:54:22 crc kubenswrapper[4791]: I1210 22:54:22.244465 4791 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Dec 10 22:54:25 crc kubenswrapper[4791]: I1210 22:54:25.730530 4791 cert_rotation.go:91] certificate rotation detected, shutting down client connections to start using new credentials Dec 10 22:54:28 crc kubenswrapper[4791]: I1210 22:54:28.369523 4791 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-8x2rc"] Dec 10 22:54:28 crc kubenswrapper[4791]: I1210 22:54:28.370440 4791 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/certified-operators-8x2rc" podUID="009efc35-5943-429f-98fd-3307a5f557f2" containerName="registry-server" containerID="cri-o://716a7c482ddb51c0626fcc775f7f3b405fbe182e7fd48c9aa126321e8e6267d9" gracePeriod=30 Dec 10 22:54:28 crc kubenswrapper[4791]: I1210 22:54:28.379484 4791 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-gskzr"] Dec 10 22:54:28 crc kubenswrapper[4791]: I1210 22:54:28.379830 4791 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/community-operators-gskzr" podUID="08d50c37-7a4d-4731-beb8-fbc6d2e10bb0" containerName="registry-server" containerID="cri-o://b95dc6273b4eb68303d46abd4b7a567379de20b6c589ae921a68bc99c91fd48e" gracePeriod=30 Dec 10 22:54:28 crc kubenswrapper[4791]: I1210 22:54:28.384862 4791 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/marketplace-operator-79b997595-7g828"] Dec 10 22:54:28 crc kubenswrapper[4791]: I1210 22:54:28.385083 4791 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/marketplace-operator-79b997595-7g828" podUID="a9185ee1-adaf-48e8-b9a9-90ef658f8212" containerName="marketplace-operator" containerID="cri-o://b4b2f13d92bcf32cde11b26e411412c60deeba2f579aca917b1aea04f10c82e2" gracePeriod=30 Dec 10 22:54:28 crc kubenswrapper[4791]: I1210 22:54:28.397385 4791 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-wkwbl"] Dec 10 22:54:28 crc kubenswrapper[4791]: I1210 22:54:28.398398 4791 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-marketplace-wkwbl" podUID="396f55d6-c96d-4520-8165-862b5905a8c0" containerName="registry-server" containerID="cri-o://bc477107547cc777dce54e21ac00704217e22d0bd8695c6a9bac537e5daa72f1" gracePeriod=30 Dec 10 22:54:28 crc kubenswrapper[4791]: I1210 22:54:28.432029 4791 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/marketplace-operator-79b997595-8b9dm"] Dec 10 22:54:28 crc kubenswrapper[4791]: E1210 22:54:28.432422 4791 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f85e55b1a89d02b0cb034b1ea31ed45a" containerName="startup-monitor" Dec 10 22:54:28 crc kubenswrapper[4791]: I1210 22:54:28.432446 4791 state_mem.go:107] "Deleted CPUSet assignment" podUID="f85e55b1a89d02b0cb034b1ea31ed45a" containerName="startup-monitor" Dec 10 22:54:28 crc kubenswrapper[4791]: I1210 22:54:28.432585 4791 memory_manager.go:354] "RemoveStaleState removing state" podUID="f85e55b1a89d02b0cb034b1ea31ed45a" containerName="startup-monitor" Dec 10 22:54:28 crc kubenswrapper[4791]: I1210 22:54:28.433072 4791 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/marketplace-operator-79b997595-8b9dm" Dec 10 22:54:28 crc kubenswrapper[4791]: I1210 22:54:28.436618 4791 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-tj8hn"] Dec 10 22:54:28 crc kubenswrapper[4791]: I1210 22:54:28.436914 4791 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-operators-tj8hn" podUID="ab4c6e9a-3413-49c1-8f49-fb22c8366cb6" containerName="registry-server" containerID="cri-o://ec98715ef0499713aacf43ea6bdd8b414c5f4931f812cc72b9d0f6235bd510cf" gracePeriod=30 Dec 10 22:54:28 crc kubenswrapper[4791]: I1210 22:54:28.450494 4791 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/marketplace-operator-79b997595-8b9dm"] Dec 10 22:54:28 crc kubenswrapper[4791]: I1210 22:54:28.534638 4791 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"marketplace-trusted-ca\" (UniqueName: \"kubernetes.io/configmap/c2479901-785c-4a0a-b494-90675f64b71d-marketplace-trusted-ca\") pod \"marketplace-operator-79b997595-8b9dm\" (UID: \"c2479901-785c-4a0a-b494-90675f64b71d\") " pod="openshift-marketplace/marketplace-operator-79b997595-8b9dm" Dec 10 22:54:28 crc kubenswrapper[4791]: I1210 22:54:28.534870 4791 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-2hn28\" (UniqueName: \"kubernetes.io/projected/c2479901-785c-4a0a-b494-90675f64b71d-kube-api-access-2hn28\") pod \"marketplace-operator-79b997595-8b9dm\" (UID: \"c2479901-785c-4a0a-b494-90675f64b71d\") " pod="openshift-marketplace/marketplace-operator-79b997595-8b9dm" Dec 10 22:54:28 crc kubenswrapper[4791]: I1210 22:54:28.534964 4791 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"marketplace-operator-metrics\" (UniqueName: \"kubernetes.io/secret/c2479901-785c-4a0a-b494-90675f64b71d-marketplace-operator-metrics\") pod \"marketplace-operator-79b997595-8b9dm\" (UID: \"c2479901-785c-4a0a-b494-90675f64b71d\") " pod="openshift-marketplace/marketplace-operator-79b997595-8b9dm" Dec 10 22:54:28 crc kubenswrapper[4791]: I1210 22:54:28.636308 4791 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"marketplace-operator-metrics\" (UniqueName: \"kubernetes.io/secret/c2479901-785c-4a0a-b494-90675f64b71d-marketplace-operator-metrics\") pod \"marketplace-operator-79b997595-8b9dm\" (UID: \"c2479901-785c-4a0a-b494-90675f64b71d\") " pod="openshift-marketplace/marketplace-operator-79b997595-8b9dm" Dec 10 22:54:28 crc kubenswrapper[4791]: I1210 22:54:28.636509 4791 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"marketplace-trusted-ca\" (UniqueName: \"kubernetes.io/configmap/c2479901-785c-4a0a-b494-90675f64b71d-marketplace-trusted-ca\") pod \"marketplace-operator-79b997595-8b9dm\" (UID: \"c2479901-785c-4a0a-b494-90675f64b71d\") " pod="openshift-marketplace/marketplace-operator-79b997595-8b9dm" Dec 10 22:54:28 crc kubenswrapper[4791]: I1210 22:54:28.636564 4791 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-2hn28\" (UniqueName: \"kubernetes.io/projected/c2479901-785c-4a0a-b494-90675f64b71d-kube-api-access-2hn28\") pod \"marketplace-operator-79b997595-8b9dm\" (UID: \"c2479901-785c-4a0a-b494-90675f64b71d\") " pod="openshift-marketplace/marketplace-operator-79b997595-8b9dm" Dec 10 22:54:28 crc kubenswrapper[4791]: I1210 22:54:28.637959 4791 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"marketplace-trusted-ca\" (UniqueName: \"kubernetes.io/configmap/c2479901-785c-4a0a-b494-90675f64b71d-marketplace-trusted-ca\") pod \"marketplace-operator-79b997595-8b9dm\" (UID: \"c2479901-785c-4a0a-b494-90675f64b71d\") " pod="openshift-marketplace/marketplace-operator-79b997595-8b9dm" Dec 10 22:54:28 crc kubenswrapper[4791]: I1210 22:54:28.644000 4791 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"marketplace-operator-metrics\" (UniqueName: \"kubernetes.io/secret/c2479901-785c-4a0a-b494-90675f64b71d-marketplace-operator-metrics\") pod \"marketplace-operator-79b997595-8b9dm\" (UID: \"c2479901-785c-4a0a-b494-90675f64b71d\") " pod="openshift-marketplace/marketplace-operator-79b997595-8b9dm" Dec 10 22:54:28 crc kubenswrapper[4791]: I1210 22:54:28.654632 4791 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-2hn28\" (UniqueName: \"kubernetes.io/projected/c2479901-785c-4a0a-b494-90675f64b71d-kube-api-access-2hn28\") pod \"marketplace-operator-79b997595-8b9dm\" (UID: \"c2479901-785c-4a0a-b494-90675f64b71d\") " pod="openshift-marketplace/marketplace-operator-79b997595-8b9dm" Dec 10 22:54:28 crc kubenswrapper[4791]: I1210 22:54:28.749869 4791 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/marketplace-operator-79b997595-8b9dm" Dec 10 22:54:29 crc kubenswrapper[4791]: I1210 22:54:29.050371 4791 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/marketplace-operator-79b997595-8b9dm"] Dec 10 22:54:29 crc kubenswrapper[4791]: W1210 22:54:29.081898 4791 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podc2479901_785c_4a0a_b494_90675f64b71d.slice/crio-b072dfa332b9cc39cb22297d168cd2ad3ea3cb4aa52bf737938e7d9f3119710c WatchSource:0}: Error finding container b072dfa332b9cc39cb22297d168cd2ad3ea3cb4aa52bf737938e7d9f3119710c: Status 404 returned error can't find the container with id b072dfa332b9cc39cb22297d168cd2ad3ea3cb4aa52bf737938e7d9f3119710c Dec 10 22:54:29 crc kubenswrapper[4791]: I1210 22:54:29.296060 4791 generic.go:334] "Generic (PLEG): container finished" podID="396f55d6-c96d-4520-8165-862b5905a8c0" containerID="bc477107547cc777dce54e21ac00704217e22d0bd8695c6a9bac537e5daa72f1" exitCode=0 Dec 10 22:54:29 crc kubenswrapper[4791]: I1210 22:54:29.296141 4791 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-wkwbl" event={"ID":"396f55d6-c96d-4520-8165-862b5905a8c0","Type":"ContainerDied","Data":"bc477107547cc777dce54e21ac00704217e22d0bd8695c6a9bac537e5daa72f1"} Dec 10 22:54:29 crc kubenswrapper[4791]: I1210 22:54:29.298732 4791 generic.go:334] "Generic (PLEG): container finished" podID="a9185ee1-adaf-48e8-b9a9-90ef658f8212" containerID="b4b2f13d92bcf32cde11b26e411412c60deeba2f579aca917b1aea04f10c82e2" exitCode=0 Dec 10 22:54:29 crc kubenswrapper[4791]: I1210 22:54:29.298810 4791 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/marketplace-operator-79b997595-7g828" event={"ID":"a9185ee1-adaf-48e8-b9a9-90ef658f8212","Type":"ContainerDied","Data":"b4b2f13d92bcf32cde11b26e411412c60deeba2f579aca917b1aea04f10c82e2"} Dec 10 22:54:29 crc kubenswrapper[4791]: I1210 22:54:29.300741 4791 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/marketplace-operator-79b997595-8b9dm" event={"ID":"c2479901-785c-4a0a-b494-90675f64b71d","Type":"ContainerStarted","Data":"b072dfa332b9cc39cb22297d168cd2ad3ea3cb4aa52bf737938e7d9f3119710c"} Dec 10 22:54:29 crc kubenswrapper[4791]: I1210 22:54:29.303451 4791 generic.go:334] "Generic (PLEG): container finished" podID="08d50c37-7a4d-4731-beb8-fbc6d2e10bb0" containerID="b95dc6273b4eb68303d46abd4b7a567379de20b6c589ae921a68bc99c91fd48e" exitCode=0 Dec 10 22:54:29 crc kubenswrapper[4791]: I1210 22:54:29.303481 4791 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-gskzr" event={"ID":"08d50c37-7a4d-4731-beb8-fbc6d2e10bb0","Type":"ContainerDied","Data":"b95dc6273b4eb68303d46abd4b7a567379de20b6c589ae921a68bc99c91fd48e"} Dec 10 22:54:29 crc kubenswrapper[4791]: E1210 22:54:29.666456 4791 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = NotFound desc = container is not created or running: checking if PID of 716a7c482ddb51c0626fcc775f7f3b405fbe182e7fd48c9aa126321e8e6267d9 is running failed: container process not found" containerID="716a7c482ddb51c0626fcc775f7f3b405fbe182e7fd48c9aa126321e8e6267d9" cmd=["grpc_health_probe","-addr=:50051"] Dec 10 22:54:29 crc kubenswrapper[4791]: E1210 22:54:29.666931 4791 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = NotFound desc = container is not created or running: checking if PID of 716a7c482ddb51c0626fcc775f7f3b405fbe182e7fd48c9aa126321e8e6267d9 is running failed: container process not found" containerID="716a7c482ddb51c0626fcc775f7f3b405fbe182e7fd48c9aa126321e8e6267d9" cmd=["grpc_health_probe","-addr=:50051"] Dec 10 22:54:29 crc kubenswrapper[4791]: E1210 22:54:29.667471 4791 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = NotFound desc = container is not created or running: checking if PID of 716a7c482ddb51c0626fcc775f7f3b405fbe182e7fd48c9aa126321e8e6267d9 is running failed: container process not found" containerID="716a7c482ddb51c0626fcc775f7f3b405fbe182e7fd48c9aa126321e8e6267d9" cmd=["grpc_health_probe","-addr=:50051"] Dec 10 22:54:29 crc kubenswrapper[4791]: E1210 22:54:29.667526 4791 prober.go:104] "Probe errored" err="rpc error: code = NotFound desc = container is not created or running: checking if PID of 716a7c482ddb51c0626fcc775f7f3b405fbe182e7fd48c9aa126321e8e6267d9 is running failed: container process not found" probeType="Readiness" pod="openshift-marketplace/certified-operators-8x2rc" podUID="009efc35-5943-429f-98fd-3307a5f557f2" containerName="registry-server" Dec 10 22:54:29 crc kubenswrapper[4791]: E1210 22:54:29.911524 4791 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = NotFound desc = container is not created or running: checking if PID of b95dc6273b4eb68303d46abd4b7a567379de20b6c589ae921a68bc99c91fd48e is running failed: container process not found" containerID="b95dc6273b4eb68303d46abd4b7a567379de20b6c589ae921a68bc99c91fd48e" cmd=["grpc_health_probe","-addr=:50051"] Dec 10 22:54:29 crc kubenswrapper[4791]: E1210 22:54:29.911984 4791 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = NotFound desc = container is not created or running: checking if PID of b95dc6273b4eb68303d46abd4b7a567379de20b6c589ae921a68bc99c91fd48e is running failed: container process not found" containerID="b95dc6273b4eb68303d46abd4b7a567379de20b6c589ae921a68bc99c91fd48e" cmd=["grpc_health_probe","-addr=:50051"] Dec 10 22:54:29 crc kubenswrapper[4791]: E1210 22:54:29.912418 4791 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = NotFound desc = container is not created or running: checking if PID of b95dc6273b4eb68303d46abd4b7a567379de20b6c589ae921a68bc99c91fd48e is running failed: container process not found" containerID="b95dc6273b4eb68303d46abd4b7a567379de20b6c589ae921a68bc99c91fd48e" cmd=["grpc_health_probe","-addr=:50051"] Dec 10 22:54:29 crc kubenswrapper[4791]: E1210 22:54:29.912466 4791 prober.go:104] "Probe errored" err="rpc error: code = NotFound desc = container is not created or running: checking if PID of b95dc6273b4eb68303d46abd4b7a567379de20b6c589ae921a68bc99c91fd48e is running failed: container process not found" probeType="Readiness" pod="openshift-marketplace/community-operators-gskzr" podUID="08d50c37-7a4d-4731-beb8-fbc6d2e10bb0" containerName="registry-server" Dec 10 22:54:30 crc kubenswrapper[4791]: I1210 22:54:30.314700 4791 generic.go:334] "Generic (PLEG): container finished" podID="ab4c6e9a-3413-49c1-8f49-fb22c8366cb6" containerID="ec98715ef0499713aacf43ea6bdd8b414c5f4931f812cc72b9d0f6235bd510cf" exitCode=0 Dec 10 22:54:30 crc kubenswrapper[4791]: I1210 22:54:30.314780 4791 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-tj8hn" event={"ID":"ab4c6e9a-3413-49c1-8f49-fb22c8366cb6","Type":"ContainerDied","Data":"ec98715ef0499713aacf43ea6bdd8b414c5f4931f812cc72b9d0f6235bd510cf"} Dec 10 22:54:30 crc kubenswrapper[4791]: I1210 22:54:30.318384 4791 generic.go:334] "Generic (PLEG): container finished" podID="009efc35-5943-429f-98fd-3307a5f557f2" containerID="716a7c482ddb51c0626fcc775f7f3b405fbe182e7fd48c9aa126321e8e6267d9" exitCode=0 Dec 10 22:54:30 crc kubenswrapper[4791]: I1210 22:54:30.318428 4791 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-8x2rc" event={"ID":"009efc35-5943-429f-98fd-3307a5f557f2","Type":"ContainerDied","Data":"716a7c482ddb51c0626fcc775f7f3b405fbe182e7fd48c9aa126321e8e6267d9"} Dec 10 22:54:31 crc kubenswrapper[4791]: I1210 22:54:31.327592 4791 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/marketplace-operator-79b997595-8b9dm" event={"ID":"c2479901-785c-4a0a-b494-90675f64b71d","Type":"ContainerStarted","Data":"f517f0bc880ceb3d464991f393f754f2798aa665208395eee09d11b483bfb2fd"} Dec 10 22:54:31 crc kubenswrapper[4791]: I1210 22:54:31.551816 4791 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-gskzr" Dec 10 22:54:31 crc kubenswrapper[4791]: I1210 22:54:31.675606 4791 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/08d50c37-7a4d-4731-beb8-fbc6d2e10bb0-utilities\") pod \"08d50c37-7a4d-4731-beb8-fbc6d2e10bb0\" (UID: \"08d50c37-7a4d-4731-beb8-fbc6d2e10bb0\") " Dec 10 22:54:31 crc kubenswrapper[4791]: I1210 22:54:31.675715 4791 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-qm95v\" (UniqueName: \"kubernetes.io/projected/08d50c37-7a4d-4731-beb8-fbc6d2e10bb0-kube-api-access-qm95v\") pod \"08d50c37-7a4d-4731-beb8-fbc6d2e10bb0\" (UID: \"08d50c37-7a4d-4731-beb8-fbc6d2e10bb0\") " Dec 10 22:54:31 crc kubenswrapper[4791]: I1210 22:54:31.675743 4791 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/08d50c37-7a4d-4731-beb8-fbc6d2e10bb0-catalog-content\") pod \"08d50c37-7a4d-4731-beb8-fbc6d2e10bb0\" (UID: \"08d50c37-7a4d-4731-beb8-fbc6d2e10bb0\") " Dec 10 22:54:31 crc kubenswrapper[4791]: I1210 22:54:31.678677 4791 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/08d50c37-7a4d-4731-beb8-fbc6d2e10bb0-utilities" (OuterVolumeSpecName: "utilities") pod "08d50c37-7a4d-4731-beb8-fbc6d2e10bb0" (UID: "08d50c37-7a4d-4731-beb8-fbc6d2e10bb0"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 10 22:54:31 crc kubenswrapper[4791]: I1210 22:54:31.690235 4791 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/08d50c37-7a4d-4731-beb8-fbc6d2e10bb0-kube-api-access-qm95v" (OuterVolumeSpecName: "kube-api-access-qm95v") pod "08d50c37-7a4d-4731-beb8-fbc6d2e10bb0" (UID: "08d50c37-7a4d-4731-beb8-fbc6d2e10bb0"). InnerVolumeSpecName "kube-api-access-qm95v". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 10 22:54:31 crc kubenswrapper[4791]: I1210 22:54:31.740632 4791 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/08d50c37-7a4d-4731-beb8-fbc6d2e10bb0-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "08d50c37-7a4d-4731-beb8-fbc6d2e10bb0" (UID: "08d50c37-7a4d-4731-beb8-fbc6d2e10bb0"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 10 22:54:31 crc kubenswrapper[4791]: I1210 22:54:31.744105 4791 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-wkwbl" Dec 10 22:54:31 crc kubenswrapper[4791]: I1210 22:54:31.748271 4791 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-8x2rc" Dec 10 22:54:31 crc kubenswrapper[4791]: I1210 22:54:31.752871 4791 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/marketplace-operator-79b997595-7g828" Dec 10 22:54:31 crc kubenswrapper[4791]: I1210 22:54:31.758227 4791 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-tj8hn" Dec 10 22:54:31 crc kubenswrapper[4791]: I1210 22:54:31.778010 4791 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/08d50c37-7a4d-4731-beb8-fbc6d2e10bb0-utilities\") on node \"crc\" DevicePath \"\"" Dec 10 22:54:31 crc kubenswrapper[4791]: I1210 22:54:31.778041 4791 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/08d50c37-7a4d-4731-beb8-fbc6d2e10bb0-catalog-content\") on node \"crc\" DevicePath \"\"" Dec 10 22:54:31 crc kubenswrapper[4791]: I1210 22:54:31.778054 4791 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-qm95v\" (UniqueName: \"kubernetes.io/projected/08d50c37-7a4d-4731-beb8-fbc6d2e10bb0-kube-api-access-qm95v\") on node \"crc\" DevicePath \"\"" Dec 10 22:54:31 crc kubenswrapper[4791]: I1210 22:54:31.879460 4791 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/396f55d6-c96d-4520-8165-862b5905a8c0-catalog-content\") pod \"396f55d6-c96d-4520-8165-862b5905a8c0\" (UID: \"396f55d6-c96d-4520-8165-862b5905a8c0\") " Dec 10 22:54:31 crc kubenswrapper[4791]: I1210 22:54:31.879593 4791 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-z2kjq\" (UniqueName: \"kubernetes.io/projected/009efc35-5943-429f-98fd-3307a5f557f2-kube-api-access-z2kjq\") pod \"009efc35-5943-429f-98fd-3307a5f557f2\" (UID: \"009efc35-5943-429f-98fd-3307a5f557f2\") " Dec 10 22:54:31 crc kubenswrapper[4791]: I1210 22:54:31.879819 4791 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/009efc35-5943-429f-98fd-3307a5f557f2-catalog-content\") pod \"009efc35-5943-429f-98fd-3307a5f557f2\" (UID: \"009efc35-5943-429f-98fd-3307a5f557f2\") " Dec 10 22:54:31 crc kubenswrapper[4791]: I1210 22:54:31.879852 4791 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"marketplace-trusted-ca\" (UniqueName: \"kubernetes.io/configmap/a9185ee1-adaf-48e8-b9a9-90ef658f8212-marketplace-trusted-ca\") pod \"a9185ee1-adaf-48e8-b9a9-90ef658f8212\" (UID: \"a9185ee1-adaf-48e8-b9a9-90ef658f8212\") " Dec 10 22:54:31 crc kubenswrapper[4791]: I1210 22:54:31.879897 4791 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-mq8d4\" (UniqueName: \"kubernetes.io/projected/396f55d6-c96d-4520-8165-862b5905a8c0-kube-api-access-mq8d4\") pod \"396f55d6-c96d-4520-8165-862b5905a8c0\" (UID: \"396f55d6-c96d-4520-8165-862b5905a8c0\") " Dec 10 22:54:31 crc kubenswrapper[4791]: I1210 22:54:31.879920 4791 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/ab4c6e9a-3413-49c1-8f49-fb22c8366cb6-utilities\") pod \"ab4c6e9a-3413-49c1-8f49-fb22c8366cb6\" (UID: \"ab4c6e9a-3413-49c1-8f49-fb22c8366cb6\") " Dec 10 22:54:31 crc kubenswrapper[4791]: I1210 22:54:31.879961 4791 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-hsmgd\" (UniqueName: \"kubernetes.io/projected/a9185ee1-adaf-48e8-b9a9-90ef658f8212-kube-api-access-hsmgd\") pod \"a9185ee1-adaf-48e8-b9a9-90ef658f8212\" (UID: \"a9185ee1-adaf-48e8-b9a9-90ef658f8212\") " Dec 10 22:54:31 crc kubenswrapper[4791]: I1210 22:54:31.879992 4791 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/009efc35-5943-429f-98fd-3307a5f557f2-utilities\") pod \"009efc35-5943-429f-98fd-3307a5f557f2\" (UID: \"009efc35-5943-429f-98fd-3307a5f557f2\") " Dec 10 22:54:31 crc kubenswrapper[4791]: I1210 22:54:31.880021 4791 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"marketplace-operator-metrics\" (UniqueName: \"kubernetes.io/secret/a9185ee1-adaf-48e8-b9a9-90ef658f8212-marketplace-operator-metrics\") pod \"a9185ee1-adaf-48e8-b9a9-90ef658f8212\" (UID: \"a9185ee1-adaf-48e8-b9a9-90ef658f8212\") " Dec 10 22:54:31 crc kubenswrapper[4791]: I1210 22:54:31.880083 4791 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/396f55d6-c96d-4520-8165-862b5905a8c0-utilities\") pod \"396f55d6-c96d-4520-8165-862b5905a8c0\" (UID: \"396f55d6-c96d-4520-8165-862b5905a8c0\") " Dec 10 22:54:31 crc kubenswrapper[4791]: I1210 22:54:31.880119 4791 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-ck7dp\" (UniqueName: \"kubernetes.io/projected/ab4c6e9a-3413-49c1-8f49-fb22c8366cb6-kube-api-access-ck7dp\") pod \"ab4c6e9a-3413-49c1-8f49-fb22c8366cb6\" (UID: \"ab4c6e9a-3413-49c1-8f49-fb22c8366cb6\") " Dec 10 22:54:31 crc kubenswrapper[4791]: I1210 22:54:31.880154 4791 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/ab4c6e9a-3413-49c1-8f49-fb22c8366cb6-catalog-content\") pod \"ab4c6e9a-3413-49c1-8f49-fb22c8366cb6\" (UID: \"ab4c6e9a-3413-49c1-8f49-fb22c8366cb6\") " Dec 10 22:54:31 crc kubenswrapper[4791]: I1210 22:54:31.880963 4791 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/a9185ee1-adaf-48e8-b9a9-90ef658f8212-marketplace-trusted-ca" (OuterVolumeSpecName: "marketplace-trusted-ca") pod "a9185ee1-adaf-48e8-b9a9-90ef658f8212" (UID: "a9185ee1-adaf-48e8-b9a9-90ef658f8212"). InnerVolumeSpecName "marketplace-trusted-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 10 22:54:31 crc kubenswrapper[4791]: I1210 22:54:31.881235 4791 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/396f55d6-c96d-4520-8165-862b5905a8c0-utilities" (OuterVolumeSpecName: "utilities") pod "396f55d6-c96d-4520-8165-862b5905a8c0" (UID: "396f55d6-c96d-4520-8165-862b5905a8c0"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 10 22:54:31 crc kubenswrapper[4791]: I1210 22:54:31.881328 4791 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/009efc35-5943-429f-98fd-3307a5f557f2-utilities" (OuterVolumeSpecName: "utilities") pod "009efc35-5943-429f-98fd-3307a5f557f2" (UID: "009efc35-5943-429f-98fd-3307a5f557f2"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 10 22:54:31 crc kubenswrapper[4791]: I1210 22:54:31.881509 4791 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/ab4c6e9a-3413-49c1-8f49-fb22c8366cb6-utilities" (OuterVolumeSpecName: "utilities") pod "ab4c6e9a-3413-49c1-8f49-fb22c8366cb6" (UID: "ab4c6e9a-3413-49c1-8f49-fb22c8366cb6"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 10 22:54:31 crc kubenswrapper[4791]: I1210 22:54:31.882909 4791 reconciler_common.go:293] "Volume detached for volume \"marketplace-trusted-ca\" (UniqueName: \"kubernetes.io/configmap/a9185ee1-adaf-48e8-b9a9-90ef658f8212-marketplace-trusted-ca\") on node \"crc\" DevicePath \"\"" Dec 10 22:54:31 crc kubenswrapper[4791]: I1210 22:54:31.883070 4791 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/ab4c6e9a-3413-49c1-8f49-fb22c8366cb6-utilities\") on node \"crc\" DevicePath \"\"" Dec 10 22:54:31 crc kubenswrapper[4791]: I1210 22:54:31.883176 4791 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/009efc35-5943-429f-98fd-3307a5f557f2-utilities\") on node \"crc\" DevicePath \"\"" Dec 10 22:54:31 crc kubenswrapper[4791]: I1210 22:54:31.883284 4791 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/396f55d6-c96d-4520-8165-862b5905a8c0-utilities\") on node \"crc\" DevicePath \"\"" Dec 10 22:54:31 crc kubenswrapper[4791]: I1210 22:54:31.884465 4791 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/396f55d6-c96d-4520-8165-862b5905a8c0-kube-api-access-mq8d4" (OuterVolumeSpecName: "kube-api-access-mq8d4") pod "396f55d6-c96d-4520-8165-862b5905a8c0" (UID: "396f55d6-c96d-4520-8165-862b5905a8c0"). InnerVolumeSpecName "kube-api-access-mq8d4". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 10 22:54:31 crc kubenswrapper[4791]: I1210 22:54:31.884778 4791 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/a9185ee1-adaf-48e8-b9a9-90ef658f8212-kube-api-access-hsmgd" (OuterVolumeSpecName: "kube-api-access-hsmgd") pod "a9185ee1-adaf-48e8-b9a9-90ef658f8212" (UID: "a9185ee1-adaf-48e8-b9a9-90ef658f8212"). InnerVolumeSpecName "kube-api-access-hsmgd". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 10 22:54:31 crc kubenswrapper[4791]: I1210 22:54:31.885011 4791 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/a9185ee1-adaf-48e8-b9a9-90ef658f8212-marketplace-operator-metrics" (OuterVolumeSpecName: "marketplace-operator-metrics") pod "a9185ee1-adaf-48e8-b9a9-90ef658f8212" (UID: "a9185ee1-adaf-48e8-b9a9-90ef658f8212"). InnerVolumeSpecName "marketplace-operator-metrics". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 10 22:54:31 crc kubenswrapper[4791]: I1210 22:54:31.885202 4791 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/009efc35-5943-429f-98fd-3307a5f557f2-kube-api-access-z2kjq" (OuterVolumeSpecName: "kube-api-access-z2kjq") pod "009efc35-5943-429f-98fd-3307a5f557f2" (UID: "009efc35-5943-429f-98fd-3307a5f557f2"). InnerVolumeSpecName "kube-api-access-z2kjq". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 10 22:54:31 crc kubenswrapper[4791]: I1210 22:54:31.886149 4791 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/ab4c6e9a-3413-49c1-8f49-fb22c8366cb6-kube-api-access-ck7dp" (OuterVolumeSpecName: "kube-api-access-ck7dp") pod "ab4c6e9a-3413-49c1-8f49-fb22c8366cb6" (UID: "ab4c6e9a-3413-49c1-8f49-fb22c8366cb6"). InnerVolumeSpecName "kube-api-access-ck7dp". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 10 22:54:31 crc kubenswrapper[4791]: I1210 22:54:31.915016 4791 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/396f55d6-c96d-4520-8165-862b5905a8c0-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "396f55d6-c96d-4520-8165-862b5905a8c0" (UID: "396f55d6-c96d-4520-8165-862b5905a8c0"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 10 22:54:31 crc kubenswrapper[4791]: I1210 22:54:31.950301 4791 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/009efc35-5943-429f-98fd-3307a5f557f2-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "009efc35-5943-429f-98fd-3307a5f557f2" (UID: "009efc35-5943-429f-98fd-3307a5f557f2"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 10 22:54:31 crc kubenswrapper[4791]: I1210 22:54:31.984762 4791 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-z2kjq\" (UniqueName: \"kubernetes.io/projected/009efc35-5943-429f-98fd-3307a5f557f2-kube-api-access-z2kjq\") on node \"crc\" DevicePath \"\"" Dec 10 22:54:31 crc kubenswrapper[4791]: I1210 22:54:31.984791 4791 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/009efc35-5943-429f-98fd-3307a5f557f2-catalog-content\") on node \"crc\" DevicePath \"\"" Dec 10 22:54:31 crc kubenswrapper[4791]: I1210 22:54:31.984800 4791 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-mq8d4\" (UniqueName: \"kubernetes.io/projected/396f55d6-c96d-4520-8165-862b5905a8c0-kube-api-access-mq8d4\") on node \"crc\" DevicePath \"\"" Dec 10 22:54:31 crc kubenswrapper[4791]: I1210 22:54:31.984809 4791 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-hsmgd\" (UniqueName: \"kubernetes.io/projected/a9185ee1-adaf-48e8-b9a9-90ef658f8212-kube-api-access-hsmgd\") on node \"crc\" DevicePath \"\"" Dec 10 22:54:31 crc kubenswrapper[4791]: I1210 22:54:31.984818 4791 reconciler_common.go:293] "Volume detached for volume \"marketplace-operator-metrics\" (UniqueName: \"kubernetes.io/secret/a9185ee1-adaf-48e8-b9a9-90ef658f8212-marketplace-operator-metrics\") on node \"crc\" DevicePath \"\"" Dec 10 22:54:31 crc kubenswrapper[4791]: I1210 22:54:31.984827 4791 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-ck7dp\" (UniqueName: \"kubernetes.io/projected/ab4c6e9a-3413-49c1-8f49-fb22c8366cb6-kube-api-access-ck7dp\") on node \"crc\" DevicePath \"\"" Dec 10 22:54:31 crc kubenswrapper[4791]: I1210 22:54:31.984836 4791 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/396f55d6-c96d-4520-8165-862b5905a8c0-catalog-content\") on node \"crc\" DevicePath \"\"" Dec 10 22:54:32 crc kubenswrapper[4791]: I1210 22:54:32.009745 4791 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/ab4c6e9a-3413-49c1-8f49-fb22c8366cb6-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "ab4c6e9a-3413-49c1-8f49-fb22c8366cb6" (UID: "ab4c6e9a-3413-49c1-8f49-fb22c8366cb6"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 10 22:54:32 crc kubenswrapper[4791]: I1210 22:54:32.086078 4791 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/ab4c6e9a-3413-49c1-8f49-fb22c8366cb6-catalog-content\") on node \"crc\" DevicePath \"\"" Dec 10 22:54:32 crc kubenswrapper[4791]: I1210 22:54:32.325843 4791 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-oauth-apiserver"/"etcd-client" Dec 10 22:54:32 crc kubenswrapper[4791]: I1210 22:54:32.336013 4791 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-tj8hn" event={"ID":"ab4c6e9a-3413-49c1-8f49-fb22c8366cb6","Type":"ContainerDied","Data":"44af7aee753f8d061171bf3f5302bff67aec21202f33f7425d7c967c105223c0"} Dec 10 22:54:32 crc kubenswrapper[4791]: I1210 22:54:32.336079 4791 scope.go:117] "RemoveContainer" containerID="ec98715ef0499713aacf43ea6bdd8b414c5f4931f812cc72b9d0f6235bd510cf" Dec 10 22:54:32 crc kubenswrapper[4791]: I1210 22:54:32.336025 4791 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-tj8hn" Dec 10 22:54:32 crc kubenswrapper[4791]: I1210 22:54:32.341359 4791 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-gskzr" Dec 10 22:54:32 crc kubenswrapper[4791]: I1210 22:54:32.341461 4791 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-gskzr" event={"ID":"08d50c37-7a4d-4731-beb8-fbc6d2e10bb0","Type":"ContainerDied","Data":"fb24af949a64f3aba6e54c13b2a56bcabd7e6f909d37593e4fc58035bb20bf7e"} Dec 10 22:54:32 crc kubenswrapper[4791]: I1210 22:54:32.348407 4791 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-wkwbl" Dec 10 22:54:32 crc kubenswrapper[4791]: I1210 22:54:32.348400 4791 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-wkwbl" event={"ID":"396f55d6-c96d-4520-8165-862b5905a8c0","Type":"ContainerDied","Data":"867bd37f5f9510139aaf03f18fd25afbac0100646d83d88e2ae1c3b62c657cff"} Dec 10 22:54:32 crc kubenswrapper[4791]: I1210 22:54:32.354226 4791 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-8x2rc" event={"ID":"009efc35-5943-429f-98fd-3307a5f557f2","Type":"ContainerDied","Data":"d76e6bdaae80e7f5f3b052f64ad503cba5daecead3682b207dbfbbbb7671bc6a"} Dec 10 22:54:32 crc kubenswrapper[4791]: I1210 22:54:32.354405 4791 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-8x2rc" Dec 10 22:54:32 crc kubenswrapper[4791]: I1210 22:54:32.357374 4791 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/marketplace-operator-79b997595-7g828" event={"ID":"a9185ee1-adaf-48e8-b9a9-90ef658f8212","Type":"ContainerDied","Data":"a2010281d4a0a094eb30b9d1a4f837090035e9f671f22891a41b5288b0f507fc"} Dec 10 22:54:32 crc kubenswrapper[4791]: I1210 22:54:32.357391 4791 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/marketplace-operator-79b997595-7g828" Dec 10 22:54:32 crc kubenswrapper[4791]: I1210 22:54:32.357828 4791 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/marketplace-operator-79b997595-8b9dm" Dec 10 22:54:32 crc kubenswrapper[4791]: I1210 22:54:32.358415 4791 scope.go:117] "RemoveContainer" containerID="f62e716ba96d2eeb3fa3d37a632ba2bf20f383b5d039007cb3bb88040fa60ac5" Dec 10 22:54:32 crc kubenswrapper[4791]: I1210 22:54:32.364786 4791 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/marketplace-operator-79b997595-8b9dm" Dec 10 22:54:32 crc kubenswrapper[4791]: I1210 22:54:32.368152 4791 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-gskzr"] Dec 10 22:54:32 crc kubenswrapper[4791]: I1210 22:54:32.374458 4791 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/community-operators-gskzr"] Dec 10 22:54:32 crc kubenswrapper[4791]: I1210 22:54:32.380419 4791 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-tj8hn"] Dec 10 22:54:32 crc kubenswrapper[4791]: I1210 22:54:32.385247 4791 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-operators-tj8hn"] Dec 10 22:54:32 crc kubenswrapper[4791]: I1210 22:54:32.406999 4791 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/marketplace-operator-79b997595-8b9dm" podStartSLOduration=4.406975759 podStartE2EDuration="4.406975759s" podCreationTimestamp="2025-12-10 22:54:28 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-10 22:54:32.400372545 +0000 UTC m=+306.829990158" watchObservedRunningTime="2025-12-10 22:54:32.406975759 +0000 UTC m=+306.836593372" Dec 10 22:54:32 crc kubenswrapper[4791]: I1210 22:54:32.414127 4791 scope.go:117] "RemoveContainer" containerID="8cb71106b36181af379ea2e38b1fd99f7c46ecb2e6edd181c06339af142843e1" Dec 10 22:54:32 crc kubenswrapper[4791]: I1210 22:54:32.439921 4791 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/marketplace-operator-79b997595-7g828"] Dec 10 22:54:32 crc kubenswrapper[4791]: I1210 22:54:32.443369 4791 scope.go:117] "RemoveContainer" containerID="b95dc6273b4eb68303d46abd4b7a567379de20b6c589ae921a68bc99c91fd48e" Dec 10 22:54:32 crc kubenswrapper[4791]: I1210 22:54:32.443809 4791 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/marketplace-operator-79b997595-7g828"] Dec 10 22:54:32 crc kubenswrapper[4791]: I1210 22:54:32.453450 4791 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-8x2rc"] Dec 10 22:54:32 crc kubenswrapper[4791]: I1210 22:54:32.457833 4791 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/certified-operators-8x2rc"] Dec 10 22:54:32 crc kubenswrapper[4791]: I1210 22:54:32.476837 4791 scope.go:117] "RemoveContainer" containerID="58d691ba8d536c11c04f63e8e4cf0ff5bafc842c07ce006053f9244a17a81b0d" Dec 10 22:54:32 crc kubenswrapper[4791]: I1210 22:54:32.478126 4791 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-wkwbl"] Dec 10 22:54:32 crc kubenswrapper[4791]: I1210 22:54:32.493578 4791 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-marketplace-wkwbl"] Dec 10 22:54:32 crc kubenswrapper[4791]: I1210 22:54:32.499608 4791 scope.go:117] "RemoveContainer" containerID="efaa4cb838ed737e6af746d2de57003b1499ef7bbcbd43dcd52edbb01871d25b" Dec 10 22:54:32 crc kubenswrapper[4791]: I1210 22:54:32.513850 4791 scope.go:117] "RemoveContainer" containerID="bc477107547cc777dce54e21ac00704217e22d0bd8695c6a9bac537e5daa72f1" Dec 10 22:54:32 crc kubenswrapper[4791]: I1210 22:54:32.558115 4791 scope.go:117] "RemoveContainer" containerID="5f188d26e1f99d267391af0a2204832707a2d80f1b5747cb124469c3048ea6d8" Dec 10 22:54:32 crc kubenswrapper[4791]: I1210 22:54:32.573654 4791 scope.go:117] "RemoveContainer" containerID="2168264d99228a03c59b448c46e5492deb2f9d8b5c2dc0f6b1d727fe35dd6679" Dec 10 22:54:32 crc kubenswrapper[4791]: I1210 22:54:32.595118 4791 scope.go:117] "RemoveContainer" containerID="716a7c482ddb51c0626fcc775f7f3b405fbe182e7fd48c9aa126321e8e6267d9" Dec 10 22:54:32 crc kubenswrapper[4791]: I1210 22:54:32.613263 4791 scope.go:117] "RemoveContainer" containerID="d82a5b0dd8195ec88a52c56bfdb70f420c467627cf24a4ed0e5c7aa05f60366b" Dec 10 22:54:32 crc kubenswrapper[4791]: I1210 22:54:32.625984 4791 scope.go:117] "RemoveContainer" containerID="200d18941348a8e2c514d3071c544e59138565cf3001822a3d80bf82b72a1861" Dec 10 22:54:32 crc kubenswrapper[4791]: I1210 22:54:32.639276 4791 scope.go:117] "RemoveContainer" containerID="b4b2f13d92bcf32cde11b26e411412c60deeba2f579aca917b1aea04f10c82e2" Dec 10 22:54:33 crc kubenswrapper[4791]: I1210 22:54:33.891236 4791 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="009efc35-5943-429f-98fd-3307a5f557f2" path="/var/lib/kubelet/pods/009efc35-5943-429f-98fd-3307a5f557f2/volumes" Dec 10 22:54:33 crc kubenswrapper[4791]: I1210 22:54:33.892296 4791 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="08d50c37-7a4d-4731-beb8-fbc6d2e10bb0" path="/var/lib/kubelet/pods/08d50c37-7a4d-4731-beb8-fbc6d2e10bb0/volumes" Dec 10 22:54:33 crc kubenswrapper[4791]: I1210 22:54:33.893017 4791 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="396f55d6-c96d-4520-8165-862b5905a8c0" path="/var/lib/kubelet/pods/396f55d6-c96d-4520-8165-862b5905a8c0/volumes" Dec 10 22:54:33 crc kubenswrapper[4791]: I1210 22:54:33.894307 4791 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="a9185ee1-adaf-48e8-b9a9-90ef658f8212" path="/var/lib/kubelet/pods/a9185ee1-adaf-48e8-b9a9-90ef658f8212/volumes" Dec 10 22:54:33 crc kubenswrapper[4791]: I1210 22:54:33.895059 4791 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="ab4c6e9a-3413-49c1-8f49-fb22c8366cb6" path="/var/lib/kubelet/pods/ab4c6e9a-3413-49c1-8f49-fb22c8366cb6/volumes" Dec 10 22:54:34 crc kubenswrapper[4791]: I1210 22:54:34.659776 4791 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-dns-operator"/"openshift-service-ca.crt" Dec 10 22:54:36 crc kubenswrapper[4791]: I1210 22:54:36.380598 4791 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-marketplace"/"kube-root-ca.crt" Dec 10 22:54:38 crc kubenswrapper[4791]: I1210 22:54:38.744749 4791 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver-operator"/"openshift-apiserver-operator-config" Dec 10 22:54:42 crc kubenswrapper[4791]: I1210 22:54:42.552361 4791 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-controller-manager/controller-manager-879f6c89f-9tg4p"] Dec 10 22:54:42 crc kubenswrapper[4791]: I1210 22:54:42.553106 4791 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-controller-manager/controller-manager-879f6c89f-9tg4p" podUID="a5827e8a-4cb3-494d-87a1-556d920c2a4a" containerName="controller-manager" containerID="cri-o://1b567163092c2059ebeb396c427945f52d9972a1c85abdccb0bd739aac9e995f" gracePeriod=30 Dec 10 22:54:42 crc kubenswrapper[4791]: I1210 22:54:42.659750 4791 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-6576b87f9c-vsnsh"] Dec 10 22:54:42 crc kubenswrapper[4791]: I1210 22:54:42.660152 4791 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-vsnsh" podUID="5369b8df-3c36-4108-b4e7-cfb11478b3a7" containerName="route-controller-manager" containerID="cri-o://4b9f827d94a26dadac2b0de34bb8c127f77e2fb7e7156bfd1f63d2ea2036e664" gracePeriod=30 Dec 10 22:54:42 crc kubenswrapper[4791]: I1210 22:54:42.770861 4791 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver-operator"/"kube-root-ca.crt" Dec 10 22:54:42 crc kubenswrapper[4791]: I1210 22:54:42.945069 4791 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager/controller-manager-879f6c89f-9tg4p" Dec 10 22:54:43 crc kubenswrapper[4791]: I1210 22:54:43.013062 4791 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/a5827e8a-4cb3-494d-87a1-556d920c2a4a-config\") pod \"a5827e8a-4cb3-494d-87a1-556d920c2a4a\" (UID: \"a5827e8a-4cb3-494d-87a1-556d920c2a4a\") " Dec 10 22:54:43 crc kubenswrapper[4791]: I1210 22:54:43.013355 4791 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/a5827e8a-4cb3-494d-87a1-556d920c2a4a-proxy-ca-bundles\") pod \"a5827e8a-4cb3-494d-87a1-556d920c2a4a\" (UID: \"a5827e8a-4cb3-494d-87a1-556d920c2a4a\") " Dec 10 22:54:43 crc kubenswrapper[4791]: I1210 22:54:43.013381 4791 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/a5827e8a-4cb3-494d-87a1-556d920c2a4a-client-ca\") pod \"a5827e8a-4cb3-494d-87a1-556d920c2a4a\" (UID: \"a5827e8a-4cb3-494d-87a1-556d920c2a4a\") " Dec 10 22:54:43 crc kubenswrapper[4791]: I1210 22:54:43.013428 4791 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-xhsjq\" (UniqueName: \"kubernetes.io/projected/a5827e8a-4cb3-494d-87a1-556d920c2a4a-kube-api-access-xhsjq\") pod \"a5827e8a-4cb3-494d-87a1-556d920c2a4a\" (UID: \"a5827e8a-4cb3-494d-87a1-556d920c2a4a\") " Dec 10 22:54:43 crc kubenswrapper[4791]: I1210 22:54:43.013455 4791 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/a5827e8a-4cb3-494d-87a1-556d920c2a4a-serving-cert\") pod \"a5827e8a-4cb3-494d-87a1-556d920c2a4a\" (UID: \"a5827e8a-4cb3-494d-87a1-556d920c2a4a\") " Dec 10 22:54:43 crc kubenswrapper[4791]: I1210 22:54:43.013896 4791 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/a5827e8a-4cb3-494d-87a1-556d920c2a4a-client-ca" (OuterVolumeSpecName: "client-ca") pod "a5827e8a-4cb3-494d-87a1-556d920c2a4a" (UID: "a5827e8a-4cb3-494d-87a1-556d920c2a4a"). InnerVolumeSpecName "client-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 10 22:54:43 crc kubenswrapper[4791]: I1210 22:54:43.013962 4791 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/a5827e8a-4cb3-494d-87a1-556d920c2a4a-config" (OuterVolumeSpecName: "config") pod "a5827e8a-4cb3-494d-87a1-556d920c2a4a" (UID: "a5827e8a-4cb3-494d-87a1-556d920c2a4a"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 10 22:54:43 crc kubenswrapper[4791]: I1210 22:54:43.014641 4791 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/a5827e8a-4cb3-494d-87a1-556d920c2a4a-proxy-ca-bundles" (OuterVolumeSpecName: "proxy-ca-bundles") pod "a5827e8a-4cb3-494d-87a1-556d920c2a4a" (UID: "a5827e8a-4cb3-494d-87a1-556d920c2a4a"). InnerVolumeSpecName "proxy-ca-bundles". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 10 22:54:43 crc kubenswrapper[4791]: I1210 22:54:43.019105 4791 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/a5827e8a-4cb3-494d-87a1-556d920c2a4a-kube-api-access-xhsjq" (OuterVolumeSpecName: "kube-api-access-xhsjq") pod "a5827e8a-4cb3-494d-87a1-556d920c2a4a" (UID: "a5827e8a-4cb3-494d-87a1-556d920c2a4a"). InnerVolumeSpecName "kube-api-access-xhsjq". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 10 22:54:43 crc kubenswrapper[4791]: I1210 22:54:43.019137 4791 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/a5827e8a-4cb3-494d-87a1-556d920c2a4a-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "a5827e8a-4cb3-494d-87a1-556d920c2a4a" (UID: "a5827e8a-4cb3-494d-87a1-556d920c2a4a"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 10 22:54:43 crc kubenswrapper[4791]: I1210 22:54:43.027833 4791 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-vsnsh" Dec 10 22:54:43 crc kubenswrapper[4791]: I1210 22:54:43.115370 4791 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/5369b8df-3c36-4108-b4e7-cfb11478b3a7-serving-cert\") pod \"5369b8df-3c36-4108-b4e7-cfb11478b3a7\" (UID: \"5369b8df-3c36-4108-b4e7-cfb11478b3a7\") " Dec 10 22:54:43 crc kubenswrapper[4791]: I1210 22:54:43.115477 4791 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-z5dwf\" (UniqueName: \"kubernetes.io/projected/5369b8df-3c36-4108-b4e7-cfb11478b3a7-kube-api-access-z5dwf\") pod \"5369b8df-3c36-4108-b4e7-cfb11478b3a7\" (UID: \"5369b8df-3c36-4108-b4e7-cfb11478b3a7\") " Dec 10 22:54:43 crc kubenswrapper[4791]: I1210 22:54:43.115497 4791 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/5369b8df-3c36-4108-b4e7-cfb11478b3a7-config\") pod \"5369b8df-3c36-4108-b4e7-cfb11478b3a7\" (UID: \"5369b8df-3c36-4108-b4e7-cfb11478b3a7\") " Dec 10 22:54:43 crc kubenswrapper[4791]: I1210 22:54:43.115547 4791 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/5369b8df-3c36-4108-b4e7-cfb11478b3a7-client-ca\") pod \"5369b8df-3c36-4108-b4e7-cfb11478b3a7\" (UID: \"5369b8df-3c36-4108-b4e7-cfb11478b3a7\") " Dec 10 22:54:43 crc kubenswrapper[4791]: I1210 22:54:43.115894 4791 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/a5827e8a-4cb3-494d-87a1-556d920c2a4a-config\") on node \"crc\" DevicePath \"\"" Dec 10 22:54:43 crc kubenswrapper[4791]: I1210 22:54:43.115912 4791 reconciler_common.go:293] "Volume detached for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/a5827e8a-4cb3-494d-87a1-556d920c2a4a-proxy-ca-bundles\") on node \"crc\" DevicePath \"\"" Dec 10 22:54:43 crc kubenswrapper[4791]: I1210 22:54:43.115924 4791 reconciler_common.go:293] "Volume detached for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/a5827e8a-4cb3-494d-87a1-556d920c2a4a-client-ca\") on node \"crc\" DevicePath \"\"" Dec 10 22:54:43 crc kubenswrapper[4791]: I1210 22:54:43.115938 4791 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-xhsjq\" (UniqueName: \"kubernetes.io/projected/a5827e8a-4cb3-494d-87a1-556d920c2a4a-kube-api-access-xhsjq\") on node \"crc\" DevicePath \"\"" Dec 10 22:54:43 crc kubenswrapper[4791]: I1210 22:54:43.115949 4791 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/a5827e8a-4cb3-494d-87a1-556d920c2a4a-serving-cert\") on node \"crc\" DevicePath \"\"" Dec 10 22:54:43 crc kubenswrapper[4791]: I1210 22:54:43.116387 4791 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/5369b8df-3c36-4108-b4e7-cfb11478b3a7-client-ca" (OuterVolumeSpecName: "client-ca") pod "5369b8df-3c36-4108-b4e7-cfb11478b3a7" (UID: "5369b8df-3c36-4108-b4e7-cfb11478b3a7"). InnerVolumeSpecName "client-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 10 22:54:43 crc kubenswrapper[4791]: I1210 22:54:43.116413 4791 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/5369b8df-3c36-4108-b4e7-cfb11478b3a7-config" (OuterVolumeSpecName: "config") pod "5369b8df-3c36-4108-b4e7-cfb11478b3a7" (UID: "5369b8df-3c36-4108-b4e7-cfb11478b3a7"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 10 22:54:43 crc kubenswrapper[4791]: I1210 22:54:43.118494 4791 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/5369b8df-3c36-4108-b4e7-cfb11478b3a7-kube-api-access-z5dwf" (OuterVolumeSpecName: "kube-api-access-z5dwf") pod "5369b8df-3c36-4108-b4e7-cfb11478b3a7" (UID: "5369b8df-3c36-4108-b4e7-cfb11478b3a7"). InnerVolumeSpecName "kube-api-access-z5dwf". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 10 22:54:43 crc kubenswrapper[4791]: I1210 22:54:43.119540 4791 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/5369b8df-3c36-4108-b4e7-cfb11478b3a7-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "5369b8df-3c36-4108-b4e7-cfb11478b3a7" (UID: "5369b8df-3c36-4108-b4e7-cfb11478b3a7"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 10 22:54:43 crc kubenswrapper[4791]: I1210 22:54:43.217517 4791 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-z5dwf\" (UniqueName: \"kubernetes.io/projected/5369b8df-3c36-4108-b4e7-cfb11478b3a7-kube-api-access-z5dwf\") on node \"crc\" DevicePath \"\"" Dec 10 22:54:43 crc kubenswrapper[4791]: I1210 22:54:43.217561 4791 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/5369b8df-3c36-4108-b4e7-cfb11478b3a7-config\") on node \"crc\" DevicePath \"\"" Dec 10 22:54:43 crc kubenswrapper[4791]: I1210 22:54:43.217576 4791 reconciler_common.go:293] "Volume detached for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/5369b8df-3c36-4108-b4e7-cfb11478b3a7-client-ca\") on node \"crc\" DevicePath \"\"" Dec 10 22:54:43 crc kubenswrapper[4791]: I1210 22:54:43.217588 4791 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/5369b8df-3c36-4108-b4e7-cfb11478b3a7-serving-cert\") on node \"crc\" DevicePath \"\"" Dec 10 22:54:43 crc kubenswrapper[4791]: I1210 22:54:43.424173 4791 generic.go:334] "Generic (PLEG): container finished" podID="5369b8df-3c36-4108-b4e7-cfb11478b3a7" containerID="4b9f827d94a26dadac2b0de34bb8c127f77e2fb7e7156bfd1f63d2ea2036e664" exitCode=0 Dec 10 22:54:43 crc kubenswrapper[4791]: I1210 22:54:43.424213 4791 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-vsnsh" event={"ID":"5369b8df-3c36-4108-b4e7-cfb11478b3a7","Type":"ContainerDied","Data":"4b9f827d94a26dadac2b0de34bb8c127f77e2fb7e7156bfd1f63d2ea2036e664"} Dec 10 22:54:43 crc kubenswrapper[4791]: I1210 22:54:43.424243 4791 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-vsnsh" Dec 10 22:54:43 crc kubenswrapper[4791]: I1210 22:54:43.424277 4791 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-vsnsh" event={"ID":"5369b8df-3c36-4108-b4e7-cfb11478b3a7","Type":"ContainerDied","Data":"99a5d5b9b8638ab2472e67f0037b48896f9464090eb279f1bfb13addf1ffd77d"} Dec 10 22:54:43 crc kubenswrapper[4791]: I1210 22:54:43.424302 4791 scope.go:117] "RemoveContainer" containerID="4b9f827d94a26dadac2b0de34bb8c127f77e2fb7e7156bfd1f63d2ea2036e664" Dec 10 22:54:43 crc kubenswrapper[4791]: I1210 22:54:43.427205 4791 generic.go:334] "Generic (PLEG): container finished" podID="a5827e8a-4cb3-494d-87a1-556d920c2a4a" containerID="1b567163092c2059ebeb396c427945f52d9972a1c85abdccb0bd739aac9e995f" exitCode=0 Dec 10 22:54:43 crc kubenswrapper[4791]: I1210 22:54:43.427258 4791 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager/controller-manager-879f6c89f-9tg4p" event={"ID":"a5827e8a-4cb3-494d-87a1-556d920c2a4a","Type":"ContainerDied","Data":"1b567163092c2059ebeb396c427945f52d9972a1c85abdccb0bd739aac9e995f"} Dec 10 22:54:43 crc kubenswrapper[4791]: I1210 22:54:43.427280 4791 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager/controller-manager-879f6c89f-9tg4p" Dec 10 22:54:43 crc kubenswrapper[4791]: I1210 22:54:43.427296 4791 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager/controller-manager-879f6c89f-9tg4p" event={"ID":"a5827e8a-4cb3-494d-87a1-556d920c2a4a","Type":"ContainerDied","Data":"f323620800ff77b675c44d18a09436762bdf76b08f3551d85e0a8cd459422268"} Dec 10 22:54:43 crc kubenswrapper[4791]: I1210 22:54:43.444626 4791 scope.go:117] "RemoveContainer" containerID="4b9f827d94a26dadac2b0de34bb8c127f77e2fb7e7156bfd1f63d2ea2036e664" Dec 10 22:54:43 crc kubenswrapper[4791]: E1210 22:54:43.445082 4791 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"4b9f827d94a26dadac2b0de34bb8c127f77e2fb7e7156bfd1f63d2ea2036e664\": container with ID starting with 4b9f827d94a26dadac2b0de34bb8c127f77e2fb7e7156bfd1f63d2ea2036e664 not found: ID does not exist" containerID="4b9f827d94a26dadac2b0de34bb8c127f77e2fb7e7156bfd1f63d2ea2036e664" Dec 10 22:54:43 crc kubenswrapper[4791]: I1210 22:54:43.445124 4791 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"4b9f827d94a26dadac2b0de34bb8c127f77e2fb7e7156bfd1f63d2ea2036e664"} err="failed to get container status \"4b9f827d94a26dadac2b0de34bb8c127f77e2fb7e7156bfd1f63d2ea2036e664\": rpc error: code = NotFound desc = could not find container \"4b9f827d94a26dadac2b0de34bb8c127f77e2fb7e7156bfd1f63d2ea2036e664\": container with ID starting with 4b9f827d94a26dadac2b0de34bb8c127f77e2fb7e7156bfd1f63d2ea2036e664 not found: ID does not exist" Dec 10 22:54:43 crc kubenswrapper[4791]: I1210 22:54:43.445197 4791 scope.go:117] "RemoveContainer" containerID="1b567163092c2059ebeb396c427945f52d9972a1c85abdccb0bd739aac9e995f" Dec 10 22:54:43 crc kubenswrapper[4791]: I1210 22:54:43.456595 4791 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-6576b87f9c-vsnsh"] Dec 10 22:54:43 crc kubenswrapper[4791]: I1210 22:54:43.461433 4791 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-6576b87f9c-vsnsh"] Dec 10 22:54:43 crc kubenswrapper[4791]: I1210 22:54:43.464115 4791 scope.go:117] "RemoveContainer" containerID="1b567163092c2059ebeb396c427945f52d9972a1c85abdccb0bd739aac9e995f" Dec 10 22:54:43 crc kubenswrapper[4791]: E1210 22:54:43.464567 4791 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"1b567163092c2059ebeb396c427945f52d9972a1c85abdccb0bd739aac9e995f\": container with ID starting with 1b567163092c2059ebeb396c427945f52d9972a1c85abdccb0bd739aac9e995f not found: ID does not exist" containerID="1b567163092c2059ebeb396c427945f52d9972a1c85abdccb0bd739aac9e995f" Dec 10 22:54:43 crc kubenswrapper[4791]: I1210 22:54:43.464637 4791 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"1b567163092c2059ebeb396c427945f52d9972a1c85abdccb0bd739aac9e995f"} err="failed to get container status \"1b567163092c2059ebeb396c427945f52d9972a1c85abdccb0bd739aac9e995f\": rpc error: code = NotFound desc = could not find container \"1b567163092c2059ebeb396c427945f52d9972a1c85abdccb0bd739aac9e995f\": container with ID starting with 1b567163092c2059ebeb396c427945f52d9972a1c85abdccb0bd739aac9e995f not found: ID does not exist" Dec 10 22:54:43 crc kubenswrapper[4791]: I1210 22:54:43.470726 4791 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-controller-manager/controller-manager-879f6c89f-9tg4p"] Dec 10 22:54:43 crc kubenswrapper[4791]: I1210 22:54:43.474989 4791 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-controller-manager/controller-manager-879f6c89f-9tg4p"] Dec 10 22:54:43 crc kubenswrapper[4791]: I1210 22:54:43.828704 4791 patch_prober.go:28] interesting pod/controller-manager-879f6c89f-9tg4p container/controller-manager namespace/openshift-controller-manager: Readiness probe status=failure output="Get \"https://10.217.0.6:8443/healthz\": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers)" start-of-body= Dec 10 22:54:43 crc kubenswrapper[4791]: I1210 22:54:43.829076 4791 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-controller-manager/controller-manager-879f6c89f-9tg4p" podUID="a5827e8a-4cb3-494d-87a1-556d920c2a4a" containerName="controller-manager" probeResult="failure" output="Get \"https://10.217.0.6:8443/healthz\": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers)" Dec 10 22:54:43 crc kubenswrapper[4791]: I1210 22:54:43.891109 4791 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="5369b8df-3c36-4108-b4e7-cfb11478b3a7" path="/var/lib/kubelet/pods/5369b8df-3c36-4108-b4e7-cfb11478b3a7/volumes" Dec 10 22:54:43 crc kubenswrapper[4791]: I1210 22:54:43.891795 4791 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="a5827e8a-4cb3-494d-87a1-556d920c2a4a" path="/var/lib/kubelet/pods/a5827e8a-4cb3-494d-87a1-556d920c2a4a/volumes" Dec 10 22:54:44 crc kubenswrapper[4791]: I1210 22:54:44.199854 4791 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-controller-manager/controller-manager-649fb98567-gzgvx"] Dec 10 22:54:44 crc kubenswrapper[4791]: E1210 22:54:44.200123 4791 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="ab4c6e9a-3413-49c1-8f49-fb22c8366cb6" containerName="extract-utilities" Dec 10 22:54:44 crc kubenswrapper[4791]: I1210 22:54:44.200140 4791 state_mem.go:107] "Deleted CPUSet assignment" podUID="ab4c6e9a-3413-49c1-8f49-fb22c8366cb6" containerName="extract-utilities" Dec 10 22:54:44 crc kubenswrapper[4791]: E1210 22:54:44.200152 4791 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="a5827e8a-4cb3-494d-87a1-556d920c2a4a" containerName="controller-manager" Dec 10 22:54:44 crc kubenswrapper[4791]: I1210 22:54:44.200161 4791 state_mem.go:107] "Deleted CPUSet assignment" podUID="a5827e8a-4cb3-494d-87a1-556d920c2a4a" containerName="controller-manager" Dec 10 22:54:44 crc kubenswrapper[4791]: E1210 22:54:44.200171 4791 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="396f55d6-c96d-4520-8165-862b5905a8c0" containerName="extract-content" Dec 10 22:54:44 crc kubenswrapper[4791]: I1210 22:54:44.200178 4791 state_mem.go:107] "Deleted CPUSet assignment" podUID="396f55d6-c96d-4520-8165-862b5905a8c0" containerName="extract-content" Dec 10 22:54:44 crc kubenswrapper[4791]: E1210 22:54:44.200189 4791 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="08d50c37-7a4d-4731-beb8-fbc6d2e10bb0" containerName="extract-content" Dec 10 22:54:44 crc kubenswrapper[4791]: I1210 22:54:44.200196 4791 state_mem.go:107] "Deleted CPUSet assignment" podUID="08d50c37-7a4d-4731-beb8-fbc6d2e10bb0" containerName="extract-content" Dec 10 22:54:44 crc kubenswrapper[4791]: E1210 22:54:44.200205 4791 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="396f55d6-c96d-4520-8165-862b5905a8c0" containerName="extract-utilities" Dec 10 22:54:44 crc kubenswrapper[4791]: I1210 22:54:44.200210 4791 state_mem.go:107] "Deleted CPUSet assignment" podUID="396f55d6-c96d-4520-8165-862b5905a8c0" containerName="extract-utilities" Dec 10 22:54:44 crc kubenswrapper[4791]: E1210 22:54:44.200223 4791 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="ab4c6e9a-3413-49c1-8f49-fb22c8366cb6" containerName="registry-server" Dec 10 22:54:44 crc kubenswrapper[4791]: I1210 22:54:44.200229 4791 state_mem.go:107] "Deleted CPUSet assignment" podUID="ab4c6e9a-3413-49c1-8f49-fb22c8366cb6" containerName="registry-server" Dec 10 22:54:44 crc kubenswrapper[4791]: E1210 22:54:44.200235 4791 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="a9185ee1-adaf-48e8-b9a9-90ef658f8212" containerName="marketplace-operator" Dec 10 22:54:44 crc kubenswrapper[4791]: I1210 22:54:44.200241 4791 state_mem.go:107] "Deleted CPUSet assignment" podUID="a9185ee1-adaf-48e8-b9a9-90ef658f8212" containerName="marketplace-operator" Dec 10 22:54:44 crc kubenswrapper[4791]: E1210 22:54:44.200250 4791 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="08d50c37-7a4d-4731-beb8-fbc6d2e10bb0" containerName="extract-utilities" Dec 10 22:54:44 crc kubenswrapper[4791]: I1210 22:54:44.200256 4791 state_mem.go:107] "Deleted CPUSet assignment" podUID="08d50c37-7a4d-4731-beb8-fbc6d2e10bb0" containerName="extract-utilities" Dec 10 22:54:44 crc kubenswrapper[4791]: E1210 22:54:44.200265 4791 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="009efc35-5943-429f-98fd-3307a5f557f2" containerName="registry-server" Dec 10 22:54:44 crc kubenswrapper[4791]: I1210 22:54:44.200273 4791 state_mem.go:107] "Deleted CPUSet assignment" podUID="009efc35-5943-429f-98fd-3307a5f557f2" containerName="registry-server" Dec 10 22:54:44 crc kubenswrapper[4791]: E1210 22:54:44.200282 4791 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="ab4c6e9a-3413-49c1-8f49-fb22c8366cb6" containerName="extract-content" Dec 10 22:54:44 crc kubenswrapper[4791]: I1210 22:54:44.200288 4791 state_mem.go:107] "Deleted CPUSet assignment" podUID="ab4c6e9a-3413-49c1-8f49-fb22c8366cb6" containerName="extract-content" Dec 10 22:54:44 crc kubenswrapper[4791]: E1210 22:54:44.200298 4791 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="009efc35-5943-429f-98fd-3307a5f557f2" containerName="extract-utilities" Dec 10 22:54:44 crc kubenswrapper[4791]: I1210 22:54:44.200306 4791 state_mem.go:107] "Deleted CPUSet assignment" podUID="009efc35-5943-429f-98fd-3307a5f557f2" containerName="extract-utilities" Dec 10 22:54:44 crc kubenswrapper[4791]: E1210 22:54:44.200313 4791 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="009efc35-5943-429f-98fd-3307a5f557f2" containerName="extract-content" Dec 10 22:54:44 crc kubenswrapper[4791]: I1210 22:54:44.200319 4791 state_mem.go:107] "Deleted CPUSet assignment" podUID="009efc35-5943-429f-98fd-3307a5f557f2" containerName="extract-content" Dec 10 22:54:44 crc kubenswrapper[4791]: E1210 22:54:44.200326 4791 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="5369b8df-3c36-4108-b4e7-cfb11478b3a7" containerName="route-controller-manager" Dec 10 22:54:44 crc kubenswrapper[4791]: I1210 22:54:44.200332 4791 state_mem.go:107] "Deleted CPUSet assignment" podUID="5369b8df-3c36-4108-b4e7-cfb11478b3a7" containerName="route-controller-manager" Dec 10 22:54:44 crc kubenswrapper[4791]: E1210 22:54:44.200365 4791 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="396f55d6-c96d-4520-8165-862b5905a8c0" containerName="registry-server" Dec 10 22:54:44 crc kubenswrapper[4791]: I1210 22:54:44.200371 4791 state_mem.go:107] "Deleted CPUSet assignment" podUID="396f55d6-c96d-4520-8165-862b5905a8c0" containerName="registry-server" Dec 10 22:54:44 crc kubenswrapper[4791]: E1210 22:54:44.200379 4791 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="08d50c37-7a4d-4731-beb8-fbc6d2e10bb0" containerName="registry-server" Dec 10 22:54:44 crc kubenswrapper[4791]: I1210 22:54:44.200387 4791 state_mem.go:107] "Deleted CPUSet assignment" podUID="08d50c37-7a4d-4731-beb8-fbc6d2e10bb0" containerName="registry-server" Dec 10 22:54:44 crc kubenswrapper[4791]: I1210 22:54:44.200464 4791 memory_manager.go:354] "RemoveStaleState removing state" podUID="ab4c6e9a-3413-49c1-8f49-fb22c8366cb6" containerName="registry-server" Dec 10 22:54:44 crc kubenswrapper[4791]: I1210 22:54:44.200475 4791 memory_manager.go:354] "RemoveStaleState removing state" podUID="08d50c37-7a4d-4731-beb8-fbc6d2e10bb0" containerName="registry-server" Dec 10 22:54:44 crc kubenswrapper[4791]: I1210 22:54:44.200482 4791 memory_manager.go:354] "RemoveStaleState removing state" podUID="009efc35-5943-429f-98fd-3307a5f557f2" containerName="registry-server" Dec 10 22:54:44 crc kubenswrapper[4791]: I1210 22:54:44.200489 4791 memory_manager.go:354] "RemoveStaleState removing state" podUID="396f55d6-c96d-4520-8165-862b5905a8c0" containerName="registry-server" Dec 10 22:54:44 crc kubenswrapper[4791]: I1210 22:54:44.200496 4791 memory_manager.go:354] "RemoveStaleState removing state" podUID="a9185ee1-adaf-48e8-b9a9-90ef658f8212" containerName="marketplace-operator" Dec 10 22:54:44 crc kubenswrapper[4791]: I1210 22:54:44.200507 4791 memory_manager.go:354] "RemoveStaleState removing state" podUID="a5827e8a-4cb3-494d-87a1-556d920c2a4a" containerName="controller-manager" Dec 10 22:54:44 crc kubenswrapper[4791]: I1210 22:54:44.200513 4791 memory_manager.go:354] "RemoveStaleState removing state" podUID="5369b8df-3c36-4108-b4e7-cfb11478b3a7" containerName="route-controller-manager" Dec 10 22:54:44 crc kubenswrapper[4791]: I1210 22:54:44.200900 4791 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager/controller-manager-649fb98567-gzgvx" Dec 10 22:54:44 crc kubenswrapper[4791]: I1210 22:54:44.203936 4791 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-route-controller-manager/route-controller-manager-66c8f7db4d-6582d"] Dec 10 22:54:44 crc kubenswrapper[4791]: I1210 22:54:44.204014 4791 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-controller-manager"/"serving-cert" Dec 10 22:54:44 crc kubenswrapper[4791]: I1210 22:54:44.204122 4791 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"kube-root-ca.crt" Dec 10 22:54:44 crc kubenswrapper[4791]: I1210 22:54:44.204280 4791 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"openshift-service-ca.crt" Dec 10 22:54:44 crc kubenswrapper[4791]: I1210 22:54:44.204293 4791 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"config" Dec 10 22:54:44 crc kubenswrapper[4791]: I1210 22:54:44.204604 4791 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"client-ca" Dec 10 22:54:44 crc kubenswrapper[4791]: I1210 22:54:44.204706 4791 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-route-controller-manager/route-controller-manager-66c8f7db4d-6582d" Dec 10 22:54:44 crc kubenswrapper[4791]: I1210 22:54:44.204839 4791 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-controller-manager"/"openshift-controller-manager-sa-dockercfg-msq4c" Dec 10 22:54:44 crc kubenswrapper[4791]: I1210 22:54:44.209654 4791 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"openshift-service-ca.crt" Dec 10 22:54:44 crc kubenswrapper[4791]: I1210 22:54:44.210377 4791 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"config" Dec 10 22:54:44 crc kubenswrapper[4791]: I1210 22:54:44.211161 4791 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"kube-root-ca.crt" Dec 10 22:54:44 crc kubenswrapper[4791]: I1210 22:54:44.211304 4791 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-route-controller-manager"/"serving-cert" Dec 10 22:54:44 crc kubenswrapper[4791]: I1210 22:54:44.211483 4791 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"client-ca" Dec 10 22:54:44 crc kubenswrapper[4791]: I1210 22:54:44.211573 4791 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-route-controller-manager"/"route-controller-manager-sa-dockercfg-h2zr2" Dec 10 22:54:44 crc kubenswrapper[4791]: I1210 22:54:44.211837 4791 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-controller-manager/controller-manager-649fb98567-gzgvx"] Dec 10 22:54:44 crc kubenswrapper[4791]: I1210 22:54:44.214702 4791 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"openshift-global-ca" Dec 10 22:54:44 crc kubenswrapper[4791]: I1210 22:54:44.216395 4791 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-66c8f7db4d-6582d"] Dec 10 22:54:44 crc kubenswrapper[4791]: I1210 22:54:44.331821 4791 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/c39bed63-baf9-42a6-820e-07199de0212b-proxy-ca-bundles\") pod \"controller-manager-649fb98567-gzgvx\" (UID: \"c39bed63-baf9-42a6-820e-07199de0212b\") " pod="openshift-controller-manager/controller-manager-649fb98567-gzgvx" Dec 10 22:54:44 crc kubenswrapper[4791]: I1210 22:54:44.332117 4791 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-wnfx4\" (UniqueName: \"kubernetes.io/projected/c39bed63-baf9-42a6-820e-07199de0212b-kube-api-access-wnfx4\") pod \"controller-manager-649fb98567-gzgvx\" (UID: \"c39bed63-baf9-42a6-820e-07199de0212b\") " pod="openshift-controller-manager/controller-manager-649fb98567-gzgvx" Dec 10 22:54:44 crc kubenswrapper[4791]: I1210 22:54:44.332304 4791 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/fca6ee01-4d09-4b77-89c3-281a84c4f3a6-config\") pod \"route-controller-manager-66c8f7db4d-6582d\" (UID: \"fca6ee01-4d09-4b77-89c3-281a84c4f3a6\") " pod="openshift-route-controller-manager/route-controller-manager-66c8f7db4d-6582d" Dec 10 22:54:44 crc kubenswrapper[4791]: I1210 22:54:44.332521 4791 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/c39bed63-baf9-42a6-820e-07199de0212b-serving-cert\") pod \"controller-manager-649fb98567-gzgvx\" (UID: \"c39bed63-baf9-42a6-820e-07199de0212b\") " pod="openshift-controller-manager/controller-manager-649fb98567-gzgvx" Dec 10 22:54:44 crc kubenswrapper[4791]: I1210 22:54:44.332575 4791 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/fca6ee01-4d09-4b77-89c3-281a84c4f3a6-client-ca\") pod \"route-controller-manager-66c8f7db4d-6582d\" (UID: \"fca6ee01-4d09-4b77-89c3-281a84c4f3a6\") " pod="openshift-route-controller-manager/route-controller-manager-66c8f7db4d-6582d" Dec 10 22:54:44 crc kubenswrapper[4791]: I1210 22:54:44.332601 4791 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/c39bed63-baf9-42a6-820e-07199de0212b-client-ca\") pod \"controller-manager-649fb98567-gzgvx\" (UID: \"c39bed63-baf9-42a6-820e-07199de0212b\") " pod="openshift-controller-manager/controller-manager-649fb98567-gzgvx" Dec 10 22:54:44 crc kubenswrapper[4791]: I1210 22:54:44.332619 4791 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/fca6ee01-4d09-4b77-89c3-281a84c4f3a6-serving-cert\") pod \"route-controller-manager-66c8f7db4d-6582d\" (UID: \"fca6ee01-4d09-4b77-89c3-281a84c4f3a6\") " pod="openshift-route-controller-manager/route-controller-manager-66c8f7db4d-6582d" Dec 10 22:54:44 crc kubenswrapper[4791]: I1210 22:54:44.332638 4791 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-8xcc9\" (UniqueName: \"kubernetes.io/projected/fca6ee01-4d09-4b77-89c3-281a84c4f3a6-kube-api-access-8xcc9\") pod \"route-controller-manager-66c8f7db4d-6582d\" (UID: \"fca6ee01-4d09-4b77-89c3-281a84c4f3a6\") " pod="openshift-route-controller-manager/route-controller-manager-66c8f7db4d-6582d" Dec 10 22:54:44 crc kubenswrapper[4791]: I1210 22:54:44.332653 4791 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/c39bed63-baf9-42a6-820e-07199de0212b-config\") pod \"controller-manager-649fb98567-gzgvx\" (UID: \"c39bed63-baf9-42a6-820e-07199de0212b\") " pod="openshift-controller-manager/controller-manager-649fb98567-gzgvx" Dec 10 22:54:44 crc kubenswrapper[4791]: I1210 22:54:44.433220 4791 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/fca6ee01-4d09-4b77-89c3-281a84c4f3a6-serving-cert\") pod \"route-controller-manager-66c8f7db4d-6582d\" (UID: \"fca6ee01-4d09-4b77-89c3-281a84c4f3a6\") " pod="openshift-route-controller-manager/route-controller-manager-66c8f7db4d-6582d" Dec 10 22:54:44 crc kubenswrapper[4791]: I1210 22:54:44.433535 4791 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-8xcc9\" (UniqueName: \"kubernetes.io/projected/fca6ee01-4d09-4b77-89c3-281a84c4f3a6-kube-api-access-8xcc9\") pod \"route-controller-manager-66c8f7db4d-6582d\" (UID: \"fca6ee01-4d09-4b77-89c3-281a84c4f3a6\") " pod="openshift-route-controller-manager/route-controller-manager-66c8f7db4d-6582d" Dec 10 22:54:44 crc kubenswrapper[4791]: I1210 22:54:44.433679 4791 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/c39bed63-baf9-42a6-820e-07199de0212b-config\") pod \"controller-manager-649fb98567-gzgvx\" (UID: \"c39bed63-baf9-42a6-820e-07199de0212b\") " pod="openshift-controller-manager/controller-manager-649fb98567-gzgvx" Dec 10 22:54:44 crc kubenswrapper[4791]: I1210 22:54:44.433808 4791 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/c39bed63-baf9-42a6-820e-07199de0212b-proxy-ca-bundles\") pod \"controller-manager-649fb98567-gzgvx\" (UID: \"c39bed63-baf9-42a6-820e-07199de0212b\") " pod="openshift-controller-manager/controller-manager-649fb98567-gzgvx" Dec 10 22:54:44 crc kubenswrapper[4791]: I1210 22:54:44.433950 4791 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-wnfx4\" (UniqueName: \"kubernetes.io/projected/c39bed63-baf9-42a6-820e-07199de0212b-kube-api-access-wnfx4\") pod \"controller-manager-649fb98567-gzgvx\" (UID: \"c39bed63-baf9-42a6-820e-07199de0212b\") " pod="openshift-controller-manager/controller-manager-649fb98567-gzgvx" Dec 10 22:54:44 crc kubenswrapper[4791]: I1210 22:54:44.434078 4791 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/fca6ee01-4d09-4b77-89c3-281a84c4f3a6-config\") pod \"route-controller-manager-66c8f7db4d-6582d\" (UID: \"fca6ee01-4d09-4b77-89c3-281a84c4f3a6\") " pod="openshift-route-controller-manager/route-controller-manager-66c8f7db4d-6582d" Dec 10 22:54:44 crc kubenswrapper[4791]: I1210 22:54:44.434192 4791 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/c39bed63-baf9-42a6-820e-07199de0212b-serving-cert\") pod \"controller-manager-649fb98567-gzgvx\" (UID: \"c39bed63-baf9-42a6-820e-07199de0212b\") " pod="openshift-controller-manager/controller-manager-649fb98567-gzgvx" Dec 10 22:54:44 crc kubenswrapper[4791]: I1210 22:54:44.434395 4791 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/fca6ee01-4d09-4b77-89c3-281a84c4f3a6-client-ca\") pod \"route-controller-manager-66c8f7db4d-6582d\" (UID: \"fca6ee01-4d09-4b77-89c3-281a84c4f3a6\") " pod="openshift-route-controller-manager/route-controller-manager-66c8f7db4d-6582d" Dec 10 22:54:44 crc kubenswrapper[4791]: I1210 22:54:44.434504 4791 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/c39bed63-baf9-42a6-820e-07199de0212b-client-ca\") pod \"controller-manager-649fb98567-gzgvx\" (UID: \"c39bed63-baf9-42a6-820e-07199de0212b\") " pod="openshift-controller-manager/controller-manager-649fb98567-gzgvx" Dec 10 22:54:44 crc kubenswrapper[4791]: I1210 22:54:44.435085 4791 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/c39bed63-baf9-42a6-820e-07199de0212b-proxy-ca-bundles\") pod \"controller-manager-649fb98567-gzgvx\" (UID: \"c39bed63-baf9-42a6-820e-07199de0212b\") " pod="openshift-controller-manager/controller-manager-649fb98567-gzgvx" Dec 10 22:54:44 crc kubenswrapper[4791]: I1210 22:54:44.435172 4791 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/c39bed63-baf9-42a6-820e-07199de0212b-config\") pod \"controller-manager-649fb98567-gzgvx\" (UID: \"c39bed63-baf9-42a6-820e-07199de0212b\") " pod="openshift-controller-manager/controller-manager-649fb98567-gzgvx" Dec 10 22:54:44 crc kubenswrapper[4791]: I1210 22:54:44.435398 4791 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/fca6ee01-4d09-4b77-89c3-281a84c4f3a6-config\") pod \"route-controller-manager-66c8f7db4d-6582d\" (UID: \"fca6ee01-4d09-4b77-89c3-281a84c4f3a6\") " pod="openshift-route-controller-manager/route-controller-manager-66c8f7db4d-6582d" Dec 10 22:54:44 crc kubenswrapper[4791]: I1210 22:54:44.435693 4791 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/c39bed63-baf9-42a6-820e-07199de0212b-client-ca\") pod \"controller-manager-649fb98567-gzgvx\" (UID: \"c39bed63-baf9-42a6-820e-07199de0212b\") " pod="openshift-controller-manager/controller-manager-649fb98567-gzgvx" Dec 10 22:54:44 crc kubenswrapper[4791]: I1210 22:54:44.435930 4791 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/fca6ee01-4d09-4b77-89c3-281a84c4f3a6-client-ca\") pod \"route-controller-manager-66c8f7db4d-6582d\" (UID: \"fca6ee01-4d09-4b77-89c3-281a84c4f3a6\") " pod="openshift-route-controller-manager/route-controller-manager-66c8f7db4d-6582d" Dec 10 22:54:44 crc kubenswrapper[4791]: I1210 22:54:44.440724 4791 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/c39bed63-baf9-42a6-820e-07199de0212b-serving-cert\") pod \"controller-manager-649fb98567-gzgvx\" (UID: \"c39bed63-baf9-42a6-820e-07199de0212b\") " pod="openshift-controller-manager/controller-manager-649fb98567-gzgvx" Dec 10 22:54:44 crc kubenswrapper[4791]: I1210 22:54:44.445061 4791 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/fca6ee01-4d09-4b77-89c3-281a84c4f3a6-serving-cert\") pod \"route-controller-manager-66c8f7db4d-6582d\" (UID: \"fca6ee01-4d09-4b77-89c3-281a84c4f3a6\") " pod="openshift-route-controller-manager/route-controller-manager-66c8f7db4d-6582d" Dec 10 22:54:44 crc kubenswrapper[4791]: I1210 22:54:44.461372 4791 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-wnfx4\" (UniqueName: \"kubernetes.io/projected/c39bed63-baf9-42a6-820e-07199de0212b-kube-api-access-wnfx4\") pod \"controller-manager-649fb98567-gzgvx\" (UID: \"c39bed63-baf9-42a6-820e-07199de0212b\") " pod="openshift-controller-manager/controller-manager-649fb98567-gzgvx" Dec 10 22:54:44 crc kubenswrapper[4791]: I1210 22:54:44.468382 4791 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-8xcc9\" (UniqueName: \"kubernetes.io/projected/fca6ee01-4d09-4b77-89c3-281a84c4f3a6-kube-api-access-8xcc9\") pod \"route-controller-manager-66c8f7db4d-6582d\" (UID: \"fca6ee01-4d09-4b77-89c3-281a84c4f3a6\") " pod="openshift-route-controller-manager/route-controller-manager-66c8f7db4d-6582d" Dec 10 22:54:44 crc kubenswrapper[4791]: I1210 22:54:44.527186 4791 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager/controller-manager-649fb98567-gzgvx" Dec 10 22:54:44 crc kubenswrapper[4791]: I1210 22:54:44.536622 4791 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-route-controller-manager/route-controller-manager-66c8f7db4d-6582d" Dec 10 22:54:44 crc kubenswrapper[4791]: I1210 22:54:44.797811 4791 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-controller-manager/controller-manager-649fb98567-gzgvx"] Dec 10 22:54:44 crc kubenswrapper[4791]: I1210 22:54:44.946680 4791 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-66c8f7db4d-6582d"] Dec 10 22:54:44 crc kubenswrapper[4791]: W1210 22:54:44.951556 4791 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podfca6ee01_4d09_4b77_89c3_281a84c4f3a6.slice/crio-956e3e39fb50bb4a38b8cb11a229fa8a5ae0949381c0d96dcbd50e0a483a6b01 WatchSource:0}: Error finding container 956e3e39fb50bb4a38b8cb11a229fa8a5ae0949381c0d96dcbd50e0a483a6b01: Status 404 returned error can't find the container with id 956e3e39fb50bb4a38b8cb11a229fa8a5ae0949381c0d96dcbd50e0a483a6b01 Dec 10 22:54:45 crc kubenswrapper[4791]: I1210 22:54:45.443827 4791 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager/controller-manager-649fb98567-gzgvx" event={"ID":"c39bed63-baf9-42a6-820e-07199de0212b","Type":"ContainerStarted","Data":"4482a5363e33bf45d6f632e66fb6d1da9a1ef452a5a98c8882574069e9520d50"} Dec 10 22:54:45 crc kubenswrapper[4791]: I1210 22:54:45.444193 4791 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager/controller-manager-649fb98567-gzgvx" event={"ID":"c39bed63-baf9-42a6-820e-07199de0212b","Type":"ContainerStarted","Data":"c9aded4fcedee3de19be60b6b9d8b6ba834b343b443652988555ee51c0094409"} Dec 10 22:54:45 crc kubenswrapper[4791]: I1210 22:54:45.444999 4791 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-controller-manager/controller-manager-649fb98567-gzgvx" Dec 10 22:54:45 crc kubenswrapper[4791]: I1210 22:54:45.446542 4791 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-route-controller-manager/route-controller-manager-66c8f7db4d-6582d" event={"ID":"fca6ee01-4d09-4b77-89c3-281a84c4f3a6","Type":"ContainerStarted","Data":"24139c671ed7e44b2d95f26420006f600a27788c56af9bf724fc7b8e6d25e2e8"} Dec 10 22:54:45 crc kubenswrapper[4791]: I1210 22:54:45.446578 4791 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-route-controller-manager/route-controller-manager-66c8f7db4d-6582d" event={"ID":"fca6ee01-4d09-4b77-89c3-281a84c4f3a6","Type":"ContainerStarted","Data":"956e3e39fb50bb4a38b8cb11a229fa8a5ae0949381c0d96dcbd50e0a483a6b01"} Dec 10 22:54:45 crc kubenswrapper[4791]: I1210 22:54:45.446760 4791 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-route-controller-manager/route-controller-manager-66c8f7db4d-6582d" Dec 10 22:54:45 crc kubenswrapper[4791]: I1210 22:54:45.449866 4791 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-controller-manager/controller-manager-649fb98567-gzgvx" Dec 10 22:54:45 crc kubenswrapper[4791]: I1210 22:54:45.451169 4791 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-route-controller-manager/route-controller-manager-66c8f7db4d-6582d" Dec 10 22:54:45 crc kubenswrapper[4791]: I1210 22:54:45.461607 4791 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-controller-manager/controller-manager-649fb98567-gzgvx" podStartSLOduration=3.46159169 podStartE2EDuration="3.46159169s" podCreationTimestamp="2025-12-10 22:54:42 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-10 22:54:45.460448317 +0000 UTC m=+319.890065940" watchObservedRunningTime="2025-12-10 22:54:45.46159169 +0000 UTC m=+319.891209303" Dec 10 22:54:45 crc kubenswrapper[4791]: I1210 22:54:45.514223 4791 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-route-controller-manager/route-controller-manager-66c8f7db4d-6582d" podStartSLOduration=3.514195925 podStartE2EDuration="3.514195925s" podCreationTimestamp="2025-12-10 22:54:42 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-10 22:54:45.484623967 +0000 UTC m=+319.914241580" watchObservedRunningTime="2025-12-10 22:54:45.514195925 +0000 UTC m=+319.943813538" Dec 10 22:54:45 crc kubenswrapper[4791]: I1210 22:54:45.933479 4791 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ovn-kubernetes"/"ovnkube-script-lib" Dec 10 22:54:47 crc kubenswrapper[4791]: I1210 22:54:47.206901 4791 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-marketplace-fcnzr"] Dec 10 22:54:47 crc kubenswrapper[4791]: I1210 22:54:47.208477 4791 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-fcnzr" Dec 10 22:54:47 crc kubenswrapper[4791]: I1210 22:54:47.210421 4791 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"redhat-marketplace-dockercfg-x2ctb" Dec 10 22:54:47 crc kubenswrapper[4791]: I1210 22:54:47.218125 4791 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-fcnzr"] Dec 10 22:54:47 crc kubenswrapper[4791]: I1210 22:54:47.311528 4791 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication-operator"/"authentication-operator-dockercfg-mz9bj" Dec 10 22:54:47 crc kubenswrapper[4791]: I1210 22:54:47.368936 4791 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/44085f7f-f0e9-4dec-a816-a687a0dd06ec-catalog-content\") pod \"redhat-marketplace-fcnzr\" (UID: \"44085f7f-f0e9-4dec-a816-a687a0dd06ec\") " pod="openshift-marketplace/redhat-marketplace-fcnzr" Dec 10 22:54:47 crc kubenswrapper[4791]: I1210 22:54:47.368994 4791 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-wc76s\" (UniqueName: \"kubernetes.io/projected/44085f7f-f0e9-4dec-a816-a687a0dd06ec-kube-api-access-wc76s\") pod \"redhat-marketplace-fcnzr\" (UID: \"44085f7f-f0e9-4dec-a816-a687a0dd06ec\") " pod="openshift-marketplace/redhat-marketplace-fcnzr" Dec 10 22:54:47 crc kubenswrapper[4791]: I1210 22:54:47.369107 4791 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/44085f7f-f0e9-4dec-a816-a687a0dd06ec-utilities\") pod \"redhat-marketplace-fcnzr\" (UID: \"44085f7f-f0e9-4dec-a816-a687a0dd06ec\") " pod="openshift-marketplace/redhat-marketplace-fcnzr" Dec 10 22:54:47 crc kubenswrapper[4791]: I1210 22:54:47.407176 4791 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-operators-f8n6m"] Dec 10 22:54:47 crc kubenswrapper[4791]: I1210 22:54:47.408203 4791 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-f8n6m" Dec 10 22:54:47 crc kubenswrapper[4791]: I1210 22:54:47.410502 4791 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"redhat-operators-dockercfg-ct8rh" Dec 10 22:54:47 crc kubenswrapper[4791]: I1210 22:54:47.423647 4791 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-f8n6m"] Dec 10 22:54:47 crc kubenswrapper[4791]: I1210 22:54:47.470165 4791 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/44085f7f-f0e9-4dec-a816-a687a0dd06ec-catalog-content\") pod \"redhat-marketplace-fcnzr\" (UID: \"44085f7f-f0e9-4dec-a816-a687a0dd06ec\") " pod="openshift-marketplace/redhat-marketplace-fcnzr" Dec 10 22:54:47 crc kubenswrapper[4791]: I1210 22:54:47.470219 4791 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-wc76s\" (UniqueName: \"kubernetes.io/projected/44085f7f-f0e9-4dec-a816-a687a0dd06ec-kube-api-access-wc76s\") pod \"redhat-marketplace-fcnzr\" (UID: \"44085f7f-f0e9-4dec-a816-a687a0dd06ec\") " pod="openshift-marketplace/redhat-marketplace-fcnzr" Dec 10 22:54:47 crc kubenswrapper[4791]: I1210 22:54:47.470261 4791 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/44085f7f-f0e9-4dec-a816-a687a0dd06ec-utilities\") pod \"redhat-marketplace-fcnzr\" (UID: \"44085f7f-f0e9-4dec-a816-a687a0dd06ec\") " pod="openshift-marketplace/redhat-marketplace-fcnzr" Dec 10 22:54:47 crc kubenswrapper[4791]: I1210 22:54:47.470328 4791 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/a68872ea-89d3-42a0-b2da-877e2d37218b-catalog-content\") pod \"redhat-operators-f8n6m\" (UID: \"a68872ea-89d3-42a0-b2da-877e2d37218b\") " pod="openshift-marketplace/redhat-operators-f8n6m" Dec 10 22:54:47 crc kubenswrapper[4791]: I1210 22:54:47.470395 4791 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/a68872ea-89d3-42a0-b2da-877e2d37218b-utilities\") pod \"redhat-operators-f8n6m\" (UID: \"a68872ea-89d3-42a0-b2da-877e2d37218b\") " pod="openshift-marketplace/redhat-operators-f8n6m" Dec 10 22:54:47 crc kubenswrapper[4791]: I1210 22:54:47.470414 4791 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-jgdsn\" (UniqueName: \"kubernetes.io/projected/a68872ea-89d3-42a0-b2da-877e2d37218b-kube-api-access-jgdsn\") pod \"redhat-operators-f8n6m\" (UID: \"a68872ea-89d3-42a0-b2da-877e2d37218b\") " pod="openshift-marketplace/redhat-operators-f8n6m" Dec 10 22:54:47 crc kubenswrapper[4791]: I1210 22:54:47.470602 4791 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/44085f7f-f0e9-4dec-a816-a687a0dd06ec-catalog-content\") pod \"redhat-marketplace-fcnzr\" (UID: \"44085f7f-f0e9-4dec-a816-a687a0dd06ec\") " pod="openshift-marketplace/redhat-marketplace-fcnzr" Dec 10 22:54:47 crc kubenswrapper[4791]: I1210 22:54:47.470908 4791 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/44085f7f-f0e9-4dec-a816-a687a0dd06ec-utilities\") pod \"redhat-marketplace-fcnzr\" (UID: \"44085f7f-f0e9-4dec-a816-a687a0dd06ec\") " pod="openshift-marketplace/redhat-marketplace-fcnzr" Dec 10 22:54:47 crc kubenswrapper[4791]: I1210 22:54:47.489024 4791 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-wc76s\" (UniqueName: \"kubernetes.io/projected/44085f7f-f0e9-4dec-a816-a687a0dd06ec-kube-api-access-wc76s\") pod \"redhat-marketplace-fcnzr\" (UID: \"44085f7f-f0e9-4dec-a816-a687a0dd06ec\") " pod="openshift-marketplace/redhat-marketplace-fcnzr" Dec 10 22:54:47 crc kubenswrapper[4791]: I1210 22:54:47.531926 4791 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-fcnzr" Dec 10 22:54:47 crc kubenswrapper[4791]: I1210 22:54:47.571773 4791 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/a68872ea-89d3-42a0-b2da-877e2d37218b-catalog-content\") pod \"redhat-operators-f8n6m\" (UID: \"a68872ea-89d3-42a0-b2da-877e2d37218b\") " pod="openshift-marketplace/redhat-operators-f8n6m" Dec 10 22:54:47 crc kubenswrapper[4791]: I1210 22:54:47.571943 4791 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/a68872ea-89d3-42a0-b2da-877e2d37218b-utilities\") pod \"redhat-operators-f8n6m\" (UID: \"a68872ea-89d3-42a0-b2da-877e2d37218b\") " pod="openshift-marketplace/redhat-operators-f8n6m" Dec 10 22:54:47 crc kubenswrapper[4791]: I1210 22:54:47.571985 4791 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-jgdsn\" (UniqueName: \"kubernetes.io/projected/a68872ea-89d3-42a0-b2da-877e2d37218b-kube-api-access-jgdsn\") pod \"redhat-operators-f8n6m\" (UID: \"a68872ea-89d3-42a0-b2da-877e2d37218b\") " pod="openshift-marketplace/redhat-operators-f8n6m" Dec 10 22:54:47 crc kubenswrapper[4791]: I1210 22:54:47.572828 4791 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/a68872ea-89d3-42a0-b2da-877e2d37218b-catalog-content\") pod \"redhat-operators-f8n6m\" (UID: \"a68872ea-89d3-42a0-b2da-877e2d37218b\") " pod="openshift-marketplace/redhat-operators-f8n6m" Dec 10 22:54:47 crc kubenswrapper[4791]: I1210 22:54:47.572864 4791 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/a68872ea-89d3-42a0-b2da-877e2d37218b-utilities\") pod \"redhat-operators-f8n6m\" (UID: \"a68872ea-89d3-42a0-b2da-877e2d37218b\") " pod="openshift-marketplace/redhat-operators-f8n6m" Dec 10 22:54:47 crc kubenswrapper[4791]: I1210 22:54:47.591963 4791 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-jgdsn\" (UniqueName: \"kubernetes.io/projected/a68872ea-89d3-42a0-b2da-877e2d37218b-kube-api-access-jgdsn\") pod \"redhat-operators-f8n6m\" (UID: \"a68872ea-89d3-42a0-b2da-877e2d37218b\") " pod="openshift-marketplace/redhat-operators-f8n6m" Dec 10 22:54:47 crc kubenswrapper[4791]: I1210 22:54:47.721920 4791 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-f8n6m" Dec 10 22:54:47 crc kubenswrapper[4791]: I1210 22:54:47.761455 4791 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-fcnzr"] Dec 10 22:54:47 crc kubenswrapper[4791]: W1210 22:54:47.766148 4791 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod44085f7f_f0e9_4dec_a816_a687a0dd06ec.slice/crio-e34881f1dd37e6d12fd7380d0f71efbc6011f893e385b4609fde8f0b718d1a42 WatchSource:0}: Error finding container e34881f1dd37e6d12fd7380d0f71efbc6011f893e385b4609fde8f0b718d1a42: Status 404 returned error can't find the container with id e34881f1dd37e6d12fd7380d0f71efbc6011f893e385b4609fde8f0b718d1a42 Dec 10 22:54:47 crc kubenswrapper[4791]: I1210 22:54:47.897700 4791 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-controller-manager/controller-manager-649fb98567-gzgvx"] Dec 10 22:54:47 crc kubenswrapper[4791]: I1210 22:54:47.919124 4791 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-dns"/"openshift-service-ca.crt" Dec 10 22:54:47 crc kubenswrapper[4791]: I1210 22:54:47.925890 4791 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-66c8f7db4d-6582d"] Dec 10 22:54:47 crc kubenswrapper[4791]: I1210 22:54:47.936941 4791 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-f8n6m"] Dec 10 22:54:48 crc kubenswrapper[4791]: I1210 22:54:48.463613 4791 generic.go:334] "Generic (PLEG): container finished" podID="a68872ea-89d3-42a0-b2da-877e2d37218b" containerID="3be29dab82bd910f549d4a2daabc612755cd3823d86fedc94f3dd21f5ca43bda" exitCode=0 Dec 10 22:54:48 crc kubenswrapper[4791]: I1210 22:54:48.463693 4791 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-f8n6m" event={"ID":"a68872ea-89d3-42a0-b2da-877e2d37218b","Type":"ContainerDied","Data":"3be29dab82bd910f549d4a2daabc612755cd3823d86fedc94f3dd21f5ca43bda"} Dec 10 22:54:48 crc kubenswrapper[4791]: I1210 22:54:48.463722 4791 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-f8n6m" event={"ID":"a68872ea-89d3-42a0-b2da-877e2d37218b","Type":"ContainerStarted","Data":"12174d646b63159a5b52a01fb573b37204fcae223f498fe9de000d5296829e04"} Dec 10 22:54:48 crc kubenswrapper[4791]: I1210 22:54:48.466769 4791 generic.go:334] "Generic (PLEG): container finished" podID="44085f7f-f0e9-4dec-a816-a687a0dd06ec" containerID="b0148afd6def4ad6fdbdf05478967cf5e33a3ee8559c227696da0f3a7d5876f9" exitCode=0 Dec 10 22:54:48 crc kubenswrapper[4791]: I1210 22:54:48.466815 4791 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-fcnzr" event={"ID":"44085f7f-f0e9-4dec-a816-a687a0dd06ec","Type":"ContainerDied","Data":"b0148afd6def4ad6fdbdf05478967cf5e33a3ee8559c227696da0f3a7d5876f9"} Dec 10 22:54:48 crc kubenswrapper[4791]: I1210 22:54:48.466851 4791 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-fcnzr" event={"ID":"44085f7f-f0e9-4dec-a816-a687a0dd06ec","Type":"ContainerStarted","Data":"e34881f1dd37e6d12fd7380d0f71efbc6011f893e385b4609fde8f0b718d1a42"} Dec 10 22:54:48 crc kubenswrapper[4791]: I1210 22:54:48.466935 4791 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-route-controller-manager/route-controller-manager-66c8f7db4d-6582d" podUID="fca6ee01-4d09-4b77-89c3-281a84c4f3a6" containerName="route-controller-manager" containerID="cri-o://24139c671ed7e44b2d95f26420006f600a27788c56af9bf724fc7b8e6d25e2e8" gracePeriod=30 Dec 10 22:54:48 crc kubenswrapper[4791]: I1210 22:54:48.467274 4791 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-controller-manager/controller-manager-649fb98567-gzgvx" podUID="c39bed63-baf9-42a6-820e-07199de0212b" containerName="controller-manager" containerID="cri-o://4482a5363e33bf45d6f632e66fb6d1da9a1ef452a5a98c8882574069e9520d50" gracePeriod=30 Dec 10 22:54:48 crc kubenswrapper[4791]: I1210 22:54:48.944117 4791 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-route-controller-manager/route-controller-manager-66c8f7db4d-6582d" Dec 10 22:54:48 crc kubenswrapper[4791]: I1210 22:54:48.995673 4791 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager/controller-manager-649fb98567-gzgvx" Dec 10 22:54:49 crc kubenswrapper[4791]: I1210 22:54:49.128852 4791 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/fca6ee01-4d09-4b77-89c3-281a84c4f3a6-client-ca\") pod \"fca6ee01-4d09-4b77-89c3-281a84c4f3a6\" (UID: \"fca6ee01-4d09-4b77-89c3-281a84c4f3a6\") " Dec 10 22:54:49 crc kubenswrapper[4791]: I1210 22:54:49.128938 4791 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/fca6ee01-4d09-4b77-89c3-281a84c4f3a6-serving-cert\") pod \"fca6ee01-4d09-4b77-89c3-281a84c4f3a6\" (UID: \"fca6ee01-4d09-4b77-89c3-281a84c4f3a6\") " Dec 10 22:54:49 crc kubenswrapper[4791]: I1210 22:54:49.128993 4791 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/c39bed63-baf9-42a6-820e-07199de0212b-proxy-ca-bundles\") pod \"c39bed63-baf9-42a6-820e-07199de0212b\" (UID: \"c39bed63-baf9-42a6-820e-07199de0212b\") " Dec 10 22:54:49 crc kubenswrapper[4791]: I1210 22:54:49.129080 4791 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-wnfx4\" (UniqueName: \"kubernetes.io/projected/c39bed63-baf9-42a6-820e-07199de0212b-kube-api-access-wnfx4\") pod \"c39bed63-baf9-42a6-820e-07199de0212b\" (UID: \"c39bed63-baf9-42a6-820e-07199de0212b\") " Dec 10 22:54:49 crc kubenswrapper[4791]: I1210 22:54:49.129121 4791 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/fca6ee01-4d09-4b77-89c3-281a84c4f3a6-config\") pod \"fca6ee01-4d09-4b77-89c3-281a84c4f3a6\" (UID: \"fca6ee01-4d09-4b77-89c3-281a84c4f3a6\") " Dec 10 22:54:49 crc kubenswrapper[4791]: I1210 22:54:49.129145 4791 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/c39bed63-baf9-42a6-820e-07199de0212b-serving-cert\") pod \"c39bed63-baf9-42a6-820e-07199de0212b\" (UID: \"c39bed63-baf9-42a6-820e-07199de0212b\") " Dec 10 22:54:49 crc kubenswrapper[4791]: I1210 22:54:49.129179 4791 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-8xcc9\" (UniqueName: \"kubernetes.io/projected/fca6ee01-4d09-4b77-89c3-281a84c4f3a6-kube-api-access-8xcc9\") pod \"fca6ee01-4d09-4b77-89c3-281a84c4f3a6\" (UID: \"fca6ee01-4d09-4b77-89c3-281a84c4f3a6\") " Dec 10 22:54:49 crc kubenswrapper[4791]: I1210 22:54:49.129205 4791 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/c39bed63-baf9-42a6-820e-07199de0212b-client-ca\") pod \"c39bed63-baf9-42a6-820e-07199de0212b\" (UID: \"c39bed63-baf9-42a6-820e-07199de0212b\") " Dec 10 22:54:49 crc kubenswrapper[4791]: I1210 22:54:49.129225 4791 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/c39bed63-baf9-42a6-820e-07199de0212b-config\") pod \"c39bed63-baf9-42a6-820e-07199de0212b\" (UID: \"c39bed63-baf9-42a6-820e-07199de0212b\") " Dec 10 22:54:49 crc kubenswrapper[4791]: I1210 22:54:49.129886 4791 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/fca6ee01-4d09-4b77-89c3-281a84c4f3a6-client-ca" (OuterVolumeSpecName: "client-ca") pod "fca6ee01-4d09-4b77-89c3-281a84c4f3a6" (UID: "fca6ee01-4d09-4b77-89c3-281a84c4f3a6"). InnerVolumeSpecName "client-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 10 22:54:49 crc kubenswrapper[4791]: I1210 22:54:49.130223 4791 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/c39bed63-baf9-42a6-820e-07199de0212b-config" (OuterVolumeSpecName: "config") pod "c39bed63-baf9-42a6-820e-07199de0212b" (UID: "c39bed63-baf9-42a6-820e-07199de0212b"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 10 22:54:49 crc kubenswrapper[4791]: I1210 22:54:49.130261 4791 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/c39bed63-baf9-42a6-820e-07199de0212b-proxy-ca-bundles" (OuterVolumeSpecName: "proxy-ca-bundles") pod "c39bed63-baf9-42a6-820e-07199de0212b" (UID: "c39bed63-baf9-42a6-820e-07199de0212b"). InnerVolumeSpecName "proxy-ca-bundles". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 10 22:54:49 crc kubenswrapper[4791]: I1210 22:54:49.130272 4791 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/c39bed63-baf9-42a6-820e-07199de0212b-client-ca" (OuterVolumeSpecName: "client-ca") pod "c39bed63-baf9-42a6-820e-07199de0212b" (UID: "c39bed63-baf9-42a6-820e-07199de0212b"). InnerVolumeSpecName "client-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 10 22:54:49 crc kubenswrapper[4791]: I1210 22:54:49.130472 4791 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/fca6ee01-4d09-4b77-89c3-281a84c4f3a6-config" (OuterVolumeSpecName: "config") pod "fca6ee01-4d09-4b77-89c3-281a84c4f3a6" (UID: "fca6ee01-4d09-4b77-89c3-281a84c4f3a6"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 10 22:54:49 crc kubenswrapper[4791]: I1210 22:54:49.134650 4791 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/fca6ee01-4d09-4b77-89c3-281a84c4f3a6-kube-api-access-8xcc9" (OuterVolumeSpecName: "kube-api-access-8xcc9") pod "fca6ee01-4d09-4b77-89c3-281a84c4f3a6" (UID: "fca6ee01-4d09-4b77-89c3-281a84c4f3a6"). InnerVolumeSpecName "kube-api-access-8xcc9". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 10 22:54:49 crc kubenswrapper[4791]: I1210 22:54:49.134699 4791 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/fca6ee01-4d09-4b77-89c3-281a84c4f3a6-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "fca6ee01-4d09-4b77-89c3-281a84c4f3a6" (UID: "fca6ee01-4d09-4b77-89c3-281a84c4f3a6"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 10 22:54:49 crc kubenswrapper[4791]: I1210 22:54:49.134769 4791 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/c39bed63-baf9-42a6-820e-07199de0212b-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "c39bed63-baf9-42a6-820e-07199de0212b" (UID: "c39bed63-baf9-42a6-820e-07199de0212b"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 10 22:54:49 crc kubenswrapper[4791]: I1210 22:54:49.134903 4791 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/c39bed63-baf9-42a6-820e-07199de0212b-kube-api-access-wnfx4" (OuterVolumeSpecName: "kube-api-access-wnfx4") pod "c39bed63-baf9-42a6-820e-07199de0212b" (UID: "c39bed63-baf9-42a6-820e-07199de0212b"). InnerVolumeSpecName "kube-api-access-wnfx4". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 10 22:54:49 crc kubenswrapper[4791]: I1210 22:54:49.230368 4791 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-wnfx4\" (UniqueName: \"kubernetes.io/projected/c39bed63-baf9-42a6-820e-07199de0212b-kube-api-access-wnfx4\") on node \"crc\" DevicePath \"\"" Dec 10 22:54:49 crc kubenswrapper[4791]: I1210 22:54:49.230440 4791 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/fca6ee01-4d09-4b77-89c3-281a84c4f3a6-config\") on node \"crc\" DevicePath \"\"" Dec 10 22:54:49 crc kubenswrapper[4791]: I1210 22:54:49.230476 4791 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/c39bed63-baf9-42a6-820e-07199de0212b-serving-cert\") on node \"crc\" DevicePath \"\"" Dec 10 22:54:49 crc kubenswrapper[4791]: I1210 22:54:49.230504 4791 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-8xcc9\" (UniqueName: \"kubernetes.io/projected/fca6ee01-4d09-4b77-89c3-281a84c4f3a6-kube-api-access-8xcc9\") on node \"crc\" DevicePath \"\"" Dec 10 22:54:49 crc kubenswrapper[4791]: I1210 22:54:49.230529 4791 reconciler_common.go:293] "Volume detached for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/c39bed63-baf9-42a6-820e-07199de0212b-client-ca\") on node \"crc\" DevicePath \"\"" Dec 10 22:54:49 crc kubenswrapper[4791]: I1210 22:54:49.230548 4791 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/c39bed63-baf9-42a6-820e-07199de0212b-config\") on node \"crc\" DevicePath \"\"" Dec 10 22:54:49 crc kubenswrapper[4791]: I1210 22:54:49.230569 4791 reconciler_common.go:293] "Volume detached for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/fca6ee01-4d09-4b77-89c3-281a84c4f3a6-client-ca\") on node \"crc\" DevicePath \"\"" Dec 10 22:54:49 crc kubenswrapper[4791]: I1210 22:54:49.230590 4791 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/fca6ee01-4d09-4b77-89c3-281a84c4f3a6-serving-cert\") on node \"crc\" DevicePath \"\"" Dec 10 22:54:49 crc kubenswrapper[4791]: I1210 22:54:49.230611 4791 reconciler_common.go:293] "Volume detached for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/c39bed63-baf9-42a6-820e-07199de0212b-proxy-ca-bundles\") on node \"crc\" DevicePath \"\"" Dec 10 22:54:49 crc kubenswrapper[4791]: I1210 22:54:49.376299 4791 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-api"/"openshift-service-ca.crt" Dec 10 22:54:49 crc kubenswrapper[4791]: I1210 22:54:49.473671 4791 generic.go:334] "Generic (PLEG): container finished" podID="c39bed63-baf9-42a6-820e-07199de0212b" containerID="4482a5363e33bf45d6f632e66fb6d1da9a1ef452a5a98c8882574069e9520d50" exitCode=0 Dec 10 22:54:49 crc kubenswrapper[4791]: I1210 22:54:49.473745 4791 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager/controller-manager-649fb98567-gzgvx" event={"ID":"c39bed63-baf9-42a6-820e-07199de0212b","Type":"ContainerDied","Data":"4482a5363e33bf45d6f632e66fb6d1da9a1ef452a5a98c8882574069e9520d50"} Dec 10 22:54:49 crc kubenswrapper[4791]: I1210 22:54:49.473782 4791 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager/controller-manager-649fb98567-gzgvx" event={"ID":"c39bed63-baf9-42a6-820e-07199de0212b","Type":"ContainerDied","Data":"c9aded4fcedee3de19be60b6b9d8b6ba834b343b443652988555ee51c0094409"} Dec 10 22:54:49 crc kubenswrapper[4791]: I1210 22:54:49.473807 4791 scope.go:117] "RemoveContainer" containerID="4482a5363e33bf45d6f632e66fb6d1da9a1ef452a5a98c8882574069e9520d50" Dec 10 22:54:49 crc kubenswrapper[4791]: I1210 22:54:49.473950 4791 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager/controller-manager-649fb98567-gzgvx" Dec 10 22:54:49 crc kubenswrapper[4791]: I1210 22:54:49.478960 4791 generic.go:334] "Generic (PLEG): container finished" podID="fca6ee01-4d09-4b77-89c3-281a84c4f3a6" containerID="24139c671ed7e44b2d95f26420006f600a27788c56af9bf724fc7b8e6d25e2e8" exitCode=0 Dec 10 22:54:49 crc kubenswrapper[4791]: I1210 22:54:49.479012 4791 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-route-controller-manager/route-controller-manager-66c8f7db4d-6582d" event={"ID":"fca6ee01-4d09-4b77-89c3-281a84c4f3a6","Type":"ContainerDied","Data":"24139c671ed7e44b2d95f26420006f600a27788c56af9bf724fc7b8e6d25e2e8"} Dec 10 22:54:49 crc kubenswrapper[4791]: I1210 22:54:49.479048 4791 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-route-controller-manager/route-controller-manager-66c8f7db4d-6582d" event={"ID":"fca6ee01-4d09-4b77-89c3-281a84c4f3a6","Type":"ContainerDied","Data":"956e3e39fb50bb4a38b8cb11a229fa8a5ae0949381c0d96dcbd50e0a483a6b01"} Dec 10 22:54:49 crc kubenswrapper[4791]: I1210 22:54:49.479232 4791 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-route-controller-manager/route-controller-manager-66c8f7db4d-6582d" Dec 10 22:54:49 crc kubenswrapper[4791]: I1210 22:54:49.808404 4791 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/certified-operators-ttrzg"] Dec 10 22:54:49 crc kubenswrapper[4791]: E1210 22:54:49.808951 4791 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="c39bed63-baf9-42a6-820e-07199de0212b" containerName="controller-manager" Dec 10 22:54:49 crc kubenswrapper[4791]: I1210 22:54:49.808967 4791 state_mem.go:107] "Deleted CPUSet assignment" podUID="c39bed63-baf9-42a6-820e-07199de0212b" containerName="controller-manager" Dec 10 22:54:49 crc kubenswrapper[4791]: E1210 22:54:49.808987 4791 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="fca6ee01-4d09-4b77-89c3-281a84c4f3a6" containerName="route-controller-manager" Dec 10 22:54:49 crc kubenswrapper[4791]: I1210 22:54:49.808995 4791 state_mem.go:107] "Deleted CPUSet assignment" podUID="fca6ee01-4d09-4b77-89c3-281a84c4f3a6" containerName="route-controller-manager" Dec 10 22:54:49 crc kubenswrapper[4791]: I1210 22:54:49.809094 4791 memory_manager.go:354] "RemoveStaleState removing state" podUID="fca6ee01-4d09-4b77-89c3-281a84c4f3a6" containerName="route-controller-manager" Dec 10 22:54:49 crc kubenswrapper[4791]: I1210 22:54:49.809115 4791 memory_manager.go:354] "RemoveStaleState removing state" podUID="c39bed63-baf9-42a6-820e-07199de0212b" containerName="controller-manager" Dec 10 22:54:49 crc kubenswrapper[4791]: I1210 22:54:49.810940 4791 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-ttrzg" Dec 10 22:54:49 crc kubenswrapper[4791]: I1210 22:54:49.819670 4791 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-ttrzg"] Dec 10 22:54:49 crc kubenswrapper[4791]: I1210 22:54:49.820052 4791 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"certified-operators-dockercfg-4rs5g" Dec 10 22:54:50 crc kubenswrapper[4791]: I1210 22:54:50.012035 4791 scope.go:117] "RemoveContainer" containerID="4482a5363e33bf45d6f632e66fb6d1da9a1ef452a5a98c8882574069e9520d50" Dec 10 22:54:50 crc kubenswrapper[4791]: E1210 22:54:50.014168 4791 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"4482a5363e33bf45d6f632e66fb6d1da9a1ef452a5a98c8882574069e9520d50\": container with ID starting with 4482a5363e33bf45d6f632e66fb6d1da9a1ef452a5a98c8882574069e9520d50 not found: ID does not exist" containerID="4482a5363e33bf45d6f632e66fb6d1da9a1ef452a5a98c8882574069e9520d50" Dec 10 22:54:50 crc kubenswrapper[4791]: I1210 22:54:50.014244 4791 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"4482a5363e33bf45d6f632e66fb6d1da9a1ef452a5a98c8882574069e9520d50"} err="failed to get container status \"4482a5363e33bf45d6f632e66fb6d1da9a1ef452a5a98c8882574069e9520d50\": rpc error: code = NotFound desc = could not find container \"4482a5363e33bf45d6f632e66fb6d1da9a1ef452a5a98c8882574069e9520d50\": container with ID starting with 4482a5363e33bf45d6f632e66fb6d1da9a1ef452a5a98c8882574069e9520d50 not found: ID does not exist" Dec 10 22:54:50 crc kubenswrapper[4791]: I1210 22:54:50.014277 4791 scope.go:117] "RemoveContainer" containerID="24139c671ed7e44b2d95f26420006f600a27788c56af9bf724fc7b8e6d25e2e8" Dec 10 22:54:50 crc kubenswrapper[4791]: I1210 22:54:50.029467 4791 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/community-operators-x582q"] Dec 10 22:54:50 crc kubenswrapper[4791]: I1210 22:54:50.041459 4791 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-x582q"] Dec 10 22:54:50 crc kubenswrapper[4791]: I1210 22:54:50.041649 4791 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-x582q" Dec 10 22:54:50 crc kubenswrapper[4791]: I1210 22:54:50.042845 4791 scope.go:117] "RemoveContainer" containerID="24139c671ed7e44b2d95f26420006f600a27788c56af9bf724fc7b8e6d25e2e8" Dec 10 22:54:50 crc kubenswrapper[4791]: E1210 22:54:50.043265 4791 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"24139c671ed7e44b2d95f26420006f600a27788c56af9bf724fc7b8e6d25e2e8\": container with ID starting with 24139c671ed7e44b2d95f26420006f600a27788c56af9bf724fc7b8e6d25e2e8 not found: ID does not exist" containerID="24139c671ed7e44b2d95f26420006f600a27788c56af9bf724fc7b8e6d25e2e8" Dec 10 22:54:50 crc kubenswrapper[4791]: I1210 22:54:50.043298 4791 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"24139c671ed7e44b2d95f26420006f600a27788c56af9bf724fc7b8e6d25e2e8"} err="failed to get container status \"24139c671ed7e44b2d95f26420006f600a27788c56af9bf724fc7b8e6d25e2e8\": rpc error: code = NotFound desc = could not find container \"24139c671ed7e44b2d95f26420006f600a27788c56af9bf724fc7b8e6d25e2e8\": container with ID starting with 24139c671ed7e44b2d95f26420006f600a27788c56af9bf724fc7b8e6d25e2e8 not found: ID does not exist" Dec 10 22:54:50 crc kubenswrapper[4791]: I1210 22:54:50.044105 4791 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"community-operators-dockercfg-dmngl" Dec 10 22:54:50 crc kubenswrapper[4791]: I1210 22:54:50.048076 4791 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/f00159a6-00e0-4975-87a0-9d8b2b67b157-catalog-content\") pod \"certified-operators-ttrzg\" (UID: \"f00159a6-00e0-4975-87a0-9d8b2b67b157\") " pod="openshift-marketplace/certified-operators-ttrzg" Dec 10 22:54:50 crc kubenswrapper[4791]: I1210 22:54:50.048131 4791 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-v9bhf\" (UniqueName: \"kubernetes.io/projected/f00159a6-00e0-4975-87a0-9d8b2b67b157-kube-api-access-v9bhf\") pod \"certified-operators-ttrzg\" (UID: \"f00159a6-00e0-4975-87a0-9d8b2b67b157\") " pod="openshift-marketplace/certified-operators-ttrzg" Dec 10 22:54:50 crc kubenswrapper[4791]: I1210 22:54:50.048173 4791 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/f00159a6-00e0-4975-87a0-9d8b2b67b157-utilities\") pod \"certified-operators-ttrzg\" (UID: \"f00159a6-00e0-4975-87a0-9d8b2b67b157\") " pod="openshift-marketplace/certified-operators-ttrzg" Dec 10 22:54:50 crc kubenswrapper[4791]: I1210 22:54:50.073679 4791 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-controller-manager/controller-manager-649fb98567-gzgvx"] Dec 10 22:54:50 crc kubenswrapper[4791]: I1210 22:54:50.077304 4791 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-controller-manager/controller-manager-649fb98567-gzgvx"] Dec 10 22:54:50 crc kubenswrapper[4791]: I1210 22:54:50.150403 4791 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/f00159a6-00e0-4975-87a0-9d8b2b67b157-utilities\") pod \"certified-operators-ttrzg\" (UID: \"f00159a6-00e0-4975-87a0-9d8b2b67b157\") " pod="openshift-marketplace/certified-operators-ttrzg" Dec 10 22:54:50 crc kubenswrapper[4791]: I1210 22:54:50.150490 4791 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/d02b36c2-1b51-4466-af31-2e69357acc9c-utilities\") pod \"community-operators-x582q\" (UID: \"d02b36c2-1b51-4466-af31-2e69357acc9c\") " pod="openshift-marketplace/community-operators-x582q" Dec 10 22:54:50 crc kubenswrapper[4791]: I1210 22:54:50.150587 4791 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/d02b36c2-1b51-4466-af31-2e69357acc9c-catalog-content\") pod \"community-operators-x582q\" (UID: \"d02b36c2-1b51-4466-af31-2e69357acc9c\") " pod="openshift-marketplace/community-operators-x582q" Dec 10 22:54:50 crc kubenswrapper[4791]: I1210 22:54:50.150613 4791 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-mnr2t\" (UniqueName: \"kubernetes.io/projected/d02b36c2-1b51-4466-af31-2e69357acc9c-kube-api-access-mnr2t\") pod \"community-operators-x582q\" (UID: \"d02b36c2-1b51-4466-af31-2e69357acc9c\") " pod="openshift-marketplace/community-operators-x582q" Dec 10 22:54:50 crc kubenswrapper[4791]: I1210 22:54:50.150744 4791 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/f00159a6-00e0-4975-87a0-9d8b2b67b157-catalog-content\") pod \"certified-operators-ttrzg\" (UID: \"f00159a6-00e0-4975-87a0-9d8b2b67b157\") " pod="openshift-marketplace/certified-operators-ttrzg" Dec 10 22:54:50 crc kubenswrapper[4791]: I1210 22:54:50.150828 4791 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-v9bhf\" (UniqueName: \"kubernetes.io/projected/f00159a6-00e0-4975-87a0-9d8b2b67b157-kube-api-access-v9bhf\") pod \"certified-operators-ttrzg\" (UID: \"f00159a6-00e0-4975-87a0-9d8b2b67b157\") " pod="openshift-marketplace/certified-operators-ttrzg" Dec 10 22:54:50 crc kubenswrapper[4791]: I1210 22:54:50.151012 4791 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/f00159a6-00e0-4975-87a0-9d8b2b67b157-utilities\") pod \"certified-operators-ttrzg\" (UID: \"f00159a6-00e0-4975-87a0-9d8b2b67b157\") " pod="openshift-marketplace/certified-operators-ttrzg" Dec 10 22:54:50 crc kubenswrapper[4791]: I1210 22:54:50.151579 4791 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/f00159a6-00e0-4975-87a0-9d8b2b67b157-catalog-content\") pod \"certified-operators-ttrzg\" (UID: \"f00159a6-00e0-4975-87a0-9d8b2b67b157\") " pod="openshift-marketplace/certified-operators-ttrzg" Dec 10 22:54:50 crc kubenswrapper[4791]: I1210 22:54:50.168365 4791 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-v9bhf\" (UniqueName: \"kubernetes.io/projected/f00159a6-00e0-4975-87a0-9d8b2b67b157-kube-api-access-v9bhf\") pod \"certified-operators-ttrzg\" (UID: \"f00159a6-00e0-4975-87a0-9d8b2b67b157\") " pod="openshift-marketplace/certified-operators-ttrzg" Dec 10 22:54:50 crc kubenswrapper[4791]: I1210 22:54:50.207024 4791 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-route-controller-manager/route-controller-manager-66c8f7db4d-q648r"] Dec 10 22:54:50 crc kubenswrapper[4791]: I1210 22:54:50.207813 4791 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-route-controller-manager/route-controller-manager-66c8f7db4d-q648r" Dec 10 22:54:50 crc kubenswrapper[4791]: I1210 22:54:50.210149 4791 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"config" Dec 10 22:54:50 crc kubenswrapper[4791]: I1210 22:54:50.210176 4791 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"client-ca" Dec 10 22:54:50 crc kubenswrapper[4791]: I1210 22:54:50.210459 4791 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"openshift-service-ca.crt" Dec 10 22:54:50 crc kubenswrapper[4791]: I1210 22:54:50.210660 4791 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"kube-root-ca.crt" Dec 10 22:54:50 crc kubenswrapper[4791]: I1210 22:54:50.210803 4791 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-route-controller-manager"/"route-controller-manager-sa-dockercfg-h2zr2" Dec 10 22:54:50 crc kubenswrapper[4791]: I1210 22:54:50.211843 4791 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-route-controller-manager"/"serving-cert" Dec 10 22:54:50 crc kubenswrapper[4791]: I1210 22:54:50.218305 4791 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-66c8f7db4d-q648r"] Dec 10 22:54:50 crc kubenswrapper[4791]: I1210 22:54:50.252363 4791 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/d02b36c2-1b51-4466-af31-2e69357acc9c-catalog-content\") pod \"community-operators-x582q\" (UID: \"d02b36c2-1b51-4466-af31-2e69357acc9c\") " pod="openshift-marketplace/community-operators-x582q" Dec 10 22:54:50 crc kubenswrapper[4791]: I1210 22:54:50.252407 4791 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-mnr2t\" (UniqueName: \"kubernetes.io/projected/d02b36c2-1b51-4466-af31-2e69357acc9c-kube-api-access-mnr2t\") pod \"community-operators-x582q\" (UID: \"d02b36c2-1b51-4466-af31-2e69357acc9c\") " pod="openshift-marketplace/community-operators-x582q" Dec 10 22:54:50 crc kubenswrapper[4791]: I1210 22:54:50.252472 4791 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/d02b36c2-1b51-4466-af31-2e69357acc9c-utilities\") pod \"community-operators-x582q\" (UID: \"d02b36c2-1b51-4466-af31-2e69357acc9c\") " pod="openshift-marketplace/community-operators-x582q" Dec 10 22:54:50 crc kubenswrapper[4791]: I1210 22:54:50.252939 4791 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/d02b36c2-1b51-4466-af31-2e69357acc9c-catalog-content\") pod \"community-operators-x582q\" (UID: \"d02b36c2-1b51-4466-af31-2e69357acc9c\") " pod="openshift-marketplace/community-operators-x582q" Dec 10 22:54:50 crc kubenswrapper[4791]: I1210 22:54:50.253183 4791 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/d02b36c2-1b51-4466-af31-2e69357acc9c-utilities\") pod \"community-operators-x582q\" (UID: \"d02b36c2-1b51-4466-af31-2e69357acc9c\") " pod="openshift-marketplace/community-operators-x582q" Dec 10 22:54:50 crc kubenswrapper[4791]: I1210 22:54:50.273425 4791 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-mnr2t\" (UniqueName: \"kubernetes.io/projected/d02b36c2-1b51-4466-af31-2e69357acc9c-kube-api-access-mnr2t\") pod \"community-operators-x582q\" (UID: \"d02b36c2-1b51-4466-af31-2e69357acc9c\") " pod="openshift-marketplace/community-operators-x582q" Dec 10 22:54:50 crc kubenswrapper[4791]: I1210 22:54:50.342206 4791 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-ttrzg" Dec 10 22:54:50 crc kubenswrapper[4791]: I1210 22:54:50.354841 4791 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/d4ca2665-1d0c-44b2-8956-336444b2cea9-serving-cert\") pod \"route-controller-manager-66c8f7db4d-q648r\" (UID: \"d4ca2665-1d0c-44b2-8956-336444b2cea9\") " pod="openshift-route-controller-manager/route-controller-manager-66c8f7db4d-q648r" Dec 10 22:54:50 crc kubenswrapper[4791]: I1210 22:54:50.354905 4791 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/d4ca2665-1d0c-44b2-8956-336444b2cea9-client-ca\") pod \"route-controller-manager-66c8f7db4d-q648r\" (UID: \"d4ca2665-1d0c-44b2-8956-336444b2cea9\") " pod="openshift-route-controller-manager/route-controller-manager-66c8f7db4d-q648r" Dec 10 22:54:50 crc kubenswrapper[4791]: I1210 22:54:50.354982 4791 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/d4ca2665-1d0c-44b2-8956-336444b2cea9-config\") pod \"route-controller-manager-66c8f7db4d-q648r\" (UID: \"d4ca2665-1d0c-44b2-8956-336444b2cea9\") " pod="openshift-route-controller-manager/route-controller-manager-66c8f7db4d-q648r" Dec 10 22:54:50 crc kubenswrapper[4791]: I1210 22:54:50.355008 4791 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-j26jx\" (UniqueName: \"kubernetes.io/projected/d4ca2665-1d0c-44b2-8956-336444b2cea9-kube-api-access-j26jx\") pod \"route-controller-manager-66c8f7db4d-q648r\" (UID: \"d4ca2665-1d0c-44b2-8956-336444b2cea9\") " pod="openshift-route-controller-manager/route-controller-manager-66c8f7db4d-q648r" Dec 10 22:54:50 crc kubenswrapper[4791]: I1210 22:54:50.394622 4791 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-x582q" Dec 10 22:54:50 crc kubenswrapper[4791]: I1210 22:54:50.459970 4791 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/d4ca2665-1d0c-44b2-8956-336444b2cea9-serving-cert\") pod \"route-controller-manager-66c8f7db4d-q648r\" (UID: \"d4ca2665-1d0c-44b2-8956-336444b2cea9\") " pod="openshift-route-controller-manager/route-controller-manager-66c8f7db4d-q648r" Dec 10 22:54:50 crc kubenswrapper[4791]: I1210 22:54:50.460402 4791 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/d4ca2665-1d0c-44b2-8956-336444b2cea9-client-ca\") pod \"route-controller-manager-66c8f7db4d-q648r\" (UID: \"d4ca2665-1d0c-44b2-8956-336444b2cea9\") " pod="openshift-route-controller-manager/route-controller-manager-66c8f7db4d-q648r" Dec 10 22:54:50 crc kubenswrapper[4791]: I1210 22:54:50.461334 4791 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/d4ca2665-1d0c-44b2-8956-336444b2cea9-client-ca\") pod \"route-controller-manager-66c8f7db4d-q648r\" (UID: \"d4ca2665-1d0c-44b2-8956-336444b2cea9\") " pod="openshift-route-controller-manager/route-controller-manager-66c8f7db4d-q648r" Dec 10 22:54:50 crc kubenswrapper[4791]: I1210 22:54:50.462150 4791 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/d4ca2665-1d0c-44b2-8956-336444b2cea9-config\") pod \"route-controller-manager-66c8f7db4d-q648r\" (UID: \"d4ca2665-1d0c-44b2-8956-336444b2cea9\") " pod="openshift-route-controller-manager/route-controller-manager-66c8f7db4d-q648r" Dec 10 22:54:50 crc kubenswrapper[4791]: I1210 22:54:50.462184 4791 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/d4ca2665-1d0c-44b2-8956-336444b2cea9-config\") pod \"route-controller-manager-66c8f7db4d-q648r\" (UID: \"d4ca2665-1d0c-44b2-8956-336444b2cea9\") " pod="openshift-route-controller-manager/route-controller-manager-66c8f7db4d-q648r" Dec 10 22:54:50 crc kubenswrapper[4791]: I1210 22:54:50.462219 4791 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-j26jx\" (UniqueName: \"kubernetes.io/projected/d4ca2665-1d0c-44b2-8956-336444b2cea9-kube-api-access-j26jx\") pod \"route-controller-manager-66c8f7db4d-q648r\" (UID: \"d4ca2665-1d0c-44b2-8956-336444b2cea9\") " pod="openshift-route-controller-manager/route-controller-manager-66c8f7db4d-q648r" Dec 10 22:54:50 crc kubenswrapper[4791]: I1210 22:54:50.467292 4791 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/d4ca2665-1d0c-44b2-8956-336444b2cea9-serving-cert\") pod \"route-controller-manager-66c8f7db4d-q648r\" (UID: \"d4ca2665-1d0c-44b2-8956-336444b2cea9\") " pod="openshift-route-controller-manager/route-controller-manager-66c8f7db4d-q648r" Dec 10 22:54:50 crc kubenswrapper[4791]: I1210 22:54:50.488327 4791 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-j26jx\" (UniqueName: \"kubernetes.io/projected/d4ca2665-1d0c-44b2-8956-336444b2cea9-kube-api-access-j26jx\") pod \"route-controller-manager-66c8f7db4d-q648r\" (UID: \"d4ca2665-1d0c-44b2-8956-336444b2cea9\") " pod="openshift-route-controller-manager/route-controller-manager-66c8f7db4d-q648r" Dec 10 22:54:50 crc kubenswrapper[4791]: I1210 22:54:50.491473 4791 generic.go:334] "Generic (PLEG): container finished" podID="44085f7f-f0e9-4dec-a816-a687a0dd06ec" containerID="8e6ba64f65c39988b97340e897b237956d7abe9734049f816baa98585242c40e" exitCode=0 Dec 10 22:54:50 crc kubenswrapper[4791]: I1210 22:54:50.491532 4791 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-fcnzr" event={"ID":"44085f7f-f0e9-4dec-a816-a687a0dd06ec","Type":"ContainerDied","Data":"8e6ba64f65c39988b97340e897b237956d7abe9734049f816baa98585242c40e"} Dec 10 22:54:50 crc kubenswrapper[4791]: I1210 22:54:50.512062 4791 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-f8n6m" event={"ID":"a68872ea-89d3-42a0-b2da-877e2d37218b","Type":"ContainerStarted","Data":"078a3fd1aca90b373113bd2c7c046fb1675dc77fa9d277babd800239c04aa00c"} Dec 10 22:54:50 crc kubenswrapper[4791]: I1210 22:54:50.602181 4791 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-route-controller-manager/route-controller-manager-66c8f7db4d-q648r" Dec 10 22:54:50 crc kubenswrapper[4791]: I1210 22:54:50.790903 4791 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-66c8f7db4d-q648r"] Dec 10 22:54:50 crc kubenswrapper[4791]: I1210 22:54:50.794253 4791 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-ttrzg"] Dec 10 22:54:50 crc kubenswrapper[4791]: W1210 22:54:50.804514 4791 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podf00159a6_00e0_4975_87a0_9d8b2b67b157.slice/crio-a144abeebcd2dbd886fa194cfaad11b723fbdd129a7b14b88655666921aa98c9 WatchSource:0}: Error finding container a144abeebcd2dbd886fa194cfaad11b723fbdd129a7b14b88655666921aa98c9: Status 404 returned error can't find the container with id a144abeebcd2dbd886fa194cfaad11b723fbdd129a7b14b88655666921aa98c9 Dec 10 22:54:50 crc kubenswrapper[4791]: I1210 22:54:50.835173 4791 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-x582q"] Dec 10 22:54:50 crc kubenswrapper[4791]: W1210 22:54:50.884468 4791 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podd02b36c2_1b51_4466_af31_2e69357acc9c.slice/crio-a1289af07c82266f5a3c2561ca77ddaffcfb2e0f9a274d3b024bd37e9b3ea1e2 WatchSource:0}: Error finding container a1289af07c82266f5a3c2561ca77ddaffcfb2e0f9a274d3b024bd37e9b3ea1e2: Status 404 returned error can't find the container with id a1289af07c82266f5a3c2561ca77ddaffcfb2e0f9a274d3b024bd37e9b3ea1e2 Dec 10 22:54:51 crc kubenswrapper[4791]: I1210 22:54:51.210475 4791 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-controller-manager/controller-manager-5747cbd54d-9mwk9"] Dec 10 22:54:51 crc kubenswrapper[4791]: I1210 22:54:51.211877 4791 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager/controller-manager-5747cbd54d-9mwk9" Dec 10 22:54:51 crc kubenswrapper[4791]: I1210 22:54:51.218551 4791 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-controller-manager/controller-manager-5747cbd54d-9mwk9"] Dec 10 22:54:51 crc kubenswrapper[4791]: I1210 22:54:51.220310 4791 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"client-ca" Dec 10 22:54:51 crc kubenswrapper[4791]: I1210 22:54:51.220463 4791 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"config" Dec 10 22:54:51 crc kubenswrapper[4791]: I1210 22:54:51.220566 4791 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"openshift-service-ca.crt" Dec 10 22:54:51 crc kubenswrapper[4791]: I1210 22:54:51.220737 4791 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-controller-manager"/"openshift-controller-manager-sa-dockercfg-msq4c" Dec 10 22:54:51 crc kubenswrapper[4791]: I1210 22:54:51.221256 4791 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"kube-root-ca.crt" Dec 10 22:54:51 crc kubenswrapper[4791]: I1210 22:54:51.223841 4791 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-controller-manager"/"serving-cert" Dec 10 22:54:51 crc kubenswrapper[4791]: I1210 22:54:51.226105 4791 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"openshift-global-ca" Dec 10 22:54:51 crc kubenswrapper[4791]: I1210 22:54:51.373605 4791 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/b9b00f47-2658-4115-8c86-5d57ad0632df-serving-cert\") pod \"controller-manager-5747cbd54d-9mwk9\" (UID: \"b9b00f47-2658-4115-8c86-5d57ad0632df\") " pod="openshift-controller-manager/controller-manager-5747cbd54d-9mwk9" Dec 10 22:54:51 crc kubenswrapper[4791]: I1210 22:54:51.373683 4791 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-2stvb\" (UniqueName: \"kubernetes.io/projected/b9b00f47-2658-4115-8c86-5d57ad0632df-kube-api-access-2stvb\") pod \"controller-manager-5747cbd54d-9mwk9\" (UID: \"b9b00f47-2658-4115-8c86-5d57ad0632df\") " pod="openshift-controller-manager/controller-manager-5747cbd54d-9mwk9" Dec 10 22:54:51 crc kubenswrapper[4791]: I1210 22:54:51.373712 4791 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/b9b00f47-2658-4115-8c86-5d57ad0632df-proxy-ca-bundles\") pod \"controller-manager-5747cbd54d-9mwk9\" (UID: \"b9b00f47-2658-4115-8c86-5d57ad0632df\") " pod="openshift-controller-manager/controller-manager-5747cbd54d-9mwk9" Dec 10 22:54:51 crc kubenswrapper[4791]: I1210 22:54:51.373736 4791 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/b9b00f47-2658-4115-8c86-5d57ad0632df-client-ca\") pod \"controller-manager-5747cbd54d-9mwk9\" (UID: \"b9b00f47-2658-4115-8c86-5d57ad0632df\") " pod="openshift-controller-manager/controller-manager-5747cbd54d-9mwk9" Dec 10 22:54:51 crc kubenswrapper[4791]: I1210 22:54:51.373777 4791 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/b9b00f47-2658-4115-8c86-5d57ad0632df-config\") pod \"controller-manager-5747cbd54d-9mwk9\" (UID: \"b9b00f47-2658-4115-8c86-5d57ad0632df\") " pod="openshift-controller-manager/controller-manager-5747cbd54d-9mwk9" Dec 10 22:54:51 crc kubenswrapper[4791]: I1210 22:54:51.474892 4791 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/b9b00f47-2658-4115-8c86-5d57ad0632df-serving-cert\") pod \"controller-manager-5747cbd54d-9mwk9\" (UID: \"b9b00f47-2658-4115-8c86-5d57ad0632df\") " pod="openshift-controller-manager/controller-manager-5747cbd54d-9mwk9" Dec 10 22:54:51 crc kubenswrapper[4791]: I1210 22:54:51.474944 4791 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-2stvb\" (UniqueName: \"kubernetes.io/projected/b9b00f47-2658-4115-8c86-5d57ad0632df-kube-api-access-2stvb\") pod \"controller-manager-5747cbd54d-9mwk9\" (UID: \"b9b00f47-2658-4115-8c86-5d57ad0632df\") " pod="openshift-controller-manager/controller-manager-5747cbd54d-9mwk9" Dec 10 22:54:51 crc kubenswrapper[4791]: I1210 22:54:51.474964 4791 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/b9b00f47-2658-4115-8c86-5d57ad0632df-proxy-ca-bundles\") pod \"controller-manager-5747cbd54d-9mwk9\" (UID: \"b9b00f47-2658-4115-8c86-5d57ad0632df\") " pod="openshift-controller-manager/controller-manager-5747cbd54d-9mwk9" Dec 10 22:54:51 crc kubenswrapper[4791]: I1210 22:54:51.474982 4791 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/b9b00f47-2658-4115-8c86-5d57ad0632df-client-ca\") pod \"controller-manager-5747cbd54d-9mwk9\" (UID: \"b9b00f47-2658-4115-8c86-5d57ad0632df\") " pod="openshift-controller-manager/controller-manager-5747cbd54d-9mwk9" Dec 10 22:54:51 crc kubenswrapper[4791]: I1210 22:54:51.475024 4791 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/b9b00f47-2658-4115-8c86-5d57ad0632df-config\") pod \"controller-manager-5747cbd54d-9mwk9\" (UID: \"b9b00f47-2658-4115-8c86-5d57ad0632df\") " pod="openshift-controller-manager/controller-manager-5747cbd54d-9mwk9" Dec 10 22:54:51 crc kubenswrapper[4791]: I1210 22:54:51.476398 4791 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/b9b00f47-2658-4115-8c86-5d57ad0632df-client-ca\") pod \"controller-manager-5747cbd54d-9mwk9\" (UID: \"b9b00f47-2658-4115-8c86-5d57ad0632df\") " pod="openshift-controller-manager/controller-manager-5747cbd54d-9mwk9" Dec 10 22:54:51 crc kubenswrapper[4791]: I1210 22:54:51.476558 4791 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/b9b00f47-2658-4115-8c86-5d57ad0632df-proxy-ca-bundles\") pod \"controller-manager-5747cbd54d-9mwk9\" (UID: \"b9b00f47-2658-4115-8c86-5d57ad0632df\") " pod="openshift-controller-manager/controller-manager-5747cbd54d-9mwk9" Dec 10 22:54:51 crc kubenswrapper[4791]: I1210 22:54:51.476721 4791 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/b9b00f47-2658-4115-8c86-5d57ad0632df-config\") pod \"controller-manager-5747cbd54d-9mwk9\" (UID: \"b9b00f47-2658-4115-8c86-5d57ad0632df\") " pod="openshift-controller-manager/controller-manager-5747cbd54d-9mwk9" Dec 10 22:54:51 crc kubenswrapper[4791]: I1210 22:54:51.481572 4791 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/b9b00f47-2658-4115-8c86-5d57ad0632df-serving-cert\") pod \"controller-manager-5747cbd54d-9mwk9\" (UID: \"b9b00f47-2658-4115-8c86-5d57ad0632df\") " pod="openshift-controller-manager/controller-manager-5747cbd54d-9mwk9" Dec 10 22:54:51 crc kubenswrapper[4791]: I1210 22:54:51.495489 4791 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-2stvb\" (UniqueName: \"kubernetes.io/projected/b9b00f47-2658-4115-8c86-5d57ad0632df-kube-api-access-2stvb\") pod \"controller-manager-5747cbd54d-9mwk9\" (UID: \"b9b00f47-2658-4115-8c86-5d57ad0632df\") " pod="openshift-controller-manager/controller-manager-5747cbd54d-9mwk9" Dec 10 22:54:51 crc kubenswrapper[4791]: I1210 22:54:51.522476 4791 generic.go:334] "Generic (PLEG): container finished" podID="a68872ea-89d3-42a0-b2da-877e2d37218b" containerID="078a3fd1aca90b373113bd2c7c046fb1675dc77fa9d277babd800239c04aa00c" exitCode=0 Dec 10 22:54:51 crc kubenswrapper[4791]: I1210 22:54:51.522552 4791 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-f8n6m" event={"ID":"a68872ea-89d3-42a0-b2da-877e2d37218b","Type":"ContainerDied","Data":"078a3fd1aca90b373113bd2c7c046fb1675dc77fa9d277babd800239c04aa00c"} Dec 10 22:54:51 crc kubenswrapper[4791]: I1210 22:54:51.524509 4791 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-route-controller-manager/route-controller-manager-66c8f7db4d-q648r" event={"ID":"d4ca2665-1d0c-44b2-8956-336444b2cea9","Type":"ContainerStarted","Data":"87d1e951b1b19fcdfa218158ad2f4b0d4c10eb447c09fee1a0bcb941411e765f"} Dec 10 22:54:51 crc kubenswrapper[4791]: I1210 22:54:51.524552 4791 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-route-controller-manager/route-controller-manager-66c8f7db4d-q648r" event={"ID":"d4ca2665-1d0c-44b2-8956-336444b2cea9","Type":"ContainerStarted","Data":"a8348b8ec491736115e165c1f70b7a229ffff96075dc19581805163b96a93700"} Dec 10 22:54:51 crc kubenswrapper[4791]: I1210 22:54:51.524684 4791 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-route-controller-manager/route-controller-manager-66c8f7db4d-q648r" Dec 10 22:54:51 crc kubenswrapper[4791]: I1210 22:54:51.528537 4791 generic.go:334] "Generic (PLEG): container finished" podID="f00159a6-00e0-4975-87a0-9d8b2b67b157" containerID="03d58c5c8454efc83b466213d9227a8c922b9de1db055bcd3da961cae612f83a" exitCode=0 Dec 10 22:54:51 crc kubenswrapper[4791]: I1210 22:54:51.528604 4791 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-ttrzg" event={"ID":"f00159a6-00e0-4975-87a0-9d8b2b67b157","Type":"ContainerDied","Data":"03d58c5c8454efc83b466213d9227a8c922b9de1db055bcd3da961cae612f83a"} Dec 10 22:54:51 crc kubenswrapper[4791]: I1210 22:54:51.528630 4791 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-ttrzg" event={"ID":"f00159a6-00e0-4975-87a0-9d8b2b67b157","Type":"ContainerStarted","Data":"a144abeebcd2dbd886fa194cfaad11b723fbdd129a7b14b88655666921aa98c9"} Dec 10 22:54:51 crc kubenswrapper[4791]: I1210 22:54:51.530282 4791 generic.go:334] "Generic (PLEG): container finished" podID="d02b36c2-1b51-4466-af31-2e69357acc9c" containerID="3fdb17aac96abc1f7640904b5a304c77bde56e9a7b6b3df88c5701080b4d5392" exitCode=0 Dec 10 22:54:51 crc kubenswrapper[4791]: I1210 22:54:51.530313 4791 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-x582q" event={"ID":"d02b36c2-1b51-4466-af31-2e69357acc9c","Type":"ContainerDied","Data":"3fdb17aac96abc1f7640904b5a304c77bde56e9a7b6b3df88c5701080b4d5392"} Dec 10 22:54:51 crc kubenswrapper[4791]: I1210 22:54:51.530334 4791 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-x582q" event={"ID":"d02b36c2-1b51-4466-af31-2e69357acc9c","Type":"ContainerStarted","Data":"a1289af07c82266f5a3c2561ca77ddaffcfb2e0f9a274d3b024bd37e9b3ea1e2"} Dec 10 22:54:51 crc kubenswrapper[4791]: I1210 22:54:51.532127 4791 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager/controller-manager-5747cbd54d-9mwk9" Dec 10 22:54:51 crc kubenswrapper[4791]: I1210 22:54:51.592219 4791 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-route-controller-manager/route-controller-manager-66c8f7db4d-q648r" podStartSLOduration=3.59219503 podStartE2EDuration="3.59219503s" podCreationTimestamp="2025-12-10 22:54:48 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-10 22:54:51.588617235 +0000 UTC m=+326.018234878" watchObservedRunningTime="2025-12-10 22:54:51.59219503 +0000 UTC m=+326.021812643" Dec 10 22:54:51 crc kubenswrapper[4791]: I1210 22:54:51.769269 4791 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-controller-manager/controller-manager-5747cbd54d-9mwk9"] Dec 10 22:54:51 crc kubenswrapper[4791]: I1210 22:54:51.892351 4791 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="c39bed63-baf9-42a6-820e-07199de0212b" path="/var/lib/kubelet/pods/c39bed63-baf9-42a6-820e-07199de0212b/volumes" Dec 10 22:54:51 crc kubenswrapper[4791]: I1210 22:54:51.934769 4791 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-route-controller-manager/route-controller-manager-66c8f7db4d-q648r" Dec 10 22:54:52 crc kubenswrapper[4791]: I1210 22:54:52.535754 4791 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager/controller-manager-5747cbd54d-9mwk9" event={"ID":"b9b00f47-2658-4115-8c86-5d57ad0632df","Type":"ContainerStarted","Data":"018136a3268c789d9ddd9dd74003b3e01e9f3efae97293726a35d6582353bb26"} Dec 10 22:54:52 crc kubenswrapper[4791]: I1210 22:54:52.536056 4791 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager/controller-manager-5747cbd54d-9mwk9" event={"ID":"b9b00f47-2658-4115-8c86-5d57ad0632df","Type":"ContainerStarted","Data":"c6df9b1d71ae0fe7a2ae82393e128670b8b6b046d598d7e61bc305e885a1df04"} Dec 10 22:54:52 crc kubenswrapper[4791]: I1210 22:54:52.536071 4791 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-controller-manager/controller-manager-5747cbd54d-9mwk9" Dec 10 22:54:52 crc kubenswrapper[4791]: I1210 22:54:52.540333 4791 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-fcnzr" event={"ID":"44085f7f-f0e9-4dec-a816-a687a0dd06ec","Type":"ContainerStarted","Data":"86aa9b9d1b642a1101d531b4c8eeaea161f9039b532fd32343b6315414331f84"} Dec 10 22:54:52 crc kubenswrapper[4791]: I1210 22:54:52.544885 4791 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-controller-manager/controller-manager-5747cbd54d-9mwk9" Dec 10 22:54:52 crc kubenswrapper[4791]: I1210 22:54:52.554251 4791 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-controller-manager/controller-manager-5747cbd54d-9mwk9" podStartSLOduration=5.55423459 podStartE2EDuration="5.55423459s" podCreationTimestamp="2025-12-10 22:54:47 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-10 22:54:52.55123581 +0000 UTC m=+326.980853423" watchObservedRunningTime="2025-12-10 22:54:52.55423459 +0000 UTC m=+326.983852203" Dec 10 22:54:52 crc kubenswrapper[4791]: I1210 22:54:52.617711 4791 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-marketplace-fcnzr" podStartSLOduration=2.7323373220000002 podStartE2EDuration="5.617693597s" podCreationTimestamp="2025-12-10 22:54:47 +0000 UTC" firstStartedPulling="2025-12-10 22:54:48.469933807 +0000 UTC m=+322.899551430" lastFinishedPulling="2025-12-10 22:54:51.355290072 +0000 UTC m=+325.784907705" observedRunningTime="2025-12-10 22:54:52.615936134 +0000 UTC m=+327.045553747" watchObservedRunningTime="2025-12-10 22:54:52.617693597 +0000 UTC m=+327.047311210" Dec 10 22:54:53 crc kubenswrapper[4791]: I1210 22:54:53.557441 4791 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-f8n6m" event={"ID":"a68872ea-89d3-42a0-b2da-877e2d37218b","Type":"ContainerStarted","Data":"352c956d7b46fa20d806bf22363b1b6bdd5db1ef0ae09abd16d40893c5feee5f"} Dec 10 22:54:53 crc kubenswrapper[4791]: I1210 22:54:53.559684 4791 generic.go:334] "Generic (PLEG): container finished" podID="f00159a6-00e0-4975-87a0-9d8b2b67b157" containerID="c364bedbb065523a05d7f6e6823a440f9098a3dd7c9baff34c2b16acb2d6458d" exitCode=0 Dec 10 22:54:53 crc kubenswrapper[4791]: I1210 22:54:53.559742 4791 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-ttrzg" event={"ID":"f00159a6-00e0-4975-87a0-9d8b2b67b157","Type":"ContainerDied","Data":"c364bedbb065523a05d7f6e6823a440f9098a3dd7c9baff34c2b16acb2d6458d"} Dec 10 22:54:53 crc kubenswrapper[4791]: I1210 22:54:53.567764 4791 generic.go:334] "Generic (PLEG): container finished" podID="d02b36c2-1b51-4466-af31-2e69357acc9c" containerID="6c0e2131fa38031bdb808a2299181b6946b5286ec4e9cac0dee68f887c359d9e" exitCode=0 Dec 10 22:54:53 crc kubenswrapper[4791]: I1210 22:54:53.568044 4791 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-x582q" event={"ID":"d02b36c2-1b51-4466-af31-2e69357acc9c","Type":"ContainerDied","Data":"6c0e2131fa38031bdb808a2299181b6946b5286ec4e9cac0dee68f887c359d9e"} Dec 10 22:54:53 crc kubenswrapper[4791]: I1210 22:54:53.582783 4791 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-operators-f8n6m" podStartSLOduration=2.209260045 podStartE2EDuration="6.582764741s" podCreationTimestamp="2025-12-10 22:54:47 +0000 UTC" firstStartedPulling="2025-12-10 22:54:48.465679032 +0000 UTC m=+322.895296645" lastFinishedPulling="2025-12-10 22:54:52.839183728 +0000 UTC m=+327.268801341" observedRunningTime="2025-12-10 22:54:53.576393508 +0000 UTC m=+328.006011121" watchObservedRunningTime="2025-12-10 22:54:53.582764741 +0000 UTC m=+328.012382354" Dec 10 22:54:54 crc kubenswrapper[4791]: I1210 22:54:54.576925 4791 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-x582q" event={"ID":"d02b36c2-1b51-4466-af31-2e69357acc9c","Type":"ContainerStarted","Data":"a26e090813a0adcd5973879c8c04cabe77785a90d3f23c76241441f36858d804"} Dec 10 22:54:54 crc kubenswrapper[4791]: I1210 22:54:54.579631 4791 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-ttrzg" event={"ID":"f00159a6-00e0-4975-87a0-9d8b2b67b157","Type":"ContainerStarted","Data":"a510617cedd04f0ed52a6e86f460ffc4061bdd0b9f5d38be528eaebf33162a4e"} Dec 10 22:54:54 crc kubenswrapper[4791]: I1210 22:54:54.617507 4791 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/community-operators-x582q" podStartSLOduration=3.131866124 podStartE2EDuration="5.617490568s" podCreationTimestamp="2025-12-10 22:54:49 +0000 UTC" firstStartedPulling="2025-12-10 22:54:51.531362043 +0000 UTC m=+325.960979656" lastFinishedPulling="2025-12-10 22:54:54.016986487 +0000 UTC m=+328.446604100" observedRunningTime="2025-12-10 22:54:54.615313312 +0000 UTC m=+329.044930925" watchObservedRunningTime="2025-12-10 22:54:54.617490568 +0000 UTC m=+329.047108181" Dec 10 22:54:55 crc kubenswrapper[4791]: I1210 22:54:55.038464 4791 patch_prober.go:28] interesting pod/machine-config-daemon-5rb5l container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 10 22:54:55 crc kubenswrapper[4791]: I1210 22:54:55.038529 4791 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-5rb5l" podUID="ba35653c-6e06-4cee-a4d6-137764090d18" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 10 22:54:57 crc kubenswrapper[4791]: I1210 22:54:57.532420 4791 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-marketplace-fcnzr" Dec 10 22:54:57 crc kubenswrapper[4791]: I1210 22:54:57.532470 4791 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-marketplace-fcnzr" Dec 10 22:54:57 crc kubenswrapper[4791]: I1210 22:54:57.585847 4791 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-marketplace-fcnzr" Dec 10 22:54:57 crc kubenswrapper[4791]: I1210 22:54:57.604684 4791 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/certified-operators-ttrzg" podStartSLOduration=6.06996176 podStartE2EDuration="8.604664356s" podCreationTimestamp="2025-12-10 22:54:49 +0000 UTC" firstStartedPulling="2025-12-10 22:54:51.529728595 +0000 UTC m=+325.959346208" lastFinishedPulling="2025-12-10 22:54:54.064431181 +0000 UTC m=+328.494048804" observedRunningTime="2025-12-10 22:54:54.6499868 +0000 UTC m=+329.079604413" watchObservedRunningTime="2025-12-10 22:54:57.604664356 +0000 UTC m=+332.034281969" Dec 10 22:54:57 crc kubenswrapper[4791]: I1210 22:54:57.631762 4791 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-marketplace-fcnzr" Dec 10 22:54:57 crc kubenswrapper[4791]: I1210 22:54:57.722857 4791 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-operators-f8n6m" Dec 10 22:54:57 crc kubenswrapper[4791]: I1210 22:54:57.722912 4791 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-operators-f8n6m" Dec 10 22:54:58 crc kubenswrapper[4791]: I1210 22:54:58.771896 4791 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-marketplace/redhat-operators-f8n6m" podUID="a68872ea-89d3-42a0-b2da-877e2d37218b" containerName="registry-server" probeResult="failure" output=< Dec 10 22:54:58 crc kubenswrapper[4791]: timeout: failed to connect service ":50051" within 1s Dec 10 22:54:58 crc kubenswrapper[4791]: > Dec 10 22:55:00 crc kubenswrapper[4791]: I1210 22:55:00.342735 4791 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/certified-operators-ttrzg" Dec 10 22:55:00 crc kubenswrapper[4791]: I1210 22:55:00.343057 4791 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/certified-operators-ttrzg" Dec 10 22:55:00 crc kubenswrapper[4791]: I1210 22:55:00.379355 4791 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/certified-operators-ttrzg" Dec 10 22:55:00 crc kubenswrapper[4791]: I1210 22:55:00.394785 4791 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/community-operators-x582q" Dec 10 22:55:00 crc kubenswrapper[4791]: I1210 22:55:00.394835 4791 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/community-operators-x582q" Dec 10 22:55:00 crc kubenswrapper[4791]: I1210 22:55:00.436059 4791 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/community-operators-x582q" Dec 10 22:55:00 crc kubenswrapper[4791]: I1210 22:55:00.638473 4791 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/community-operators-x582q" Dec 10 22:55:00 crc kubenswrapper[4791]: I1210 22:55:00.663101 4791 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/certified-operators-ttrzg" Dec 10 22:55:07 crc kubenswrapper[4791]: I1210 22:55:07.780558 4791 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-operators-f8n6m" Dec 10 22:55:07 crc kubenswrapper[4791]: I1210 22:55:07.824581 4791 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-operators-f8n6m" Dec 10 22:55:20 crc kubenswrapper[4791]: I1210 22:55:20.014058 4791 pod_container_manager_linux.go:210] "Failed to delete cgroup paths" cgroupName=["kubepods","burstable","podfca6ee01-4d09-4b77-89c3-281a84c4f3a6"] err="unable to destroy cgroup paths for cgroup [kubepods burstable podfca6ee01-4d09-4b77-89c3-281a84c4f3a6] : Timed out while waiting for systemd to remove kubepods-burstable-podfca6ee01_4d09_4b77_89c3_281a84c4f3a6.slice" Dec 10 22:55:20 crc kubenswrapper[4791]: E1210 22:55:20.014581 4791 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to delete cgroup paths for [kubepods burstable podfca6ee01-4d09-4b77-89c3-281a84c4f3a6] : unable to destroy cgroup paths for cgroup [kubepods burstable podfca6ee01-4d09-4b77-89c3-281a84c4f3a6] : Timed out while waiting for systemd to remove kubepods-burstable-podfca6ee01_4d09_4b77_89c3_281a84c4f3a6.slice" pod="openshift-route-controller-manager/route-controller-manager-66c8f7db4d-6582d" podUID="fca6ee01-4d09-4b77-89c3-281a84c4f3a6" Dec 10 22:55:20 crc kubenswrapper[4791]: I1210 22:55:20.730836 4791 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-route-controller-manager/route-controller-manager-66c8f7db4d-6582d" Dec 10 22:55:20 crc kubenswrapper[4791]: I1210 22:55:20.747861 4791 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-66c8f7db4d-6582d"] Dec 10 22:55:20 crc kubenswrapper[4791]: I1210 22:55:20.750952 4791 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-66c8f7db4d-6582d"] Dec 10 22:55:21 crc kubenswrapper[4791]: I1210 22:55:21.891259 4791 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="fca6ee01-4d09-4b77-89c3-281a84c4f3a6" path="/var/lib/kubelet/pods/fca6ee01-4d09-4b77-89c3-281a84c4f3a6/volumes" Dec 10 22:55:25 crc kubenswrapper[4791]: I1210 22:55:25.038328 4791 patch_prober.go:28] interesting pod/machine-config-daemon-5rb5l container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 10 22:55:25 crc kubenswrapper[4791]: I1210 22:55:25.038920 4791 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-5rb5l" podUID="ba35653c-6e06-4cee-a4d6-137764090d18" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 10 22:55:34 crc kubenswrapper[4791]: I1210 22:55:34.275178 4791 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-image-registry/image-registry-66df7c8f76-r6pgs"] Dec 10 22:55:34 crc kubenswrapper[4791]: I1210 22:55:34.276179 4791 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-image-registry/image-registry-66df7c8f76-r6pgs" Dec 10 22:55:34 crc kubenswrapper[4791]: I1210 22:55:34.290994 4791 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-image-registry/image-registry-66df7c8f76-r6pgs"] Dec 10 22:55:34 crc kubenswrapper[4791]: I1210 22:55:34.471353 4791 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-66df7c8f76-r6pgs\" (UID: \"aa5628f5-c7ea-4793-acca-dd3923bffe6d\") " pod="openshift-image-registry/image-registry-66df7c8f76-r6pgs" Dec 10 22:55:34 crc kubenswrapper[4791]: I1210 22:55:34.471412 4791 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/aa5628f5-c7ea-4793-acca-dd3923bffe6d-trusted-ca\") pod \"image-registry-66df7c8f76-r6pgs\" (UID: \"aa5628f5-c7ea-4793-acca-dd3923bffe6d\") " pod="openshift-image-registry/image-registry-66df7c8f76-r6pgs" Dec 10 22:55:34 crc kubenswrapper[4791]: I1210 22:55:34.471463 4791 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"registry-certificates\" (UniqueName: \"kubernetes.io/configmap/aa5628f5-c7ea-4793-acca-dd3923bffe6d-registry-certificates\") pod \"image-registry-66df7c8f76-r6pgs\" (UID: \"aa5628f5-c7ea-4793-acca-dd3923bffe6d\") " pod="openshift-image-registry/image-registry-66df7c8f76-r6pgs" Dec 10 22:55:34 crc kubenswrapper[4791]: I1210 22:55:34.471494 4791 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"installation-pull-secrets\" (UniqueName: \"kubernetes.io/secret/aa5628f5-c7ea-4793-acca-dd3923bffe6d-installation-pull-secrets\") pod \"image-registry-66df7c8f76-r6pgs\" (UID: \"aa5628f5-c7ea-4793-acca-dd3923bffe6d\") " pod="openshift-image-registry/image-registry-66df7c8f76-r6pgs" Dec 10 22:55:34 crc kubenswrapper[4791]: I1210 22:55:34.471644 4791 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-g9rxs\" (UniqueName: \"kubernetes.io/projected/aa5628f5-c7ea-4793-acca-dd3923bffe6d-kube-api-access-g9rxs\") pod \"image-registry-66df7c8f76-r6pgs\" (UID: \"aa5628f5-c7ea-4793-acca-dd3923bffe6d\") " pod="openshift-image-registry/image-registry-66df7c8f76-r6pgs" Dec 10 22:55:34 crc kubenswrapper[4791]: I1210 22:55:34.471689 4791 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/aa5628f5-c7ea-4793-acca-dd3923bffe6d-bound-sa-token\") pod \"image-registry-66df7c8f76-r6pgs\" (UID: \"aa5628f5-c7ea-4793-acca-dd3923bffe6d\") " pod="openshift-image-registry/image-registry-66df7c8f76-r6pgs" Dec 10 22:55:34 crc kubenswrapper[4791]: I1210 22:55:34.471724 4791 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"registry-tls\" (UniqueName: \"kubernetes.io/projected/aa5628f5-c7ea-4793-acca-dd3923bffe6d-registry-tls\") pod \"image-registry-66df7c8f76-r6pgs\" (UID: \"aa5628f5-c7ea-4793-acca-dd3923bffe6d\") " pod="openshift-image-registry/image-registry-66df7c8f76-r6pgs" Dec 10 22:55:34 crc kubenswrapper[4791]: I1210 22:55:34.471761 4791 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ca-trust-extracted\" (UniqueName: \"kubernetes.io/empty-dir/aa5628f5-c7ea-4793-acca-dd3923bffe6d-ca-trust-extracted\") pod \"image-registry-66df7c8f76-r6pgs\" (UID: \"aa5628f5-c7ea-4793-acca-dd3923bffe6d\") " pod="openshift-image-registry/image-registry-66df7c8f76-r6pgs" Dec 10 22:55:34 crc kubenswrapper[4791]: I1210 22:55:34.489598 4791 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-66df7c8f76-r6pgs\" (UID: \"aa5628f5-c7ea-4793-acca-dd3923bffe6d\") " pod="openshift-image-registry/image-registry-66df7c8f76-r6pgs" Dec 10 22:55:34 crc kubenswrapper[4791]: I1210 22:55:34.573056 4791 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/aa5628f5-c7ea-4793-acca-dd3923bffe6d-trusted-ca\") pod \"image-registry-66df7c8f76-r6pgs\" (UID: \"aa5628f5-c7ea-4793-acca-dd3923bffe6d\") " pod="openshift-image-registry/image-registry-66df7c8f76-r6pgs" Dec 10 22:55:34 crc kubenswrapper[4791]: I1210 22:55:34.573115 4791 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"registry-certificates\" (UniqueName: \"kubernetes.io/configmap/aa5628f5-c7ea-4793-acca-dd3923bffe6d-registry-certificates\") pod \"image-registry-66df7c8f76-r6pgs\" (UID: \"aa5628f5-c7ea-4793-acca-dd3923bffe6d\") " pod="openshift-image-registry/image-registry-66df7c8f76-r6pgs" Dec 10 22:55:34 crc kubenswrapper[4791]: I1210 22:55:34.573137 4791 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"installation-pull-secrets\" (UniqueName: \"kubernetes.io/secret/aa5628f5-c7ea-4793-acca-dd3923bffe6d-installation-pull-secrets\") pod \"image-registry-66df7c8f76-r6pgs\" (UID: \"aa5628f5-c7ea-4793-acca-dd3923bffe6d\") " pod="openshift-image-registry/image-registry-66df7c8f76-r6pgs" Dec 10 22:55:34 crc kubenswrapper[4791]: I1210 22:55:34.573175 4791 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-g9rxs\" (UniqueName: \"kubernetes.io/projected/aa5628f5-c7ea-4793-acca-dd3923bffe6d-kube-api-access-g9rxs\") pod \"image-registry-66df7c8f76-r6pgs\" (UID: \"aa5628f5-c7ea-4793-acca-dd3923bffe6d\") " pod="openshift-image-registry/image-registry-66df7c8f76-r6pgs" Dec 10 22:55:34 crc kubenswrapper[4791]: I1210 22:55:34.573207 4791 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/aa5628f5-c7ea-4793-acca-dd3923bffe6d-bound-sa-token\") pod \"image-registry-66df7c8f76-r6pgs\" (UID: \"aa5628f5-c7ea-4793-acca-dd3923bffe6d\") " pod="openshift-image-registry/image-registry-66df7c8f76-r6pgs" Dec 10 22:55:34 crc kubenswrapper[4791]: I1210 22:55:34.573242 4791 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"registry-tls\" (UniqueName: \"kubernetes.io/projected/aa5628f5-c7ea-4793-acca-dd3923bffe6d-registry-tls\") pod \"image-registry-66df7c8f76-r6pgs\" (UID: \"aa5628f5-c7ea-4793-acca-dd3923bffe6d\") " pod="openshift-image-registry/image-registry-66df7c8f76-r6pgs" Dec 10 22:55:34 crc kubenswrapper[4791]: I1210 22:55:34.573264 4791 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ca-trust-extracted\" (UniqueName: \"kubernetes.io/empty-dir/aa5628f5-c7ea-4793-acca-dd3923bffe6d-ca-trust-extracted\") pod \"image-registry-66df7c8f76-r6pgs\" (UID: \"aa5628f5-c7ea-4793-acca-dd3923bffe6d\") " pod="openshift-image-registry/image-registry-66df7c8f76-r6pgs" Dec 10 22:55:34 crc kubenswrapper[4791]: I1210 22:55:34.574049 4791 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ca-trust-extracted\" (UniqueName: \"kubernetes.io/empty-dir/aa5628f5-c7ea-4793-acca-dd3923bffe6d-ca-trust-extracted\") pod \"image-registry-66df7c8f76-r6pgs\" (UID: \"aa5628f5-c7ea-4793-acca-dd3923bffe6d\") " pod="openshift-image-registry/image-registry-66df7c8f76-r6pgs" Dec 10 22:55:34 crc kubenswrapper[4791]: I1210 22:55:34.574587 4791 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/aa5628f5-c7ea-4793-acca-dd3923bffe6d-trusted-ca\") pod \"image-registry-66df7c8f76-r6pgs\" (UID: \"aa5628f5-c7ea-4793-acca-dd3923bffe6d\") " pod="openshift-image-registry/image-registry-66df7c8f76-r6pgs" Dec 10 22:55:34 crc kubenswrapper[4791]: I1210 22:55:34.574611 4791 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"registry-certificates\" (UniqueName: \"kubernetes.io/configmap/aa5628f5-c7ea-4793-acca-dd3923bffe6d-registry-certificates\") pod \"image-registry-66df7c8f76-r6pgs\" (UID: \"aa5628f5-c7ea-4793-acca-dd3923bffe6d\") " pod="openshift-image-registry/image-registry-66df7c8f76-r6pgs" Dec 10 22:55:34 crc kubenswrapper[4791]: I1210 22:55:34.579751 4791 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"installation-pull-secrets\" (UniqueName: \"kubernetes.io/secret/aa5628f5-c7ea-4793-acca-dd3923bffe6d-installation-pull-secrets\") pod \"image-registry-66df7c8f76-r6pgs\" (UID: \"aa5628f5-c7ea-4793-acca-dd3923bffe6d\") " pod="openshift-image-registry/image-registry-66df7c8f76-r6pgs" Dec 10 22:55:34 crc kubenswrapper[4791]: I1210 22:55:34.581542 4791 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"registry-tls\" (UniqueName: \"kubernetes.io/projected/aa5628f5-c7ea-4793-acca-dd3923bffe6d-registry-tls\") pod \"image-registry-66df7c8f76-r6pgs\" (UID: \"aa5628f5-c7ea-4793-acca-dd3923bffe6d\") " pod="openshift-image-registry/image-registry-66df7c8f76-r6pgs" Dec 10 22:55:34 crc kubenswrapper[4791]: I1210 22:55:34.588721 4791 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/aa5628f5-c7ea-4793-acca-dd3923bffe6d-bound-sa-token\") pod \"image-registry-66df7c8f76-r6pgs\" (UID: \"aa5628f5-c7ea-4793-acca-dd3923bffe6d\") " pod="openshift-image-registry/image-registry-66df7c8f76-r6pgs" Dec 10 22:55:34 crc kubenswrapper[4791]: I1210 22:55:34.589437 4791 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-g9rxs\" (UniqueName: \"kubernetes.io/projected/aa5628f5-c7ea-4793-acca-dd3923bffe6d-kube-api-access-g9rxs\") pod \"image-registry-66df7c8f76-r6pgs\" (UID: \"aa5628f5-c7ea-4793-acca-dd3923bffe6d\") " pod="openshift-image-registry/image-registry-66df7c8f76-r6pgs" Dec 10 22:55:34 crc kubenswrapper[4791]: I1210 22:55:34.598695 4791 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-image-registry/image-registry-66df7c8f76-r6pgs" Dec 10 22:55:34 crc kubenswrapper[4791]: I1210 22:55:34.983958 4791 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-image-registry/image-registry-66df7c8f76-r6pgs"] Dec 10 22:55:35 crc kubenswrapper[4791]: I1210 22:55:35.808492 4791 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-image-registry/image-registry-66df7c8f76-r6pgs" event={"ID":"aa5628f5-c7ea-4793-acca-dd3923bffe6d","Type":"ContainerStarted","Data":"9d43edac567dbbd1631edbe2dcf092417343beedae940b8d8bcb4254a81f684d"} Dec 10 22:55:35 crc kubenswrapper[4791]: I1210 22:55:35.808550 4791 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-image-registry/image-registry-66df7c8f76-r6pgs" event={"ID":"aa5628f5-c7ea-4793-acca-dd3923bffe6d","Type":"ContainerStarted","Data":"76138a7898805d1ecb0a245f79510a9fd1e3ecf0551e35132a5dd5ae761a1f7a"} Dec 10 22:55:35 crc kubenswrapper[4791]: I1210 22:55:35.808811 4791 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-image-registry/image-registry-66df7c8f76-r6pgs" Dec 10 22:55:35 crc kubenswrapper[4791]: I1210 22:55:35.827680 4791 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-image-registry/image-registry-66df7c8f76-r6pgs" podStartSLOduration=1.827664124 podStartE2EDuration="1.827664124s" podCreationTimestamp="2025-12-10 22:55:34 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-10 22:55:35.824284082 +0000 UTC m=+370.253901695" watchObservedRunningTime="2025-12-10 22:55:35.827664124 +0000 UTC m=+370.257281727" Dec 10 22:55:54 crc kubenswrapper[4791]: I1210 22:55:54.607976 4791 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-image-registry/image-registry-66df7c8f76-r6pgs" Dec 10 22:55:54 crc kubenswrapper[4791]: I1210 22:55:54.661553 4791 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-image-registry/image-registry-697d97f7c8-srd52"] Dec 10 22:55:55 crc kubenswrapper[4791]: I1210 22:55:55.038834 4791 patch_prober.go:28] interesting pod/machine-config-daemon-5rb5l container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 10 22:55:55 crc kubenswrapper[4791]: I1210 22:55:55.038920 4791 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-5rb5l" podUID="ba35653c-6e06-4cee-a4d6-137764090d18" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 10 22:55:55 crc kubenswrapper[4791]: I1210 22:55:55.038980 4791 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-5rb5l" Dec 10 22:55:55 crc kubenswrapper[4791]: I1210 22:55:55.039923 4791 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"05fce03d9a939befc9dac71cbc9750ba888a8f61825ef9c21bf847dd73aa10dc"} pod="openshift-machine-config-operator/machine-config-daemon-5rb5l" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Dec 10 22:55:55 crc kubenswrapper[4791]: I1210 22:55:55.040037 4791 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-5rb5l" podUID="ba35653c-6e06-4cee-a4d6-137764090d18" containerName="machine-config-daemon" containerID="cri-o://05fce03d9a939befc9dac71cbc9750ba888a8f61825ef9c21bf847dd73aa10dc" gracePeriod=600 Dec 10 22:55:55 crc kubenswrapper[4791]: I1210 22:55:55.916430 4791 generic.go:334] "Generic (PLEG): container finished" podID="ba35653c-6e06-4cee-a4d6-137764090d18" containerID="05fce03d9a939befc9dac71cbc9750ba888a8f61825ef9c21bf847dd73aa10dc" exitCode=0 Dec 10 22:55:55 crc kubenswrapper[4791]: I1210 22:55:55.916527 4791 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-5rb5l" event={"ID":"ba35653c-6e06-4cee-a4d6-137764090d18","Type":"ContainerDied","Data":"05fce03d9a939befc9dac71cbc9750ba888a8f61825ef9c21bf847dd73aa10dc"} Dec 10 22:55:55 crc kubenswrapper[4791]: I1210 22:55:55.916761 4791 scope.go:117] "RemoveContainer" containerID="185eb2ff44221a514878283ec873773a6c9a9073dc64e296f982e60dee1338e4" Dec 10 22:55:56 crc kubenswrapper[4791]: I1210 22:55:56.925550 4791 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-5rb5l" event={"ID":"ba35653c-6e06-4cee-a4d6-137764090d18","Type":"ContainerStarted","Data":"c09d66dfdd1ff68cb29fe7936cd36ccb2e44990bd887b93547e1487b370efa4e"} Dec 10 22:56:02 crc kubenswrapper[4791]: I1210 22:56:02.576823 4791 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-controller-manager/controller-manager-5747cbd54d-9mwk9"] Dec 10 22:56:02 crc kubenswrapper[4791]: I1210 22:56:02.577674 4791 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-controller-manager/controller-manager-5747cbd54d-9mwk9" podUID="b9b00f47-2658-4115-8c86-5d57ad0632df" containerName="controller-manager" containerID="cri-o://018136a3268c789d9ddd9dd74003b3e01e9f3efae97293726a35d6582353bb26" gracePeriod=30 Dec 10 22:56:02 crc kubenswrapper[4791]: I1210 22:56:02.970450 4791 generic.go:334] "Generic (PLEG): container finished" podID="b9b00f47-2658-4115-8c86-5d57ad0632df" containerID="018136a3268c789d9ddd9dd74003b3e01e9f3efae97293726a35d6582353bb26" exitCode=0 Dec 10 22:56:02 crc kubenswrapper[4791]: I1210 22:56:02.970537 4791 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager/controller-manager-5747cbd54d-9mwk9" event={"ID":"b9b00f47-2658-4115-8c86-5d57ad0632df","Type":"ContainerDied","Data":"018136a3268c789d9ddd9dd74003b3e01e9f3efae97293726a35d6582353bb26"} Dec 10 22:56:02 crc kubenswrapper[4791]: I1210 22:56:02.970751 4791 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager/controller-manager-5747cbd54d-9mwk9" event={"ID":"b9b00f47-2658-4115-8c86-5d57ad0632df","Type":"ContainerDied","Data":"c6df9b1d71ae0fe7a2ae82393e128670b8b6b046d598d7e61bc305e885a1df04"} Dec 10 22:56:02 crc kubenswrapper[4791]: I1210 22:56:02.970766 4791 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="c6df9b1d71ae0fe7a2ae82393e128670b8b6b046d598d7e61bc305e885a1df04" Dec 10 22:56:02 crc kubenswrapper[4791]: I1210 22:56:02.973790 4791 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager/controller-manager-5747cbd54d-9mwk9" Dec 10 22:56:03 crc kubenswrapper[4791]: I1210 22:56:03.097933 4791 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/b9b00f47-2658-4115-8c86-5d57ad0632df-proxy-ca-bundles\") pod \"b9b00f47-2658-4115-8c86-5d57ad0632df\" (UID: \"b9b00f47-2658-4115-8c86-5d57ad0632df\") " Dec 10 22:56:03 crc kubenswrapper[4791]: I1210 22:56:03.098009 4791 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-2stvb\" (UniqueName: \"kubernetes.io/projected/b9b00f47-2658-4115-8c86-5d57ad0632df-kube-api-access-2stvb\") pod \"b9b00f47-2658-4115-8c86-5d57ad0632df\" (UID: \"b9b00f47-2658-4115-8c86-5d57ad0632df\") " Dec 10 22:56:03 crc kubenswrapper[4791]: I1210 22:56:03.098053 4791 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/b9b00f47-2658-4115-8c86-5d57ad0632df-client-ca\") pod \"b9b00f47-2658-4115-8c86-5d57ad0632df\" (UID: \"b9b00f47-2658-4115-8c86-5d57ad0632df\") " Dec 10 22:56:03 crc kubenswrapper[4791]: I1210 22:56:03.098088 4791 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/b9b00f47-2658-4115-8c86-5d57ad0632df-serving-cert\") pod \"b9b00f47-2658-4115-8c86-5d57ad0632df\" (UID: \"b9b00f47-2658-4115-8c86-5d57ad0632df\") " Dec 10 22:56:03 crc kubenswrapper[4791]: I1210 22:56:03.098145 4791 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/b9b00f47-2658-4115-8c86-5d57ad0632df-config\") pod \"b9b00f47-2658-4115-8c86-5d57ad0632df\" (UID: \"b9b00f47-2658-4115-8c86-5d57ad0632df\") " Dec 10 22:56:03 crc kubenswrapper[4791]: I1210 22:56:03.099604 4791 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/b9b00f47-2658-4115-8c86-5d57ad0632df-config" (OuterVolumeSpecName: "config") pod "b9b00f47-2658-4115-8c86-5d57ad0632df" (UID: "b9b00f47-2658-4115-8c86-5d57ad0632df"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 10 22:56:03 crc kubenswrapper[4791]: I1210 22:56:03.099982 4791 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/b9b00f47-2658-4115-8c86-5d57ad0632df-client-ca" (OuterVolumeSpecName: "client-ca") pod "b9b00f47-2658-4115-8c86-5d57ad0632df" (UID: "b9b00f47-2658-4115-8c86-5d57ad0632df"). InnerVolumeSpecName "client-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 10 22:56:03 crc kubenswrapper[4791]: I1210 22:56:03.100425 4791 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/b9b00f47-2658-4115-8c86-5d57ad0632df-proxy-ca-bundles" (OuterVolumeSpecName: "proxy-ca-bundles") pod "b9b00f47-2658-4115-8c86-5d57ad0632df" (UID: "b9b00f47-2658-4115-8c86-5d57ad0632df"). InnerVolumeSpecName "proxy-ca-bundles". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 10 22:56:03 crc kubenswrapper[4791]: I1210 22:56:03.103575 4791 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/b9b00f47-2658-4115-8c86-5d57ad0632df-kube-api-access-2stvb" (OuterVolumeSpecName: "kube-api-access-2stvb") pod "b9b00f47-2658-4115-8c86-5d57ad0632df" (UID: "b9b00f47-2658-4115-8c86-5d57ad0632df"). InnerVolumeSpecName "kube-api-access-2stvb". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 10 22:56:03 crc kubenswrapper[4791]: I1210 22:56:03.105903 4791 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/b9b00f47-2658-4115-8c86-5d57ad0632df-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "b9b00f47-2658-4115-8c86-5d57ad0632df" (UID: "b9b00f47-2658-4115-8c86-5d57ad0632df"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 10 22:56:03 crc kubenswrapper[4791]: I1210 22:56:03.200025 4791 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/b9b00f47-2658-4115-8c86-5d57ad0632df-config\") on node \"crc\" DevicePath \"\"" Dec 10 22:56:03 crc kubenswrapper[4791]: I1210 22:56:03.200076 4791 reconciler_common.go:293] "Volume detached for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/b9b00f47-2658-4115-8c86-5d57ad0632df-proxy-ca-bundles\") on node \"crc\" DevicePath \"\"" Dec 10 22:56:03 crc kubenswrapper[4791]: I1210 22:56:03.200090 4791 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-2stvb\" (UniqueName: \"kubernetes.io/projected/b9b00f47-2658-4115-8c86-5d57ad0632df-kube-api-access-2stvb\") on node \"crc\" DevicePath \"\"" Dec 10 22:56:03 crc kubenswrapper[4791]: I1210 22:56:03.200101 4791 reconciler_common.go:293] "Volume detached for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/b9b00f47-2658-4115-8c86-5d57ad0632df-client-ca\") on node \"crc\" DevicePath \"\"" Dec 10 22:56:03 crc kubenswrapper[4791]: I1210 22:56:03.200113 4791 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/b9b00f47-2658-4115-8c86-5d57ad0632df-serving-cert\") on node \"crc\" DevicePath \"\"" Dec 10 22:56:03 crc kubenswrapper[4791]: I1210 22:56:03.975069 4791 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager/controller-manager-5747cbd54d-9mwk9" Dec 10 22:56:03 crc kubenswrapper[4791]: I1210 22:56:03.990596 4791 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-controller-manager/controller-manager-5747cbd54d-9mwk9"] Dec 10 22:56:03 crc kubenswrapper[4791]: I1210 22:56:03.996495 4791 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-controller-manager/controller-manager-5747cbd54d-9mwk9"] Dec 10 22:56:04 crc kubenswrapper[4791]: I1210 22:56:04.265323 4791 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-controller-manager/controller-manager-649fb98567-9gx2g"] Dec 10 22:56:04 crc kubenswrapper[4791]: E1210 22:56:04.265865 4791 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="b9b00f47-2658-4115-8c86-5d57ad0632df" containerName="controller-manager" Dec 10 22:56:04 crc kubenswrapper[4791]: I1210 22:56:04.265951 4791 state_mem.go:107] "Deleted CPUSet assignment" podUID="b9b00f47-2658-4115-8c86-5d57ad0632df" containerName="controller-manager" Dec 10 22:56:04 crc kubenswrapper[4791]: I1210 22:56:04.266128 4791 memory_manager.go:354] "RemoveStaleState removing state" podUID="b9b00f47-2658-4115-8c86-5d57ad0632df" containerName="controller-manager" Dec 10 22:56:04 crc kubenswrapper[4791]: I1210 22:56:04.266582 4791 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager/controller-manager-649fb98567-9gx2g" Dec 10 22:56:04 crc kubenswrapper[4791]: I1210 22:56:04.270185 4791 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"client-ca" Dec 10 22:56:04 crc kubenswrapper[4791]: I1210 22:56:04.270197 4791 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"config" Dec 10 22:56:04 crc kubenswrapper[4791]: I1210 22:56:04.270778 4791 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"openshift-service-ca.crt" Dec 10 22:56:04 crc kubenswrapper[4791]: I1210 22:56:04.271675 4791 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"kube-root-ca.crt" Dec 10 22:56:04 crc kubenswrapper[4791]: I1210 22:56:04.271723 4791 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-controller-manager"/"openshift-controller-manager-sa-dockercfg-msq4c" Dec 10 22:56:04 crc kubenswrapper[4791]: I1210 22:56:04.271949 4791 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-controller-manager"/"serving-cert" Dec 10 22:56:04 crc kubenswrapper[4791]: I1210 22:56:04.281283 4791 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-controller-manager/controller-manager-649fb98567-9gx2g"] Dec 10 22:56:04 crc kubenswrapper[4791]: I1210 22:56:04.283888 4791 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"openshift-global-ca" Dec 10 22:56:04 crc kubenswrapper[4791]: I1210 22:56:04.417704 4791 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/825e8ad7-31bb-437f-a907-e14f51eaa477-serving-cert\") pod \"controller-manager-649fb98567-9gx2g\" (UID: \"825e8ad7-31bb-437f-a907-e14f51eaa477\") " pod="openshift-controller-manager/controller-manager-649fb98567-9gx2g" Dec 10 22:56:04 crc kubenswrapper[4791]: I1210 22:56:04.417767 4791 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/825e8ad7-31bb-437f-a907-e14f51eaa477-client-ca\") pod \"controller-manager-649fb98567-9gx2g\" (UID: \"825e8ad7-31bb-437f-a907-e14f51eaa477\") " pod="openshift-controller-manager/controller-manager-649fb98567-9gx2g" Dec 10 22:56:04 crc kubenswrapper[4791]: I1210 22:56:04.417871 4791 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/825e8ad7-31bb-437f-a907-e14f51eaa477-proxy-ca-bundles\") pod \"controller-manager-649fb98567-9gx2g\" (UID: \"825e8ad7-31bb-437f-a907-e14f51eaa477\") " pod="openshift-controller-manager/controller-manager-649fb98567-9gx2g" Dec 10 22:56:04 crc kubenswrapper[4791]: I1210 22:56:04.417898 4791 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/825e8ad7-31bb-437f-a907-e14f51eaa477-config\") pod \"controller-manager-649fb98567-9gx2g\" (UID: \"825e8ad7-31bb-437f-a907-e14f51eaa477\") " pod="openshift-controller-manager/controller-manager-649fb98567-9gx2g" Dec 10 22:56:04 crc kubenswrapper[4791]: I1210 22:56:04.417931 4791 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-sv5zg\" (UniqueName: \"kubernetes.io/projected/825e8ad7-31bb-437f-a907-e14f51eaa477-kube-api-access-sv5zg\") pod \"controller-manager-649fb98567-9gx2g\" (UID: \"825e8ad7-31bb-437f-a907-e14f51eaa477\") " pod="openshift-controller-manager/controller-manager-649fb98567-9gx2g" Dec 10 22:56:04 crc kubenswrapper[4791]: I1210 22:56:04.519682 4791 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/825e8ad7-31bb-437f-a907-e14f51eaa477-serving-cert\") pod \"controller-manager-649fb98567-9gx2g\" (UID: \"825e8ad7-31bb-437f-a907-e14f51eaa477\") " pod="openshift-controller-manager/controller-manager-649fb98567-9gx2g" Dec 10 22:56:04 crc kubenswrapper[4791]: I1210 22:56:04.519805 4791 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/825e8ad7-31bb-437f-a907-e14f51eaa477-client-ca\") pod \"controller-manager-649fb98567-9gx2g\" (UID: \"825e8ad7-31bb-437f-a907-e14f51eaa477\") " pod="openshift-controller-manager/controller-manager-649fb98567-9gx2g" Dec 10 22:56:04 crc kubenswrapper[4791]: I1210 22:56:04.519906 4791 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/825e8ad7-31bb-437f-a907-e14f51eaa477-proxy-ca-bundles\") pod \"controller-manager-649fb98567-9gx2g\" (UID: \"825e8ad7-31bb-437f-a907-e14f51eaa477\") " pod="openshift-controller-manager/controller-manager-649fb98567-9gx2g" Dec 10 22:56:04 crc kubenswrapper[4791]: I1210 22:56:04.519948 4791 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/825e8ad7-31bb-437f-a907-e14f51eaa477-config\") pod \"controller-manager-649fb98567-9gx2g\" (UID: \"825e8ad7-31bb-437f-a907-e14f51eaa477\") " pod="openshift-controller-manager/controller-manager-649fb98567-9gx2g" Dec 10 22:56:04 crc kubenswrapper[4791]: I1210 22:56:04.519993 4791 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-sv5zg\" (UniqueName: \"kubernetes.io/projected/825e8ad7-31bb-437f-a907-e14f51eaa477-kube-api-access-sv5zg\") pod \"controller-manager-649fb98567-9gx2g\" (UID: \"825e8ad7-31bb-437f-a907-e14f51eaa477\") " pod="openshift-controller-manager/controller-manager-649fb98567-9gx2g" Dec 10 22:56:04 crc kubenswrapper[4791]: I1210 22:56:04.521806 4791 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/825e8ad7-31bb-437f-a907-e14f51eaa477-proxy-ca-bundles\") pod \"controller-manager-649fb98567-9gx2g\" (UID: \"825e8ad7-31bb-437f-a907-e14f51eaa477\") " pod="openshift-controller-manager/controller-manager-649fb98567-9gx2g" Dec 10 22:56:04 crc kubenswrapper[4791]: I1210 22:56:04.521943 4791 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/825e8ad7-31bb-437f-a907-e14f51eaa477-config\") pod \"controller-manager-649fb98567-9gx2g\" (UID: \"825e8ad7-31bb-437f-a907-e14f51eaa477\") " pod="openshift-controller-manager/controller-manager-649fb98567-9gx2g" Dec 10 22:56:04 crc kubenswrapper[4791]: I1210 22:56:04.522198 4791 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/825e8ad7-31bb-437f-a907-e14f51eaa477-client-ca\") pod \"controller-manager-649fb98567-9gx2g\" (UID: \"825e8ad7-31bb-437f-a907-e14f51eaa477\") " pod="openshift-controller-manager/controller-manager-649fb98567-9gx2g" Dec 10 22:56:04 crc kubenswrapper[4791]: I1210 22:56:04.527998 4791 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/825e8ad7-31bb-437f-a907-e14f51eaa477-serving-cert\") pod \"controller-manager-649fb98567-9gx2g\" (UID: \"825e8ad7-31bb-437f-a907-e14f51eaa477\") " pod="openshift-controller-manager/controller-manager-649fb98567-9gx2g" Dec 10 22:56:04 crc kubenswrapper[4791]: I1210 22:56:04.551693 4791 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-sv5zg\" (UniqueName: \"kubernetes.io/projected/825e8ad7-31bb-437f-a907-e14f51eaa477-kube-api-access-sv5zg\") pod \"controller-manager-649fb98567-9gx2g\" (UID: \"825e8ad7-31bb-437f-a907-e14f51eaa477\") " pod="openshift-controller-manager/controller-manager-649fb98567-9gx2g" Dec 10 22:56:04 crc kubenswrapper[4791]: I1210 22:56:04.585351 4791 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager/controller-manager-649fb98567-9gx2g" Dec 10 22:56:05 crc kubenswrapper[4791]: I1210 22:56:05.005086 4791 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-controller-manager/controller-manager-649fb98567-9gx2g"] Dec 10 22:56:05 crc kubenswrapper[4791]: I1210 22:56:05.897847 4791 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="b9b00f47-2658-4115-8c86-5d57ad0632df" path="/var/lib/kubelet/pods/b9b00f47-2658-4115-8c86-5d57ad0632df/volumes" Dec 10 22:56:05 crc kubenswrapper[4791]: I1210 22:56:05.992324 4791 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager/controller-manager-649fb98567-9gx2g" event={"ID":"825e8ad7-31bb-437f-a907-e14f51eaa477","Type":"ContainerStarted","Data":"042424f555f4e5d1d67d50d5a5124878e6a84f1ce2358ab6180d96d4fbf1f106"} Dec 10 22:56:05 crc kubenswrapper[4791]: I1210 22:56:05.992403 4791 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager/controller-manager-649fb98567-9gx2g" event={"ID":"825e8ad7-31bb-437f-a907-e14f51eaa477","Type":"ContainerStarted","Data":"bd3f4e226027781ef1b1db73a8393b8adb82c5ed068339b146dfc2582f4302eb"} Dec 10 22:56:05 crc kubenswrapper[4791]: I1210 22:56:05.992687 4791 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-controller-manager/controller-manager-649fb98567-9gx2g" Dec 10 22:56:06 crc kubenswrapper[4791]: I1210 22:56:06.002308 4791 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-controller-manager/controller-manager-649fb98567-9gx2g" Dec 10 22:56:06 crc kubenswrapper[4791]: I1210 22:56:06.027950 4791 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-controller-manager/controller-manager-649fb98567-9gx2g" podStartSLOduration=4.027922963 podStartE2EDuration="4.027922963s" podCreationTimestamp="2025-12-10 22:56:02 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-10 22:56:06.026226953 +0000 UTC m=+400.455844576" watchObservedRunningTime="2025-12-10 22:56:06.027922963 +0000 UTC m=+400.457540616" Dec 10 22:56:19 crc kubenswrapper[4791]: I1210 22:56:19.706231 4791 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-image-registry/image-registry-697d97f7c8-srd52" podUID="e93dbd5d-0ca7-4a15-9eb3-0d71595fada0" containerName="registry" containerID="cri-o://e6eab22df9424a9028c3071cbd5cae655435fbbba257f037a06dfc92facc867d" gracePeriod=30 Dec 10 22:56:20 crc kubenswrapper[4791]: I1210 22:56:20.073845 4791 generic.go:334] "Generic (PLEG): container finished" podID="e93dbd5d-0ca7-4a15-9eb3-0d71595fada0" containerID="e6eab22df9424a9028c3071cbd5cae655435fbbba257f037a06dfc92facc867d" exitCode=0 Dec 10 22:56:20 crc kubenswrapper[4791]: I1210 22:56:20.073889 4791 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-image-registry/image-registry-697d97f7c8-srd52" event={"ID":"e93dbd5d-0ca7-4a15-9eb3-0d71595fada0","Type":"ContainerDied","Data":"e6eab22df9424a9028c3071cbd5cae655435fbbba257f037a06dfc92facc867d"} Dec 10 22:56:20 crc kubenswrapper[4791]: I1210 22:56:20.153501 4791 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-image-registry/image-registry-697d97f7c8-srd52" Dec 10 22:56:20 crc kubenswrapper[4791]: I1210 22:56:20.245461 4791 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ca-trust-extracted\" (UniqueName: \"kubernetes.io/empty-dir/e93dbd5d-0ca7-4a15-9eb3-0d71595fada0-ca-trust-extracted\") pod \"e93dbd5d-0ca7-4a15-9eb3-0d71595fada0\" (UID: \"e93dbd5d-0ca7-4a15-9eb3-0d71595fada0\") " Dec 10 22:56:20 crc kubenswrapper[4791]: I1210 22:56:20.245980 4791 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"registry-storage\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"e93dbd5d-0ca7-4a15-9eb3-0d71595fada0\" (UID: \"e93dbd5d-0ca7-4a15-9eb3-0d71595fada0\") " Dec 10 22:56:20 crc kubenswrapper[4791]: I1210 22:56:20.246083 4791 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"registry-certificates\" (UniqueName: \"kubernetes.io/configmap/e93dbd5d-0ca7-4a15-9eb3-0d71595fada0-registry-certificates\") pod \"e93dbd5d-0ca7-4a15-9eb3-0d71595fada0\" (UID: \"e93dbd5d-0ca7-4a15-9eb3-0d71595fada0\") " Dec 10 22:56:20 crc kubenswrapper[4791]: I1210 22:56:20.246132 4791 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"registry-tls\" (UniqueName: \"kubernetes.io/projected/e93dbd5d-0ca7-4a15-9eb3-0d71595fada0-registry-tls\") pod \"e93dbd5d-0ca7-4a15-9eb3-0d71595fada0\" (UID: \"e93dbd5d-0ca7-4a15-9eb3-0d71595fada0\") " Dec 10 22:56:20 crc kubenswrapper[4791]: I1210 22:56:20.247130 4791 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/e93dbd5d-0ca7-4a15-9eb3-0d71595fada0-trusted-ca\") pod \"e93dbd5d-0ca7-4a15-9eb3-0d71595fada0\" (UID: \"e93dbd5d-0ca7-4a15-9eb3-0d71595fada0\") " Dec 10 22:56:20 crc kubenswrapper[4791]: I1210 22:56:20.246860 4791 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/e93dbd5d-0ca7-4a15-9eb3-0d71595fada0-registry-certificates" (OuterVolumeSpecName: "registry-certificates") pod "e93dbd5d-0ca7-4a15-9eb3-0d71595fada0" (UID: "e93dbd5d-0ca7-4a15-9eb3-0d71595fada0"). InnerVolumeSpecName "registry-certificates". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 10 22:56:20 crc kubenswrapper[4791]: I1210 22:56:20.247169 4791 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/e93dbd5d-0ca7-4a15-9eb3-0d71595fada0-bound-sa-token\") pod \"e93dbd5d-0ca7-4a15-9eb3-0d71595fada0\" (UID: \"e93dbd5d-0ca7-4a15-9eb3-0d71595fada0\") " Dec 10 22:56:20 crc kubenswrapper[4791]: I1210 22:56:20.247232 4791 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-97w86\" (UniqueName: \"kubernetes.io/projected/e93dbd5d-0ca7-4a15-9eb3-0d71595fada0-kube-api-access-97w86\") pod \"e93dbd5d-0ca7-4a15-9eb3-0d71595fada0\" (UID: \"e93dbd5d-0ca7-4a15-9eb3-0d71595fada0\") " Dec 10 22:56:20 crc kubenswrapper[4791]: I1210 22:56:20.247316 4791 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"installation-pull-secrets\" (UniqueName: \"kubernetes.io/secret/e93dbd5d-0ca7-4a15-9eb3-0d71595fada0-installation-pull-secrets\") pod \"e93dbd5d-0ca7-4a15-9eb3-0d71595fada0\" (UID: \"e93dbd5d-0ca7-4a15-9eb3-0d71595fada0\") " Dec 10 22:56:20 crc kubenswrapper[4791]: I1210 22:56:20.247564 4791 reconciler_common.go:293] "Volume detached for volume \"registry-certificates\" (UniqueName: \"kubernetes.io/configmap/e93dbd5d-0ca7-4a15-9eb3-0d71595fada0-registry-certificates\") on node \"crc\" DevicePath \"\"" Dec 10 22:56:20 crc kubenswrapper[4791]: I1210 22:56:20.250297 4791 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/e93dbd5d-0ca7-4a15-9eb3-0d71595fada0-trusted-ca" (OuterVolumeSpecName: "trusted-ca") pod "e93dbd5d-0ca7-4a15-9eb3-0d71595fada0" (UID: "e93dbd5d-0ca7-4a15-9eb3-0d71595fada0"). InnerVolumeSpecName "trusted-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 10 22:56:20 crc kubenswrapper[4791]: I1210 22:56:20.253448 4791 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/e93dbd5d-0ca7-4a15-9eb3-0d71595fada0-bound-sa-token" (OuterVolumeSpecName: "bound-sa-token") pod "e93dbd5d-0ca7-4a15-9eb3-0d71595fada0" (UID: "e93dbd5d-0ca7-4a15-9eb3-0d71595fada0"). InnerVolumeSpecName "bound-sa-token". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 10 22:56:20 crc kubenswrapper[4791]: I1210 22:56:20.254838 4791 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/e93dbd5d-0ca7-4a15-9eb3-0d71595fada0-installation-pull-secrets" (OuterVolumeSpecName: "installation-pull-secrets") pod "e93dbd5d-0ca7-4a15-9eb3-0d71595fada0" (UID: "e93dbd5d-0ca7-4a15-9eb3-0d71595fada0"). InnerVolumeSpecName "installation-pull-secrets". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 10 22:56:20 crc kubenswrapper[4791]: I1210 22:56:20.255303 4791 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/e93dbd5d-0ca7-4a15-9eb3-0d71595fada0-registry-tls" (OuterVolumeSpecName: "registry-tls") pod "e93dbd5d-0ca7-4a15-9eb3-0d71595fada0" (UID: "e93dbd5d-0ca7-4a15-9eb3-0d71595fada0"). InnerVolumeSpecName "registry-tls". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 10 22:56:20 crc kubenswrapper[4791]: I1210 22:56:20.255640 4791 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/e93dbd5d-0ca7-4a15-9eb3-0d71595fada0-kube-api-access-97w86" (OuterVolumeSpecName: "kube-api-access-97w86") pod "e93dbd5d-0ca7-4a15-9eb3-0d71595fada0" (UID: "e93dbd5d-0ca7-4a15-9eb3-0d71595fada0"). InnerVolumeSpecName "kube-api-access-97w86". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 10 22:56:20 crc kubenswrapper[4791]: I1210 22:56:20.257126 4791 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (OuterVolumeSpecName: "registry-storage") pod "e93dbd5d-0ca7-4a15-9eb3-0d71595fada0" (UID: "e93dbd5d-0ca7-4a15-9eb3-0d71595fada0"). InnerVolumeSpecName "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8". PluginName "kubernetes.io/csi", VolumeGidValue "" Dec 10 22:56:20 crc kubenswrapper[4791]: I1210 22:56:20.263222 4791 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/e93dbd5d-0ca7-4a15-9eb3-0d71595fada0-ca-trust-extracted" (OuterVolumeSpecName: "ca-trust-extracted") pod "e93dbd5d-0ca7-4a15-9eb3-0d71595fada0" (UID: "e93dbd5d-0ca7-4a15-9eb3-0d71595fada0"). InnerVolumeSpecName "ca-trust-extracted". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 10 22:56:20 crc kubenswrapper[4791]: I1210 22:56:20.348481 4791 reconciler_common.go:293] "Volume detached for volume \"installation-pull-secrets\" (UniqueName: \"kubernetes.io/secret/e93dbd5d-0ca7-4a15-9eb3-0d71595fada0-installation-pull-secrets\") on node \"crc\" DevicePath \"\"" Dec 10 22:56:20 crc kubenswrapper[4791]: I1210 22:56:20.348512 4791 reconciler_common.go:293] "Volume detached for volume \"ca-trust-extracted\" (UniqueName: \"kubernetes.io/empty-dir/e93dbd5d-0ca7-4a15-9eb3-0d71595fada0-ca-trust-extracted\") on node \"crc\" DevicePath \"\"" Dec 10 22:56:20 crc kubenswrapper[4791]: I1210 22:56:20.348522 4791 reconciler_common.go:293] "Volume detached for volume \"registry-tls\" (UniqueName: \"kubernetes.io/projected/e93dbd5d-0ca7-4a15-9eb3-0d71595fada0-registry-tls\") on node \"crc\" DevicePath \"\"" Dec 10 22:56:20 crc kubenswrapper[4791]: I1210 22:56:20.348530 4791 reconciler_common.go:293] "Volume detached for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/e93dbd5d-0ca7-4a15-9eb3-0d71595fada0-trusted-ca\") on node \"crc\" DevicePath \"\"" Dec 10 22:56:20 crc kubenswrapper[4791]: I1210 22:56:20.348538 4791 reconciler_common.go:293] "Volume detached for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/e93dbd5d-0ca7-4a15-9eb3-0d71595fada0-bound-sa-token\") on node \"crc\" DevicePath \"\"" Dec 10 22:56:20 crc kubenswrapper[4791]: I1210 22:56:20.348545 4791 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-97w86\" (UniqueName: \"kubernetes.io/projected/e93dbd5d-0ca7-4a15-9eb3-0d71595fada0-kube-api-access-97w86\") on node \"crc\" DevicePath \"\"" Dec 10 22:56:21 crc kubenswrapper[4791]: I1210 22:56:21.081184 4791 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-image-registry/image-registry-697d97f7c8-srd52" event={"ID":"e93dbd5d-0ca7-4a15-9eb3-0d71595fada0","Type":"ContainerDied","Data":"188e6bd81aba3d08f6a97b4a555211ed5d1b7ae3905df653258fe88d8a31fc5e"} Dec 10 22:56:21 crc kubenswrapper[4791]: I1210 22:56:21.081241 4791 scope.go:117] "RemoveContainer" containerID="e6eab22df9424a9028c3071cbd5cae655435fbbba257f037a06dfc92facc867d" Dec 10 22:56:21 crc kubenswrapper[4791]: I1210 22:56:21.081278 4791 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-image-registry/image-registry-697d97f7c8-srd52" Dec 10 22:56:21 crc kubenswrapper[4791]: I1210 22:56:21.110205 4791 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-image-registry/image-registry-697d97f7c8-srd52"] Dec 10 22:56:21 crc kubenswrapper[4791]: I1210 22:56:21.113520 4791 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-image-registry/image-registry-697d97f7c8-srd52"] Dec 10 22:56:21 crc kubenswrapper[4791]: I1210 22:56:21.905577 4791 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="e93dbd5d-0ca7-4a15-9eb3-0d71595fada0" path="/var/lib/kubelet/pods/e93dbd5d-0ca7-4a15-9eb3-0d71595fada0/volumes" Dec 10 22:58:25 crc kubenswrapper[4791]: I1210 22:58:25.037916 4791 patch_prober.go:28] interesting pod/machine-config-daemon-5rb5l container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 10 22:58:25 crc kubenswrapper[4791]: I1210 22:58:25.038605 4791 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-5rb5l" podUID="ba35653c-6e06-4cee-a4d6-137764090d18" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 10 22:58:55 crc kubenswrapper[4791]: I1210 22:58:55.038690 4791 patch_prober.go:28] interesting pod/machine-config-daemon-5rb5l container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 10 22:58:55 crc kubenswrapper[4791]: I1210 22:58:55.039411 4791 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-5rb5l" podUID="ba35653c-6e06-4cee-a4d6-137764090d18" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 10 22:59:25 crc kubenswrapper[4791]: I1210 22:59:25.037957 4791 patch_prober.go:28] interesting pod/machine-config-daemon-5rb5l container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 10 22:59:25 crc kubenswrapper[4791]: I1210 22:59:25.038588 4791 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-5rb5l" podUID="ba35653c-6e06-4cee-a4d6-137764090d18" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 10 22:59:25 crc kubenswrapper[4791]: I1210 22:59:25.038641 4791 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-5rb5l" Dec 10 22:59:25 crc kubenswrapper[4791]: I1210 22:59:25.039273 4791 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"c09d66dfdd1ff68cb29fe7936cd36ccb2e44990bd887b93547e1487b370efa4e"} pod="openshift-machine-config-operator/machine-config-daemon-5rb5l" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Dec 10 22:59:25 crc kubenswrapper[4791]: I1210 22:59:25.039372 4791 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-5rb5l" podUID="ba35653c-6e06-4cee-a4d6-137764090d18" containerName="machine-config-daemon" containerID="cri-o://c09d66dfdd1ff68cb29fe7936cd36ccb2e44990bd887b93547e1487b370efa4e" gracePeriod=600 Dec 10 22:59:26 crc kubenswrapper[4791]: I1210 22:59:26.152773 4791 generic.go:334] "Generic (PLEG): container finished" podID="ba35653c-6e06-4cee-a4d6-137764090d18" containerID="c09d66dfdd1ff68cb29fe7936cd36ccb2e44990bd887b93547e1487b370efa4e" exitCode=0 Dec 10 22:59:26 crc kubenswrapper[4791]: I1210 22:59:26.152909 4791 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-5rb5l" event={"ID":"ba35653c-6e06-4cee-a4d6-137764090d18","Type":"ContainerDied","Data":"c09d66dfdd1ff68cb29fe7936cd36ccb2e44990bd887b93547e1487b370efa4e"} Dec 10 22:59:26 crc kubenswrapper[4791]: I1210 22:59:26.153140 4791 scope.go:117] "RemoveContainer" containerID="05fce03d9a939befc9dac71cbc9750ba888a8f61825ef9c21bf847dd73aa10dc" Dec 10 22:59:27 crc kubenswrapper[4791]: I1210 22:59:27.159753 4791 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-5rb5l" event={"ID":"ba35653c-6e06-4cee-a4d6-137764090d18","Type":"ContainerStarted","Data":"cb32d752921f31f5faa2ab64f1cbd0e43f7e3f3e95a29f0f2fbc2ed42452d743"} Dec 10 23:00:00 crc kubenswrapper[4791]: I1210 23:00:00.171499 4791 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29423460-rp94k"] Dec 10 23:00:00 crc kubenswrapper[4791]: E1210 23:00:00.172394 4791 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="e93dbd5d-0ca7-4a15-9eb3-0d71595fada0" containerName="registry" Dec 10 23:00:00 crc kubenswrapper[4791]: I1210 23:00:00.172413 4791 state_mem.go:107] "Deleted CPUSet assignment" podUID="e93dbd5d-0ca7-4a15-9eb3-0d71595fada0" containerName="registry" Dec 10 23:00:00 crc kubenswrapper[4791]: I1210 23:00:00.172561 4791 memory_manager.go:354] "RemoveStaleState removing state" podUID="e93dbd5d-0ca7-4a15-9eb3-0d71595fada0" containerName="registry" Dec 10 23:00:00 crc kubenswrapper[4791]: I1210 23:00:00.173103 4791 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29423460-rp94k" Dec 10 23:00:00 crc kubenswrapper[4791]: I1210 23:00:00.175469 4791 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"collect-profiles-dockercfg-kzf4t" Dec 10 23:00:00 crc kubenswrapper[4791]: I1210 23:00:00.175776 4791 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-operator-lifecycle-manager"/"collect-profiles-config" Dec 10 23:00:00 crc kubenswrapper[4791]: I1210 23:00:00.182182 4791 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29423460-rp94k"] Dec 10 23:00:00 crc kubenswrapper[4791]: I1210 23:00:00.233770 4791 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/262dcaa0-a67b-421d-ad84-95eaaa3d6a7f-secret-volume\") pod \"collect-profiles-29423460-rp94k\" (UID: \"262dcaa0-a67b-421d-ad84-95eaaa3d6a7f\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29423460-rp94k" Dec 10 23:00:00 crc kubenswrapper[4791]: I1210 23:00:00.233856 4791 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/262dcaa0-a67b-421d-ad84-95eaaa3d6a7f-config-volume\") pod \"collect-profiles-29423460-rp94k\" (UID: \"262dcaa0-a67b-421d-ad84-95eaaa3d6a7f\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29423460-rp94k" Dec 10 23:00:00 crc kubenswrapper[4791]: I1210 23:00:00.233897 4791 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-ct85f\" (UniqueName: \"kubernetes.io/projected/262dcaa0-a67b-421d-ad84-95eaaa3d6a7f-kube-api-access-ct85f\") pod \"collect-profiles-29423460-rp94k\" (UID: \"262dcaa0-a67b-421d-ad84-95eaaa3d6a7f\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29423460-rp94k" Dec 10 23:00:00 crc kubenswrapper[4791]: I1210 23:00:00.334712 4791 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/262dcaa0-a67b-421d-ad84-95eaaa3d6a7f-secret-volume\") pod \"collect-profiles-29423460-rp94k\" (UID: \"262dcaa0-a67b-421d-ad84-95eaaa3d6a7f\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29423460-rp94k" Dec 10 23:00:00 crc kubenswrapper[4791]: I1210 23:00:00.334757 4791 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/262dcaa0-a67b-421d-ad84-95eaaa3d6a7f-config-volume\") pod \"collect-profiles-29423460-rp94k\" (UID: \"262dcaa0-a67b-421d-ad84-95eaaa3d6a7f\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29423460-rp94k" Dec 10 23:00:00 crc kubenswrapper[4791]: I1210 23:00:00.334776 4791 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-ct85f\" (UniqueName: \"kubernetes.io/projected/262dcaa0-a67b-421d-ad84-95eaaa3d6a7f-kube-api-access-ct85f\") pod \"collect-profiles-29423460-rp94k\" (UID: \"262dcaa0-a67b-421d-ad84-95eaaa3d6a7f\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29423460-rp94k" Dec 10 23:00:00 crc kubenswrapper[4791]: I1210 23:00:00.336092 4791 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/262dcaa0-a67b-421d-ad84-95eaaa3d6a7f-config-volume\") pod \"collect-profiles-29423460-rp94k\" (UID: \"262dcaa0-a67b-421d-ad84-95eaaa3d6a7f\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29423460-rp94k" Dec 10 23:00:00 crc kubenswrapper[4791]: I1210 23:00:00.345720 4791 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/262dcaa0-a67b-421d-ad84-95eaaa3d6a7f-secret-volume\") pod \"collect-profiles-29423460-rp94k\" (UID: \"262dcaa0-a67b-421d-ad84-95eaaa3d6a7f\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29423460-rp94k" Dec 10 23:00:00 crc kubenswrapper[4791]: I1210 23:00:00.360984 4791 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-ct85f\" (UniqueName: \"kubernetes.io/projected/262dcaa0-a67b-421d-ad84-95eaaa3d6a7f-kube-api-access-ct85f\") pod \"collect-profiles-29423460-rp94k\" (UID: \"262dcaa0-a67b-421d-ad84-95eaaa3d6a7f\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29423460-rp94k" Dec 10 23:00:00 crc kubenswrapper[4791]: I1210 23:00:00.494508 4791 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29423460-rp94k" Dec 10 23:00:00 crc kubenswrapper[4791]: I1210 23:00:00.720201 4791 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29423460-rp94k"] Dec 10 23:00:01 crc kubenswrapper[4791]: I1210 23:00:01.365013 4791 generic.go:334] "Generic (PLEG): container finished" podID="262dcaa0-a67b-421d-ad84-95eaaa3d6a7f" containerID="0ff11197939d88bcbbaa9c417821a4a555a4f94ceb603408b0d6283461991b2c" exitCode=0 Dec 10 23:00:01 crc kubenswrapper[4791]: I1210 23:00:01.365053 4791 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29423460-rp94k" event={"ID":"262dcaa0-a67b-421d-ad84-95eaaa3d6a7f","Type":"ContainerDied","Data":"0ff11197939d88bcbbaa9c417821a4a555a4f94ceb603408b0d6283461991b2c"} Dec 10 23:00:01 crc kubenswrapper[4791]: I1210 23:00:01.365080 4791 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29423460-rp94k" event={"ID":"262dcaa0-a67b-421d-ad84-95eaaa3d6a7f","Type":"ContainerStarted","Data":"6d4007b0454c616c7fcac210e46af630869e143d61cf77b43c5bb19b83319892"} Dec 10 23:00:02 crc kubenswrapper[4791]: I1210 23:00:02.618296 4791 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29423460-rp94k" Dec 10 23:00:02 crc kubenswrapper[4791]: I1210 23:00:02.767601 4791 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-ct85f\" (UniqueName: \"kubernetes.io/projected/262dcaa0-a67b-421d-ad84-95eaaa3d6a7f-kube-api-access-ct85f\") pod \"262dcaa0-a67b-421d-ad84-95eaaa3d6a7f\" (UID: \"262dcaa0-a67b-421d-ad84-95eaaa3d6a7f\") " Dec 10 23:00:02 crc kubenswrapper[4791]: I1210 23:00:02.768283 4791 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/262dcaa0-a67b-421d-ad84-95eaaa3d6a7f-secret-volume\") pod \"262dcaa0-a67b-421d-ad84-95eaaa3d6a7f\" (UID: \"262dcaa0-a67b-421d-ad84-95eaaa3d6a7f\") " Dec 10 23:00:02 crc kubenswrapper[4791]: I1210 23:00:02.768433 4791 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/262dcaa0-a67b-421d-ad84-95eaaa3d6a7f-config-volume\") pod \"262dcaa0-a67b-421d-ad84-95eaaa3d6a7f\" (UID: \"262dcaa0-a67b-421d-ad84-95eaaa3d6a7f\") " Dec 10 23:00:02 crc kubenswrapper[4791]: I1210 23:00:02.769293 4791 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/262dcaa0-a67b-421d-ad84-95eaaa3d6a7f-config-volume" (OuterVolumeSpecName: "config-volume") pod "262dcaa0-a67b-421d-ad84-95eaaa3d6a7f" (UID: "262dcaa0-a67b-421d-ad84-95eaaa3d6a7f"). InnerVolumeSpecName "config-volume". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 10 23:00:02 crc kubenswrapper[4791]: I1210 23:00:02.774173 4791 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/262dcaa0-a67b-421d-ad84-95eaaa3d6a7f-kube-api-access-ct85f" (OuterVolumeSpecName: "kube-api-access-ct85f") pod "262dcaa0-a67b-421d-ad84-95eaaa3d6a7f" (UID: "262dcaa0-a67b-421d-ad84-95eaaa3d6a7f"). InnerVolumeSpecName "kube-api-access-ct85f". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 10 23:00:02 crc kubenswrapper[4791]: I1210 23:00:02.775522 4791 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/262dcaa0-a67b-421d-ad84-95eaaa3d6a7f-secret-volume" (OuterVolumeSpecName: "secret-volume") pod "262dcaa0-a67b-421d-ad84-95eaaa3d6a7f" (UID: "262dcaa0-a67b-421d-ad84-95eaaa3d6a7f"). InnerVolumeSpecName "secret-volume". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 10 23:00:02 crc kubenswrapper[4791]: I1210 23:00:02.869765 4791 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-ct85f\" (UniqueName: \"kubernetes.io/projected/262dcaa0-a67b-421d-ad84-95eaaa3d6a7f-kube-api-access-ct85f\") on node \"crc\" DevicePath \"\"" Dec 10 23:00:02 crc kubenswrapper[4791]: I1210 23:00:02.869825 4791 reconciler_common.go:293] "Volume detached for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/262dcaa0-a67b-421d-ad84-95eaaa3d6a7f-secret-volume\") on node \"crc\" DevicePath \"\"" Dec 10 23:00:02 crc kubenswrapper[4791]: I1210 23:00:02.869847 4791 reconciler_common.go:293] "Volume detached for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/262dcaa0-a67b-421d-ad84-95eaaa3d6a7f-config-volume\") on node \"crc\" DevicePath \"\"" Dec 10 23:00:03 crc kubenswrapper[4791]: I1210 23:00:03.380171 4791 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29423460-rp94k" event={"ID":"262dcaa0-a67b-421d-ad84-95eaaa3d6a7f","Type":"ContainerDied","Data":"6d4007b0454c616c7fcac210e46af630869e143d61cf77b43c5bb19b83319892"} Dec 10 23:00:03 crc kubenswrapper[4791]: I1210 23:00:03.380217 4791 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="6d4007b0454c616c7fcac210e46af630869e143d61cf77b43c5bb19b83319892" Dec 10 23:00:03 crc kubenswrapper[4791]: I1210 23:00:03.380266 4791 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29423460-rp94k" Dec 10 23:00:37 crc kubenswrapper[4791]: I1210 23:00:37.248652 4791 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["cert-manager/cert-manager-cainjector-7f985d654d-ntt86"] Dec 10 23:00:37 crc kubenswrapper[4791]: E1210 23:00:37.249530 4791 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="262dcaa0-a67b-421d-ad84-95eaaa3d6a7f" containerName="collect-profiles" Dec 10 23:00:37 crc kubenswrapper[4791]: I1210 23:00:37.249549 4791 state_mem.go:107] "Deleted CPUSet assignment" podUID="262dcaa0-a67b-421d-ad84-95eaaa3d6a7f" containerName="collect-profiles" Dec 10 23:00:37 crc kubenswrapper[4791]: I1210 23:00:37.249659 4791 memory_manager.go:354] "RemoveStaleState removing state" podUID="262dcaa0-a67b-421d-ad84-95eaaa3d6a7f" containerName="collect-profiles" Dec 10 23:00:37 crc kubenswrapper[4791]: I1210 23:00:37.250113 4791 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="cert-manager/cert-manager-cainjector-7f985d654d-ntt86" Dec 10 23:00:37 crc kubenswrapper[4791]: I1210 23:00:37.254195 4791 reflector.go:368] Caches populated for *v1.ConfigMap from object-"cert-manager"/"kube-root-ca.crt" Dec 10 23:00:37 crc kubenswrapper[4791]: I1210 23:00:37.256785 4791 reflector.go:368] Caches populated for *v1.ConfigMap from object-"cert-manager"/"openshift-service-ca.crt" Dec 10 23:00:37 crc kubenswrapper[4791]: I1210 23:00:37.257052 4791 reflector.go:368] Caches populated for *v1.Secret from object-"cert-manager"/"cert-manager-cainjector-dockercfg-pnkv8" Dec 10 23:00:37 crc kubenswrapper[4791]: I1210 23:00:37.258289 4791 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["cert-manager/cert-manager-5b446d88c5-w9jv9"] Dec 10 23:00:37 crc kubenswrapper[4791]: I1210 23:00:37.259327 4791 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="cert-manager/cert-manager-5b446d88c5-w9jv9" Dec 10 23:00:37 crc kubenswrapper[4791]: I1210 23:00:37.260978 4791 reflector.go:368] Caches populated for *v1.Secret from object-"cert-manager"/"cert-manager-dockercfg-7xw4v" Dec 10 23:00:37 crc kubenswrapper[4791]: I1210 23:00:37.269690 4791 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["cert-manager/cert-manager-cainjector-7f985d654d-ntt86"] Dec 10 23:00:37 crc kubenswrapper[4791]: I1210 23:00:37.274668 4791 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["cert-manager/cert-manager-webhook-5655c58dd6-xzg6b"] Dec 10 23:00:37 crc kubenswrapper[4791]: I1210 23:00:37.275297 4791 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="cert-manager/cert-manager-webhook-5655c58dd6-xzg6b" Dec 10 23:00:37 crc kubenswrapper[4791]: I1210 23:00:37.279378 4791 reflector.go:368] Caches populated for *v1.Secret from object-"cert-manager"/"cert-manager-webhook-dockercfg-gjlst" Dec 10 23:00:37 crc kubenswrapper[4791]: I1210 23:00:37.283638 4791 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["cert-manager/cert-manager-webhook-5655c58dd6-xzg6b"] Dec 10 23:00:37 crc kubenswrapper[4791]: I1210 23:00:37.288576 4791 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["cert-manager/cert-manager-5b446d88c5-w9jv9"] Dec 10 23:00:37 crc kubenswrapper[4791]: I1210 23:00:37.298096 4791 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-v4kq7\" (UniqueName: \"kubernetes.io/projected/5f070f43-21a8-43bf-80db-9038ed2dd9f5-kube-api-access-v4kq7\") pod \"cert-manager-5b446d88c5-w9jv9\" (UID: \"5f070f43-21a8-43bf-80db-9038ed2dd9f5\") " pod="cert-manager/cert-manager-5b446d88c5-w9jv9" Dec 10 23:00:37 crc kubenswrapper[4791]: I1210 23:00:37.298139 4791 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-gbmp5\" (UniqueName: \"kubernetes.io/projected/33717f3f-c6f5-4be9-a7c5-d6a7ae599dc9-kube-api-access-gbmp5\") pod \"cert-manager-webhook-5655c58dd6-xzg6b\" (UID: \"33717f3f-c6f5-4be9-a7c5-d6a7ae599dc9\") " pod="cert-manager/cert-manager-webhook-5655c58dd6-xzg6b" Dec 10 23:00:37 crc kubenswrapper[4791]: I1210 23:00:37.298173 4791 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-t5x56\" (UniqueName: \"kubernetes.io/projected/18fded69-b7d4-42c9-80de-1c78e6389edf-kube-api-access-t5x56\") pod \"cert-manager-cainjector-7f985d654d-ntt86\" (UID: \"18fded69-b7d4-42c9-80de-1c78e6389edf\") " pod="cert-manager/cert-manager-cainjector-7f985d654d-ntt86" Dec 10 23:00:37 crc kubenswrapper[4791]: I1210 23:00:37.399379 4791 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-v4kq7\" (UniqueName: \"kubernetes.io/projected/5f070f43-21a8-43bf-80db-9038ed2dd9f5-kube-api-access-v4kq7\") pod \"cert-manager-5b446d88c5-w9jv9\" (UID: \"5f070f43-21a8-43bf-80db-9038ed2dd9f5\") " pod="cert-manager/cert-manager-5b446d88c5-w9jv9" Dec 10 23:00:37 crc kubenswrapper[4791]: I1210 23:00:37.399444 4791 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-gbmp5\" (UniqueName: \"kubernetes.io/projected/33717f3f-c6f5-4be9-a7c5-d6a7ae599dc9-kube-api-access-gbmp5\") pod \"cert-manager-webhook-5655c58dd6-xzg6b\" (UID: \"33717f3f-c6f5-4be9-a7c5-d6a7ae599dc9\") " pod="cert-manager/cert-manager-webhook-5655c58dd6-xzg6b" Dec 10 23:00:37 crc kubenswrapper[4791]: I1210 23:00:37.399488 4791 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-t5x56\" (UniqueName: \"kubernetes.io/projected/18fded69-b7d4-42c9-80de-1c78e6389edf-kube-api-access-t5x56\") pod \"cert-manager-cainjector-7f985d654d-ntt86\" (UID: \"18fded69-b7d4-42c9-80de-1c78e6389edf\") " pod="cert-manager/cert-manager-cainjector-7f985d654d-ntt86" Dec 10 23:00:37 crc kubenswrapper[4791]: I1210 23:00:37.419161 4791 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-v4kq7\" (UniqueName: \"kubernetes.io/projected/5f070f43-21a8-43bf-80db-9038ed2dd9f5-kube-api-access-v4kq7\") pod \"cert-manager-5b446d88c5-w9jv9\" (UID: \"5f070f43-21a8-43bf-80db-9038ed2dd9f5\") " pod="cert-manager/cert-manager-5b446d88c5-w9jv9" Dec 10 23:00:37 crc kubenswrapper[4791]: I1210 23:00:37.420554 4791 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-t5x56\" (UniqueName: \"kubernetes.io/projected/18fded69-b7d4-42c9-80de-1c78e6389edf-kube-api-access-t5x56\") pod \"cert-manager-cainjector-7f985d654d-ntt86\" (UID: \"18fded69-b7d4-42c9-80de-1c78e6389edf\") " pod="cert-manager/cert-manager-cainjector-7f985d654d-ntt86" Dec 10 23:00:37 crc kubenswrapper[4791]: I1210 23:00:37.421126 4791 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-gbmp5\" (UniqueName: \"kubernetes.io/projected/33717f3f-c6f5-4be9-a7c5-d6a7ae599dc9-kube-api-access-gbmp5\") pod \"cert-manager-webhook-5655c58dd6-xzg6b\" (UID: \"33717f3f-c6f5-4be9-a7c5-d6a7ae599dc9\") " pod="cert-manager/cert-manager-webhook-5655c58dd6-xzg6b" Dec 10 23:00:37 crc kubenswrapper[4791]: I1210 23:00:37.571108 4791 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="cert-manager/cert-manager-cainjector-7f985d654d-ntt86" Dec 10 23:00:37 crc kubenswrapper[4791]: I1210 23:00:37.583888 4791 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="cert-manager/cert-manager-5b446d88c5-w9jv9" Dec 10 23:00:37 crc kubenswrapper[4791]: I1210 23:00:37.594866 4791 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="cert-manager/cert-manager-webhook-5655c58dd6-xzg6b" Dec 10 23:00:37 crc kubenswrapper[4791]: I1210 23:00:37.760660 4791 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["cert-manager/cert-manager-cainjector-7f985d654d-ntt86"] Dec 10 23:00:37 crc kubenswrapper[4791]: I1210 23:00:37.777767 4791 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Dec 10 23:00:38 crc kubenswrapper[4791]: I1210 23:00:38.027870 4791 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["cert-manager/cert-manager-5b446d88c5-w9jv9"] Dec 10 23:00:38 crc kubenswrapper[4791]: W1210 23:00:38.036109 4791 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod33717f3f_c6f5_4be9_a7c5_d6a7ae599dc9.slice/crio-34b1ac3452a3550c80c603b559fb78f128d7867102a108756ec219dea754dd70 WatchSource:0}: Error finding container 34b1ac3452a3550c80c603b559fb78f128d7867102a108756ec219dea754dd70: Status 404 returned error can't find the container with id 34b1ac3452a3550c80c603b559fb78f128d7867102a108756ec219dea754dd70 Dec 10 23:00:38 crc kubenswrapper[4791]: I1210 23:00:38.043613 4791 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["cert-manager/cert-manager-webhook-5655c58dd6-xzg6b"] Dec 10 23:00:38 crc kubenswrapper[4791]: I1210 23:00:38.587162 4791 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="cert-manager/cert-manager-5b446d88c5-w9jv9" event={"ID":"5f070f43-21a8-43bf-80db-9038ed2dd9f5","Type":"ContainerStarted","Data":"92b9e3094ca4e10efd25b8fa1cb132eb9d85b69d3562cb6048343b9d81bc7657"} Dec 10 23:00:38 crc kubenswrapper[4791]: I1210 23:00:38.590032 4791 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="cert-manager/cert-manager-cainjector-7f985d654d-ntt86" event={"ID":"18fded69-b7d4-42c9-80de-1c78e6389edf","Type":"ContainerStarted","Data":"d3880715fd1b846d569d8d0410840406e5d08f3bff7d616a91b68c869ef84f32"} Dec 10 23:00:38 crc kubenswrapper[4791]: I1210 23:00:38.592153 4791 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="cert-manager/cert-manager-webhook-5655c58dd6-xzg6b" event={"ID":"33717f3f-c6f5-4be9-a7c5-d6a7ae599dc9","Type":"ContainerStarted","Data":"34b1ac3452a3550c80c603b559fb78f128d7867102a108756ec219dea754dd70"} Dec 10 23:00:44 crc kubenswrapper[4791]: I1210 23:00:44.626032 4791 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="cert-manager/cert-manager-cainjector-7f985d654d-ntt86" event={"ID":"18fded69-b7d4-42c9-80de-1c78e6389edf","Type":"ContainerStarted","Data":"ecff6e2aabff64872e6b317da4115839e9d094ddbdc0f7ac2cf3e951528fd54d"} Dec 10 23:00:44 crc kubenswrapper[4791]: I1210 23:00:44.628154 4791 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="cert-manager/cert-manager-webhook-5655c58dd6-xzg6b" event={"ID":"33717f3f-c6f5-4be9-a7c5-d6a7ae599dc9","Type":"ContainerStarted","Data":"a5ecba7f51d249e55385eac452978ee33e993f69a72e7b59556248bb230963af"} Dec 10 23:00:44 crc kubenswrapper[4791]: I1210 23:00:44.628324 4791 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="cert-manager/cert-manager-webhook-5655c58dd6-xzg6b" Dec 10 23:00:44 crc kubenswrapper[4791]: I1210 23:00:44.629864 4791 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="cert-manager/cert-manager-5b446d88c5-w9jv9" event={"ID":"5f070f43-21a8-43bf-80db-9038ed2dd9f5","Type":"ContainerStarted","Data":"44e0743a07c89992ec510fd73e04fc8d0f4ef96638af41afc40b6c7e29916c74"} Dec 10 23:00:44 crc kubenswrapper[4791]: I1210 23:00:44.643676 4791 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="cert-manager/cert-manager-cainjector-7f985d654d-ntt86" podStartSLOduration=1.753046038 podStartE2EDuration="7.643651708s" podCreationTimestamp="2025-12-10 23:00:37 +0000 UTC" firstStartedPulling="2025-12-10 23:00:37.777519422 +0000 UTC m=+672.207137035" lastFinishedPulling="2025-12-10 23:00:43.668125082 +0000 UTC m=+678.097742705" observedRunningTime="2025-12-10 23:00:44.641753744 +0000 UTC m=+679.071371367" watchObservedRunningTime="2025-12-10 23:00:44.643651708 +0000 UTC m=+679.073269361" Dec 10 23:00:44 crc kubenswrapper[4791]: I1210 23:00:44.658407 4791 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="cert-manager/cert-manager-webhook-5655c58dd6-xzg6b" podStartSLOduration=1.968165336 podStartE2EDuration="7.65838357s" podCreationTimestamp="2025-12-10 23:00:37 +0000 UTC" firstStartedPulling="2025-12-10 23:00:38.039918604 +0000 UTC m=+672.469536217" lastFinishedPulling="2025-12-10 23:00:43.730136838 +0000 UTC m=+678.159754451" observedRunningTime="2025-12-10 23:00:44.656163816 +0000 UTC m=+679.085781439" watchObservedRunningTime="2025-12-10 23:00:44.65838357 +0000 UTC m=+679.088001183" Dec 10 23:00:44 crc kubenswrapper[4791]: I1210 23:00:44.676065 4791 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="cert-manager/cert-manager-5b446d88c5-w9jv9" podStartSLOduration=2.042842333 podStartE2EDuration="7.676046305s" podCreationTimestamp="2025-12-10 23:00:37 +0000 UTC" firstStartedPulling="2025-12-10 23:00:38.035423465 +0000 UTC m=+672.465041088" lastFinishedPulling="2025-12-10 23:00:43.668627437 +0000 UTC m=+678.098245060" observedRunningTime="2025-12-10 23:00:44.674230563 +0000 UTC m=+679.103848196" watchObservedRunningTime="2025-12-10 23:00:44.676046305 +0000 UTC m=+679.105663918" Dec 10 23:00:47 crc kubenswrapper[4791]: I1210 23:00:47.664176 4791 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-ovn-kubernetes/ovnkube-node-zhq64"] Dec 10 23:00:47 crc kubenswrapper[4791]: I1210 23:00:47.664908 4791 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-ovn-kubernetes/ovnkube-node-zhq64" podUID="3cd47739-0fa9-4321-aff1-220f8721a0b3" containerName="ovn-controller" containerID="cri-o://1a52d19b7671e8b143017c8a981a6f8d3ee3b84650e4b223b814a531ac76f364" gracePeriod=30 Dec 10 23:00:47 crc kubenswrapper[4791]: I1210 23:00:47.665001 4791 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-ovn-kubernetes/ovnkube-node-zhq64" podUID="3cd47739-0fa9-4321-aff1-220f8721a0b3" containerName="kube-rbac-proxy-node" containerID="cri-o://68795a29944e56870ffe94fa732e9dcf0bfb344c0d85d1e590f58a876129c61f" gracePeriod=30 Dec 10 23:00:47 crc kubenswrapper[4791]: I1210 23:00:47.665040 4791 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-ovn-kubernetes/ovnkube-node-zhq64" podUID="3cd47739-0fa9-4321-aff1-220f8721a0b3" containerName="ovn-acl-logging" containerID="cri-o://937137ae9d7480cdc7c59bc4a1df0a4c10e72b3b63693af285278789953ed218" gracePeriod=30 Dec 10 23:00:47 crc kubenswrapper[4791]: I1210 23:00:47.665088 4791 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-ovn-kubernetes/ovnkube-node-zhq64" podUID="3cd47739-0fa9-4321-aff1-220f8721a0b3" containerName="kube-rbac-proxy-ovn-metrics" containerID="cri-o://bc721b530c08b433756e978ec0423cc18aabf6771b61eb96f903b4d57ea335ff" gracePeriod=30 Dec 10 23:00:47 crc kubenswrapper[4791]: I1210 23:00:47.665049 4791 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-ovn-kubernetes/ovnkube-node-zhq64" podUID="3cd47739-0fa9-4321-aff1-220f8721a0b3" containerName="nbdb" containerID="cri-o://152a7618be8f38fadf4e564b3948980503cedcdb466c37344ccc20f4a3173a5f" gracePeriod=30 Dec 10 23:00:47 crc kubenswrapper[4791]: I1210 23:00:47.665075 4791 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-ovn-kubernetes/ovnkube-node-zhq64" podUID="3cd47739-0fa9-4321-aff1-220f8721a0b3" containerName="sbdb" containerID="cri-o://dd3f80e5713221f96ed50da7f2e7fd5b9fabfcd5c1656da8d8a86d818a5fdc3f" gracePeriod=30 Dec 10 23:00:47 crc kubenswrapper[4791]: I1210 23:00:47.665218 4791 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-ovn-kubernetes/ovnkube-node-zhq64" podUID="3cd47739-0fa9-4321-aff1-220f8721a0b3" containerName="northd" containerID="cri-o://3d18f9cb7207efb1e32de75a30d33e024f656a08f0503db95e5488bc5280971f" gracePeriod=30 Dec 10 23:00:47 crc kubenswrapper[4791]: I1210 23:00:47.700102 4791 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-ovn-kubernetes/ovnkube-node-zhq64" podUID="3cd47739-0fa9-4321-aff1-220f8721a0b3" containerName="ovnkube-controller" containerID="cri-o://8839756dfbc89d902c809e3b2b4c8dd5de8ef52ec1009fb2aac5a06a16de4b66" gracePeriod=30 Dec 10 23:00:47 crc kubenswrapper[4791]: I1210 23:00:47.947548 4791 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-zhq64_3cd47739-0fa9-4321-aff1-220f8721a0b3/ovnkube-controller/3.log" Dec 10 23:00:47 crc kubenswrapper[4791]: I1210 23:00:47.950302 4791 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-zhq64_3cd47739-0fa9-4321-aff1-220f8721a0b3/ovn-acl-logging/0.log" Dec 10 23:00:47 crc kubenswrapper[4791]: I1210 23:00:47.951379 4791 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-zhq64_3cd47739-0fa9-4321-aff1-220f8721a0b3/ovn-controller/0.log" Dec 10 23:00:47 crc kubenswrapper[4791]: I1210 23:00:47.952131 4791 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-ovn-kubernetes/ovnkube-node-zhq64" Dec 10 23:00:48 crc kubenswrapper[4791]: I1210 23:00:48.005548 4791 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-ovn-kubernetes/ovnkube-node-bdhcv"] Dec 10 23:00:48 crc kubenswrapper[4791]: E1210 23:00:48.006008 4791 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="3cd47739-0fa9-4321-aff1-220f8721a0b3" containerName="sbdb" Dec 10 23:00:48 crc kubenswrapper[4791]: I1210 23:00:48.006068 4791 state_mem.go:107] "Deleted CPUSet assignment" podUID="3cd47739-0fa9-4321-aff1-220f8721a0b3" containerName="sbdb" Dec 10 23:00:48 crc kubenswrapper[4791]: E1210 23:00:48.006128 4791 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="3cd47739-0fa9-4321-aff1-220f8721a0b3" containerName="ovn-acl-logging" Dec 10 23:00:48 crc kubenswrapper[4791]: I1210 23:00:48.006178 4791 state_mem.go:107] "Deleted CPUSet assignment" podUID="3cd47739-0fa9-4321-aff1-220f8721a0b3" containerName="ovn-acl-logging" Dec 10 23:00:48 crc kubenswrapper[4791]: E1210 23:00:48.006229 4791 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="3cd47739-0fa9-4321-aff1-220f8721a0b3" containerName="kube-rbac-proxy-ovn-metrics" Dec 10 23:00:48 crc kubenswrapper[4791]: I1210 23:00:48.006283 4791 state_mem.go:107] "Deleted CPUSet assignment" podUID="3cd47739-0fa9-4321-aff1-220f8721a0b3" containerName="kube-rbac-proxy-ovn-metrics" Dec 10 23:00:48 crc kubenswrapper[4791]: E1210 23:00:48.006354 4791 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="3cd47739-0fa9-4321-aff1-220f8721a0b3" containerName="ovnkube-controller" Dec 10 23:00:48 crc kubenswrapper[4791]: I1210 23:00:48.006415 4791 state_mem.go:107] "Deleted CPUSet assignment" podUID="3cd47739-0fa9-4321-aff1-220f8721a0b3" containerName="ovnkube-controller" Dec 10 23:00:48 crc kubenswrapper[4791]: E1210 23:00:48.006468 4791 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="3cd47739-0fa9-4321-aff1-220f8721a0b3" containerName="kube-rbac-proxy-node" Dec 10 23:00:48 crc kubenswrapper[4791]: I1210 23:00:48.006516 4791 state_mem.go:107] "Deleted CPUSet assignment" podUID="3cd47739-0fa9-4321-aff1-220f8721a0b3" containerName="kube-rbac-proxy-node" Dec 10 23:00:48 crc kubenswrapper[4791]: E1210 23:00:48.006564 4791 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="3cd47739-0fa9-4321-aff1-220f8721a0b3" containerName="kubecfg-setup" Dec 10 23:00:48 crc kubenswrapper[4791]: I1210 23:00:48.006614 4791 state_mem.go:107] "Deleted CPUSet assignment" podUID="3cd47739-0fa9-4321-aff1-220f8721a0b3" containerName="kubecfg-setup" Dec 10 23:00:48 crc kubenswrapper[4791]: E1210 23:00:48.006668 4791 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="3cd47739-0fa9-4321-aff1-220f8721a0b3" containerName="ovnkube-controller" Dec 10 23:00:48 crc kubenswrapper[4791]: I1210 23:00:48.006718 4791 state_mem.go:107] "Deleted CPUSet assignment" podUID="3cd47739-0fa9-4321-aff1-220f8721a0b3" containerName="ovnkube-controller" Dec 10 23:00:48 crc kubenswrapper[4791]: E1210 23:00:48.006767 4791 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="3cd47739-0fa9-4321-aff1-220f8721a0b3" containerName="ovn-controller" Dec 10 23:00:48 crc kubenswrapper[4791]: I1210 23:00:48.006814 4791 state_mem.go:107] "Deleted CPUSet assignment" podUID="3cd47739-0fa9-4321-aff1-220f8721a0b3" containerName="ovn-controller" Dec 10 23:00:48 crc kubenswrapper[4791]: E1210 23:00:48.006866 4791 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="3cd47739-0fa9-4321-aff1-220f8721a0b3" containerName="northd" Dec 10 23:00:48 crc kubenswrapper[4791]: I1210 23:00:48.006917 4791 state_mem.go:107] "Deleted CPUSet assignment" podUID="3cd47739-0fa9-4321-aff1-220f8721a0b3" containerName="northd" Dec 10 23:00:48 crc kubenswrapper[4791]: E1210 23:00:48.006967 4791 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="3cd47739-0fa9-4321-aff1-220f8721a0b3" containerName="ovnkube-controller" Dec 10 23:00:48 crc kubenswrapper[4791]: I1210 23:00:48.007015 4791 state_mem.go:107] "Deleted CPUSet assignment" podUID="3cd47739-0fa9-4321-aff1-220f8721a0b3" containerName="ovnkube-controller" Dec 10 23:00:48 crc kubenswrapper[4791]: E1210 23:00:48.007065 4791 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="3cd47739-0fa9-4321-aff1-220f8721a0b3" containerName="ovnkube-controller" Dec 10 23:00:48 crc kubenswrapper[4791]: I1210 23:00:48.007112 4791 state_mem.go:107] "Deleted CPUSet assignment" podUID="3cd47739-0fa9-4321-aff1-220f8721a0b3" containerName="ovnkube-controller" Dec 10 23:00:48 crc kubenswrapper[4791]: E1210 23:00:48.007156 4791 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="3cd47739-0fa9-4321-aff1-220f8721a0b3" containerName="nbdb" Dec 10 23:00:48 crc kubenswrapper[4791]: I1210 23:00:48.007209 4791 state_mem.go:107] "Deleted CPUSet assignment" podUID="3cd47739-0fa9-4321-aff1-220f8721a0b3" containerName="nbdb" Dec 10 23:00:48 crc kubenswrapper[4791]: I1210 23:00:48.007392 4791 memory_manager.go:354] "RemoveStaleState removing state" podUID="3cd47739-0fa9-4321-aff1-220f8721a0b3" containerName="ovnkube-controller" Dec 10 23:00:48 crc kubenswrapper[4791]: I1210 23:00:48.007483 4791 memory_manager.go:354] "RemoveStaleState removing state" podUID="3cd47739-0fa9-4321-aff1-220f8721a0b3" containerName="ovnkube-controller" Dec 10 23:00:48 crc kubenswrapper[4791]: I1210 23:00:48.007543 4791 memory_manager.go:354] "RemoveStaleState removing state" podUID="3cd47739-0fa9-4321-aff1-220f8721a0b3" containerName="ovnkube-controller" Dec 10 23:00:48 crc kubenswrapper[4791]: I1210 23:00:48.007593 4791 memory_manager.go:354] "RemoveStaleState removing state" podUID="3cd47739-0fa9-4321-aff1-220f8721a0b3" containerName="ovnkube-controller" Dec 10 23:00:48 crc kubenswrapper[4791]: I1210 23:00:48.007644 4791 memory_manager.go:354] "RemoveStaleState removing state" podUID="3cd47739-0fa9-4321-aff1-220f8721a0b3" containerName="ovn-controller" Dec 10 23:00:48 crc kubenswrapper[4791]: I1210 23:00:48.007699 4791 memory_manager.go:354] "RemoveStaleState removing state" podUID="3cd47739-0fa9-4321-aff1-220f8721a0b3" containerName="kube-rbac-proxy-node" Dec 10 23:00:48 crc kubenswrapper[4791]: I1210 23:00:48.007753 4791 memory_manager.go:354] "RemoveStaleState removing state" podUID="3cd47739-0fa9-4321-aff1-220f8721a0b3" containerName="sbdb" Dec 10 23:00:48 crc kubenswrapper[4791]: I1210 23:00:48.007802 4791 memory_manager.go:354] "RemoveStaleState removing state" podUID="3cd47739-0fa9-4321-aff1-220f8721a0b3" containerName="ovn-acl-logging" Dec 10 23:00:48 crc kubenswrapper[4791]: I1210 23:00:48.007853 4791 memory_manager.go:354] "RemoveStaleState removing state" podUID="3cd47739-0fa9-4321-aff1-220f8721a0b3" containerName="northd" Dec 10 23:00:48 crc kubenswrapper[4791]: I1210 23:00:48.007913 4791 memory_manager.go:354] "RemoveStaleState removing state" podUID="3cd47739-0fa9-4321-aff1-220f8721a0b3" containerName="nbdb" Dec 10 23:00:48 crc kubenswrapper[4791]: I1210 23:00:48.007965 4791 memory_manager.go:354] "RemoveStaleState removing state" podUID="3cd47739-0fa9-4321-aff1-220f8721a0b3" containerName="kube-rbac-proxy-ovn-metrics" Dec 10 23:00:48 crc kubenswrapper[4791]: E1210 23:00:48.008103 4791 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="3cd47739-0fa9-4321-aff1-220f8721a0b3" containerName="ovnkube-controller" Dec 10 23:00:48 crc kubenswrapper[4791]: I1210 23:00:48.008159 4791 state_mem.go:107] "Deleted CPUSet assignment" podUID="3cd47739-0fa9-4321-aff1-220f8721a0b3" containerName="ovnkube-controller" Dec 10 23:00:48 crc kubenswrapper[4791]: I1210 23:00:48.008419 4791 memory_manager.go:354] "RemoveStaleState removing state" podUID="3cd47739-0fa9-4321-aff1-220f8721a0b3" containerName="ovnkube-controller" Dec 10 23:00:48 crc kubenswrapper[4791]: I1210 23:00:48.010069 4791 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-ovn-kubernetes/ovnkube-node-bdhcv" Dec 10 23:00:48 crc kubenswrapper[4791]: I1210 23:00:48.130189 4791 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/3cd47739-0fa9-4321-aff1-220f8721a0b3-ovnkube-config\") pod \"3cd47739-0fa9-4321-aff1-220f8721a0b3\" (UID: \"3cd47739-0fa9-4321-aff1-220f8721a0b3\") " Dec 10 23:00:48 crc kubenswrapper[4791]: I1210 23:00:48.130447 4791 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"run-systemd\" (UniqueName: \"kubernetes.io/host-path/3cd47739-0fa9-4321-aff1-220f8721a0b3-run-systemd\") pod \"3cd47739-0fa9-4321-aff1-220f8721a0b3\" (UID: \"3cd47739-0fa9-4321-aff1-220f8721a0b3\") " Dec 10 23:00:48 crc kubenswrapper[4791]: I1210 23:00:48.130553 4791 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"var-lib-openvswitch\" (UniqueName: \"kubernetes.io/host-path/3cd47739-0fa9-4321-aff1-220f8721a0b3-var-lib-openvswitch\") pod \"3cd47739-0fa9-4321-aff1-220f8721a0b3\" (UID: \"3cd47739-0fa9-4321-aff1-220f8721a0b3\") " Dec 10 23:00:48 crc kubenswrapper[4791]: I1210 23:00:48.130623 4791 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"run-openvswitch\" (UniqueName: \"kubernetes.io/host-path/3cd47739-0fa9-4321-aff1-220f8721a0b3-run-openvswitch\") pod \"3cd47739-0fa9-4321-aff1-220f8721a0b3\" (UID: \"3cd47739-0fa9-4321-aff1-220f8721a0b3\") " Dec 10 23:00:48 crc kubenswrapper[4791]: I1210 23:00:48.130685 4791 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"host-cni-bin\" (UniqueName: \"kubernetes.io/host-path/3cd47739-0fa9-4321-aff1-220f8721a0b3-host-cni-bin\") pod \"3cd47739-0fa9-4321-aff1-220f8721a0b3\" (UID: \"3cd47739-0fa9-4321-aff1-220f8721a0b3\") " Dec 10 23:00:48 crc kubenswrapper[4791]: I1210 23:00:48.130623 4791 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/3cd47739-0fa9-4321-aff1-220f8721a0b3-var-lib-openvswitch" (OuterVolumeSpecName: "var-lib-openvswitch") pod "3cd47739-0fa9-4321-aff1-220f8721a0b3" (UID: "3cd47739-0fa9-4321-aff1-220f8721a0b3"). InnerVolumeSpecName "var-lib-openvswitch". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 10 23:00:48 crc kubenswrapper[4791]: I1210 23:00:48.130714 4791 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/3cd47739-0fa9-4321-aff1-220f8721a0b3-run-openvswitch" (OuterVolumeSpecName: "run-openvswitch") pod "3cd47739-0fa9-4321-aff1-220f8721a0b3" (UID: "3cd47739-0fa9-4321-aff1-220f8721a0b3"). InnerVolumeSpecName "run-openvswitch". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 10 23:00:48 crc kubenswrapper[4791]: I1210 23:00:48.130747 4791 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovn-node-metrics-cert\" (UniqueName: \"kubernetes.io/secret/3cd47739-0fa9-4321-aff1-220f8721a0b3-ovn-node-metrics-cert\") pod \"3cd47739-0fa9-4321-aff1-220f8721a0b3\" (UID: \"3cd47739-0fa9-4321-aff1-220f8721a0b3\") " Dec 10 23:00:48 crc kubenswrapper[4791]: I1210 23:00:48.130838 4791 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/3cd47739-0fa9-4321-aff1-220f8721a0b3-env-overrides\") pod \"3cd47739-0fa9-4321-aff1-220f8721a0b3\" (UID: \"3cd47739-0fa9-4321-aff1-220f8721a0b3\") " Dec 10 23:00:48 crc kubenswrapper[4791]: I1210 23:00:48.130883 4791 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"host-run-netns\" (UniqueName: \"kubernetes.io/host-path/3cd47739-0fa9-4321-aff1-220f8721a0b3-host-run-netns\") pod \"3cd47739-0fa9-4321-aff1-220f8721a0b3\" (UID: \"3cd47739-0fa9-4321-aff1-220f8721a0b3\") " Dec 10 23:00:48 crc kubenswrapper[4791]: I1210 23:00:48.130891 4791 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/3cd47739-0fa9-4321-aff1-220f8721a0b3-host-cni-bin" (OuterVolumeSpecName: "host-cni-bin") pod "3cd47739-0fa9-4321-aff1-220f8721a0b3" (UID: "3cd47739-0fa9-4321-aff1-220f8721a0b3"). InnerVolumeSpecName "host-cni-bin". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 10 23:00:48 crc kubenswrapper[4791]: I1210 23:00:48.130916 4791 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/3cd47739-0fa9-4321-aff1-220f8721a0b3-host-kubelet" (OuterVolumeSpecName: "host-kubelet") pod "3cd47739-0fa9-4321-aff1-220f8721a0b3" (UID: "3cd47739-0fa9-4321-aff1-220f8721a0b3"). InnerVolumeSpecName "host-kubelet". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 10 23:00:48 crc kubenswrapper[4791]: I1210 23:00:48.130899 4791 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"host-kubelet\" (UniqueName: \"kubernetes.io/host-path/3cd47739-0fa9-4321-aff1-220f8721a0b3-host-kubelet\") pod \"3cd47739-0fa9-4321-aff1-220f8721a0b3\" (UID: \"3cd47739-0fa9-4321-aff1-220f8721a0b3\") " Dec 10 23:00:48 crc kubenswrapper[4791]: I1210 23:00:48.130945 4791 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovnkube-script-lib\" (UniqueName: \"kubernetes.io/configmap/3cd47739-0fa9-4321-aff1-220f8721a0b3-ovnkube-script-lib\") pod \"3cd47739-0fa9-4321-aff1-220f8721a0b3\" (UID: \"3cd47739-0fa9-4321-aff1-220f8721a0b3\") " Dec 10 23:00:48 crc kubenswrapper[4791]: I1210 23:00:48.130961 4791 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/3cd47739-0fa9-4321-aff1-220f8721a0b3-host-run-netns" (OuterVolumeSpecName: "host-run-netns") pod "3cd47739-0fa9-4321-aff1-220f8721a0b3" (UID: "3cd47739-0fa9-4321-aff1-220f8721a0b3"). InnerVolumeSpecName "host-run-netns". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 10 23:00:48 crc kubenswrapper[4791]: I1210 23:00:48.130970 4791 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-wvpw8\" (UniqueName: \"kubernetes.io/projected/3cd47739-0fa9-4321-aff1-220f8721a0b3-kube-api-access-wvpw8\") pod \"3cd47739-0fa9-4321-aff1-220f8721a0b3\" (UID: \"3cd47739-0fa9-4321-aff1-220f8721a0b3\") " Dec 10 23:00:48 crc kubenswrapper[4791]: I1210 23:00:48.131042 4791 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/3cd47739-0fa9-4321-aff1-220f8721a0b3-ovnkube-config" (OuterVolumeSpecName: "ovnkube-config") pod "3cd47739-0fa9-4321-aff1-220f8721a0b3" (UID: "3cd47739-0fa9-4321-aff1-220f8721a0b3"). InnerVolumeSpecName "ovnkube-config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 10 23:00:48 crc kubenswrapper[4791]: I1210 23:00:48.131054 4791 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"host-var-lib-cni-networks-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/3cd47739-0fa9-4321-aff1-220f8721a0b3-host-var-lib-cni-networks-ovn-kubernetes\") pod \"3cd47739-0fa9-4321-aff1-220f8721a0b3\" (UID: \"3cd47739-0fa9-4321-aff1-220f8721a0b3\") " Dec 10 23:00:48 crc kubenswrapper[4791]: I1210 23:00:48.131081 4791 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/3cd47739-0fa9-4321-aff1-220f8721a0b3-host-var-lib-cni-networks-ovn-kubernetes" (OuterVolumeSpecName: "host-var-lib-cni-networks-ovn-kubernetes") pod "3cd47739-0fa9-4321-aff1-220f8721a0b3" (UID: "3cd47739-0fa9-4321-aff1-220f8721a0b3"). InnerVolumeSpecName "host-var-lib-cni-networks-ovn-kubernetes". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 10 23:00:48 crc kubenswrapper[4791]: I1210 23:00:48.131224 4791 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/3cd47739-0fa9-4321-aff1-220f8721a0b3-host-slash" (OuterVolumeSpecName: "host-slash") pod "3cd47739-0fa9-4321-aff1-220f8721a0b3" (UID: "3cd47739-0fa9-4321-aff1-220f8721a0b3"). InnerVolumeSpecName "host-slash". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 10 23:00:48 crc kubenswrapper[4791]: I1210 23:00:48.131267 4791 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"host-slash\" (UniqueName: \"kubernetes.io/host-path/3cd47739-0fa9-4321-aff1-220f8721a0b3-host-slash\") pod \"3cd47739-0fa9-4321-aff1-220f8721a0b3\" (UID: \"3cd47739-0fa9-4321-aff1-220f8721a0b3\") " Dec 10 23:00:48 crc kubenswrapper[4791]: I1210 23:00:48.131304 4791 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"run-ovn\" (UniqueName: \"kubernetes.io/host-path/3cd47739-0fa9-4321-aff1-220f8721a0b3-run-ovn\") pod \"3cd47739-0fa9-4321-aff1-220f8721a0b3\" (UID: \"3cd47739-0fa9-4321-aff1-220f8721a0b3\") " Dec 10 23:00:48 crc kubenswrapper[4791]: I1210 23:00:48.131365 4791 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"log-socket\" (UniqueName: \"kubernetes.io/host-path/3cd47739-0fa9-4321-aff1-220f8721a0b3-log-socket\") pod \"3cd47739-0fa9-4321-aff1-220f8721a0b3\" (UID: \"3cd47739-0fa9-4321-aff1-220f8721a0b3\") " Dec 10 23:00:48 crc kubenswrapper[4791]: I1210 23:00:48.131548 4791 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"host-run-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/3cd47739-0fa9-4321-aff1-220f8721a0b3-host-run-ovn-kubernetes\") pod \"3cd47739-0fa9-4321-aff1-220f8721a0b3\" (UID: \"3cd47739-0fa9-4321-aff1-220f8721a0b3\") " Dec 10 23:00:48 crc kubenswrapper[4791]: I1210 23:00:48.131514 4791 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/3cd47739-0fa9-4321-aff1-220f8721a0b3-run-ovn" (OuterVolumeSpecName: "run-ovn") pod "3cd47739-0fa9-4321-aff1-220f8721a0b3" (UID: "3cd47739-0fa9-4321-aff1-220f8721a0b3"). InnerVolumeSpecName "run-ovn". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 10 23:00:48 crc kubenswrapper[4791]: I1210 23:00:48.131592 4791 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"host-cni-netd\" (UniqueName: \"kubernetes.io/host-path/3cd47739-0fa9-4321-aff1-220f8721a0b3-host-cni-netd\") pod \"3cd47739-0fa9-4321-aff1-220f8721a0b3\" (UID: \"3cd47739-0fa9-4321-aff1-220f8721a0b3\") " Dec 10 23:00:48 crc kubenswrapper[4791]: I1210 23:00:48.131608 4791 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/3cd47739-0fa9-4321-aff1-220f8721a0b3-env-overrides" (OuterVolumeSpecName: "env-overrides") pod "3cd47739-0fa9-4321-aff1-220f8721a0b3" (UID: "3cd47739-0fa9-4321-aff1-220f8721a0b3"). InnerVolumeSpecName "env-overrides". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 10 23:00:48 crc kubenswrapper[4791]: I1210 23:00:48.131626 4791 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etc-openvswitch\" (UniqueName: \"kubernetes.io/host-path/3cd47739-0fa9-4321-aff1-220f8721a0b3-etc-openvswitch\") pod \"3cd47739-0fa9-4321-aff1-220f8721a0b3\" (UID: \"3cd47739-0fa9-4321-aff1-220f8721a0b3\") " Dec 10 23:00:48 crc kubenswrapper[4791]: I1210 23:00:48.131641 4791 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/3cd47739-0fa9-4321-aff1-220f8721a0b3-log-socket" (OuterVolumeSpecName: "log-socket") pod "3cd47739-0fa9-4321-aff1-220f8721a0b3" (UID: "3cd47739-0fa9-4321-aff1-220f8721a0b3"). InnerVolumeSpecName "log-socket". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 10 23:00:48 crc kubenswrapper[4791]: I1210 23:00:48.131671 4791 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/3cd47739-0fa9-4321-aff1-220f8721a0b3-etc-openvswitch" (OuterVolumeSpecName: "etc-openvswitch") pod "3cd47739-0fa9-4321-aff1-220f8721a0b3" (UID: "3cd47739-0fa9-4321-aff1-220f8721a0b3"). InnerVolumeSpecName "etc-openvswitch". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 10 23:00:48 crc kubenswrapper[4791]: I1210 23:00:48.131690 4791 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/3cd47739-0fa9-4321-aff1-220f8721a0b3-host-run-ovn-kubernetes" (OuterVolumeSpecName: "host-run-ovn-kubernetes") pod "3cd47739-0fa9-4321-aff1-220f8721a0b3" (UID: "3cd47739-0fa9-4321-aff1-220f8721a0b3"). InnerVolumeSpecName "host-run-ovn-kubernetes". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 10 23:00:48 crc kubenswrapper[4791]: I1210 23:00:48.131700 4791 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/3cd47739-0fa9-4321-aff1-220f8721a0b3-host-cni-netd" (OuterVolumeSpecName: "host-cni-netd") pod "3cd47739-0fa9-4321-aff1-220f8721a0b3" (UID: "3cd47739-0fa9-4321-aff1-220f8721a0b3"). InnerVolumeSpecName "host-cni-netd". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 10 23:00:48 crc kubenswrapper[4791]: I1210 23:00:48.131730 4791 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"systemd-units\" (UniqueName: \"kubernetes.io/host-path/3cd47739-0fa9-4321-aff1-220f8721a0b3-systemd-units\") pod \"3cd47739-0fa9-4321-aff1-220f8721a0b3\" (UID: \"3cd47739-0fa9-4321-aff1-220f8721a0b3\") " Dec 10 23:00:48 crc kubenswrapper[4791]: I1210 23:00:48.131783 4791 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/3cd47739-0fa9-4321-aff1-220f8721a0b3-systemd-units" (OuterVolumeSpecName: "systemd-units") pod "3cd47739-0fa9-4321-aff1-220f8721a0b3" (UID: "3cd47739-0fa9-4321-aff1-220f8721a0b3"). InnerVolumeSpecName "systemd-units". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 10 23:00:48 crc kubenswrapper[4791]: I1210 23:00:48.131829 4791 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"node-log\" (UniqueName: \"kubernetes.io/host-path/3cd47739-0fa9-4321-aff1-220f8721a0b3-node-log\") pod \"3cd47739-0fa9-4321-aff1-220f8721a0b3\" (UID: \"3cd47739-0fa9-4321-aff1-220f8721a0b3\") " Dec 10 23:00:48 crc kubenswrapper[4791]: I1210 23:00:48.132008 4791 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/3cd47739-0fa9-4321-aff1-220f8721a0b3-node-log" (OuterVolumeSpecName: "node-log") pod "3cd47739-0fa9-4321-aff1-220f8721a0b3" (UID: "3cd47739-0fa9-4321-aff1-220f8721a0b3"). InnerVolumeSpecName "node-log". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 10 23:00:48 crc kubenswrapper[4791]: I1210 23:00:48.132038 4791 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/3cd47739-0fa9-4321-aff1-220f8721a0b3-ovnkube-script-lib" (OuterVolumeSpecName: "ovnkube-script-lib") pod "3cd47739-0fa9-4321-aff1-220f8721a0b3" (UID: "3cd47739-0fa9-4321-aff1-220f8721a0b3"). InnerVolumeSpecName "ovnkube-script-lib". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 10 23:00:48 crc kubenswrapper[4791]: I1210 23:00:48.132122 4791 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-var-lib-cni-networks-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/b1289472-e6e4-468c-a54c-0a83ff1474e8-host-var-lib-cni-networks-ovn-kubernetes\") pod \"ovnkube-node-bdhcv\" (UID: \"b1289472-e6e4-468c-a54c-0a83ff1474e8\") " pod="openshift-ovn-kubernetes/ovnkube-node-bdhcv" Dec 10 23:00:48 crc kubenswrapper[4791]: I1210 23:00:48.132191 4791 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-lib-openvswitch\" (UniqueName: \"kubernetes.io/host-path/b1289472-e6e4-468c-a54c-0a83ff1474e8-var-lib-openvswitch\") pod \"ovnkube-node-bdhcv\" (UID: \"b1289472-e6e4-468c-a54c-0a83ff1474e8\") " pod="openshift-ovn-kubernetes/ovnkube-node-bdhcv" Dec 10 23:00:48 crc kubenswrapper[4791]: I1210 23:00:48.132311 4791 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovnkube-script-lib\" (UniqueName: \"kubernetes.io/configmap/b1289472-e6e4-468c-a54c-0a83ff1474e8-ovnkube-script-lib\") pod \"ovnkube-node-bdhcv\" (UID: \"b1289472-e6e4-468c-a54c-0a83ff1474e8\") " pod="openshift-ovn-kubernetes/ovnkube-node-bdhcv" Dec 10 23:00:48 crc kubenswrapper[4791]: I1210 23:00:48.132394 4791 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/b1289472-e6e4-468c-a54c-0a83ff1474e8-ovnkube-config\") pod \"ovnkube-node-bdhcv\" (UID: \"b1289472-e6e4-468c-a54c-0a83ff1474e8\") " pod="openshift-ovn-kubernetes/ovnkube-node-bdhcv" Dec 10 23:00:48 crc kubenswrapper[4791]: I1210 23:00:48.132432 4791 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-kubelet\" (UniqueName: \"kubernetes.io/host-path/b1289472-e6e4-468c-a54c-0a83ff1474e8-host-kubelet\") pod \"ovnkube-node-bdhcv\" (UID: \"b1289472-e6e4-468c-a54c-0a83ff1474e8\") " pod="openshift-ovn-kubernetes/ovnkube-node-bdhcv" Dec 10 23:00:48 crc kubenswrapper[4791]: I1210 23:00:48.132464 4791 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"log-socket\" (UniqueName: \"kubernetes.io/host-path/b1289472-e6e4-468c-a54c-0a83ff1474e8-log-socket\") pod \"ovnkube-node-bdhcv\" (UID: \"b1289472-e6e4-468c-a54c-0a83ff1474e8\") " pod="openshift-ovn-kubernetes/ovnkube-node-bdhcv" Dec 10 23:00:48 crc kubenswrapper[4791]: I1210 23:00:48.132500 4791 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/b1289472-e6e4-468c-a54c-0a83ff1474e8-env-overrides\") pod \"ovnkube-node-bdhcv\" (UID: \"b1289472-e6e4-468c-a54c-0a83ff1474e8\") " pod="openshift-ovn-kubernetes/ovnkube-node-bdhcv" Dec 10 23:00:48 crc kubenswrapper[4791]: I1210 23:00:48.132533 4791 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-openvswitch\" (UniqueName: \"kubernetes.io/host-path/b1289472-e6e4-468c-a54c-0a83ff1474e8-etc-openvswitch\") pod \"ovnkube-node-bdhcv\" (UID: \"b1289472-e6e4-468c-a54c-0a83ff1474e8\") " pod="openshift-ovn-kubernetes/ovnkube-node-bdhcv" Dec 10 23:00:48 crc kubenswrapper[4791]: I1210 23:00:48.132568 4791 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-run-netns\" (UniqueName: \"kubernetes.io/host-path/b1289472-e6e4-468c-a54c-0a83ff1474e8-host-run-netns\") pod \"ovnkube-node-bdhcv\" (UID: \"b1289472-e6e4-468c-a54c-0a83ff1474e8\") " pod="openshift-ovn-kubernetes/ovnkube-node-bdhcv" Dec 10 23:00:48 crc kubenswrapper[4791]: I1210 23:00:48.132598 4791 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-openvswitch\" (UniqueName: \"kubernetes.io/host-path/b1289472-e6e4-468c-a54c-0a83ff1474e8-run-openvswitch\") pod \"ovnkube-node-bdhcv\" (UID: \"b1289472-e6e4-468c-a54c-0a83ff1474e8\") " pod="openshift-ovn-kubernetes/ovnkube-node-bdhcv" Dec 10 23:00:48 crc kubenswrapper[4791]: I1210 23:00:48.132640 4791 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-cni-netd\" (UniqueName: \"kubernetes.io/host-path/b1289472-e6e4-468c-a54c-0a83ff1474e8-host-cni-netd\") pod \"ovnkube-node-bdhcv\" (UID: \"b1289472-e6e4-468c-a54c-0a83ff1474e8\") " pod="openshift-ovn-kubernetes/ovnkube-node-bdhcv" Dec 10 23:00:48 crc kubenswrapper[4791]: I1210 23:00:48.132672 4791 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-slash\" (UniqueName: \"kubernetes.io/host-path/b1289472-e6e4-468c-a54c-0a83ff1474e8-host-slash\") pod \"ovnkube-node-bdhcv\" (UID: \"b1289472-e6e4-468c-a54c-0a83ff1474e8\") " pod="openshift-ovn-kubernetes/ovnkube-node-bdhcv" Dec 10 23:00:48 crc kubenswrapper[4791]: I1210 23:00:48.132700 4791 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-systemd\" (UniqueName: \"kubernetes.io/host-path/b1289472-e6e4-468c-a54c-0a83ff1474e8-run-systemd\") pod \"ovnkube-node-bdhcv\" (UID: \"b1289472-e6e4-468c-a54c-0a83ff1474e8\") " pod="openshift-ovn-kubernetes/ovnkube-node-bdhcv" Dec 10 23:00:48 crc kubenswrapper[4791]: I1210 23:00:48.132736 4791 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-r9gbs\" (UniqueName: \"kubernetes.io/projected/b1289472-e6e4-468c-a54c-0a83ff1474e8-kube-api-access-r9gbs\") pod \"ovnkube-node-bdhcv\" (UID: \"b1289472-e6e4-468c-a54c-0a83ff1474e8\") " pod="openshift-ovn-kubernetes/ovnkube-node-bdhcv" Dec 10 23:00:48 crc kubenswrapper[4791]: I1210 23:00:48.132776 4791 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"systemd-units\" (UniqueName: \"kubernetes.io/host-path/b1289472-e6e4-468c-a54c-0a83ff1474e8-systemd-units\") pod \"ovnkube-node-bdhcv\" (UID: \"b1289472-e6e4-468c-a54c-0a83ff1474e8\") " pod="openshift-ovn-kubernetes/ovnkube-node-bdhcv" Dec 10 23:00:48 crc kubenswrapper[4791]: I1210 23:00:48.132814 4791 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-run-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/b1289472-e6e4-468c-a54c-0a83ff1474e8-host-run-ovn-kubernetes\") pod \"ovnkube-node-bdhcv\" (UID: \"b1289472-e6e4-468c-a54c-0a83ff1474e8\") " pod="openshift-ovn-kubernetes/ovnkube-node-bdhcv" Dec 10 23:00:48 crc kubenswrapper[4791]: I1210 23:00:48.132906 4791 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovn-node-metrics-cert\" (UniqueName: \"kubernetes.io/secret/b1289472-e6e4-468c-a54c-0a83ff1474e8-ovn-node-metrics-cert\") pod \"ovnkube-node-bdhcv\" (UID: \"b1289472-e6e4-468c-a54c-0a83ff1474e8\") " pod="openshift-ovn-kubernetes/ovnkube-node-bdhcv" Dec 10 23:00:48 crc kubenswrapper[4791]: I1210 23:00:48.133011 4791 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-ovn\" (UniqueName: \"kubernetes.io/host-path/b1289472-e6e4-468c-a54c-0a83ff1474e8-run-ovn\") pod \"ovnkube-node-bdhcv\" (UID: \"b1289472-e6e4-468c-a54c-0a83ff1474e8\") " pod="openshift-ovn-kubernetes/ovnkube-node-bdhcv" Dec 10 23:00:48 crc kubenswrapper[4791]: I1210 23:00:48.133063 4791 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-cni-bin\" (UniqueName: \"kubernetes.io/host-path/b1289472-e6e4-468c-a54c-0a83ff1474e8-host-cni-bin\") pod \"ovnkube-node-bdhcv\" (UID: \"b1289472-e6e4-468c-a54c-0a83ff1474e8\") " pod="openshift-ovn-kubernetes/ovnkube-node-bdhcv" Dec 10 23:00:48 crc kubenswrapper[4791]: I1210 23:00:48.133101 4791 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"node-log\" (UniqueName: \"kubernetes.io/host-path/b1289472-e6e4-468c-a54c-0a83ff1474e8-node-log\") pod \"ovnkube-node-bdhcv\" (UID: \"b1289472-e6e4-468c-a54c-0a83ff1474e8\") " pod="openshift-ovn-kubernetes/ovnkube-node-bdhcv" Dec 10 23:00:48 crc kubenswrapper[4791]: I1210 23:00:48.133228 4791 reconciler_common.go:293] "Volume detached for volume \"var-lib-openvswitch\" (UniqueName: \"kubernetes.io/host-path/3cd47739-0fa9-4321-aff1-220f8721a0b3-var-lib-openvswitch\") on node \"crc\" DevicePath \"\"" Dec 10 23:00:48 crc kubenswrapper[4791]: I1210 23:00:48.133259 4791 reconciler_common.go:293] "Volume detached for volume \"run-openvswitch\" (UniqueName: \"kubernetes.io/host-path/3cd47739-0fa9-4321-aff1-220f8721a0b3-run-openvswitch\") on node \"crc\" DevicePath \"\"" Dec 10 23:00:48 crc kubenswrapper[4791]: I1210 23:00:48.133277 4791 reconciler_common.go:293] "Volume detached for volume \"host-cni-bin\" (UniqueName: \"kubernetes.io/host-path/3cd47739-0fa9-4321-aff1-220f8721a0b3-host-cni-bin\") on node \"crc\" DevicePath \"\"" Dec 10 23:00:48 crc kubenswrapper[4791]: I1210 23:00:48.133296 4791 reconciler_common.go:293] "Volume detached for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/3cd47739-0fa9-4321-aff1-220f8721a0b3-env-overrides\") on node \"crc\" DevicePath \"\"" Dec 10 23:00:48 crc kubenswrapper[4791]: I1210 23:00:48.133313 4791 reconciler_common.go:293] "Volume detached for volume \"host-run-netns\" (UniqueName: \"kubernetes.io/host-path/3cd47739-0fa9-4321-aff1-220f8721a0b3-host-run-netns\") on node \"crc\" DevicePath \"\"" Dec 10 23:00:48 crc kubenswrapper[4791]: I1210 23:00:48.133329 4791 reconciler_common.go:293] "Volume detached for volume \"host-kubelet\" (UniqueName: \"kubernetes.io/host-path/3cd47739-0fa9-4321-aff1-220f8721a0b3-host-kubelet\") on node \"crc\" DevicePath \"\"" Dec 10 23:00:48 crc kubenswrapper[4791]: I1210 23:00:48.133569 4791 reconciler_common.go:293] "Volume detached for volume \"ovnkube-script-lib\" (UniqueName: \"kubernetes.io/configmap/3cd47739-0fa9-4321-aff1-220f8721a0b3-ovnkube-script-lib\") on node \"crc\" DevicePath \"\"" Dec 10 23:00:48 crc kubenswrapper[4791]: I1210 23:00:48.133597 4791 reconciler_common.go:293] "Volume detached for volume \"host-var-lib-cni-networks-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/3cd47739-0fa9-4321-aff1-220f8721a0b3-host-var-lib-cni-networks-ovn-kubernetes\") on node \"crc\" DevicePath \"\"" Dec 10 23:00:48 crc kubenswrapper[4791]: I1210 23:00:48.133621 4791 reconciler_common.go:293] "Volume detached for volume \"host-slash\" (UniqueName: \"kubernetes.io/host-path/3cd47739-0fa9-4321-aff1-220f8721a0b3-host-slash\") on node \"crc\" DevicePath \"\"" Dec 10 23:00:48 crc kubenswrapper[4791]: I1210 23:00:48.133644 4791 reconciler_common.go:293] "Volume detached for volume \"run-ovn\" (UniqueName: \"kubernetes.io/host-path/3cd47739-0fa9-4321-aff1-220f8721a0b3-run-ovn\") on node \"crc\" DevicePath \"\"" Dec 10 23:00:48 crc kubenswrapper[4791]: I1210 23:00:48.133665 4791 reconciler_common.go:293] "Volume detached for volume \"log-socket\" (UniqueName: \"kubernetes.io/host-path/3cd47739-0fa9-4321-aff1-220f8721a0b3-log-socket\") on node \"crc\" DevicePath \"\"" Dec 10 23:00:48 crc kubenswrapper[4791]: I1210 23:00:48.133687 4791 reconciler_common.go:293] "Volume detached for volume \"host-run-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/3cd47739-0fa9-4321-aff1-220f8721a0b3-host-run-ovn-kubernetes\") on node \"crc\" DevicePath \"\"" Dec 10 23:00:48 crc kubenswrapper[4791]: I1210 23:00:48.133710 4791 reconciler_common.go:293] "Volume detached for volume \"host-cni-netd\" (UniqueName: \"kubernetes.io/host-path/3cd47739-0fa9-4321-aff1-220f8721a0b3-host-cni-netd\") on node \"crc\" DevicePath \"\"" Dec 10 23:00:48 crc kubenswrapper[4791]: I1210 23:00:48.133730 4791 reconciler_common.go:293] "Volume detached for volume \"etc-openvswitch\" (UniqueName: \"kubernetes.io/host-path/3cd47739-0fa9-4321-aff1-220f8721a0b3-etc-openvswitch\") on node \"crc\" DevicePath \"\"" Dec 10 23:00:48 crc kubenswrapper[4791]: I1210 23:00:48.133746 4791 reconciler_common.go:293] "Volume detached for volume \"systemd-units\" (UniqueName: \"kubernetes.io/host-path/3cd47739-0fa9-4321-aff1-220f8721a0b3-systemd-units\") on node \"crc\" DevicePath \"\"" Dec 10 23:00:48 crc kubenswrapper[4791]: I1210 23:00:48.133763 4791 reconciler_common.go:293] "Volume detached for volume \"node-log\" (UniqueName: \"kubernetes.io/host-path/3cd47739-0fa9-4321-aff1-220f8721a0b3-node-log\") on node \"crc\" DevicePath \"\"" Dec 10 23:00:48 crc kubenswrapper[4791]: I1210 23:00:48.133784 4791 reconciler_common.go:293] "Volume detached for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/3cd47739-0fa9-4321-aff1-220f8721a0b3-ovnkube-config\") on node \"crc\" DevicePath \"\"" Dec 10 23:00:48 crc kubenswrapper[4791]: I1210 23:00:48.136840 4791 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/3cd47739-0fa9-4321-aff1-220f8721a0b3-kube-api-access-wvpw8" (OuterVolumeSpecName: "kube-api-access-wvpw8") pod "3cd47739-0fa9-4321-aff1-220f8721a0b3" (UID: "3cd47739-0fa9-4321-aff1-220f8721a0b3"). InnerVolumeSpecName "kube-api-access-wvpw8". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 10 23:00:48 crc kubenswrapper[4791]: I1210 23:00:48.137771 4791 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/3cd47739-0fa9-4321-aff1-220f8721a0b3-ovn-node-metrics-cert" (OuterVolumeSpecName: "ovn-node-metrics-cert") pod "3cd47739-0fa9-4321-aff1-220f8721a0b3" (UID: "3cd47739-0fa9-4321-aff1-220f8721a0b3"). InnerVolumeSpecName "ovn-node-metrics-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 10 23:00:48 crc kubenswrapper[4791]: I1210 23:00:48.146859 4791 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/3cd47739-0fa9-4321-aff1-220f8721a0b3-run-systemd" (OuterVolumeSpecName: "run-systemd") pod "3cd47739-0fa9-4321-aff1-220f8721a0b3" (UID: "3cd47739-0fa9-4321-aff1-220f8721a0b3"). InnerVolumeSpecName "run-systemd". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 10 23:00:48 crc kubenswrapper[4791]: I1210 23:00:48.234666 4791 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"node-log\" (UniqueName: \"kubernetes.io/host-path/b1289472-e6e4-468c-a54c-0a83ff1474e8-node-log\") pod \"ovnkube-node-bdhcv\" (UID: \"b1289472-e6e4-468c-a54c-0a83ff1474e8\") " pod="openshift-ovn-kubernetes/ovnkube-node-bdhcv" Dec 10 23:00:48 crc kubenswrapper[4791]: I1210 23:00:48.234725 4791 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-var-lib-cni-networks-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/b1289472-e6e4-468c-a54c-0a83ff1474e8-host-var-lib-cni-networks-ovn-kubernetes\") pod \"ovnkube-node-bdhcv\" (UID: \"b1289472-e6e4-468c-a54c-0a83ff1474e8\") " pod="openshift-ovn-kubernetes/ovnkube-node-bdhcv" Dec 10 23:00:48 crc kubenswrapper[4791]: I1210 23:00:48.234751 4791 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-lib-openvswitch\" (UniqueName: \"kubernetes.io/host-path/b1289472-e6e4-468c-a54c-0a83ff1474e8-var-lib-openvswitch\") pod \"ovnkube-node-bdhcv\" (UID: \"b1289472-e6e4-468c-a54c-0a83ff1474e8\") " pod="openshift-ovn-kubernetes/ovnkube-node-bdhcv" Dec 10 23:00:48 crc kubenswrapper[4791]: I1210 23:00:48.234779 4791 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovnkube-script-lib\" (UniqueName: \"kubernetes.io/configmap/b1289472-e6e4-468c-a54c-0a83ff1474e8-ovnkube-script-lib\") pod \"ovnkube-node-bdhcv\" (UID: \"b1289472-e6e4-468c-a54c-0a83ff1474e8\") " pod="openshift-ovn-kubernetes/ovnkube-node-bdhcv" Dec 10 23:00:48 crc kubenswrapper[4791]: I1210 23:00:48.234794 4791 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/b1289472-e6e4-468c-a54c-0a83ff1474e8-ovnkube-config\") pod \"ovnkube-node-bdhcv\" (UID: \"b1289472-e6e4-468c-a54c-0a83ff1474e8\") " pod="openshift-ovn-kubernetes/ovnkube-node-bdhcv" Dec 10 23:00:48 crc kubenswrapper[4791]: I1210 23:00:48.234810 4791 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-kubelet\" (UniqueName: \"kubernetes.io/host-path/b1289472-e6e4-468c-a54c-0a83ff1474e8-host-kubelet\") pod \"ovnkube-node-bdhcv\" (UID: \"b1289472-e6e4-468c-a54c-0a83ff1474e8\") " pod="openshift-ovn-kubernetes/ovnkube-node-bdhcv" Dec 10 23:00:48 crc kubenswrapper[4791]: I1210 23:00:48.234816 4791 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-lib-openvswitch\" (UniqueName: \"kubernetes.io/host-path/b1289472-e6e4-468c-a54c-0a83ff1474e8-var-lib-openvswitch\") pod \"ovnkube-node-bdhcv\" (UID: \"b1289472-e6e4-468c-a54c-0a83ff1474e8\") " pod="openshift-ovn-kubernetes/ovnkube-node-bdhcv" Dec 10 23:00:48 crc kubenswrapper[4791]: I1210 23:00:48.234847 4791 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"log-socket\" (UniqueName: \"kubernetes.io/host-path/b1289472-e6e4-468c-a54c-0a83ff1474e8-log-socket\") pod \"ovnkube-node-bdhcv\" (UID: \"b1289472-e6e4-468c-a54c-0a83ff1474e8\") " pod="openshift-ovn-kubernetes/ovnkube-node-bdhcv" Dec 10 23:00:48 crc kubenswrapper[4791]: I1210 23:00:48.234825 4791 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"log-socket\" (UniqueName: \"kubernetes.io/host-path/b1289472-e6e4-468c-a54c-0a83ff1474e8-log-socket\") pod \"ovnkube-node-bdhcv\" (UID: \"b1289472-e6e4-468c-a54c-0a83ff1474e8\") " pod="openshift-ovn-kubernetes/ovnkube-node-bdhcv" Dec 10 23:00:48 crc kubenswrapper[4791]: I1210 23:00:48.234874 4791 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-var-lib-cni-networks-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/b1289472-e6e4-468c-a54c-0a83ff1474e8-host-var-lib-cni-networks-ovn-kubernetes\") pod \"ovnkube-node-bdhcv\" (UID: \"b1289472-e6e4-468c-a54c-0a83ff1474e8\") " pod="openshift-ovn-kubernetes/ovnkube-node-bdhcv" Dec 10 23:00:48 crc kubenswrapper[4791]: I1210 23:00:48.234899 4791 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/b1289472-e6e4-468c-a54c-0a83ff1474e8-env-overrides\") pod \"ovnkube-node-bdhcv\" (UID: \"b1289472-e6e4-468c-a54c-0a83ff1474e8\") " pod="openshift-ovn-kubernetes/ovnkube-node-bdhcv" Dec 10 23:00:48 crc kubenswrapper[4791]: I1210 23:00:48.234998 4791 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-openvswitch\" (UniqueName: \"kubernetes.io/host-path/b1289472-e6e4-468c-a54c-0a83ff1474e8-etc-openvswitch\") pod \"ovnkube-node-bdhcv\" (UID: \"b1289472-e6e4-468c-a54c-0a83ff1474e8\") " pod="openshift-ovn-kubernetes/ovnkube-node-bdhcv" Dec 10 23:00:48 crc kubenswrapper[4791]: I1210 23:00:48.235046 4791 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-openvswitch\" (UniqueName: \"kubernetes.io/host-path/b1289472-e6e4-468c-a54c-0a83ff1474e8-run-openvswitch\") pod \"ovnkube-node-bdhcv\" (UID: \"b1289472-e6e4-468c-a54c-0a83ff1474e8\") " pod="openshift-ovn-kubernetes/ovnkube-node-bdhcv" Dec 10 23:00:48 crc kubenswrapper[4791]: I1210 23:00:48.235084 4791 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-run-netns\" (UniqueName: \"kubernetes.io/host-path/b1289472-e6e4-468c-a54c-0a83ff1474e8-host-run-netns\") pod \"ovnkube-node-bdhcv\" (UID: \"b1289472-e6e4-468c-a54c-0a83ff1474e8\") " pod="openshift-ovn-kubernetes/ovnkube-node-bdhcv" Dec 10 23:00:48 crc kubenswrapper[4791]: I1210 23:00:48.235128 4791 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-cni-netd\" (UniqueName: \"kubernetes.io/host-path/b1289472-e6e4-468c-a54c-0a83ff1474e8-host-cni-netd\") pod \"ovnkube-node-bdhcv\" (UID: \"b1289472-e6e4-468c-a54c-0a83ff1474e8\") " pod="openshift-ovn-kubernetes/ovnkube-node-bdhcv" Dec 10 23:00:48 crc kubenswrapper[4791]: I1210 23:00:48.235170 4791 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-slash\" (UniqueName: \"kubernetes.io/host-path/b1289472-e6e4-468c-a54c-0a83ff1474e8-host-slash\") pod \"ovnkube-node-bdhcv\" (UID: \"b1289472-e6e4-468c-a54c-0a83ff1474e8\") " pod="openshift-ovn-kubernetes/ovnkube-node-bdhcv" Dec 10 23:00:48 crc kubenswrapper[4791]: I1210 23:00:48.235210 4791 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-systemd\" (UniqueName: \"kubernetes.io/host-path/b1289472-e6e4-468c-a54c-0a83ff1474e8-run-systemd\") pod \"ovnkube-node-bdhcv\" (UID: \"b1289472-e6e4-468c-a54c-0a83ff1474e8\") " pod="openshift-ovn-kubernetes/ovnkube-node-bdhcv" Dec 10 23:00:48 crc kubenswrapper[4791]: I1210 23:00:48.235306 4791 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-r9gbs\" (UniqueName: \"kubernetes.io/projected/b1289472-e6e4-468c-a54c-0a83ff1474e8-kube-api-access-r9gbs\") pod \"ovnkube-node-bdhcv\" (UID: \"b1289472-e6e4-468c-a54c-0a83ff1474e8\") " pod="openshift-ovn-kubernetes/ovnkube-node-bdhcv" Dec 10 23:00:48 crc kubenswrapper[4791]: I1210 23:00:48.235395 4791 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"systemd-units\" (UniqueName: \"kubernetes.io/host-path/b1289472-e6e4-468c-a54c-0a83ff1474e8-systemd-units\") pod \"ovnkube-node-bdhcv\" (UID: \"b1289472-e6e4-468c-a54c-0a83ff1474e8\") " pod="openshift-ovn-kubernetes/ovnkube-node-bdhcv" Dec 10 23:00:48 crc kubenswrapper[4791]: I1210 23:00:48.234771 4791 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"node-log\" (UniqueName: \"kubernetes.io/host-path/b1289472-e6e4-468c-a54c-0a83ff1474e8-node-log\") pod \"ovnkube-node-bdhcv\" (UID: \"b1289472-e6e4-468c-a54c-0a83ff1474e8\") " pod="openshift-ovn-kubernetes/ovnkube-node-bdhcv" Dec 10 23:00:48 crc kubenswrapper[4791]: I1210 23:00:48.235477 4791 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-run-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/b1289472-e6e4-468c-a54c-0a83ff1474e8-host-run-ovn-kubernetes\") pod \"ovnkube-node-bdhcv\" (UID: \"b1289472-e6e4-468c-a54c-0a83ff1474e8\") " pod="openshift-ovn-kubernetes/ovnkube-node-bdhcv" Dec 10 23:00:48 crc kubenswrapper[4791]: I1210 23:00:48.235539 4791 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovn-node-metrics-cert\" (UniqueName: \"kubernetes.io/secret/b1289472-e6e4-468c-a54c-0a83ff1474e8-ovn-node-metrics-cert\") pod \"ovnkube-node-bdhcv\" (UID: \"b1289472-e6e4-468c-a54c-0a83ff1474e8\") " pod="openshift-ovn-kubernetes/ovnkube-node-bdhcv" Dec 10 23:00:48 crc kubenswrapper[4791]: I1210 23:00:48.235550 4791 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovnkube-script-lib\" (UniqueName: \"kubernetes.io/configmap/b1289472-e6e4-468c-a54c-0a83ff1474e8-ovnkube-script-lib\") pod \"ovnkube-node-bdhcv\" (UID: \"b1289472-e6e4-468c-a54c-0a83ff1474e8\") " pod="openshift-ovn-kubernetes/ovnkube-node-bdhcv" Dec 10 23:00:48 crc kubenswrapper[4791]: I1210 23:00:48.235544 4791 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-kubelet\" (UniqueName: \"kubernetes.io/host-path/b1289472-e6e4-468c-a54c-0a83ff1474e8-host-kubelet\") pod \"ovnkube-node-bdhcv\" (UID: \"b1289472-e6e4-468c-a54c-0a83ff1474e8\") " pod="openshift-ovn-kubernetes/ovnkube-node-bdhcv" Dec 10 23:00:48 crc kubenswrapper[4791]: I1210 23:00:48.235600 4791 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-ovn\" (UniqueName: \"kubernetes.io/host-path/b1289472-e6e4-468c-a54c-0a83ff1474e8-run-ovn\") pod \"ovnkube-node-bdhcv\" (UID: \"b1289472-e6e4-468c-a54c-0a83ff1474e8\") " pod="openshift-ovn-kubernetes/ovnkube-node-bdhcv" Dec 10 23:00:48 crc kubenswrapper[4791]: I1210 23:00:48.235629 4791 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"systemd-units\" (UniqueName: \"kubernetes.io/host-path/b1289472-e6e4-468c-a54c-0a83ff1474e8-systemd-units\") pod \"ovnkube-node-bdhcv\" (UID: \"b1289472-e6e4-468c-a54c-0a83ff1474e8\") " pod="openshift-ovn-kubernetes/ovnkube-node-bdhcv" Dec 10 23:00:48 crc kubenswrapper[4791]: I1210 23:00:48.235662 4791 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-systemd\" (UniqueName: \"kubernetes.io/host-path/b1289472-e6e4-468c-a54c-0a83ff1474e8-run-systemd\") pod \"ovnkube-node-bdhcv\" (UID: \"b1289472-e6e4-468c-a54c-0a83ff1474e8\") " pod="openshift-ovn-kubernetes/ovnkube-node-bdhcv" Dec 10 23:00:48 crc kubenswrapper[4791]: I1210 23:00:48.235650 4791 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-cni-bin\" (UniqueName: \"kubernetes.io/host-path/b1289472-e6e4-468c-a54c-0a83ff1474e8-host-cni-bin\") pod \"ovnkube-node-bdhcv\" (UID: \"b1289472-e6e4-468c-a54c-0a83ff1474e8\") " pod="openshift-ovn-kubernetes/ovnkube-node-bdhcv" Dec 10 23:00:48 crc kubenswrapper[4791]: I1210 23:00:48.235705 4791 reconciler_common.go:293] "Volume detached for volume \"run-systemd\" (UniqueName: \"kubernetes.io/host-path/3cd47739-0fa9-4321-aff1-220f8721a0b3-run-systemd\") on node \"crc\" DevicePath \"\"" Dec 10 23:00:48 crc kubenswrapper[4791]: I1210 23:00:48.235707 4791 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/b1289472-e6e4-468c-a54c-0a83ff1474e8-env-overrides\") pod \"ovnkube-node-bdhcv\" (UID: \"b1289472-e6e4-468c-a54c-0a83ff1474e8\") " pod="openshift-ovn-kubernetes/ovnkube-node-bdhcv" Dec 10 23:00:48 crc kubenswrapper[4791]: I1210 23:00:48.235708 4791 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-openvswitch\" (UniqueName: \"kubernetes.io/host-path/b1289472-e6e4-468c-a54c-0a83ff1474e8-run-openvswitch\") pod \"ovnkube-node-bdhcv\" (UID: \"b1289472-e6e4-468c-a54c-0a83ff1474e8\") " pod="openshift-ovn-kubernetes/ovnkube-node-bdhcv" Dec 10 23:00:48 crc kubenswrapper[4791]: I1210 23:00:48.235755 4791 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-run-netns\" (UniqueName: \"kubernetes.io/host-path/b1289472-e6e4-468c-a54c-0a83ff1474e8-host-run-netns\") pod \"ovnkube-node-bdhcv\" (UID: \"b1289472-e6e4-468c-a54c-0a83ff1474e8\") " pod="openshift-ovn-kubernetes/ovnkube-node-bdhcv" Dec 10 23:00:48 crc kubenswrapper[4791]: I1210 23:00:48.235779 4791 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-cni-netd\" (UniqueName: \"kubernetes.io/host-path/b1289472-e6e4-468c-a54c-0a83ff1474e8-host-cni-netd\") pod \"ovnkube-node-bdhcv\" (UID: \"b1289472-e6e4-468c-a54c-0a83ff1474e8\") " pod="openshift-ovn-kubernetes/ovnkube-node-bdhcv" Dec 10 23:00:48 crc kubenswrapper[4791]: I1210 23:00:48.235601 4791 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-slash\" (UniqueName: \"kubernetes.io/host-path/b1289472-e6e4-468c-a54c-0a83ff1474e8-host-slash\") pod \"ovnkube-node-bdhcv\" (UID: \"b1289472-e6e4-468c-a54c-0a83ff1474e8\") " pod="openshift-ovn-kubernetes/ovnkube-node-bdhcv" Dec 10 23:00:48 crc kubenswrapper[4791]: I1210 23:00:48.235802 4791 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-ovn\" (UniqueName: \"kubernetes.io/host-path/b1289472-e6e4-468c-a54c-0a83ff1474e8-run-ovn\") pod \"ovnkube-node-bdhcv\" (UID: \"b1289472-e6e4-468c-a54c-0a83ff1474e8\") " pod="openshift-ovn-kubernetes/ovnkube-node-bdhcv" Dec 10 23:00:48 crc kubenswrapper[4791]: I1210 23:00:48.235763 4791 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-run-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/b1289472-e6e4-468c-a54c-0a83ff1474e8-host-run-ovn-kubernetes\") pod \"ovnkube-node-bdhcv\" (UID: \"b1289472-e6e4-468c-a54c-0a83ff1474e8\") " pod="openshift-ovn-kubernetes/ovnkube-node-bdhcv" Dec 10 23:00:48 crc kubenswrapper[4791]: I1210 23:00:48.235709 4791 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-cni-bin\" (UniqueName: \"kubernetes.io/host-path/b1289472-e6e4-468c-a54c-0a83ff1474e8-host-cni-bin\") pod \"ovnkube-node-bdhcv\" (UID: \"b1289472-e6e4-468c-a54c-0a83ff1474e8\") " pod="openshift-ovn-kubernetes/ovnkube-node-bdhcv" Dec 10 23:00:48 crc kubenswrapper[4791]: I1210 23:00:48.235719 4791 reconciler_common.go:293] "Volume detached for volume \"ovn-node-metrics-cert\" (UniqueName: \"kubernetes.io/secret/3cd47739-0fa9-4321-aff1-220f8721a0b3-ovn-node-metrics-cert\") on node \"crc\" DevicePath \"\"" Dec 10 23:00:48 crc kubenswrapper[4791]: I1210 23:00:48.235950 4791 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-wvpw8\" (UniqueName: \"kubernetes.io/projected/3cd47739-0fa9-4321-aff1-220f8721a0b3-kube-api-access-wvpw8\") on node \"crc\" DevicePath \"\"" Dec 10 23:00:48 crc kubenswrapper[4791]: I1210 23:00:48.235651 4791 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-openvswitch\" (UniqueName: \"kubernetes.io/host-path/b1289472-e6e4-468c-a54c-0a83ff1474e8-etc-openvswitch\") pod \"ovnkube-node-bdhcv\" (UID: \"b1289472-e6e4-468c-a54c-0a83ff1474e8\") " pod="openshift-ovn-kubernetes/ovnkube-node-bdhcv" Dec 10 23:00:48 crc kubenswrapper[4791]: I1210 23:00:48.236293 4791 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/b1289472-e6e4-468c-a54c-0a83ff1474e8-ovnkube-config\") pod \"ovnkube-node-bdhcv\" (UID: \"b1289472-e6e4-468c-a54c-0a83ff1474e8\") " pod="openshift-ovn-kubernetes/ovnkube-node-bdhcv" Dec 10 23:00:48 crc kubenswrapper[4791]: I1210 23:00:48.239033 4791 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovn-node-metrics-cert\" (UniqueName: \"kubernetes.io/secret/b1289472-e6e4-468c-a54c-0a83ff1474e8-ovn-node-metrics-cert\") pod \"ovnkube-node-bdhcv\" (UID: \"b1289472-e6e4-468c-a54c-0a83ff1474e8\") " pod="openshift-ovn-kubernetes/ovnkube-node-bdhcv" Dec 10 23:00:48 crc kubenswrapper[4791]: I1210 23:00:48.263762 4791 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-r9gbs\" (UniqueName: \"kubernetes.io/projected/b1289472-e6e4-468c-a54c-0a83ff1474e8-kube-api-access-r9gbs\") pod \"ovnkube-node-bdhcv\" (UID: \"b1289472-e6e4-468c-a54c-0a83ff1474e8\") " pod="openshift-ovn-kubernetes/ovnkube-node-bdhcv" Dec 10 23:00:48 crc kubenswrapper[4791]: I1210 23:00:48.325700 4791 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-ovn-kubernetes/ovnkube-node-bdhcv" Dec 10 23:00:48 crc kubenswrapper[4791]: W1210 23:00:48.343786 4791 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podb1289472_e6e4_468c_a54c_0a83ff1474e8.slice/crio-d6c5ac17a2555d690273ca60f9410d1fae7af550073e3875986d976ed1de925a WatchSource:0}: Error finding container d6c5ac17a2555d690273ca60f9410d1fae7af550073e3875986d976ed1de925a: Status 404 returned error can't find the container with id d6c5ac17a2555d690273ca60f9410d1fae7af550073e3875986d976ed1de925a Dec 10 23:00:48 crc kubenswrapper[4791]: I1210 23:00:48.659133 4791 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-multus_multus-4nwkq_672aa28c-8169-49ed-87b8-21187d13a80c/kube-multus/2.log" Dec 10 23:00:48 crc kubenswrapper[4791]: I1210 23:00:48.659998 4791 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-multus_multus-4nwkq_672aa28c-8169-49ed-87b8-21187d13a80c/kube-multus/1.log" Dec 10 23:00:48 crc kubenswrapper[4791]: I1210 23:00:48.660136 4791 generic.go:334] "Generic (PLEG): container finished" podID="672aa28c-8169-49ed-87b8-21187d13a80c" containerID="ccdf9cf4480255bc2180847012278cee2a7f7e9392d14b7affe69a361b92cf4b" exitCode=2 Dec 10 23:00:48 crc kubenswrapper[4791]: I1210 23:00:48.660233 4791 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-4nwkq" event={"ID":"672aa28c-8169-49ed-87b8-21187d13a80c","Type":"ContainerDied","Data":"ccdf9cf4480255bc2180847012278cee2a7f7e9392d14b7affe69a361b92cf4b"} Dec 10 23:00:48 crc kubenswrapper[4791]: I1210 23:00:48.660286 4791 scope.go:117] "RemoveContainer" containerID="f337f881ffd71f4a2ee19e92bb0b92c76cbb9b8352511f741d0c87e7da4cdd18" Dec 10 23:00:48 crc kubenswrapper[4791]: I1210 23:00:48.661320 4791 scope.go:117] "RemoveContainer" containerID="ccdf9cf4480255bc2180847012278cee2a7f7e9392d14b7affe69a361b92cf4b" Dec 10 23:00:48 crc kubenswrapper[4791]: E1210 23:00:48.661778 4791 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"kube-multus\" with CrashLoopBackOff: \"back-off 20s restarting failed container=kube-multus pod=multus-4nwkq_openshift-multus(672aa28c-8169-49ed-87b8-21187d13a80c)\"" pod="openshift-multus/multus-4nwkq" podUID="672aa28c-8169-49ed-87b8-21187d13a80c" Dec 10 23:00:48 crc kubenswrapper[4791]: I1210 23:00:48.662761 4791 generic.go:334] "Generic (PLEG): container finished" podID="b1289472-e6e4-468c-a54c-0a83ff1474e8" containerID="185183fd44419c23f6a31c5763ce2a062c7336319cc110dfd1eae41185e5d03f" exitCode=0 Dec 10 23:00:48 crc kubenswrapper[4791]: I1210 23:00:48.662833 4791 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-bdhcv" event={"ID":"b1289472-e6e4-468c-a54c-0a83ff1474e8","Type":"ContainerDied","Data":"185183fd44419c23f6a31c5763ce2a062c7336319cc110dfd1eae41185e5d03f"} Dec 10 23:00:48 crc kubenswrapper[4791]: I1210 23:00:48.662856 4791 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-bdhcv" event={"ID":"b1289472-e6e4-468c-a54c-0a83ff1474e8","Type":"ContainerStarted","Data":"d6c5ac17a2555d690273ca60f9410d1fae7af550073e3875986d976ed1de925a"} Dec 10 23:00:48 crc kubenswrapper[4791]: I1210 23:00:48.665309 4791 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-zhq64_3cd47739-0fa9-4321-aff1-220f8721a0b3/ovnkube-controller/3.log" Dec 10 23:00:48 crc kubenswrapper[4791]: I1210 23:00:48.671542 4791 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-zhq64_3cd47739-0fa9-4321-aff1-220f8721a0b3/ovn-acl-logging/0.log" Dec 10 23:00:48 crc kubenswrapper[4791]: I1210 23:00:48.672295 4791 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-zhq64_3cd47739-0fa9-4321-aff1-220f8721a0b3/ovn-controller/0.log" Dec 10 23:00:48 crc kubenswrapper[4791]: I1210 23:00:48.675630 4791 generic.go:334] "Generic (PLEG): container finished" podID="3cd47739-0fa9-4321-aff1-220f8721a0b3" containerID="8839756dfbc89d902c809e3b2b4c8dd5de8ef52ec1009fb2aac5a06a16de4b66" exitCode=0 Dec 10 23:00:48 crc kubenswrapper[4791]: I1210 23:00:48.675657 4791 generic.go:334] "Generic (PLEG): container finished" podID="3cd47739-0fa9-4321-aff1-220f8721a0b3" containerID="dd3f80e5713221f96ed50da7f2e7fd5b9fabfcd5c1656da8d8a86d818a5fdc3f" exitCode=0 Dec 10 23:00:48 crc kubenswrapper[4791]: I1210 23:00:48.675664 4791 generic.go:334] "Generic (PLEG): container finished" podID="3cd47739-0fa9-4321-aff1-220f8721a0b3" containerID="152a7618be8f38fadf4e564b3948980503cedcdb466c37344ccc20f4a3173a5f" exitCode=0 Dec 10 23:00:48 crc kubenswrapper[4791]: I1210 23:00:48.675671 4791 generic.go:334] "Generic (PLEG): container finished" podID="3cd47739-0fa9-4321-aff1-220f8721a0b3" containerID="3d18f9cb7207efb1e32de75a30d33e024f656a08f0503db95e5488bc5280971f" exitCode=0 Dec 10 23:00:48 crc kubenswrapper[4791]: I1210 23:00:48.675677 4791 generic.go:334] "Generic (PLEG): container finished" podID="3cd47739-0fa9-4321-aff1-220f8721a0b3" containerID="bc721b530c08b433756e978ec0423cc18aabf6771b61eb96f903b4d57ea335ff" exitCode=0 Dec 10 23:00:48 crc kubenswrapper[4791]: I1210 23:00:48.675685 4791 generic.go:334] "Generic (PLEG): container finished" podID="3cd47739-0fa9-4321-aff1-220f8721a0b3" containerID="68795a29944e56870ffe94fa732e9dcf0bfb344c0d85d1e590f58a876129c61f" exitCode=0 Dec 10 23:00:48 crc kubenswrapper[4791]: I1210 23:00:48.675692 4791 generic.go:334] "Generic (PLEG): container finished" podID="3cd47739-0fa9-4321-aff1-220f8721a0b3" containerID="937137ae9d7480cdc7c59bc4a1df0a4c10e72b3b63693af285278789953ed218" exitCode=143 Dec 10 23:00:48 crc kubenswrapper[4791]: I1210 23:00:48.675700 4791 generic.go:334] "Generic (PLEG): container finished" podID="3cd47739-0fa9-4321-aff1-220f8721a0b3" containerID="1a52d19b7671e8b143017c8a981a6f8d3ee3b84650e4b223b814a531ac76f364" exitCode=143 Dec 10 23:00:48 crc kubenswrapper[4791]: I1210 23:00:48.675722 4791 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-zhq64" event={"ID":"3cd47739-0fa9-4321-aff1-220f8721a0b3","Type":"ContainerDied","Data":"8839756dfbc89d902c809e3b2b4c8dd5de8ef52ec1009fb2aac5a06a16de4b66"} Dec 10 23:00:48 crc kubenswrapper[4791]: I1210 23:00:48.675747 4791 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-zhq64" event={"ID":"3cd47739-0fa9-4321-aff1-220f8721a0b3","Type":"ContainerDied","Data":"dd3f80e5713221f96ed50da7f2e7fd5b9fabfcd5c1656da8d8a86d818a5fdc3f"} Dec 10 23:00:48 crc kubenswrapper[4791]: I1210 23:00:48.675758 4791 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-zhq64" event={"ID":"3cd47739-0fa9-4321-aff1-220f8721a0b3","Type":"ContainerDied","Data":"152a7618be8f38fadf4e564b3948980503cedcdb466c37344ccc20f4a3173a5f"} Dec 10 23:00:48 crc kubenswrapper[4791]: I1210 23:00:48.675767 4791 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-zhq64" event={"ID":"3cd47739-0fa9-4321-aff1-220f8721a0b3","Type":"ContainerDied","Data":"3d18f9cb7207efb1e32de75a30d33e024f656a08f0503db95e5488bc5280971f"} Dec 10 23:00:48 crc kubenswrapper[4791]: I1210 23:00:48.675777 4791 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-zhq64" event={"ID":"3cd47739-0fa9-4321-aff1-220f8721a0b3","Type":"ContainerDied","Data":"bc721b530c08b433756e978ec0423cc18aabf6771b61eb96f903b4d57ea335ff"} Dec 10 23:00:48 crc kubenswrapper[4791]: I1210 23:00:48.675787 4791 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-zhq64" event={"ID":"3cd47739-0fa9-4321-aff1-220f8721a0b3","Type":"ContainerDied","Data":"68795a29944e56870ffe94fa732e9dcf0bfb344c0d85d1e590f58a876129c61f"} Dec 10 23:00:48 crc kubenswrapper[4791]: I1210 23:00:48.675798 4791 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"8839756dfbc89d902c809e3b2b4c8dd5de8ef52ec1009fb2aac5a06a16de4b66"} Dec 10 23:00:48 crc kubenswrapper[4791]: I1210 23:00:48.675809 4791 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"5f299a2b60de0ef5fb17a1e463d8552d6c4d11d166ae32e490104b3d90116d25"} Dec 10 23:00:48 crc kubenswrapper[4791]: I1210 23:00:48.675815 4791 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"dd3f80e5713221f96ed50da7f2e7fd5b9fabfcd5c1656da8d8a86d818a5fdc3f"} Dec 10 23:00:48 crc kubenswrapper[4791]: I1210 23:00:48.675821 4791 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"152a7618be8f38fadf4e564b3948980503cedcdb466c37344ccc20f4a3173a5f"} Dec 10 23:00:48 crc kubenswrapper[4791]: I1210 23:00:48.675826 4791 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"3d18f9cb7207efb1e32de75a30d33e024f656a08f0503db95e5488bc5280971f"} Dec 10 23:00:48 crc kubenswrapper[4791]: I1210 23:00:48.675832 4791 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"bc721b530c08b433756e978ec0423cc18aabf6771b61eb96f903b4d57ea335ff"} Dec 10 23:00:48 crc kubenswrapper[4791]: I1210 23:00:48.675837 4791 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"68795a29944e56870ffe94fa732e9dcf0bfb344c0d85d1e590f58a876129c61f"} Dec 10 23:00:48 crc kubenswrapper[4791]: I1210 23:00:48.675843 4791 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"937137ae9d7480cdc7c59bc4a1df0a4c10e72b3b63693af285278789953ed218"} Dec 10 23:00:48 crc kubenswrapper[4791]: I1210 23:00:48.675848 4791 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"1a52d19b7671e8b143017c8a981a6f8d3ee3b84650e4b223b814a531ac76f364"} Dec 10 23:00:48 crc kubenswrapper[4791]: I1210 23:00:48.675855 4791 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"3afbd1ffcd2c4b40324aca6e52d6fc822db131c3f35a9af3a6f005c2cffdb7bf"} Dec 10 23:00:48 crc kubenswrapper[4791]: I1210 23:00:48.675861 4791 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-zhq64" event={"ID":"3cd47739-0fa9-4321-aff1-220f8721a0b3","Type":"ContainerDied","Data":"937137ae9d7480cdc7c59bc4a1df0a4c10e72b3b63693af285278789953ed218"} Dec 10 23:00:48 crc kubenswrapper[4791]: I1210 23:00:48.675869 4791 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"8839756dfbc89d902c809e3b2b4c8dd5de8ef52ec1009fb2aac5a06a16de4b66"} Dec 10 23:00:48 crc kubenswrapper[4791]: I1210 23:00:48.675875 4791 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"5f299a2b60de0ef5fb17a1e463d8552d6c4d11d166ae32e490104b3d90116d25"} Dec 10 23:00:48 crc kubenswrapper[4791]: I1210 23:00:48.675881 4791 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"dd3f80e5713221f96ed50da7f2e7fd5b9fabfcd5c1656da8d8a86d818a5fdc3f"} Dec 10 23:00:48 crc kubenswrapper[4791]: I1210 23:00:48.675886 4791 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"152a7618be8f38fadf4e564b3948980503cedcdb466c37344ccc20f4a3173a5f"} Dec 10 23:00:48 crc kubenswrapper[4791]: I1210 23:00:48.675891 4791 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"3d18f9cb7207efb1e32de75a30d33e024f656a08f0503db95e5488bc5280971f"} Dec 10 23:00:48 crc kubenswrapper[4791]: I1210 23:00:48.675896 4791 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"bc721b530c08b433756e978ec0423cc18aabf6771b61eb96f903b4d57ea335ff"} Dec 10 23:00:48 crc kubenswrapper[4791]: I1210 23:00:48.675901 4791 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"68795a29944e56870ffe94fa732e9dcf0bfb344c0d85d1e590f58a876129c61f"} Dec 10 23:00:48 crc kubenswrapper[4791]: I1210 23:00:48.675906 4791 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"937137ae9d7480cdc7c59bc4a1df0a4c10e72b3b63693af285278789953ed218"} Dec 10 23:00:48 crc kubenswrapper[4791]: I1210 23:00:48.675911 4791 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"1a52d19b7671e8b143017c8a981a6f8d3ee3b84650e4b223b814a531ac76f364"} Dec 10 23:00:48 crc kubenswrapper[4791]: I1210 23:00:48.675917 4791 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"3afbd1ffcd2c4b40324aca6e52d6fc822db131c3f35a9af3a6f005c2cffdb7bf"} Dec 10 23:00:48 crc kubenswrapper[4791]: I1210 23:00:48.675924 4791 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-zhq64" event={"ID":"3cd47739-0fa9-4321-aff1-220f8721a0b3","Type":"ContainerDied","Data":"1a52d19b7671e8b143017c8a981a6f8d3ee3b84650e4b223b814a531ac76f364"} Dec 10 23:00:48 crc kubenswrapper[4791]: I1210 23:00:48.675932 4791 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"8839756dfbc89d902c809e3b2b4c8dd5de8ef52ec1009fb2aac5a06a16de4b66"} Dec 10 23:00:48 crc kubenswrapper[4791]: I1210 23:00:48.675940 4791 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"5f299a2b60de0ef5fb17a1e463d8552d6c4d11d166ae32e490104b3d90116d25"} Dec 10 23:00:48 crc kubenswrapper[4791]: I1210 23:00:48.675946 4791 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"dd3f80e5713221f96ed50da7f2e7fd5b9fabfcd5c1656da8d8a86d818a5fdc3f"} Dec 10 23:00:48 crc kubenswrapper[4791]: I1210 23:00:48.675951 4791 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"152a7618be8f38fadf4e564b3948980503cedcdb466c37344ccc20f4a3173a5f"} Dec 10 23:00:48 crc kubenswrapper[4791]: I1210 23:00:48.675957 4791 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"3d18f9cb7207efb1e32de75a30d33e024f656a08f0503db95e5488bc5280971f"} Dec 10 23:00:48 crc kubenswrapper[4791]: I1210 23:00:48.675964 4791 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"bc721b530c08b433756e978ec0423cc18aabf6771b61eb96f903b4d57ea335ff"} Dec 10 23:00:48 crc kubenswrapper[4791]: I1210 23:00:48.675969 4791 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"68795a29944e56870ffe94fa732e9dcf0bfb344c0d85d1e590f58a876129c61f"} Dec 10 23:00:48 crc kubenswrapper[4791]: I1210 23:00:48.675974 4791 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"937137ae9d7480cdc7c59bc4a1df0a4c10e72b3b63693af285278789953ed218"} Dec 10 23:00:48 crc kubenswrapper[4791]: I1210 23:00:48.675980 4791 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"1a52d19b7671e8b143017c8a981a6f8d3ee3b84650e4b223b814a531ac76f364"} Dec 10 23:00:48 crc kubenswrapper[4791]: I1210 23:00:48.675985 4791 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"3afbd1ffcd2c4b40324aca6e52d6fc822db131c3f35a9af3a6f005c2cffdb7bf"} Dec 10 23:00:48 crc kubenswrapper[4791]: I1210 23:00:48.675992 4791 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-zhq64" event={"ID":"3cd47739-0fa9-4321-aff1-220f8721a0b3","Type":"ContainerDied","Data":"4768e24be0da8b9e4e556b30b09468d0b4d8bd68f0e468160e7300cc2828587d"} Dec 10 23:00:48 crc kubenswrapper[4791]: I1210 23:00:48.675999 4791 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"8839756dfbc89d902c809e3b2b4c8dd5de8ef52ec1009fb2aac5a06a16de4b66"} Dec 10 23:00:48 crc kubenswrapper[4791]: I1210 23:00:48.676006 4791 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"5f299a2b60de0ef5fb17a1e463d8552d6c4d11d166ae32e490104b3d90116d25"} Dec 10 23:00:48 crc kubenswrapper[4791]: I1210 23:00:48.676012 4791 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"dd3f80e5713221f96ed50da7f2e7fd5b9fabfcd5c1656da8d8a86d818a5fdc3f"} Dec 10 23:00:48 crc kubenswrapper[4791]: I1210 23:00:48.676017 4791 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"152a7618be8f38fadf4e564b3948980503cedcdb466c37344ccc20f4a3173a5f"} Dec 10 23:00:48 crc kubenswrapper[4791]: I1210 23:00:48.676022 4791 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"3d18f9cb7207efb1e32de75a30d33e024f656a08f0503db95e5488bc5280971f"} Dec 10 23:00:48 crc kubenswrapper[4791]: I1210 23:00:48.676027 4791 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"bc721b530c08b433756e978ec0423cc18aabf6771b61eb96f903b4d57ea335ff"} Dec 10 23:00:48 crc kubenswrapper[4791]: I1210 23:00:48.676032 4791 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"68795a29944e56870ffe94fa732e9dcf0bfb344c0d85d1e590f58a876129c61f"} Dec 10 23:00:48 crc kubenswrapper[4791]: I1210 23:00:48.676038 4791 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"937137ae9d7480cdc7c59bc4a1df0a4c10e72b3b63693af285278789953ed218"} Dec 10 23:00:48 crc kubenswrapper[4791]: I1210 23:00:48.676043 4791 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"1a52d19b7671e8b143017c8a981a6f8d3ee3b84650e4b223b814a531ac76f364"} Dec 10 23:00:48 crc kubenswrapper[4791]: I1210 23:00:48.676048 4791 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"3afbd1ffcd2c4b40324aca6e52d6fc822db131c3f35a9af3a6f005c2cffdb7bf"} Dec 10 23:00:48 crc kubenswrapper[4791]: I1210 23:00:48.675765 4791 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-ovn-kubernetes/ovnkube-node-zhq64" Dec 10 23:00:48 crc kubenswrapper[4791]: I1210 23:00:48.690654 4791 scope.go:117] "RemoveContainer" containerID="8839756dfbc89d902c809e3b2b4c8dd5de8ef52ec1009fb2aac5a06a16de4b66" Dec 10 23:00:48 crc kubenswrapper[4791]: I1210 23:00:48.731392 4791 scope.go:117] "RemoveContainer" containerID="5f299a2b60de0ef5fb17a1e463d8552d6c4d11d166ae32e490104b3d90116d25" Dec 10 23:00:48 crc kubenswrapper[4791]: I1210 23:00:48.747359 4791 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-ovn-kubernetes/ovnkube-node-zhq64"] Dec 10 23:00:48 crc kubenswrapper[4791]: I1210 23:00:48.755942 4791 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-ovn-kubernetes/ovnkube-node-zhq64"] Dec 10 23:00:48 crc kubenswrapper[4791]: I1210 23:00:48.757603 4791 scope.go:117] "RemoveContainer" containerID="dd3f80e5713221f96ed50da7f2e7fd5b9fabfcd5c1656da8d8a86d818a5fdc3f" Dec 10 23:00:48 crc kubenswrapper[4791]: I1210 23:00:48.770458 4791 scope.go:117] "RemoveContainer" containerID="152a7618be8f38fadf4e564b3948980503cedcdb466c37344ccc20f4a3173a5f" Dec 10 23:00:48 crc kubenswrapper[4791]: I1210 23:00:48.783469 4791 scope.go:117] "RemoveContainer" containerID="3d18f9cb7207efb1e32de75a30d33e024f656a08f0503db95e5488bc5280971f" Dec 10 23:00:48 crc kubenswrapper[4791]: I1210 23:00:48.818509 4791 scope.go:117] "RemoveContainer" containerID="bc721b530c08b433756e978ec0423cc18aabf6771b61eb96f903b4d57ea335ff" Dec 10 23:00:48 crc kubenswrapper[4791]: I1210 23:00:48.842434 4791 scope.go:117] "RemoveContainer" containerID="68795a29944e56870ffe94fa732e9dcf0bfb344c0d85d1e590f58a876129c61f" Dec 10 23:00:48 crc kubenswrapper[4791]: I1210 23:00:48.855668 4791 scope.go:117] "RemoveContainer" containerID="937137ae9d7480cdc7c59bc4a1df0a4c10e72b3b63693af285278789953ed218" Dec 10 23:00:48 crc kubenswrapper[4791]: I1210 23:00:48.881119 4791 scope.go:117] "RemoveContainer" containerID="1a52d19b7671e8b143017c8a981a6f8d3ee3b84650e4b223b814a531ac76f364" Dec 10 23:00:48 crc kubenswrapper[4791]: I1210 23:00:48.896470 4791 scope.go:117] "RemoveContainer" containerID="3afbd1ffcd2c4b40324aca6e52d6fc822db131c3f35a9af3a6f005c2cffdb7bf" Dec 10 23:00:48 crc kubenswrapper[4791]: I1210 23:00:48.911532 4791 scope.go:117] "RemoveContainer" containerID="8839756dfbc89d902c809e3b2b4c8dd5de8ef52ec1009fb2aac5a06a16de4b66" Dec 10 23:00:48 crc kubenswrapper[4791]: E1210 23:00:48.912832 4791 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"8839756dfbc89d902c809e3b2b4c8dd5de8ef52ec1009fb2aac5a06a16de4b66\": container with ID starting with 8839756dfbc89d902c809e3b2b4c8dd5de8ef52ec1009fb2aac5a06a16de4b66 not found: ID does not exist" containerID="8839756dfbc89d902c809e3b2b4c8dd5de8ef52ec1009fb2aac5a06a16de4b66" Dec 10 23:00:48 crc kubenswrapper[4791]: I1210 23:00:48.912880 4791 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"8839756dfbc89d902c809e3b2b4c8dd5de8ef52ec1009fb2aac5a06a16de4b66"} err="failed to get container status \"8839756dfbc89d902c809e3b2b4c8dd5de8ef52ec1009fb2aac5a06a16de4b66\": rpc error: code = NotFound desc = could not find container \"8839756dfbc89d902c809e3b2b4c8dd5de8ef52ec1009fb2aac5a06a16de4b66\": container with ID starting with 8839756dfbc89d902c809e3b2b4c8dd5de8ef52ec1009fb2aac5a06a16de4b66 not found: ID does not exist" Dec 10 23:00:48 crc kubenswrapper[4791]: I1210 23:00:48.912909 4791 scope.go:117] "RemoveContainer" containerID="5f299a2b60de0ef5fb17a1e463d8552d6c4d11d166ae32e490104b3d90116d25" Dec 10 23:00:48 crc kubenswrapper[4791]: E1210 23:00:48.913459 4791 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"5f299a2b60de0ef5fb17a1e463d8552d6c4d11d166ae32e490104b3d90116d25\": container with ID starting with 5f299a2b60de0ef5fb17a1e463d8552d6c4d11d166ae32e490104b3d90116d25 not found: ID does not exist" containerID="5f299a2b60de0ef5fb17a1e463d8552d6c4d11d166ae32e490104b3d90116d25" Dec 10 23:00:48 crc kubenswrapper[4791]: I1210 23:00:48.913494 4791 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"5f299a2b60de0ef5fb17a1e463d8552d6c4d11d166ae32e490104b3d90116d25"} err="failed to get container status \"5f299a2b60de0ef5fb17a1e463d8552d6c4d11d166ae32e490104b3d90116d25\": rpc error: code = NotFound desc = could not find container \"5f299a2b60de0ef5fb17a1e463d8552d6c4d11d166ae32e490104b3d90116d25\": container with ID starting with 5f299a2b60de0ef5fb17a1e463d8552d6c4d11d166ae32e490104b3d90116d25 not found: ID does not exist" Dec 10 23:00:48 crc kubenswrapper[4791]: I1210 23:00:48.913523 4791 scope.go:117] "RemoveContainer" containerID="dd3f80e5713221f96ed50da7f2e7fd5b9fabfcd5c1656da8d8a86d818a5fdc3f" Dec 10 23:00:48 crc kubenswrapper[4791]: E1210 23:00:48.915388 4791 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"dd3f80e5713221f96ed50da7f2e7fd5b9fabfcd5c1656da8d8a86d818a5fdc3f\": container with ID starting with dd3f80e5713221f96ed50da7f2e7fd5b9fabfcd5c1656da8d8a86d818a5fdc3f not found: ID does not exist" containerID="dd3f80e5713221f96ed50da7f2e7fd5b9fabfcd5c1656da8d8a86d818a5fdc3f" Dec 10 23:00:48 crc kubenswrapper[4791]: I1210 23:00:48.915459 4791 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"dd3f80e5713221f96ed50da7f2e7fd5b9fabfcd5c1656da8d8a86d818a5fdc3f"} err="failed to get container status \"dd3f80e5713221f96ed50da7f2e7fd5b9fabfcd5c1656da8d8a86d818a5fdc3f\": rpc error: code = NotFound desc = could not find container \"dd3f80e5713221f96ed50da7f2e7fd5b9fabfcd5c1656da8d8a86d818a5fdc3f\": container with ID starting with dd3f80e5713221f96ed50da7f2e7fd5b9fabfcd5c1656da8d8a86d818a5fdc3f not found: ID does not exist" Dec 10 23:00:48 crc kubenswrapper[4791]: I1210 23:00:48.915490 4791 scope.go:117] "RemoveContainer" containerID="152a7618be8f38fadf4e564b3948980503cedcdb466c37344ccc20f4a3173a5f" Dec 10 23:00:48 crc kubenswrapper[4791]: E1210 23:00:48.915944 4791 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"152a7618be8f38fadf4e564b3948980503cedcdb466c37344ccc20f4a3173a5f\": container with ID starting with 152a7618be8f38fadf4e564b3948980503cedcdb466c37344ccc20f4a3173a5f not found: ID does not exist" containerID="152a7618be8f38fadf4e564b3948980503cedcdb466c37344ccc20f4a3173a5f" Dec 10 23:00:48 crc kubenswrapper[4791]: I1210 23:00:48.916045 4791 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"152a7618be8f38fadf4e564b3948980503cedcdb466c37344ccc20f4a3173a5f"} err="failed to get container status \"152a7618be8f38fadf4e564b3948980503cedcdb466c37344ccc20f4a3173a5f\": rpc error: code = NotFound desc = could not find container \"152a7618be8f38fadf4e564b3948980503cedcdb466c37344ccc20f4a3173a5f\": container with ID starting with 152a7618be8f38fadf4e564b3948980503cedcdb466c37344ccc20f4a3173a5f not found: ID does not exist" Dec 10 23:00:48 crc kubenswrapper[4791]: I1210 23:00:48.916137 4791 scope.go:117] "RemoveContainer" containerID="3d18f9cb7207efb1e32de75a30d33e024f656a08f0503db95e5488bc5280971f" Dec 10 23:00:48 crc kubenswrapper[4791]: E1210 23:00:48.918646 4791 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"3d18f9cb7207efb1e32de75a30d33e024f656a08f0503db95e5488bc5280971f\": container with ID starting with 3d18f9cb7207efb1e32de75a30d33e024f656a08f0503db95e5488bc5280971f not found: ID does not exist" containerID="3d18f9cb7207efb1e32de75a30d33e024f656a08f0503db95e5488bc5280971f" Dec 10 23:00:48 crc kubenswrapper[4791]: I1210 23:00:48.918678 4791 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"3d18f9cb7207efb1e32de75a30d33e024f656a08f0503db95e5488bc5280971f"} err="failed to get container status \"3d18f9cb7207efb1e32de75a30d33e024f656a08f0503db95e5488bc5280971f\": rpc error: code = NotFound desc = could not find container \"3d18f9cb7207efb1e32de75a30d33e024f656a08f0503db95e5488bc5280971f\": container with ID starting with 3d18f9cb7207efb1e32de75a30d33e024f656a08f0503db95e5488bc5280971f not found: ID does not exist" Dec 10 23:00:48 crc kubenswrapper[4791]: I1210 23:00:48.918699 4791 scope.go:117] "RemoveContainer" containerID="bc721b530c08b433756e978ec0423cc18aabf6771b61eb96f903b4d57ea335ff" Dec 10 23:00:48 crc kubenswrapper[4791]: E1210 23:00:48.919069 4791 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"bc721b530c08b433756e978ec0423cc18aabf6771b61eb96f903b4d57ea335ff\": container with ID starting with bc721b530c08b433756e978ec0423cc18aabf6771b61eb96f903b4d57ea335ff not found: ID does not exist" containerID="bc721b530c08b433756e978ec0423cc18aabf6771b61eb96f903b4d57ea335ff" Dec 10 23:00:48 crc kubenswrapper[4791]: I1210 23:00:48.919102 4791 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"bc721b530c08b433756e978ec0423cc18aabf6771b61eb96f903b4d57ea335ff"} err="failed to get container status \"bc721b530c08b433756e978ec0423cc18aabf6771b61eb96f903b4d57ea335ff\": rpc error: code = NotFound desc = could not find container \"bc721b530c08b433756e978ec0423cc18aabf6771b61eb96f903b4d57ea335ff\": container with ID starting with bc721b530c08b433756e978ec0423cc18aabf6771b61eb96f903b4d57ea335ff not found: ID does not exist" Dec 10 23:00:48 crc kubenswrapper[4791]: I1210 23:00:48.919131 4791 scope.go:117] "RemoveContainer" containerID="68795a29944e56870ffe94fa732e9dcf0bfb344c0d85d1e590f58a876129c61f" Dec 10 23:00:48 crc kubenswrapper[4791]: E1210 23:00:48.919587 4791 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"68795a29944e56870ffe94fa732e9dcf0bfb344c0d85d1e590f58a876129c61f\": container with ID starting with 68795a29944e56870ffe94fa732e9dcf0bfb344c0d85d1e590f58a876129c61f not found: ID does not exist" containerID="68795a29944e56870ffe94fa732e9dcf0bfb344c0d85d1e590f58a876129c61f" Dec 10 23:00:48 crc kubenswrapper[4791]: I1210 23:00:48.919620 4791 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"68795a29944e56870ffe94fa732e9dcf0bfb344c0d85d1e590f58a876129c61f"} err="failed to get container status \"68795a29944e56870ffe94fa732e9dcf0bfb344c0d85d1e590f58a876129c61f\": rpc error: code = NotFound desc = could not find container \"68795a29944e56870ffe94fa732e9dcf0bfb344c0d85d1e590f58a876129c61f\": container with ID starting with 68795a29944e56870ffe94fa732e9dcf0bfb344c0d85d1e590f58a876129c61f not found: ID does not exist" Dec 10 23:00:48 crc kubenswrapper[4791]: I1210 23:00:48.919638 4791 scope.go:117] "RemoveContainer" containerID="937137ae9d7480cdc7c59bc4a1df0a4c10e72b3b63693af285278789953ed218" Dec 10 23:00:48 crc kubenswrapper[4791]: E1210 23:00:48.919940 4791 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"937137ae9d7480cdc7c59bc4a1df0a4c10e72b3b63693af285278789953ed218\": container with ID starting with 937137ae9d7480cdc7c59bc4a1df0a4c10e72b3b63693af285278789953ed218 not found: ID does not exist" containerID="937137ae9d7480cdc7c59bc4a1df0a4c10e72b3b63693af285278789953ed218" Dec 10 23:00:48 crc kubenswrapper[4791]: I1210 23:00:48.919961 4791 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"937137ae9d7480cdc7c59bc4a1df0a4c10e72b3b63693af285278789953ed218"} err="failed to get container status \"937137ae9d7480cdc7c59bc4a1df0a4c10e72b3b63693af285278789953ed218\": rpc error: code = NotFound desc = could not find container \"937137ae9d7480cdc7c59bc4a1df0a4c10e72b3b63693af285278789953ed218\": container with ID starting with 937137ae9d7480cdc7c59bc4a1df0a4c10e72b3b63693af285278789953ed218 not found: ID does not exist" Dec 10 23:00:48 crc kubenswrapper[4791]: I1210 23:00:48.919975 4791 scope.go:117] "RemoveContainer" containerID="1a52d19b7671e8b143017c8a981a6f8d3ee3b84650e4b223b814a531ac76f364" Dec 10 23:00:48 crc kubenswrapper[4791]: E1210 23:00:48.920287 4791 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"1a52d19b7671e8b143017c8a981a6f8d3ee3b84650e4b223b814a531ac76f364\": container with ID starting with 1a52d19b7671e8b143017c8a981a6f8d3ee3b84650e4b223b814a531ac76f364 not found: ID does not exist" containerID="1a52d19b7671e8b143017c8a981a6f8d3ee3b84650e4b223b814a531ac76f364" Dec 10 23:00:48 crc kubenswrapper[4791]: I1210 23:00:48.920346 4791 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"1a52d19b7671e8b143017c8a981a6f8d3ee3b84650e4b223b814a531ac76f364"} err="failed to get container status \"1a52d19b7671e8b143017c8a981a6f8d3ee3b84650e4b223b814a531ac76f364\": rpc error: code = NotFound desc = could not find container \"1a52d19b7671e8b143017c8a981a6f8d3ee3b84650e4b223b814a531ac76f364\": container with ID starting with 1a52d19b7671e8b143017c8a981a6f8d3ee3b84650e4b223b814a531ac76f364 not found: ID does not exist" Dec 10 23:00:48 crc kubenswrapper[4791]: I1210 23:00:48.920384 4791 scope.go:117] "RemoveContainer" containerID="3afbd1ffcd2c4b40324aca6e52d6fc822db131c3f35a9af3a6f005c2cffdb7bf" Dec 10 23:00:48 crc kubenswrapper[4791]: E1210 23:00:48.920722 4791 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"3afbd1ffcd2c4b40324aca6e52d6fc822db131c3f35a9af3a6f005c2cffdb7bf\": container with ID starting with 3afbd1ffcd2c4b40324aca6e52d6fc822db131c3f35a9af3a6f005c2cffdb7bf not found: ID does not exist" containerID="3afbd1ffcd2c4b40324aca6e52d6fc822db131c3f35a9af3a6f005c2cffdb7bf" Dec 10 23:00:48 crc kubenswrapper[4791]: I1210 23:00:48.920745 4791 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"3afbd1ffcd2c4b40324aca6e52d6fc822db131c3f35a9af3a6f005c2cffdb7bf"} err="failed to get container status \"3afbd1ffcd2c4b40324aca6e52d6fc822db131c3f35a9af3a6f005c2cffdb7bf\": rpc error: code = NotFound desc = could not find container \"3afbd1ffcd2c4b40324aca6e52d6fc822db131c3f35a9af3a6f005c2cffdb7bf\": container with ID starting with 3afbd1ffcd2c4b40324aca6e52d6fc822db131c3f35a9af3a6f005c2cffdb7bf not found: ID does not exist" Dec 10 23:00:48 crc kubenswrapper[4791]: I1210 23:00:48.920759 4791 scope.go:117] "RemoveContainer" containerID="8839756dfbc89d902c809e3b2b4c8dd5de8ef52ec1009fb2aac5a06a16de4b66" Dec 10 23:00:48 crc kubenswrapper[4791]: I1210 23:00:48.920999 4791 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"8839756dfbc89d902c809e3b2b4c8dd5de8ef52ec1009fb2aac5a06a16de4b66"} err="failed to get container status \"8839756dfbc89d902c809e3b2b4c8dd5de8ef52ec1009fb2aac5a06a16de4b66\": rpc error: code = NotFound desc = could not find container \"8839756dfbc89d902c809e3b2b4c8dd5de8ef52ec1009fb2aac5a06a16de4b66\": container with ID starting with 8839756dfbc89d902c809e3b2b4c8dd5de8ef52ec1009fb2aac5a06a16de4b66 not found: ID does not exist" Dec 10 23:00:48 crc kubenswrapper[4791]: I1210 23:00:48.921027 4791 scope.go:117] "RemoveContainer" containerID="5f299a2b60de0ef5fb17a1e463d8552d6c4d11d166ae32e490104b3d90116d25" Dec 10 23:00:48 crc kubenswrapper[4791]: I1210 23:00:48.921390 4791 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"5f299a2b60de0ef5fb17a1e463d8552d6c4d11d166ae32e490104b3d90116d25"} err="failed to get container status \"5f299a2b60de0ef5fb17a1e463d8552d6c4d11d166ae32e490104b3d90116d25\": rpc error: code = NotFound desc = could not find container \"5f299a2b60de0ef5fb17a1e463d8552d6c4d11d166ae32e490104b3d90116d25\": container with ID starting with 5f299a2b60de0ef5fb17a1e463d8552d6c4d11d166ae32e490104b3d90116d25 not found: ID does not exist" Dec 10 23:00:48 crc kubenswrapper[4791]: I1210 23:00:48.921420 4791 scope.go:117] "RemoveContainer" containerID="dd3f80e5713221f96ed50da7f2e7fd5b9fabfcd5c1656da8d8a86d818a5fdc3f" Dec 10 23:00:48 crc kubenswrapper[4791]: I1210 23:00:48.921714 4791 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"dd3f80e5713221f96ed50da7f2e7fd5b9fabfcd5c1656da8d8a86d818a5fdc3f"} err="failed to get container status \"dd3f80e5713221f96ed50da7f2e7fd5b9fabfcd5c1656da8d8a86d818a5fdc3f\": rpc error: code = NotFound desc = could not find container \"dd3f80e5713221f96ed50da7f2e7fd5b9fabfcd5c1656da8d8a86d818a5fdc3f\": container with ID starting with dd3f80e5713221f96ed50da7f2e7fd5b9fabfcd5c1656da8d8a86d818a5fdc3f not found: ID does not exist" Dec 10 23:00:48 crc kubenswrapper[4791]: I1210 23:00:48.921736 4791 scope.go:117] "RemoveContainer" containerID="152a7618be8f38fadf4e564b3948980503cedcdb466c37344ccc20f4a3173a5f" Dec 10 23:00:48 crc kubenswrapper[4791]: I1210 23:00:48.922184 4791 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"152a7618be8f38fadf4e564b3948980503cedcdb466c37344ccc20f4a3173a5f"} err="failed to get container status \"152a7618be8f38fadf4e564b3948980503cedcdb466c37344ccc20f4a3173a5f\": rpc error: code = NotFound desc = could not find container \"152a7618be8f38fadf4e564b3948980503cedcdb466c37344ccc20f4a3173a5f\": container with ID starting with 152a7618be8f38fadf4e564b3948980503cedcdb466c37344ccc20f4a3173a5f not found: ID does not exist" Dec 10 23:00:48 crc kubenswrapper[4791]: I1210 23:00:48.922210 4791 scope.go:117] "RemoveContainer" containerID="3d18f9cb7207efb1e32de75a30d33e024f656a08f0503db95e5488bc5280971f" Dec 10 23:00:48 crc kubenswrapper[4791]: I1210 23:00:48.922519 4791 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"3d18f9cb7207efb1e32de75a30d33e024f656a08f0503db95e5488bc5280971f"} err="failed to get container status \"3d18f9cb7207efb1e32de75a30d33e024f656a08f0503db95e5488bc5280971f\": rpc error: code = NotFound desc = could not find container \"3d18f9cb7207efb1e32de75a30d33e024f656a08f0503db95e5488bc5280971f\": container with ID starting with 3d18f9cb7207efb1e32de75a30d33e024f656a08f0503db95e5488bc5280971f not found: ID does not exist" Dec 10 23:00:48 crc kubenswrapper[4791]: I1210 23:00:48.922547 4791 scope.go:117] "RemoveContainer" containerID="bc721b530c08b433756e978ec0423cc18aabf6771b61eb96f903b4d57ea335ff" Dec 10 23:00:48 crc kubenswrapper[4791]: I1210 23:00:48.922800 4791 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"bc721b530c08b433756e978ec0423cc18aabf6771b61eb96f903b4d57ea335ff"} err="failed to get container status \"bc721b530c08b433756e978ec0423cc18aabf6771b61eb96f903b4d57ea335ff\": rpc error: code = NotFound desc = could not find container \"bc721b530c08b433756e978ec0423cc18aabf6771b61eb96f903b4d57ea335ff\": container with ID starting with bc721b530c08b433756e978ec0423cc18aabf6771b61eb96f903b4d57ea335ff not found: ID does not exist" Dec 10 23:00:48 crc kubenswrapper[4791]: I1210 23:00:48.922824 4791 scope.go:117] "RemoveContainer" containerID="68795a29944e56870ffe94fa732e9dcf0bfb344c0d85d1e590f58a876129c61f" Dec 10 23:00:48 crc kubenswrapper[4791]: I1210 23:00:48.923150 4791 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"68795a29944e56870ffe94fa732e9dcf0bfb344c0d85d1e590f58a876129c61f"} err="failed to get container status \"68795a29944e56870ffe94fa732e9dcf0bfb344c0d85d1e590f58a876129c61f\": rpc error: code = NotFound desc = could not find container \"68795a29944e56870ffe94fa732e9dcf0bfb344c0d85d1e590f58a876129c61f\": container with ID starting with 68795a29944e56870ffe94fa732e9dcf0bfb344c0d85d1e590f58a876129c61f not found: ID does not exist" Dec 10 23:00:48 crc kubenswrapper[4791]: I1210 23:00:48.923176 4791 scope.go:117] "RemoveContainer" containerID="937137ae9d7480cdc7c59bc4a1df0a4c10e72b3b63693af285278789953ed218" Dec 10 23:00:48 crc kubenswrapper[4791]: I1210 23:00:48.923473 4791 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"937137ae9d7480cdc7c59bc4a1df0a4c10e72b3b63693af285278789953ed218"} err="failed to get container status \"937137ae9d7480cdc7c59bc4a1df0a4c10e72b3b63693af285278789953ed218\": rpc error: code = NotFound desc = could not find container \"937137ae9d7480cdc7c59bc4a1df0a4c10e72b3b63693af285278789953ed218\": container with ID starting with 937137ae9d7480cdc7c59bc4a1df0a4c10e72b3b63693af285278789953ed218 not found: ID does not exist" Dec 10 23:00:48 crc kubenswrapper[4791]: I1210 23:00:48.923491 4791 scope.go:117] "RemoveContainer" containerID="1a52d19b7671e8b143017c8a981a6f8d3ee3b84650e4b223b814a531ac76f364" Dec 10 23:00:48 crc kubenswrapper[4791]: I1210 23:00:48.923791 4791 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"1a52d19b7671e8b143017c8a981a6f8d3ee3b84650e4b223b814a531ac76f364"} err="failed to get container status \"1a52d19b7671e8b143017c8a981a6f8d3ee3b84650e4b223b814a531ac76f364\": rpc error: code = NotFound desc = could not find container \"1a52d19b7671e8b143017c8a981a6f8d3ee3b84650e4b223b814a531ac76f364\": container with ID starting with 1a52d19b7671e8b143017c8a981a6f8d3ee3b84650e4b223b814a531ac76f364 not found: ID does not exist" Dec 10 23:00:48 crc kubenswrapper[4791]: I1210 23:00:48.923809 4791 scope.go:117] "RemoveContainer" containerID="3afbd1ffcd2c4b40324aca6e52d6fc822db131c3f35a9af3a6f005c2cffdb7bf" Dec 10 23:00:48 crc kubenswrapper[4791]: I1210 23:00:48.924166 4791 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"3afbd1ffcd2c4b40324aca6e52d6fc822db131c3f35a9af3a6f005c2cffdb7bf"} err="failed to get container status \"3afbd1ffcd2c4b40324aca6e52d6fc822db131c3f35a9af3a6f005c2cffdb7bf\": rpc error: code = NotFound desc = could not find container \"3afbd1ffcd2c4b40324aca6e52d6fc822db131c3f35a9af3a6f005c2cffdb7bf\": container with ID starting with 3afbd1ffcd2c4b40324aca6e52d6fc822db131c3f35a9af3a6f005c2cffdb7bf not found: ID does not exist" Dec 10 23:00:48 crc kubenswrapper[4791]: I1210 23:00:48.924181 4791 scope.go:117] "RemoveContainer" containerID="8839756dfbc89d902c809e3b2b4c8dd5de8ef52ec1009fb2aac5a06a16de4b66" Dec 10 23:00:48 crc kubenswrapper[4791]: I1210 23:00:48.924455 4791 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"8839756dfbc89d902c809e3b2b4c8dd5de8ef52ec1009fb2aac5a06a16de4b66"} err="failed to get container status \"8839756dfbc89d902c809e3b2b4c8dd5de8ef52ec1009fb2aac5a06a16de4b66\": rpc error: code = NotFound desc = could not find container \"8839756dfbc89d902c809e3b2b4c8dd5de8ef52ec1009fb2aac5a06a16de4b66\": container with ID starting with 8839756dfbc89d902c809e3b2b4c8dd5de8ef52ec1009fb2aac5a06a16de4b66 not found: ID does not exist" Dec 10 23:00:48 crc kubenswrapper[4791]: I1210 23:00:48.924480 4791 scope.go:117] "RemoveContainer" containerID="5f299a2b60de0ef5fb17a1e463d8552d6c4d11d166ae32e490104b3d90116d25" Dec 10 23:00:48 crc kubenswrapper[4791]: I1210 23:00:48.924743 4791 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"5f299a2b60de0ef5fb17a1e463d8552d6c4d11d166ae32e490104b3d90116d25"} err="failed to get container status \"5f299a2b60de0ef5fb17a1e463d8552d6c4d11d166ae32e490104b3d90116d25\": rpc error: code = NotFound desc = could not find container \"5f299a2b60de0ef5fb17a1e463d8552d6c4d11d166ae32e490104b3d90116d25\": container with ID starting with 5f299a2b60de0ef5fb17a1e463d8552d6c4d11d166ae32e490104b3d90116d25 not found: ID does not exist" Dec 10 23:00:48 crc kubenswrapper[4791]: I1210 23:00:48.924760 4791 scope.go:117] "RemoveContainer" containerID="dd3f80e5713221f96ed50da7f2e7fd5b9fabfcd5c1656da8d8a86d818a5fdc3f" Dec 10 23:00:48 crc kubenswrapper[4791]: I1210 23:00:48.924991 4791 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"dd3f80e5713221f96ed50da7f2e7fd5b9fabfcd5c1656da8d8a86d818a5fdc3f"} err="failed to get container status \"dd3f80e5713221f96ed50da7f2e7fd5b9fabfcd5c1656da8d8a86d818a5fdc3f\": rpc error: code = NotFound desc = could not find container \"dd3f80e5713221f96ed50da7f2e7fd5b9fabfcd5c1656da8d8a86d818a5fdc3f\": container with ID starting with dd3f80e5713221f96ed50da7f2e7fd5b9fabfcd5c1656da8d8a86d818a5fdc3f not found: ID does not exist" Dec 10 23:00:48 crc kubenswrapper[4791]: I1210 23:00:48.925017 4791 scope.go:117] "RemoveContainer" containerID="152a7618be8f38fadf4e564b3948980503cedcdb466c37344ccc20f4a3173a5f" Dec 10 23:00:48 crc kubenswrapper[4791]: I1210 23:00:48.925290 4791 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"152a7618be8f38fadf4e564b3948980503cedcdb466c37344ccc20f4a3173a5f"} err="failed to get container status \"152a7618be8f38fadf4e564b3948980503cedcdb466c37344ccc20f4a3173a5f\": rpc error: code = NotFound desc = could not find container \"152a7618be8f38fadf4e564b3948980503cedcdb466c37344ccc20f4a3173a5f\": container with ID starting with 152a7618be8f38fadf4e564b3948980503cedcdb466c37344ccc20f4a3173a5f not found: ID does not exist" Dec 10 23:00:48 crc kubenswrapper[4791]: I1210 23:00:48.925309 4791 scope.go:117] "RemoveContainer" containerID="3d18f9cb7207efb1e32de75a30d33e024f656a08f0503db95e5488bc5280971f" Dec 10 23:00:48 crc kubenswrapper[4791]: I1210 23:00:48.925696 4791 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"3d18f9cb7207efb1e32de75a30d33e024f656a08f0503db95e5488bc5280971f"} err="failed to get container status \"3d18f9cb7207efb1e32de75a30d33e024f656a08f0503db95e5488bc5280971f\": rpc error: code = NotFound desc = could not find container \"3d18f9cb7207efb1e32de75a30d33e024f656a08f0503db95e5488bc5280971f\": container with ID starting with 3d18f9cb7207efb1e32de75a30d33e024f656a08f0503db95e5488bc5280971f not found: ID does not exist" Dec 10 23:00:48 crc kubenswrapper[4791]: I1210 23:00:48.925717 4791 scope.go:117] "RemoveContainer" containerID="bc721b530c08b433756e978ec0423cc18aabf6771b61eb96f903b4d57ea335ff" Dec 10 23:00:48 crc kubenswrapper[4791]: I1210 23:00:48.926002 4791 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"bc721b530c08b433756e978ec0423cc18aabf6771b61eb96f903b4d57ea335ff"} err="failed to get container status \"bc721b530c08b433756e978ec0423cc18aabf6771b61eb96f903b4d57ea335ff\": rpc error: code = NotFound desc = could not find container \"bc721b530c08b433756e978ec0423cc18aabf6771b61eb96f903b4d57ea335ff\": container with ID starting with bc721b530c08b433756e978ec0423cc18aabf6771b61eb96f903b4d57ea335ff not found: ID does not exist" Dec 10 23:00:48 crc kubenswrapper[4791]: I1210 23:00:48.926029 4791 scope.go:117] "RemoveContainer" containerID="68795a29944e56870ffe94fa732e9dcf0bfb344c0d85d1e590f58a876129c61f" Dec 10 23:00:48 crc kubenswrapper[4791]: I1210 23:00:48.926242 4791 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"68795a29944e56870ffe94fa732e9dcf0bfb344c0d85d1e590f58a876129c61f"} err="failed to get container status \"68795a29944e56870ffe94fa732e9dcf0bfb344c0d85d1e590f58a876129c61f\": rpc error: code = NotFound desc = could not find container \"68795a29944e56870ffe94fa732e9dcf0bfb344c0d85d1e590f58a876129c61f\": container with ID starting with 68795a29944e56870ffe94fa732e9dcf0bfb344c0d85d1e590f58a876129c61f not found: ID does not exist" Dec 10 23:00:48 crc kubenswrapper[4791]: I1210 23:00:48.926265 4791 scope.go:117] "RemoveContainer" containerID="937137ae9d7480cdc7c59bc4a1df0a4c10e72b3b63693af285278789953ed218" Dec 10 23:00:48 crc kubenswrapper[4791]: I1210 23:00:48.926555 4791 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"937137ae9d7480cdc7c59bc4a1df0a4c10e72b3b63693af285278789953ed218"} err="failed to get container status \"937137ae9d7480cdc7c59bc4a1df0a4c10e72b3b63693af285278789953ed218\": rpc error: code = NotFound desc = could not find container \"937137ae9d7480cdc7c59bc4a1df0a4c10e72b3b63693af285278789953ed218\": container with ID starting with 937137ae9d7480cdc7c59bc4a1df0a4c10e72b3b63693af285278789953ed218 not found: ID does not exist" Dec 10 23:00:48 crc kubenswrapper[4791]: I1210 23:00:48.926579 4791 scope.go:117] "RemoveContainer" containerID="1a52d19b7671e8b143017c8a981a6f8d3ee3b84650e4b223b814a531ac76f364" Dec 10 23:00:48 crc kubenswrapper[4791]: I1210 23:00:48.926875 4791 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"1a52d19b7671e8b143017c8a981a6f8d3ee3b84650e4b223b814a531ac76f364"} err="failed to get container status \"1a52d19b7671e8b143017c8a981a6f8d3ee3b84650e4b223b814a531ac76f364\": rpc error: code = NotFound desc = could not find container \"1a52d19b7671e8b143017c8a981a6f8d3ee3b84650e4b223b814a531ac76f364\": container with ID starting with 1a52d19b7671e8b143017c8a981a6f8d3ee3b84650e4b223b814a531ac76f364 not found: ID does not exist" Dec 10 23:00:48 crc kubenswrapper[4791]: I1210 23:00:48.926902 4791 scope.go:117] "RemoveContainer" containerID="3afbd1ffcd2c4b40324aca6e52d6fc822db131c3f35a9af3a6f005c2cffdb7bf" Dec 10 23:00:48 crc kubenswrapper[4791]: I1210 23:00:48.927166 4791 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"3afbd1ffcd2c4b40324aca6e52d6fc822db131c3f35a9af3a6f005c2cffdb7bf"} err="failed to get container status \"3afbd1ffcd2c4b40324aca6e52d6fc822db131c3f35a9af3a6f005c2cffdb7bf\": rpc error: code = NotFound desc = could not find container \"3afbd1ffcd2c4b40324aca6e52d6fc822db131c3f35a9af3a6f005c2cffdb7bf\": container with ID starting with 3afbd1ffcd2c4b40324aca6e52d6fc822db131c3f35a9af3a6f005c2cffdb7bf not found: ID does not exist" Dec 10 23:00:48 crc kubenswrapper[4791]: I1210 23:00:48.927185 4791 scope.go:117] "RemoveContainer" containerID="8839756dfbc89d902c809e3b2b4c8dd5de8ef52ec1009fb2aac5a06a16de4b66" Dec 10 23:00:48 crc kubenswrapper[4791]: I1210 23:00:48.927576 4791 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"8839756dfbc89d902c809e3b2b4c8dd5de8ef52ec1009fb2aac5a06a16de4b66"} err="failed to get container status \"8839756dfbc89d902c809e3b2b4c8dd5de8ef52ec1009fb2aac5a06a16de4b66\": rpc error: code = NotFound desc = could not find container \"8839756dfbc89d902c809e3b2b4c8dd5de8ef52ec1009fb2aac5a06a16de4b66\": container with ID starting with 8839756dfbc89d902c809e3b2b4c8dd5de8ef52ec1009fb2aac5a06a16de4b66 not found: ID does not exist" Dec 10 23:00:48 crc kubenswrapper[4791]: I1210 23:00:48.927605 4791 scope.go:117] "RemoveContainer" containerID="5f299a2b60de0ef5fb17a1e463d8552d6c4d11d166ae32e490104b3d90116d25" Dec 10 23:00:48 crc kubenswrapper[4791]: I1210 23:00:48.927964 4791 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"5f299a2b60de0ef5fb17a1e463d8552d6c4d11d166ae32e490104b3d90116d25"} err="failed to get container status \"5f299a2b60de0ef5fb17a1e463d8552d6c4d11d166ae32e490104b3d90116d25\": rpc error: code = NotFound desc = could not find container \"5f299a2b60de0ef5fb17a1e463d8552d6c4d11d166ae32e490104b3d90116d25\": container with ID starting with 5f299a2b60de0ef5fb17a1e463d8552d6c4d11d166ae32e490104b3d90116d25 not found: ID does not exist" Dec 10 23:00:48 crc kubenswrapper[4791]: I1210 23:00:48.927981 4791 scope.go:117] "RemoveContainer" containerID="dd3f80e5713221f96ed50da7f2e7fd5b9fabfcd5c1656da8d8a86d818a5fdc3f" Dec 10 23:00:48 crc kubenswrapper[4791]: I1210 23:00:48.928234 4791 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"dd3f80e5713221f96ed50da7f2e7fd5b9fabfcd5c1656da8d8a86d818a5fdc3f"} err="failed to get container status \"dd3f80e5713221f96ed50da7f2e7fd5b9fabfcd5c1656da8d8a86d818a5fdc3f\": rpc error: code = NotFound desc = could not find container \"dd3f80e5713221f96ed50da7f2e7fd5b9fabfcd5c1656da8d8a86d818a5fdc3f\": container with ID starting with dd3f80e5713221f96ed50da7f2e7fd5b9fabfcd5c1656da8d8a86d818a5fdc3f not found: ID does not exist" Dec 10 23:00:48 crc kubenswrapper[4791]: I1210 23:00:48.928264 4791 scope.go:117] "RemoveContainer" containerID="152a7618be8f38fadf4e564b3948980503cedcdb466c37344ccc20f4a3173a5f" Dec 10 23:00:48 crc kubenswrapper[4791]: I1210 23:00:48.928535 4791 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"152a7618be8f38fadf4e564b3948980503cedcdb466c37344ccc20f4a3173a5f"} err="failed to get container status \"152a7618be8f38fadf4e564b3948980503cedcdb466c37344ccc20f4a3173a5f\": rpc error: code = NotFound desc = could not find container \"152a7618be8f38fadf4e564b3948980503cedcdb466c37344ccc20f4a3173a5f\": container with ID starting with 152a7618be8f38fadf4e564b3948980503cedcdb466c37344ccc20f4a3173a5f not found: ID does not exist" Dec 10 23:00:48 crc kubenswrapper[4791]: I1210 23:00:48.928563 4791 scope.go:117] "RemoveContainer" containerID="3d18f9cb7207efb1e32de75a30d33e024f656a08f0503db95e5488bc5280971f" Dec 10 23:00:48 crc kubenswrapper[4791]: I1210 23:00:48.928900 4791 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"3d18f9cb7207efb1e32de75a30d33e024f656a08f0503db95e5488bc5280971f"} err="failed to get container status \"3d18f9cb7207efb1e32de75a30d33e024f656a08f0503db95e5488bc5280971f\": rpc error: code = NotFound desc = could not find container \"3d18f9cb7207efb1e32de75a30d33e024f656a08f0503db95e5488bc5280971f\": container with ID starting with 3d18f9cb7207efb1e32de75a30d33e024f656a08f0503db95e5488bc5280971f not found: ID does not exist" Dec 10 23:00:48 crc kubenswrapper[4791]: I1210 23:00:48.928927 4791 scope.go:117] "RemoveContainer" containerID="bc721b530c08b433756e978ec0423cc18aabf6771b61eb96f903b4d57ea335ff" Dec 10 23:00:48 crc kubenswrapper[4791]: I1210 23:00:48.929184 4791 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"bc721b530c08b433756e978ec0423cc18aabf6771b61eb96f903b4d57ea335ff"} err="failed to get container status \"bc721b530c08b433756e978ec0423cc18aabf6771b61eb96f903b4d57ea335ff\": rpc error: code = NotFound desc = could not find container \"bc721b530c08b433756e978ec0423cc18aabf6771b61eb96f903b4d57ea335ff\": container with ID starting with bc721b530c08b433756e978ec0423cc18aabf6771b61eb96f903b4d57ea335ff not found: ID does not exist" Dec 10 23:00:48 crc kubenswrapper[4791]: I1210 23:00:48.929202 4791 scope.go:117] "RemoveContainer" containerID="68795a29944e56870ffe94fa732e9dcf0bfb344c0d85d1e590f58a876129c61f" Dec 10 23:00:48 crc kubenswrapper[4791]: I1210 23:00:48.929568 4791 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"68795a29944e56870ffe94fa732e9dcf0bfb344c0d85d1e590f58a876129c61f"} err="failed to get container status \"68795a29944e56870ffe94fa732e9dcf0bfb344c0d85d1e590f58a876129c61f\": rpc error: code = NotFound desc = could not find container \"68795a29944e56870ffe94fa732e9dcf0bfb344c0d85d1e590f58a876129c61f\": container with ID starting with 68795a29944e56870ffe94fa732e9dcf0bfb344c0d85d1e590f58a876129c61f not found: ID does not exist" Dec 10 23:00:48 crc kubenswrapper[4791]: I1210 23:00:48.929585 4791 scope.go:117] "RemoveContainer" containerID="937137ae9d7480cdc7c59bc4a1df0a4c10e72b3b63693af285278789953ed218" Dec 10 23:00:48 crc kubenswrapper[4791]: I1210 23:00:48.929844 4791 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"937137ae9d7480cdc7c59bc4a1df0a4c10e72b3b63693af285278789953ed218"} err="failed to get container status \"937137ae9d7480cdc7c59bc4a1df0a4c10e72b3b63693af285278789953ed218\": rpc error: code = NotFound desc = could not find container \"937137ae9d7480cdc7c59bc4a1df0a4c10e72b3b63693af285278789953ed218\": container with ID starting with 937137ae9d7480cdc7c59bc4a1df0a4c10e72b3b63693af285278789953ed218 not found: ID does not exist" Dec 10 23:00:48 crc kubenswrapper[4791]: I1210 23:00:48.929870 4791 scope.go:117] "RemoveContainer" containerID="1a52d19b7671e8b143017c8a981a6f8d3ee3b84650e4b223b814a531ac76f364" Dec 10 23:00:48 crc kubenswrapper[4791]: I1210 23:00:48.930135 4791 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"1a52d19b7671e8b143017c8a981a6f8d3ee3b84650e4b223b814a531ac76f364"} err="failed to get container status \"1a52d19b7671e8b143017c8a981a6f8d3ee3b84650e4b223b814a531ac76f364\": rpc error: code = NotFound desc = could not find container \"1a52d19b7671e8b143017c8a981a6f8d3ee3b84650e4b223b814a531ac76f364\": container with ID starting with 1a52d19b7671e8b143017c8a981a6f8d3ee3b84650e4b223b814a531ac76f364 not found: ID does not exist" Dec 10 23:00:48 crc kubenswrapper[4791]: I1210 23:00:48.930157 4791 scope.go:117] "RemoveContainer" containerID="3afbd1ffcd2c4b40324aca6e52d6fc822db131c3f35a9af3a6f005c2cffdb7bf" Dec 10 23:00:48 crc kubenswrapper[4791]: I1210 23:00:48.930375 4791 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"3afbd1ffcd2c4b40324aca6e52d6fc822db131c3f35a9af3a6f005c2cffdb7bf"} err="failed to get container status \"3afbd1ffcd2c4b40324aca6e52d6fc822db131c3f35a9af3a6f005c2cffdb7bf\": rpc error: code = NotFound desc = could not find container \"3afbd1ffcd2c4b40324aca6e52d6fc822db131c3f35a9af3a6f005c2cffdb7bf\": container with ID starting with 3afbd1ffcd2c4b40324aca6e52d6fc822db131c3f35a9af3a6f005c2cffdb7bf not found: ID does not exist" Dec 10 23:00:49 crc kubenswrapper[4791]: I1210 23:00:49.684954 4791 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-multus_multus-4nwkq_672aa28c-8169-49ed-87b8-21187d13a80c/kube-multus/2.log" Dec 10 23:00:49 crc kubenswrapper[4791]: I1210 23:00:49.690003 4791 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-bdhcv" event={"ID":"b1289472-e6e4-468c-a54c-0a83ff1474e8","Type":"ContainerStarted","Data":"b2b6ad6507690784e0a31f6749c99839a106d4534570b89dff3f6fb505418828"} Dec 10 23:00:49 crc kubenswrapper[4791]: I1210 23:00:49.690058 4791 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-bdhcv" event={"ID":"b1289472-e6e4-468c-a54c-0a83ff1474e8","Type":"ContainerStarted","Data":"8d6c4e509685f0229f2b4284b0d1efa498c3476f39a853a9cf245c9054675f3d"} Dec 10 23:00:49 crc kubenswrapper[4791]: I1210 23:00:49.690077 4791 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-bdhcv" event={"ID":"b1289472-e6e4-468c-a54c-0a83ff1474e8","Type":"ContainerStarted","Data":"d0281642ba3163deef1e106532e5ec4062d1524643ba4e0d79c9ab66a4b531d1"} Dec 10 23:00:49 crc kubenswrapper[4791]: I1210 23:00:49.690095 4791 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-bdhcv" event={"ID":"b1289472-e6e4-468c-a54c-0a83ff1474e8","Type":"ContainerStarted","Data":"a9cb7750d1e4cc2652a503b16315e14ba05d21a391be9197d9f11ff9d3c4ffd1"} Dec 10 23:00:49 crc kubenswrapper[4791]: I1210 23:00:49.690112 4791 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-bdhcv" event={"ID":"b1289472-e6e4-468c-a54c-0a83ff1474e8","Type":"ContainerStarted","Data":"4aa45dbd6a1349991d60268c2344835b355b06edf8d49bbe798813e975e16732"} Dec 10 23:00:49 crc kubenswrapper[4791]: I1210 23:00:49.690127 4791 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-bdhcv" event={"ID":"b1289472-e6e4-468c-a54c-0a83ff1474e8","Type":"ContainerStarted","Data":"2dd2de4548e2316bb54607d345e778c66c5f49350a31f11c9e10da735576bc8b"} Dec 10 23:00:49 crc kubenswrapper[4791]: I1210 23:00:49.893865 4791 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="3cd47739-0fa9-4321-aff1-220f8721a0b3" path="/var/lib/kubelet/pods/3cd47739-0fa9-4321-aff1-220f8721a0b3/volumes" Dec 10 23:00:51 crc kubenswrapper[4791]: I1210 23:00:51.708760 4791 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-bdhcv" event={"ID":"b1289472-e6e4-468c-a54c-0a83ff1474e8","Type":"ContainerStarted","Data":"2d2a973cb980d6e4a2eef232a882c9f4bf27f790e8d756f03c28a8eb5730737a"} Dec 10 23:00:52 crc kubenswrapper[4791]: I1210 23:00:52.598956 4791 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="cert-manager/cert-manager-webhook-5655c58dd6-xzg6b" Dec 10 23:00:54 crc kubenswrapper[4791]: I1210 23:00:54.737299 4791 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-bdhcv" event={"ID":"b1289472-e6e4-468c-a54c-0a83ff1474e8","Type":"ContainerStarted","Data":"0a61edaa616bb8ce88a34fdd74890e1d1c7abb9b93e7d1f8dfcba55baa32d956"} Dec 10 23:00:54 crc kubenswrapper[4791]: I1210 23:00:54.738573 4791 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-ovn-kubernetes/ovnkube-node-bdhcv" Dec 10 23:00:54 crc kubenswrapper[4791]: I1210 23:00:54.738695 4791 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-ovn-kubernetes/ovnkube-node-bdhcv" Dec 10 23:00:54 crc kubenswrapper[4791]: I1210 23:00:54.739138 4791 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-ovn-kubernetes/ovnkube-node-bdhcv" Dec 10 23:00:54 crc kubenswrapper[4791]: I1210 23:00:54.775177 4791 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-ovn-kubernetes/ovnkube-node-bdhcv" podStartSLOduration=7.775163465 podStartE2EDuration="7.775163465s" podCreationTimestamp="2025-12-10 23:00:47 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-10 23:00:54.770039338 +0000 UTC m=+689.199656961" watchObservedRunningTime="2025-12-10 23:00:54.775163465 +0000 UTC m=+689.204781078" Dec 10 23:00:54 crc kubenswrapper[4791]: I1210 23:00:54.782308 4791 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-ovn-kubernetes/ovnkube-node-bdhcv" Dec 10 23:00:54 crc kubenswrapper[4791]: I1210 23:00:54.787582 4791 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-ovn-kubernetes/ovnkube-node-bdhcv" Dec 10 23:01:01 crc kubenswrapper[4791]: I1210 23:01:01.885595 4791 scope.go:117] "RemoveContainer" containerID="ccdf9cf4480255bc2180847012278cee2a7f7e9392d14b7affe69a361b92cf4b" Dec 10 23:01:01 crc kubenswrapper[4791]: E1210 23:01:01.886617 4791 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"kube-multus\" with CrashLoopBackOff: \"back-off 20s restarting failed container=kube-multus pod=multus-4nwkq_openshift-multus(672aa28c-8169-49ed-87b8-21187d13a80c)\"" pod="openshift-multus/multus-4nwkq" podUID="672aa28c-8169-49ed-87b8-21187d13a80c" Dec 10 23:01:16 crc kubenswrapper[4791]: I1210 23:01:16.886239 4791 scope.go:117] "RemoveContainer" containerID="ccdf9cf4480255bc2180847012278cee2a7f7e9392d14b7affe69a361b92cf4b" Dec 10 23:01:17 crc kubenswrapper[4791]: I1210 23:01:17.873577 4791 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-multus_multus-4nwkq_672aa28c-8169-49ed-87b8-21187d13a80c/kube-multus/2.log" Dec 10 23:01:17 crc kubenswrapper[4791]: I1210 23:01:17.874092 4791 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-4nwkq" event={"ID":"672aa28c-8169-49ed-87b8-21187d13a80c","Type":"ContainerStarted","Data":"d8a819d9cf93286232d283b37563a4bc370b8d880800c5908294e18aebf7324a"} Dec 10 23:01:18 crc kubenswrapper[4791]: I1210 23:01:18.364168 4791 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-ovn-kubernetes/ovnkube-node-bdhcv" Dec 10 23:01:26 crc kubenswrapper[4791]: I1210 23:01:26.082909 4791 scope.go:117] "RemoveContainer" containerID="018136a3268c789d9ddd9dd74003b3e01e9f3efae97293726a35d6582353bb26" Dec 10 23:01:40 crc kubenswrapper[4791]: I1210 23:01:40.034950 4791 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/5064f9f8917b246f69f5d7fc025e7e6c34236c02bca31167615d38212fppnkc"] Dec 10 23:01:40 crc kubenswrapper[4791]: I1210 23:01:40.036537 4791 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/5064f9f8917b246f69f5d7fc025e7e6c34236c02bca31167615d38212fppnkc" Dec 10 23:01:40 crc kubenswrapper[4791]: I1210 23:01:40.039533 4791 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"default-dockercfg-vmwhc" Dec 10 23:01:40 crc kubenswrapper[4791]: I1210 23:01:40.041819 4791 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/5064f9f8917b246f69f5d7fc025e7e6c34236c02bca31167615d38212fppnkc"] Dec 10 23:01:40 crc kubenswrapper[4791]: I1210 23:01:40.070073 4791 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/4ebf6fba-353e-4d14-9e9e-f2be85033920-util\") pod \"5064f9f8917b246f69f5d7fc025e7e6c34236c02bca31167615d38212fppnkc\" (UID: \"4ebf6fba-353e-4d14-9e9e-f2be85033920\") " pod="openshift-marketplace/5064f9f8917b246f69f5d7fc025e7e6c34236c02bca31167615d38212fppnkc" Dec 10 23:01:40 crc kubenswrapper[4791]: I1210 23:01:40.070329 4791 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-8bdrf\" (UniqueName: \"kubernetes.io/projected/4ebf6fba-353e-4d14-9e9e-f2be85033920-kube-api-access-8bdrf\") pod \"5064f9f8917b246f69f5d7fc025e7e6c34236c02bca31167615d38212fppnkc\" (UID: \"4ebf6fba-353e-4d14-9e9e-f2be85033920\") " pod="openshift-marketplace/5064f9f8917b246f69f5d7fc025e7e6c34236c02bca31167615d38212fppnkc" Dec 10 23:01:40 crc kubenswrapper[4791]: I1210 23:01:40.070482 4791 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/4ebf6fba-353e-4d14-9e9e-f2be85033920-bundle\") pod \"5064f9f8917b246f69f5d7fc025e7e6c34236c02bca31167615d38212fppnkc\" (UID: \"4ebf6fba-353e-4d14-9e9e-f2be85033920\") " pod="openshift-marketplace/5064f9f8917b246f69f5d7fc025e7e6c34236c02bca31167615d38212fppnkc" Dec 10 23:01:40 crc kubenswrapper[4791]: I1210 23:01:40.170958 4791 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-8bdrf\" (UniqueName: \"kubernetes.io/projected/4ebf6fba-353e-4d14-9e9e-f2be85033920-kube-api-access-8bdrf\") pod \"5064f9f8917b246f69f5d7fc025e7e6c34236c02bca31167615d38212fppnkc\" (UID: \"4ebf6fba-353e-4d14-9e9e-f2be85033920\") " pod="openshift-marketplace/5064f9f8917b246f69f5d7fc025e7e6c34236c02bca31167615d38212fppnkc" Dec 10 23:01:40 crc kubenswrapper[4791]: I1210 23:01:40.171755 4791 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/4ebf6fba-353e-4d14-9e9e-f2be85033920-util\") pod \"5064f9f8917b246f69f5d7fc025e7e6c34236c02bca31167615d38212fppnkc\" (UID: \"4ebf6fba-353e-4d14-9e9e-f2be85033920\") " pod="openshift-marketplace/5064f9f8917b246f69f5d7fc025e7e6c34236c02bca31167615d38212fppnkc" Dec 10 23:01:40 crc kubenswrapper[4791]: I1210 23:01:40.171826 4791 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/4ebf6fba-353e-4d14-9e9e-f2be85033920-bundle\") pod \"5064f9f8917b246f69f5d7fc025e7e6c34236c02bca31167615d38212fppnkc\" (UID: \"4ebf6fba-353e-4d14-9e9e-f2be85033920\") " pod="openshift-marketplace/5064f9f8917b246f69f5d7fc025e7e6c34236c02bca31167615d38212fppnkc" Dec 10 23:01:40 crc kubenswrapper[4791]: I1210 23:01:40.172362 4791 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/4ebf6fba-353e-4d14-9e9e-f2be85033920-util\") pod \"5064f9f8917b246f69f5d7fc025e7e6c34236c02bca31167615d38212fppnkc\" (UID: \"4ebf6fba-353e-4d14-9e9e-f2be85033920\") " pod="openshift-marketplace/5064f9f8917b246f69f5d7fc025e7e6c34236c02bca31167615d38212fppnkc" Dec 10 23:01:40 crc kubenswrapper[4791]: I1210 23:01:40.172578 4791 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/4ebf6fba-353e-4d14-9e9e-f2be85033920-bundle\") pod \"5064f9f8917b246f69f5d7fc025e7e6c34236c02bca31167615d38212fppnkc\" (UID: \"4ebf6fba-353e-4d14-9e9e-f2be85033920\") " pod="openshift-marketplace/5064f9f8917b246f69f5d7fc025e7e6c34236c02bca31167615d38212fppnkc" Dec 10 23:01:40 crc kubenswrapper[4791]: I1210 23:01:40.196333 4791 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-8bdrf\" (UniqueName: \"kubernetes.io/projected/4ebf6fba-353e-4d14-9e9e-f2be85033920-kube-api-access-8bdrf\") pod \"5064f9f8917b246f69f5d7fc025e7e6c34236c02bca31167615d38212fppnkc\" (UID: \"4ebf6fba-353e-4d14-9e9e-f2be85033920\") " pod="openshift-marketplace/5064f9f8917b246f69f5d7fc025e7e6c34236c02bca31167615d38212fppnkc" Dec 10 23:01:40 crc kubenswrapper[4791]: I1210 23:01:40.355562 4791 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/5064f9f8917b246f69f5d7fc025e7e6c34236c02bca31167615d38212fppnkc" Dec 10 23:01:40 crc kubenswrapper[4791]: I1210 23:01:40.773960 4791 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/5064f9f8917b246f69f5d7fc025e7e6c34236c02bca31167615d38212fppnkc"] Dec 10 23:01:40 crc kubenswrapper[4791]: W1210 23:01:40.785159 4791 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod4ebf6fba_353e_4d14_9e9e_f2be85033920.slice/crio-759dddb688b0329b5908ec94254946c6331697c40326c1f66ee22b58269b9441 WatchSource:0}: Error finding container 759dddb688b0329b5908ec94254946c6331697c40326c1f66ee22b58269b9441: Status 404 returned error can't find the container with id 759dddb688b0329b5908ec94254946c6331697c40326c1f66ee22b58269b9441 Dec 10 23:01:41 crc kubenswrapper[4791]: I1210 23:01:41.018579 4791 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/5064f9f8917b246f69f5d7fc025e7e6c34236c02bca31167615d38212fppnkc" event={"ID":"4ebf6fba-353e-4d14-9e9e-f2be85033920","Type":"ContainerStarted","Data":"d65706b366584fc973e5af7cc3947321ad52acd04ff80f0b195f6057333a204b"} Dec 10 23:01:41 crc kubenswrapper[4791]: I1210 23:01:41.018632 4791 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/5064f9f8917b246f69f5d7fc025e7e6c34236c02bca31167615d38212fppnkc" event={"ID":"4ebf6fba-353e-4d14-9e9e-f2be85033920","Type":"ContainerStarted","Data":"759dddb688b0329b5908ec94254946c6331697c40326c1f66ee22b58269b9441"} Dec 10 23:01:42 crc kubenswrapper[4791]: I1210 23:01:42.025890 4791 generic.go:334] "Generic (PLEG): container finished" podID="4ebf6fba-353e-4d14-9e9e-f2be85033920" containerID="d65706b366584fc973e5af7cc3947321ad52acd04ff80f0b195f6057333a204b" exitCode=0 Dec 10 23:01:42 crc kubenswrapper[4791]: I1210 23:01:42.025943 4791 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/5064f9f8917b246f69f5d7fc025e7e6c34236c02bca31167615d38212fppnkc" event={"ID":"4ebf6fba-353e-4d14-9e9e-f2be85033920","Type":"ContainerDied","Data":"d65706b366584fc973e5af7cc3947321ad52acd04ff80f0b195f6057333a204b"} Dec 10 23:01:45 crc kubenswrapper[4791]: I1210 23:01:45.042787 4791 generic.go:334] "Generic (PLEG): container finished" podID="4ebf6fba-353e-4d14-9e9e-f2be85033920" containerID="652fba2a290d2382c703317428d16e35d781298e4ef640ea156984e5547bbbeb" exitCode=0 Dec 10 23:01:45 crc kubenswrapper[4791]: I1210 23:01:45.042863 4791 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/5064f9f8917b246f69f5d7fc025e7e6c34236c02bca31167615d38212fppnkc" event={"ID":"4ebf6fba-353e-4d14-9e9e-f2be85033920","Type":"ContainerDied","Data":"652fba2a290d2382c703317428d16e35d781298e4ef640ea156984e5547bbbeb"} Dec 10 23:01:46 crc kubenswrapper[4791]: I1210 23:01:46.053863 4791 generic.go:334] "Generic (PLEG): container finished" podID="4ebf6fba-353e-4d14-9e9e-f2be85033920" containerID="0017c7107aee3257389ef0078a1a30f4c932cd9200e3ee2295b01f953d8a05b3" exitCode=0 Dec 10 23:01:46 crc kubenswrapper[4791]: I1210 23:01:46.054069 4791 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/5064f9f8917b246f69f5d7fc025e7e6c34236c02bca31167615d38212fppnkc" event={"ID":"4ebf6fba-353e-4d14-9e9e-f2be85033920","Type":"ContainerDied","Data":"0017c7107aee3257389ef0078a1a30f4c932cd9200e3ee2295b01f953d8a05b3"} Dec 10 23:01:47 crc kubenswrapper[4791]: I1210 23:01:47.364141 4791 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/5064f9f8917b246f69f5d7fc025e7e6c34236c02bca31167615d38212fppnkc" Dec 10 23:01:47 crc kubenswrapper[4791]: I1210 23:01:47.469655 4791 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/4ebf6fba-353e-4d14-9e9e-f2be85033920-bundle\") pod \"4ebf6fba-353e-4d14-9e9e-f2be85033920\" (UID: \"4ebf6fba-353e-4d14-9e9e-f2be85033920\") " Dec 10 23:01:47 crc kubenswrapper[4791]: I1210 23:01:47.469711 4791 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-8bdrf\" (UniqueName: \"kubernetes.io/projected/4ebf6fba-353e-4d14-9e9e-f2be85033920-kube-api-access-8bdrf\") pod \"4ebf6fba-353e-4d14-9e9e-f2be85033920\" (UID: \"4ebf6fba-353e-4d14-9e9e-f2be85033920\") " Dec 10 23:01:47 crc kubenswrapper[4791]: I1210 23:01:47.469730 4791 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/4ebf6fba-353e-4d14-9e9e-f2be85033920-util\") pod \"4ebf6fba-353e-4d14-9e9e-f2be85033920\" (UID: \"4ebf6fba-353e-4d14-9e9e-f2be85033920\") " Dec 10 23:01:47 crc kubenswrapper[4791]: I1210 23:01:47.470823 4791 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/4ebf6fba-353e-4d14-9e9e-f2be85033920-bundle" (OuterVolumeSpecName: "bundle") pod "4ebf6fba-353e-4d14-9e9e-f2be85033920" (UID: "4ebf6fba-353e-4d14-9e9e-f2be85033920"). InnerVolumeSpecName "bundle". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 10 23:01:47 crc kubenswrapper[4791]: I1210 23:01:47.476160 4791 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/4ebf6fba-353e-4d14-9e9e-f2be85033920-kube-api-access-8bdrf" (OuterVolumeSpecName: "kube-api-access-8bdrf") pod "4ebf6fba-353e-4d14-9e9e-f2be85033920" (UID: "4ebf6fba-353e-4d14-9e9e-f2be85033920"). InnerVolumeSpecName "kube-api-access-8bdrf". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 10 23:01:47 crc kubenswrapper[4791]: I1210 23:01:47.480281 4791 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/4ebf6fba-353e-4d14-9e9e-f2be85033920-util" (OuterVolumeSpecName: "util") pod "4ebf6fba-353e-4d14-9e9e-f2be85033920" (UID: "4ebf6fba-353e-4d14-9e9e-f2be85033920"). InnerVolumeSpecName "util". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 10 23:01:47 crc kubenswrapper[4791]: I1210 23:01:47.570393 4791 reconciler_common.go:293] "Volume detached for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/4ebf6fba-353e-4d14-9e9e-f2be85033920-bundle\") on node \"crc\" DevicePath \"\"" Dec 10 23:01:47 crc kubenswrapper[4791]: I1210 23:01:47.570449 4791 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-8bdrf\" (UniqueName: \"kubernetes.io/projected/4ebf6fba-353e-4d14-9e9e-f2be85033920-kube-api-access-8bdrf\") on node \"crc\" DevicePath \"\"" Dec 10 23:01:47 crc kubenswrapper[4791]: I1210 23:01:47.570463 4791 reconciler_common.go:293] "Volume detached for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/4ebf6fba-353e-4d14-9e9e-f2be85033920-util\") on node \"crc\" DevicePath \"\"" Dec 10 23:01:48 crc kubenswrapper[4791]: I1210 23:01:48.077454 4791 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/5064f9f8917b246f69f5d7fc025e7e6c34236c02bca31167615d38212fppnkc" event={"ID":"4ebf6fba-353e-4d14-9e9e-f2be85033920","Type":"ContainerDied","Data":"759dddb688b0329b5908ec94254946c6331697c40326c1f66ee22b58269b9441"} Dec 10 23:01:48 crc kubenswrapper[4791]: I1210 23:01:48.077522 4791 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="759dddb688b0329b5908ec94254946c6331697c40326c1f66ee22b58269b9441" Dec 10 23:01:48 crc kubenswrapper[4791]: I1210 23:01:48.077540 4791 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/5064f9f8917b246f69f5d7fc025e7e6c34236c02bca31167615d38212fppnkc" Dec 10 23:01:51 crc kubenswrapper[4791]: I1210 23:01:51.384714 4791 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-nmstate/nmstate-operator-5b5b58f5c8-9qtjh"] Dec 10 23:01:51 crc kubenswrapper[4791]: E1210 23:01:51.385247 4791 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="4ebf6fba-353e-4d14-9e9e-f2be85033920" containerName="util" Dec 10 23:01:51 crc kubenswrapper[4791]: I1210 23:01:51.385261 4791 state_mem.go:107] "Deleted CPUSet assignment" podUID="4ebf6fba-353e-4d14-9e9e-f2be85033920" containerName="util" Dec 10 23:01:51 crc kubenswrapper[4791]: E1210 23:01:51.385275 4791 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="4ebf6fba-353e-4d14-9e9e-f2be85033920" containerName="pull" Dec 10 23:01:51 crc kubenswrapper[4791]: I1210 23:01:51.385284 4791 state_mem.go:107] "Deleted CPUSet assignment" podUID="4ebf6fba-353e-4d14-9e9e-f2be85033920" containerName="pull" Dec 10 23:01:51 crc kubenswrapper[4791]: E1210 23:01:51.385297 4791 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="4ebf6fba-353e-4d14-9e9e-f2be85033920" containerName="extract" Dec 10 23:01:51 crc kubenswrapper[4791]: I1210 23:01:51.385304 4791 state_mem.go:107] "Deleted CPUSet assignment" podUID="4ebf6fba-353e-4d14-9e9e-f2be85033920" containerName="extract" Dec 10 23:01:51 crc kubenswrapper[4791]: I1210 23:01:51.385442 4791 memory_manager.go:354] "RemoveStaleState removing state" podUID="4ebf6fba-353e-4d14-9e9e-f2be85033920" containerName="extract" Dec 10 23:01:51 crc kubenswrapper[4791]: I1210 23:01:51.386077 4791 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-nmstate/nmstate-operator-5b5b58f5c8-9qtjh" Dec 10 23:01:51 crc kubenswrapper[4791]: I1210 23:01:51.389314 4791 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-nmstate"/"openshift-service-ca.crt" Dec 10 23:01:51 crc kubenswrapper[4791]: I1210 23:01:51.389875 4791 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-nmstate"/"kube-root-ca.crt" Dec 10 23:01:51 crc kubenswrapper[4791]: I1210 23:01:51.390238 4791 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-nmstate"/"nmstate-operator-dockercfg-fzpsk" Dec 10 23:01:51 crc kubenswrapper[4791]: I1210 23:01:51.399223 4791 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-nmstate/nmstate-operator-5b5b58f5c8-9qtjh"] Dec 10 23:01:51 crc kubenswrapper[4791]: I1210 23:01:51.519402 4791 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-gflts\" (UniqueName: \"kubernetes.io/projected/f6780aea-2f04-4dc3-96c6-8e4e7ef4284e-kube-api-access-gflts\") pod \"nmstate-operator-5b5b58f5c8-9qtjh\" (UID: \"f6780aea-2f04-4dc3-96c6-8e4e7ef4284e\") " pod="openshift-nmstate/nmstate-operator-5b5b58f5c8-9qtjh" Dec 10 23:01:51 crc kubenswrapper[4791]: I1210 23:01:51.620624 4791 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-gflts\" (UniqueName: \"kubernetes.io/projected/f6780aea-2f04-4dc3-96c6-8e4e7ef4284e-kube-api-access-gflts\") pod \"nmstate-operator-5b5b58f5c8-9qtjh\" (UID: \"f6780aea-2f04-4dc3-96c6-8e4e7ef4284e\") " pod="openshift-nmstate/nmstate-operator-5b5b58f5c8-9qtjh" Dec 10 23:01:51 crc kubenswrapper[4791]: I1210 23:01:51.645791 4791 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-gflts\" (UniqueName: \"kubernetes.io/projected/f6780aea-2f04-4dc3-96c6-8e4e7ef4284e-kube-api-access-gflts\") pod \"nmstate-operator-5b5b58f5c8-9qtjh\" (UID: \"f6780aea-2f04-4dc3-96c6-8e4e7ef4284e\") " pod="openshift-nmstate/nmstate-operator-5b5b58f5c8-9qtjh" Dec 10 23:01:51 crc kubenswrapper[4791]: I1210 23:01:51.701323 4791 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-nmstate/nmstate-operator-5b5b58f5c8-9qtjh" Dec 10 23:01:51 crc kubenswrapper[4791]: I1210 23:01:51.914699 4791 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-nmstate/nmstate-operator-5b5b58f5c8-9qtjh"] Dec 10 23:01:52 crc kubenswrapper[4791]: I1210 23:01:52.098701 4791 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-nmstate/nmstate-operator-5b5b58f5c8-9qtjh" event={"ID":"f6780aea-2f04-4dc3-96c6-8e4e7ef4284e","Type":"ContainerStarted","Data":"9f70544b6adf5f5da319bb50d42b40afe38a66c8ea907f2c141edd8285ab2e6a"} Dec 10 23:01:55 crc kubenswrapper[4791]: I1210 23:01:55.038266 4791 patch_prober.go:28] interesting pod/machine-config-daemon-5rb5l container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 10 23:01:55 crc kubenswrapper[4791]: I1210 23:01:55.038714 4791 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-5rb5l" podUID="ba35653c-6e06-4cee-a4d6-137764090d18" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 10 23:01:56 crc kubenswrapper[4791]: I1210 23:01:56.123037 4791 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-nmstate/nmstate-operator-5b5b58f5c8-9qtjh" event={"ID":"f6780aea-2f04-4dc3-96c6-8e4e7ef4284e","Type":"ContainerStarted","Data":"a540d72506b77966ba963e52537e274691f7ae4432c60b8bcae1757bae1ef7fd"} Dec 10 23:01:56 crc kubenswrapper[4791]: I1210 23:01:56.142382 4791 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-nmstate/nmstate-operator-5b5b58f5c8-9qtjh" podStartSLOduration=1.720726231 podStartE2EDuration="5.14235509s" podCreationTimestamp="2025-12-10 23:01:51 +0000 UTC" firstStartedPulling="2025-12-10 23:01:51.920890654 +0000 UTC m=+746.350508267" lastFinishedPulling="2025-12-10 23:01:55.342519513 +0000 UTC m=+749.772137126" observedRunningTime="2025-12-10 23:01:56.137985017 +0000 UTC m=+750.567602640" watchObservedRunningTime="2025-12-10 23:01:56.14235509 +0000 UTC m=+750.571972703" Dec 10 23:02:00 crc kubenswrapper[4791]: I1210 23:02:00.305019 4791 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-nmstate/nmstate-metrics-7f946cbc9-mph8d"] Dec 10 23:02:00 crc kubenswrapper[4791]: I1210 23:02:00.306315 4791 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-nmstate/nmstate-metrics-7f946cbc9-mph8d" Dec 10 23:02:00 crc kubenswrapper[4791]: I1210 23:02:00.308145 4791 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-nmstate"/"nmstate-handler-dockercfg-8qbjq" Dec 10 23:02:00 crc kubenswrapper[4791]: I1210 23:02:00.314539 4791 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-nmstate/nmstate-webhook-5f6d4c5ccb-x65fz"] Dec 10 23:02:00 crc kubenswrapper[4791]: I1210 23:02:00.315648 4791 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-nmstate/nmstate-webhook-5f6d4c5ccb-x65fz" Dec 10 23:02:00 crc kubenswrapper[4791]: I1210 23:02:00.317160 4791 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-nmstate"/"openshift-nmstate-webhook" Dec 10 23:02:00 crc kubenswrapper[4791]: I1210 23:02:00.317802 4791 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-nmstate/nmstate-metrics-7f946cbc9-mph8d"] Dec 10 23:02:00 crc kubenswrapper[4791]: I1210 23:02:00.339468 4791 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-nmstate/nmstate-webhook-5f6d4c5ccb-x65fz"] Dec 10 23:02:00 crc kubenswrapper[4791]: I1210 23:02:00.370049 4791 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-nmstate/nmstate-handler-2j8xz"] Dec 10 23:02:00 crc kubenswrapper[4791]: I1210 23:02:00.371780 4791 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-nmstate/nmstate-handler-2j8xz" Dec 10 23:02:00 crc kubenswrapper[4791]: I1210 23:02:00.493022 4791 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"tls-key-pair\" (UniqueName: \"kubernetes.io/secret/99524090-326f-47fd-98d9-bf7e8352cc4f-tls-key-pair\") pod \"nmstate-webhook-5f6d4c5ccb-x65fz\" (UID: \"99524090-326f-47fd-98d9-bf7e8352cc4f\") " pod="openshift-nmstate/nmstate-webhook-5f6d4c5ccb-x65fz" Dec 10 23:02:00 crc kubenswrapper[4791]: I1210 23:02:00.493119 4791 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-flb4w\" (UniqueName: \"kubernetes.io/projected/e85292bb-7cee-4b20-8c97-b1d1902e6001-kube-api-access-flb4w\") pod \"nmstate-metrics-7f946cbc9-mph8d\" (UID: \"e85292bb-7cee-4b20-8c97-b1d1902e6001\") " pod="openshift-nmstate/nmstate-metrics-7f946cbc9-mph8d" Dec 10 23:02:00 crc kubenswrapper[4791]: I1210 23:02:00.493265 4791 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-rtlpd\" (UniqueName: \"kubernetes.io/projected/99524090-326f-47fd-98d9-bf7e8352cc4f-kube-api-access-rtlpd\") pod \"nmstate-webhook-5f6d4c5ccb-x65fz\" (UID: \"99524090-326f-47fd-98d9-bf7e8352cc4f\") " pod="openshift-nmstate/nmstate-webhook-5f6d4c5ccb-x65fz" Dec 10 23:02:00 crc kubenswrapper[4791]: I1210 23:02:00.514573 4791 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-nmstate/nmstate-console-plugin-7fbb5f6569-xxb9s"] Dec 10 23:02:00 crc kubenswrapper[4791]: I1210 23:02:00.515317 4791 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-nmstate/nmstate-console-plugin-7fbb5f6569-xxb9s" Dec 10 23:02:00 crc kubenswrapper[4791]: I1210 23:02:00.518930 4791 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-nmstate"/"plugin-serving-cert" Dec 10 23:02:00 crc kubenswrapper[4791]: I1210 23:02:00.519900 4791 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-nmstate"/"default-dockercfg-27xmz" Dec 10 23:02:00 crc kubenswrapper[4791]: I1210 23:02:00.520469 4791 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-nmstate"/"nginx-conf" Dec 10 23:02:00 crc kubenswrapper[4791]: I1210 23:02:00.525084 4791 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-nmstate/nmstate-console-plugin-7fbb5f6569-xxb9s"] Dec 10 23:02:00 crc kubenswrapper[4791]: I1210 23:02:00.593986 4791 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-flb4w\" (UniqueName: \"kubernetes.io/projected/e85292bb-7cee-4b20-8c97-b1d1902e6001-kube-api-access-flb4w\") pod \"nmstate-metrics-7f946cbc9-mph8d\" (UID: \"e85292bb-7cee-4b20-8c97-b1d1902e6001\") " pod="openshift-nmstate/nmstate-metrics-7f946cbc9-mph8d" Dec 10 23:02:00 crc kubenswrapper[4791]: I1210 23:02:00.594048 4791 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"nmstate-lock\" (UniqueName: \"kubernetes.io/host-path/fd7f0c56-2613-4ba7-b0a4-4ff765493973-nmstate-lock\") pod \"nmstate-handler-2j8xz\" (UID: \"fd7f0c56-2613-4ba7-b0a4-4ff765493973\") " pod="openshift-nmstate/nmstate-handler-2j8xz" Dec 10 23:02:00 crc kubenswrapper[4791]: I1210 23:02:00.594072 4791 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-vznmt\" (UniqueName: \"kubernetes.io/projected/fd7f0c56-2613-4ba7-b0a4-4ff765493973-kube-api-access-vznmt\") pod \"nmstate-handler-2j8xz\" (UID: \"fd7f0c56-2613-4ba7-b0a4-4ff765493973\") " pod="openshift-nmstate/nmstate-handler-2j8xz" Dec 10 23:02:00 crc kubenswrapper[4791]: I1210 23:02:00.594095 4791 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-rtlpd\" (UniqueName: \"kubernetes.io/projected/99524090-326f-47fd-98d9-bf7e8352cc4f-kube-api-access-rtlpd\") pod \"nmstate-webhook-5f6d4c5ccb-x65fz\" (UID: \"99524090-326f-47fd-98d9-bf7e8352cc4f\") " pod="openshift-nmstate/nmstate-webhook-5f6d4c5ccb-x65fz" Dec 10 23:02:00 crc kubenswrapper[4791]: I1210 23:02:00.594123 4791 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovs-socket\" (UniqueName: \"kubernetes.io/host-path/fd7f0c56-2613-4ba7-b0a4-4ff765493973-ovs-socket\") pod \"nmstate-handler-2j8xz\" (UID: \"fd7f0c56-2613-4ba7-b0a4-4ff765493973\") " pod="openshift-nmstate/nmstate-handler-2j8xz" Dec 10 23:02:00 crc kubenswrapper[4791]: I1210 23:02:00.594149 4791 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dbus-socket\" (UniqueName: \"kubernetes.io/host-path/fd7f0c56-2613-4ba7-b0a4-4ff765493973-dbus-socket\") pod \"nmstate-handler-2j8xz\" (UID: \"fd7f0c56-2613-4ba7-b0a4-4ff765493973\") " pod="openshift-nmstate/nmstate-handler-2j8xz" Dec 10 23:02:00 crc kubenswrapper[4791]: I1210 23:02:00.594165 4791 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"tls-key-pair\" (UniqueName: \"kubernetes.io/secret/99524090-326f-47fd-98d9-bf7e8352cc4f-tls-key-pair\") pod \"nmstate-webhook-5f6d4c5ccb-x65fz\" (UID: \"99524090-326f-47fd-98d9-bf7e8352cc4f\") " pod="openshift-nmstate/nmstate-webhook-5f6d4c5ccb-x65fz" Dec 10 23:02:00 crc kubenswrapper[4791]: I1210 23:02:00.600784 4791 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"tls-key-pair\" (UniqueName: \"kubernetes.io/secret/99524090-326f-47fd-98d9-bf7e8352cc4f-tls-key-pair\") pod \"nmstate-webhook-5f6d4c5ccb-x65fz\" (UID: \"99524090-326f-47fd-98d9-bf7e8352cc4f\") " pod="openshift-nmstate/nmstate-webhook-5f6d4c5ccb-x65fz" Dec 10 23:02:00 crc kubenswrapper[4791]: I1210 23:02:00.620648 4791 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-rtlpd\" (UniqueName: \"kubernetes.io/projected/99524090-326f-47fd-98d9-bf7e8352cc4f-kube-api-access-rtlpd\") pod \"nmstate-webhook-5f6d4c5ccb-x65fz\" (UID: \"99524090-326f-47fd-98d9-bf7e8352cc4f\") " pod="openshift-nmstate/nmstate-webhook-5f6d4c5ccb-x65fz" Dec 10 23:02:00 crc kubenswrapper[4791]: I1210 23:02:00.621387 4791 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-flb4w\" (UniqueName: \"kubernetes.io/projected/e85292bb-7cee-4b20-8c97-b1d1902e6001-kube-api-access-flb4w\") pod \"nmstate-metrics-7f946cbc9-mph8d\" (UID: \"e85292bb-7cee-4b20-8c97-b1d1902e6001\") " pod="openshift-nmstate/nmstate-metrics-7f946cbc9-mph8d" Dec 10 23:02:00 crc kubenswrapper[4791]: I1210 23:02:00.633596 4791 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-nmstate/nmstate-metrics-7f946cbc9-mph8d" Dec 10 23:02:00 crc kubenswrapper[4791]: I1210 23:02:00.656297 4791 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-nmstate/nmstate-webhook-5f6d4c5ccb-x65fz" Dec 10 23:02:00 crc kubenswrapper[4791]: I1210 23:02:00.677992 4791 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-console/console-6c597db9d6-86qkv"] Dec 10 23:02:00 crc kubenswrapper[4791]: I1210 23:02:00.682995 4791 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-console/console-6c597db9d6-86qkv" Dec 10 23:02:00 crc kubenswrapper[4791]: I1210 23:02:00.705023 4791 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-vznmt\" (UniqueName: \"kubernetes.io/projected/fd7f0c56-2613-4ba7-b0a4-4ff765493973-kube-api-access-vznmt\") pod \"nmstate-handler-2j8xz\" (UID: \"fd7f0c56-2613-4ba7-b0a4-4ff765493973\") " pod="openshift-nmstate/nmstate-handler-2j8xz" Dec 10 23:02:00 crc kubenswrapper[4791]: I1210 23:02:00.705092 4791 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovs-socket\" (UniqueName: \"kubernetes.io/host-path/fd7f0c56-2613-4ba7-b0a4-4ff765493973-ovs-socket\") pod \"nmstate-handler-2j8xz\" (UID: \"fd7f0c56-2613-4ba7-b0a4-4ff765493973\") " pod="openshift-nmstate/nmstate-handler-2j8xz" Dec 10 23:02:00 crc kubenswrapper[4791]: I1210 23:02:00.705129 4791 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"console-serving-cert\" (UniqueName: \"kubernetes.io/secret/681a9508-afbe-4ebd-80dd-473006e24d93-console-serving-cert\") pod \"console-6c597db9d6-86qkv\" (UID: \"681a9508-afbe-4ebd-80dd-473006e24d93\") " pod="openshift-console/console-6c597db9d6-86qkv" Dec 10 23:02:00 crc kubenswrapper[4791]: I1210 23:02:00.705165 4791 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"nginx-conf\" (UniqueName: \"kubernetes.io/configmap/14000c4a-8a7d-4ade-86e6-819e08eaf42a-nginx-conf\") pod \"nmstate-console-plugin-7fbb5f6569-xxb9s\" (UID: \"14000c4a-8a7d-4ade-86e6-819e08eaf42a\") " pod="openshift-nmstate/nmstate-console-plugin-7fbb5f6569-xxb9s" Dec 10 23:02:00 crc kubenswrapper[4791]: I1210 23:02:00.705204 4791 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dbus-socket\" (UniqueName: \"kubernetes.io/host-path/fd7f0c56-2613-4ba7-b0a4-4ff765493973-dbus-socket\") pod \"nmstate-handler-2j8xz\" (UID: \"fd7f0c56-2613-4ba7-b0a4-4ff765493973\") " pod="openshift-nmstate/nmstate-handler-2j8xz" Dec 10 23:02:00 crc kubenswrapper[4791]: I1210 23:02:00.705235 4791 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"console-config\" (UniqueName: \"kubernetes.io/configmap/681a9508-afbe-4ebd-80dd-473006e24d93-console-config\") pod \"console-6c597db9d6-86qkv\" (UID: \"681a9508-afbe-4ebd-80dd-473006e24d93\") " pod="openshift-console/console-6c597db9d6-86qkv" Dec 10 23:02:00 crc kubenswrapper[4791]: I1210 23:02:00.705272 4791 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-844f2\" (UniqueName: \"kubernetes.io/projected/681a9508-afbe-4ebd-80dd-473006e24d93-kube-api-access-844f2\") pod \"console-6c597db9d6-86qkv\" (UID: \"681a9508-afbe-4ebd-80dd-473006e24d93\") " pod="openshift-console/console-6c597db9d6-86qkv" Dec 10 23:02:00 crc kubenswrapper[4791]: I1210 23:02:00.705307 4791 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"oauth-serving-cert\" (UniqueName: \"kubernetes.io/configmap/681a9508-afbe-4ebd-80dd-473006e24d93-oauth-serving-cert\") pod \"console-6c597db9d6-86qkv\" (UID: \"681a9508-afbe-4ebd-80dd-473006e24d93\") " pod="openshift-console/console-6c597db9d6-86qkv" Dec 10 23:02:00 crc kubenswrapper[4791]: I1210 23:02:00.705357 4791 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/681a9508-afbe-4ebd-80dd-473006e24d93-service-ca\") pod \"console-6c597db9d6-86qkv\" (UID: \"681a9508-afbe-4ebd-80dd-473006e24d93\") " pod="openshift-console/console-6c597db9d6-86qkv" Dec 10 23:02:00 crc kubenswrapper[4791]: I1210 23:02:00.705381 4791 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/681a9508-afbe-4ebd-80dd-473006e24d93-trusted-ca-bundle\") pod \"console-6c597db9d6-86qkv\" (UID: \"681a9508-afbe-4ebd-80dd-473006e24d93\") " pod="openshift-console/console-6c597db9d6-86qkv" Dec 10 23:02:00 crc kubenswrapper[4791]: I1210 23:02:00.705410 4791 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"console-oauth-config\" (UniqueName: \"kubernetes.io/secret/681a9508-afbe-4ebd-80dd-473006e24d93-console-oauth-config\") pod \"console-6c597db9d6-86qkv\" (UID: \"681a9508-afbe-4ebd-80dd-473006e24d93\") " pod="openshift-console/console-6c597db9d6-86qkv" Dec 10 23:02:00 crc kubenswrapper[4791]: I1210 23:02:00.705437 4791 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"plugin-serving-cert\" (UniqueName: \"kubernetes.io/secret/14000c4a-8a7d-4ade-86e6-819e08eaf42a-plugin-serving-cert\") pod \"nmstate-console-plugin-7fbb5f6569-xxb9s\" (UID: \"14000c4a-8a7d-4ade-86e6-819e08eaf42a\") " pod="openshift-nmstate/nmstate-console-plugin-7fbb5f6569-xxb9s" Dec 10 23:02:00 crc kubenswrapper[4791]: I1210 23:02:00.705460 4791 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nmstate-lock\" (UniqueName: \"kubernetes.io/host-path/fd7f0c56-2613-4ba7-b0a4-4ff765493973-nmstate-lock\") pod \"nmstate-handler-2j8xz\" (UID: \"fd7f0c56-2613-4ba7-b0a4-4ff765493973\") " pod="openshift-nmstate/nmstate-handler-2j8xz" Dec 10 23:02:00 crc kubenswrapper[4791]: I1210 23:02:00.705481 4791 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-h6pqp\" (UniqueName: \"kubernetes.io/projected/14000c4a-8a7d-4ade-86e6-819e08eaf42a-kube-api-access-h6pqp\") pod \"nmstate-console-plugin-7fbb5f6569-xxb9s\" (UID: \"14000c4a-8a7d-4ade-86e6-819e08eaf42a\") " pod="openshift-nmstate/nmstate-console-plugin-7fbb5f6569-xxb9s" Dec 10 23:02:00 crc kubenswrapper[4791]: I1210 23:02:00.705889 4791 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovs-socket\" (UniqueName: \"kubernetes.io/host-path/fd7f0c56-2613-4ba7-b0a4-4ff765493973-ovs-socket\") pod \"nmstate-handler-2j8xz\" (UID: \"fd7f0c56-2613-4ba7-b0a4-4ff765493973\") " pod="openshift-nmstate/nmstate-handler-2j8xz" Dec 10 23:02:00 crc kubenswrapper[4791]: I1210 23:02:00.706168 4791 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dbus-socket\" (UniqueName: \"kubernetes.io/host-path/fd7f0c56-2613-4ba7-b0a4-4ff765493973-dbus-socket\") pod \"nmstate-handler-2j8xz\" (UID: \"fd7f0c56-2613-4ba7-b0a4-4ff765493973\") " pod="openshift-nmstate/nmstate-handler-2j8xz" Dec 10 23:02:00 crc kubenswrapper[4791]: I1210 23:02:00.706250 4791 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"nmstate-lock\" (UniqueName: \"kubernetes.io/host-path/fd7f0c56-2613-4ba7-b0a4-4ff765493973-nmstate-lock\") pod \"nmstate-handler-2j8xz\" (UID: \"fd7f0c56-2613-4ba7-b0a4-4ff765493973\") " pod="openshift-nmstate/nmstate-handler-2j8xz" Dec 10 23:02:00 crc kubenswrapper[4791]: I1210 23:02:00.712825 4791 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-console/console-6c597db9d6-86qkv"] Dec 10 23:02:00 crc kubenswrapper[4791]: I1210 23:02:00.749671 4791 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-vznmt\" (UniqueName: \"kubernetes.io/projected/fd7f0c56-2613-4ba7-b0a4-4ff765493973-kube-api-access-vznmt\") pod \"nmstate-handler-2j8xz\" (UID: \"fd7f0c56-2613-4ba7-b0a4-4ff765493973\") " pod="openshift-nmstate/nmstate-handler-2j8xz" Dec 10 23:02:00 crc kubenswrapper[4791]: I1210 23:02:00.807108 4791 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"plugin-serving-cert\" (UniqueName: \"kubernetes.io/secret/14000c4a-8a7d-4ade-86e6-819e08eaf42a-plugin-serving-cert\") pod \"nmstate-console-plugin-7fbb5f6569-xxb9s\" (UID: \"14000c4a-8a7d-4ade-86e6-819e08eaf42a\") " pod="openshift-nmstate/nmstate-console-plugin-7fbb5f6569-xxb9s" Dec 10 23:02:00 crc kubenswrapper[4791]: I1210 23:02:00.807386 4791 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-h6pqp\" (UniqueName: \"kubernetes.io/projected/14000c4a-8a7d-4ade-86e6-819e08eaf42a-kube-api-access-h6pqp\") pod \"nmstate-console-plugin-7fbb5f6569-xxb9s\" (UID: \"14000c4a-8a7d-4ade-86e6-819e08eaf42a\") " pod="openshift-nmstate/nmstate-console-plugin-7fbb5f6569-xxb9s" Dec 10 23:02:00 crc kubenswrapper[4791]: I1210 23:02:00.807425 4791 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"console-serving-cert\" (UniqueName: \"kubernetes.io/secret/681a9508-afbe-4ebd-80dd-473006e24d93-console-serving-cert\") pod \"console-6c597db9d6-86qkv\" (UID: \"681a9508-afbe-4ebd-80dd-473006e24d93\") " pod="openshift-console/console-6c597db9d6-86qkv" Dec 10 23:02:00 crc kubenswrapper[4791]: I1210 23:02:00.807444 4791 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nginx-conf\" (UniqueName: \"kubernetes.io/configmap/14000c4a-8a7d-4ade-86e6-819e08eaf42a-nginx-conf\") pod \"nmstate-console-plugin-7fbb5f6569-xxb9s\" (UID: \"14000c4a-8a7d-4ade-86e6-819e08eaf42a\") " pod="openshift-nmstate/nmstate-console-plugin-7fbb5f6569-xxb9s" Dec 10 23:02:00 crc kubenswrapper[4791]: I1210 23:02:00.807473 4791 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"console-config\" (UniqueName: \"kubernetes.io/configmap/681a9508-afbe-4ebd-80dd-473006e24d93-console-config\") pod \"console-6c597db9d6-86qkv\" (UID: \"681a9508-afbe-4ebd-80dd-473006e24d93\") " pod="openshift-console/console-6c597db9d6-86qkv" Dec 10 23:02:00 crc kubenswrapper[4791]: I1210 23:02:00.807488 4791 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-844f2\" (UniqueName: \"kubernetes.io/projected/681a9508-afbe-4ebd-80dd-473006e24d93-kube-api-access-844f2\") pod \"console-6c597db9d6-86qkv\" (UID: \"681a9508-afbe-4ebd-80dd-473006e24d93\") " pod="openshift-console/console-6c597db9d6-86qkv" Dec 10 23:02:00 crc kubenswrapper[4791]: I1210 23:02:00.807507 4791 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"oauth-serving-cert\" (UniqueName: \"kubernetes.io/configmap/681a9508-afbe-4ebd-80dd-473006e24d93-oauth-serving-cert\") pod \"console-6c597db9d6-86qkv\" (UID: \"681a9508-afbe-4ebd-80dd-473006e24d93\") " pod="openshift-console/console-6c597db9d6-86qkv" Dec 10 23:02:00 crc kubenswrapper[4791]: I1210 23:02:00.807528 4791 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/681a9508-afbe-4ebd-80dd-473006e24d93-service-ca\") pod \"console-6c597db9d6-86qkv\" (UID: \"681a9508-afbe-4ebd-80dd-473006e24d93\") " pod="openshift-console/console-6c597db9d6-86qkv" Dec 10 23:02:00 crc kubenswrapper[4791]: I1210 23:02:00.807551 4791 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/681a9508-afbe-4ebd-80dd-473006e24d93-trusted-ca-bundle\") pod \"console-6c597db9d6-86qkv\" (UID: \"681a9508-afbe-4ebd-80dd-473006e24d93\") " pod="openshift-console/console-6c597db9d6-86qkv" Dec 10 23:02:00 crc kubenswrapper[4791]: I1210 23:02:00.807577 4791 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"console-oauth-config\" (UniqueName: \"kubernetes.io/secret/681a9508-afbe-4ebd-80dd-473006e24d93-console-oauth-config\") pod \"console-6c597db9d6-86qkv\" (UID: \"681a9508-afbe-4ebd-80dd-473006e24d93\") " pod="openshift-console/console-6c597db9d6-86qkv" Dec 10 23:02:00 crc kubenswrapper[4791]: I1210 23:02:00.808791 4791 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"nginx-conf\" (UniqueName: \"kubernetes.io/configmap/14000c4a-8a7d-4ade-86e6-819e08eaf42a-nginx-conf\") pod \"nmstate-console-plugin-7fbb5f6569-xxb9s\" (UID: \"14000c4a-8a7d-4ade-86e6-819e08eaf42a\") " pod="openshift-nmstate/nmstate-console-plugin-7fbb5f6569-xxb9s" Dec 10 23:02:00 crc kubenswrapper[4791]: E1210 23:02:00.809934 4791 secret.go:188] Couldn't get secret openshift-nmstate/plugin-serving-cert: secret "plugin-serving-cert" not found Dec 10 23:02:00 crc kubenswrapper[4791]: E1210 23:02:00.810097 4791 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/14000c4a-8a7d-4ade-86e6-819e08eaf42a-plugin-serving-cert podName:14000c4a-8a7d-4ade-86e6-819e08eaf42a nodeName:}" failed. No retries permitted until 2025-12-10 23:02:01.310058544 +0000 UTC m=+755.739676157 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "plugin-serving-cert" (UniqueName: "kubernetes.io/secret/14000c4a-8a7d-4ade-86e6-819e08eaf42a-plugin-serving-cert") pod "nmstate-console-plugin-7fbb5f6569-xxb9s" (UID: "14000c4a-8a7d-4ade-86e6-819e08eaf42a") : secret "plugin-serving-cert" not found Dec 10 23:02:00 crc kubenswrapper[4791]: I1210 23:02:00.811517 4791 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"oauth-serving-cert\" (UniqueName: \"kubernetes.io/configmap/681a9508-afbe-4ebd-80dd-473006e24d93-oauth-serving-cert\") pod \"console-6c597db9d6-86qkv\" (UID: \"681a9508-afbe-4ebd-80dd-473006e24d93\") " pod="openshift-console/console-6c597db9d6-86qkv" Dec 10 23:02:00 crc kubenswrapper[4791]: I1210 23:02:00.811899 4791 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/681a9508-afbe-4ebd-80dd-473006e24d93-service-ca\") pod \"console-6c597db9d6-86qkv\" (UID: \"681a9508-afbe-4ebd-80dd-473006e24d93\") " pod="openshift-console/console-6c597db9d6-86qkv" Dec 10 23:02:00 crc kubenswrapper[4791]: I1210 23:02:00.812215 4791 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"console-config\" (UniqueName: \"kubernetes.io/configmap/681a9508-afbe-4ebd-80dd-473006e24d93-console-config\") pod \"console-6c597db9d6-86qkv\" (UID: \"681a9508-afbe-4ebd-80dd-473006e24d93\") " pod="openshift-console/console-6c597db9d6-86qkv" Dec 10 23:02:00 crc kubenswrapper[4791]: I1210 23:02:00.816270 4791 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"console-serving-cert\" (UniqueName: \"kubernetes.io/secret/681a9508-afbe-4ebd-80dd-473006e24d93-console-serving-cert\") pod \"console-6c597db9d6-86qkv\" (UID: \"681a9508-afbe-4ebd-80dd-473006e24d93\") " pod="openshift-console/console-6c597db9d6-86qkv" Dec 10 23:02:00 crc kubenswrapper[4791]: I1210 23:02:00.816307 4791 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/681a9508-afbe-4ebd-80dd-473006e24d93-trusted-ca-bundle\") pod \"console-6c597db9d6-86qkv\" (UID: \"681a9508-afbe-4ebd-80dd-473006e24d93\") " pod="openshift-console/console-6c597db9d6-86qkv" Dec 10 23:02:00 crc kubenswrapper[4791]: I1210 23:02:00.821066 4791 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"console-oauth-config\" (UniqueName: \"kubernetes.io/secret/681a9508-afbe-4ebd-80dd-473006e24d93-console-oauth-config\") pod \"console-6c597db9d6-86qkv\" (UID: \"681a9508-afbe-4ebd-80dd-473006e24d93\") " pod="openshift-console/console-6c597db9d6-86qkv" Dec 10 23:02:00 crc kubenswrapper[4791]: I1210 23:02:00.824936 4791 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-h6pqp\" (UniqueName: \"kubernetes.io/projected/14000c4a-8a7d-4ade-86e6-819e08eaf42a-kube-api-access-h6pqp\") pod \"nmstate-console-plugin-7fbb5f6569-xxb9s\" (UID: \"14000c4a-8a7d-4ade-86e6-819e08eaf42a\") " pod="openshift-nmstate/nmstate-console-plugin-7fbb5f6569-xxb9s" Dec 10 23:02:00 crc kubenswrapper[4791]: I1210 23:02:00.830463 4791 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-844f2\" (UniqueName: \"kubernetes.io/projected/681a9508-afbe-4ebd-80dd-473006e24d93-kube-api-access-844f2\") pod \"console-6c597db9d6-86qkv\" (UID: \"681a9508-afbe-4ebd-80dd-473006e24d93\") " pod="openshift-console/console-6c597db9d6-86qkv" Dec 10 23:02:00 crc kubenswrapper[4791]: I1210 23:02:00.997697 4791 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-nmstate/nmstate-handler-2j8xz" Dec 10 23:02:01 crc kubenswrapper[4791]: I1210 23:02:01.000267 4791 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-nmstate/nmstate-metrics-7f946cbc9-mph8d"] Dec 10 23:02:01 crc kubenswrapper[4791]: W1210 23:02:01.018058 4791 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podfd7f0c56_2613_4ba7_b0a4_4ff765493973.slice/crio-3feafc0e04239546291a3f37b2a968c2418556c23cd7bf8b25630f80cf85ecc5 WatchSource:0}: Error finding container 3feafc0e04239546291a3f37b2a968c2418556c23cd7bf8b25630f80cf85ecc5: Status 404 returned error can't find the container with id 3feafc0e04239546291a3f37b2a968c2418556c23cd7bf8b25630f80cf85ecc5 Dec 10 23:02:01 crc kubenswrapper[4791]: I1210 23:02:01.053410 4791 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-console/console-6c597db9d6-86qkv" Dec 10 23:02:01 crc kubenswrapper[4791]: I1210 23:02:01.155430 4791 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-nmstate/nmstate-webhook-5f6d4c5ccb-x65fz"] Dec 10 23:02:01 crc kubenswrapper[4791]: I1210 23:02:01.164491 4791 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-nmstate/nmstate-metrics-7f946cbc9-mph8d" event={"ID":"e85292bb-7cee-4b20-8c97-b1d1902e6001","Type":"ContainerStarted","Data":"a3504a9c93f1144ea5a591d02dcf67c6315f1fbdf452037a43535dde496b19d9"} Dec 10 23:02:01 crc kubenswrapper[4791]: I1210 23:02:01.167404 4791 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-nmstate/nmstate-handler-2j8xz" event={"ID":"fd7f0c56-2613-4ba7-b0a4-4ff765493973","Type":"ContainerStarted","Data":"3feafc0e04239546291a3f37b2a968c2418556c23cd7bf8b25630f80cf85ecc5"} Dec 10 23:02:01 crc kubenswrapper[4791]: I1210 23:02:01.250824 4791 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-console/console-6c597db9d6-86qkv"] Dec 10 23:02:01 crc kubenswrapper[4791]: W1210 23:02:01.252616 4791 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod681a9508_afbe_4ebd_80dd_473006e24d93.slice/crio-e77632b74acd245ef6744d96fc488b0f4873b72c26485223a278803615c3dbe2 WatchSource:0}: Error finding container e77632b74acd245ef6744d96fc488b0f4873b72c26485223a278803615c3dbe2: Status 404 returned error can't find the container with id e77632b74acd245ef6744d96fc488b0f4873b72c26485223a278803615c3dbe2 Dec 10 23:02:01 crc kubenswrapper[4791]: I1210 23:02:01.314119 4791 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"plugin-serving-cert\" (UniqueName: \"kubernetes.io/secret/14000c4a-8a7d-4ade-86e6-819e08eaf42a-plugin-serving-cert\") pod \"nmstate-console-plugin-7fbb5f6569-xxb9s\" (UID: \"14000c4a-8a7d-4ade-86e6-819e08eaf42a\") " pod="openshift-nmstate/nmstate-console-plugin-7fbb5f6569-xxb9s" Dec 10 23:02:01 crc kubenswrapper[4791]: I1210 23:02:01.318101 4791 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"plugin-serving-cert\" (UniqueName: \"kubernetes.io/secret/14000c4a-8a7d-4ade-86e6-819e08eaf42a-plugin-serving-cert\") pod \"nmstate-console-plugin-7fbb5f6569-xxb9s\" (UID: \"14000c4a-8a7d-4ade-86e6-819e08eaf42a\") " pod="openshift-nmstate/nmstate-console-plugin-7fbb5f6569-xxb9s" Dec 10 23:02:01 crc kubenswrapper[4791]: I1210 23:02:01.434577 4791 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-nmstate/nmstate-console-plugin-7fbb5f6569-xxb9s" Dec 10 23:02:01 crc kubenswrapper[4791]: I1210 23:02:01.868003 4791 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-nmstate/nmstate-console-plugin-7fbb5f6569-xxb9s"] Dec 10 23:02:01 crc kubenswrapper[4791]: W1210 23:02:01.877751 4791 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod14000c4a_8a7d_4ade_86e6_819e08eaf42a.slice/crio-f599b17b0606ab6bcb1928c8b841bb0f2b222e20c10c6382bcac73d7cf13d19b WatchSource:0}: Error finding container f599b17b0606ab6bcb1928c8b841bb0f2b222e20c10c6382bcac73d7cf13d19b: Status 404 returned error can't find the container with id f599b17b0606ab6bcb1928c8b841bb0f2b222e20c10c6382bcac73d7cf13d19b Dec 10 23:02:02 crc kubenswrapper[4791]: I1210 23:02:02.174405 4791 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-nmstate/nmstate-console-plugin-7fbb5f6569-xxb9s" event={"ID":"14000c4a-8a7d-4ade-86e6-819e08eaf42a","Type":"ContainerStarted","Data":"f599b17b0606ab6bcb1928c8b841bb0f2b222e20c10c6382bcac73d7cf13d19b"} Dec 10 23:02:02 crc kubenswrapper[4791]: I1210 23:02:02.176558 4791 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-console/console-6c597db9d6-86qkv" event={"ID":"681a9508-afbe-4ebd-80dd-473006e24d93","Type":"ContainerStarted","Data":"c4bd251a0cea67ed0b4e0546b4968575c5547c3b7d065295a51914965aa22abc"} Dec 10 23:02:02 crc kubenswrapper[4791]: I1210 23:02:02.176597 4791 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-console/console-6c597db9d6-86qkv" event={"ID":"681a9508-afbe-4ebd-80dd-473006e24d93","Type":"ContainerStarted","Data":"e77632b74acd245ef6744d96fc488b0f4873b72c26485223a278803615c3dbe2"} Dec 10 23:02:02 crc kubenswrapper[4791]: I1210 23:02:02.179539 4791 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-nmstate/nmstate-webhook-5f6d4c5ccb-x65fz" event={"ID":"99524090-326f-47fd-98d9-bf7e8352cc4f","Type":"ContainerStarted","Data":"2899be52306f4eab6508ba87c998c3bcf353b3be36ffbb1cc80a1e80850916f0"} Dec 10 23:02:02 crc kubenswrapper[4791]: I1210 23:02:02.197281 4791 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-console/console-6c597db9d6-86qkv" podStartSLOduration=2.196980786 podStartE2EDuration="2.196980786s" podCreationTimestamp="2025-12-10 23:02:00 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-10 23:02:02.195549215 +0000 UTC m=+756.625166828" watchObservedRunningTime="2025-12-10 23:02:02.196980786 +0000 UTC m=+756.626598399" Dec 10 23:02:05 crc kubenswrapper[4791]: I1210 23:02:05.201280 4791 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-nmstate/nmstate-handler-2j8xz" event={"ID":"fd7f0c56-2613-4ba7-b0a4-4ff765493973","Type":"ContainerStarted","Data":"6486cf304973046d77a59cf73f11b1934295a8158f2dceb06b10cd5ddb66186c"} Dec 10 23:02:05 crc kubenswrapper[4791]: I1210 23:02:05.201911 4791 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-nmstate/nmstate-handler-2j8xz" Dec 10 23:02:05 crc kubenswrapper[4791]: I1210 23:02:05.204083 4791 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-nmstate/nmstate-webhook-5f6d4c5ccb-x65fz" event={"ID":"99524090-326f-47fd-98d9-bf7e8352cc4f","Type":"ContainerStarted","Data":"820ec8738424102a8ffe727b64d5d9f00ec23f337730f2059e72ad58265b0078"} Dec 10 23:02:05 crc kubenswrapper[4791]: I1210 23:02:05.204201 4791 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-nmstate/nmstate-webhook-5f6d4c5ccb-x65fz" Dec 10 23:02:05 crc kubenswrapper[4791]: I1210 23:02:05.206208 4791 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-nmstate/nmstate-metrics-7f946cbc9-mph8d" event={"ID":"e85292bb-7cee-4b20-8c97-b1d1902e6001","Type":"ContainerStarted","Data":"23593054126a607a98ecf213d372c4182d0c984a01d8c2496fd160001fc7e666"} Dec 10 23:02:05 crc kubenswrapper[4791]: I1210 23:02:05.220808 4791 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-nmstate/nmstate-handler-2j8xz" podStartSLOduration=1.711037901 podStartE2EDuration="5.220788293s" podCreationTimestamp="2025-12-10 23:02:00 +0000 UTC" firstStartedPulling="2025-12-10 23:02:01.024415153 +0000 UTC m=+755.454032766" lastFinishedPulling="2025-12-10 23:02:04.534165545 +0000 UTC m=+758.963783158" observedRunningTime="2025-12-10 23:02:05.215439753 +0000 UTC m=+759.645057396" watchObservedRunningTime="2025-12-10 23:02:05.220788293 +0000 UTC m=+759.650405906" Dec 10 23:02:05 crc kubenswrapper[4791]: I1210 23:02:05.233704 4791 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-nmstate/nmstate-webhook-5f6d4c5ccb-x65fz" podStartSLOduration=1.907341022 podStartE2EDuration="5.233678045s" podCreationTimestamp="2025-12-10 23:02:00 +0000 UTC" firstStartedPulling="2025-12-10 23:02:01.164578268 +0000 UTC m=+755.594195871" lastFinishedPulling="2025-12-10 23:02:04.490915291 +0000 UTC m=+758.920532894" observedRunningTime="2025-12-10 23:02:05.233545851 +0000 UTC m=+759.663163464" watchObservedRunningTime="2025-12-10 23:02:05.233678045 +0000 UTC m=+759.663295658" Dec 10 23:02:06 crc kubenswrapper[4791]: I1210 23:02:06.217284 4791 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-nmstate/nmstate-console-plugin-7fbb5f6569-xxb9s" event={"ID":"14000c4a-8a7d-4ade-86e6-819e08eaf42a","Type":"ContainerStarted","Data":"e9881362c4e78f21e023a9592b30b848a9fee94653bbbbb8e8d4cfc113c9b603"} Dec 10 23:02:06 crc kubenswrapper[4791]: I1210 23:02:06.238812 4791 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-nmstate/nmstate-console-plugin-7fbb5f6569-xxb9s" podStartSLOduration=2.556484169 podStartE2EDuration="6.238793985s" podCreationTimestamp="2025-12-10 23:02:00 +0000 UTC" firstStartedPulling="2025-12-10 23:02:01.879906863 +0000 UTC m=+756.309524476" lastFinishedPulling="2025-12-10 23:02:05.562216679 +0000 UTC m=+759.991834292" observedRunningTime="2025-12-10 23:02:06.229792033 +0000 UTC m=+760.659409656" watchObservedRunningTime="2025-12-10 23:02:06.238793985 +0000 UTC m=+760.668411598" Dec 10 23:02:07 crc kubenswrapper[4791]: I1210 23:02:07.211754 4791 dynamic_cafile_content.go:123] "Loaded a new CA Bundle and Verifier" name="client-ca-bundle::/etc/kubernetes/kubelet-ca.crt" Dec 10 23:02:09 crc kubenswrapper[4791]: I1210 23:02:09.235494 4791 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-nmstate/nmstate-metrics-7f946cbc9-mph8d" event={"ID":"e85292bb-7cee-4b20-8c97-b1d1902e6001","Type":"ContainerStarted","Data":"f6a07e156353f9e8b2a843a8d31ed5840a727899cef8e8b1a49f00ba477a6010"} Dec 10 23:02:09 crc kubenswrapper[4791]: I1210 23:02:09.252325 4791 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-nmstate/nmstate-metrics-7f946cbc9-mph8d" podStartSLOduration=1.66613466 podStartE2EDuration="9.252301405s" podCreationTimestamp="2025-12-10 23:02:00 +0000 UTC" firstStartedPulling="2025-12-10 23:02:01.006997163 +0000 UTC m=+755.436614776" lastFinishedPulling="2025-12-10 23:02:08.593163908 +0000 UTC m=+763.022781521" observedRunningTime="2025-12-10 23:02:09.249947448 +0000 UTC m=+763.679565061" watchObservedRunningTime="2025-12-10 23:02:09.252301405 +0000 UTC m=+763.681919028" Dec 10 23:02:11 crc kubenswrapper[4791]: I1210 23:02:11.025907 4791 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-nmstate/nmstate-handler-2j8xz" Dec 10 23:02:11 crc kubenswrapper[4791]: I1210 23:02:11.053791 4791 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-console/console-6c597db9d6-86qkv" Dec 10 23:02:11 crc kubenswrapper[4791]: I1210 23:02:11.053837 4791 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-console/console-6c597db9d6-86qkv" Dec 10 23:02:11 crc kubenswrapper[4791]: I1210 23:02:11.058516 4791 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-console/console-6c597db9d6-86qkv" Dec 10 23:02:11 crc kubenswrapper[4791]: I1210 23:02:11.248871 4791 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-console/console-6c597db9d6-86qkv" Dec 10 23:02:11 crc kubenswrapper[4791]: I1210 23:02:11.304289 4791 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-console/console-f9d7485db-zkfkv"] Dec 10 23:02:20 crc kubenswrapper[4791]: I1210 23:02:20.664467 4791 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-nmstate/nmstate-webhook-5f6d4c5ccb-x65fz" Dec 10 23:02:25 crc kubenswrapper[4791]: I1210 23:02:25.037983 4791 patch_prober.go:28] interesting pod/machine-config-daemon-5rb5l container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 10 23:02:25 crc kubenswrapper[4791]: I1210 23:02:25.038688 4791 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-5rb5l" podUID="ba35653c-6e06-4cee-a4d6-137764090d18" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 10 23:02:32 crc kubenswrapper[4791]: I1210 23:02:32.675245 4791 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/af69698b82fdf008f5ff9e195cf808a654240e16b13dcd924b74994f83xxh8j"] Dec 10 23:02:32 crc kubenswrapper[4791]: I1210 23:02:32.677813 4791 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/af69698b82fdf008f5ff9e195cf808a654240e16b13dcd924b74994f83xxh8j" Dec 10 23:02:32 crc kubenswrapper[4791]: I1210 23:02:32.680300 4791 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"default-dockercfg-vmwhc" Dec 10 23:02:32 crc kubenswrapper[4791]: I1210 23:02:32.692236 4791 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/af69698b82fdf008f5ff9e195cf808a654240e16b13dcd924b74994f83xxh8j"] Dec 10 23:02:32 crc kubenswrapper[4791]: I1210 23:02:32.713200 4791 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-k9w6w\" (UniqueName: \"kubernetes.io/projected/d1f6824d-b614-4d99-a538-2917ad1c4bb3-kube-api-access-k9w6w\") pod \"af69698b82fdf008f5ff9e195cf808a654240e16b13dcd924b74994f83xxh8j\" (UID: \"d1f6824d-b614-4d99-a538-2917ad1c4bb3\") " pod="openshift-marketplace/af69698b82fdf008f5ff9e195cf808a654240e16b13dcd924b74994f83xxh8j" Dec 10 23:02:32 crc kubenswrapper[4791]: I1210 23:02:32.713265 4791 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/d1f6824d-b614-4d99-a538-2917ad1c4bb3-bundle\") pod \"af69698b82fdf008f5ff9e195cf808a654240e16b13dcd924b74994f83xxh8j\" (UID: \"d1f6824d-b614-4d99-a538-2917ad1c4bb3\") " pod="openshift-marketplace/af69698b82fdf008f5ff9e195cf808a654240e16b13dcd924b74994f83xxh8j" Dec 10 23:02:32 crc kubenswrapper[4791]: I1210 23:02:32.713313 4791 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/d1f6824d-b614-4d99-a538-2917ad1c4bb3-util\") pod \"af69698b82fdf008f5ff9e195cf808a654240e16b13dcd924b74994f83xxh8j\" (UID: \"d1f6824d-b614-4d99-a538-2917ad1c4bb3\") " pod="openshift-marketplace/af69698b82fdf008f5ff9e195cf808a654240e16b13dcd924b74994f83xxh8j" Dec 10 23:02:32 crc kubenswrapper[4791]: I1210 23:02:32.815086 4791 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-k9w6w\" (UniqueName: \"kubernetes.io/projected/d1f6824d-b614-4d99-a538-2917ad1c4bb3-kube-api-access-k9w6w\") pod \"af69698b82fdf008f5ff9e195cf808a654240e16b13dcd924b74994f83xxh8j\" (UID: \"d1f6824d-b614-4d99-a538-2917ad1c4bb3\") " pod="openshift-marketplace/af69698b82fdf008f5ff9e195cf808a654240e16b13dcd924b74994f83xxh8j" Dec 10 23:02:32 crc kubenswrapper[4791]: I1210 23:02:32.815139 4791 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/d1f6824d-b614-4d99-a538-2917ad1c4bb3-bundle\") pod \"af69698b82fdf008f5ff9e195cf808a654240e16b13dcd924b74994f83xxh8j\" (UID: \"d1f6824d-b614-4d99-a538-2917ad1c4bb3\") " pod="openshift-marketplace/af69698b82fdf008f5ff9e195cf808a654240e16b13dcd924b74994f83xxh8j" Dec 10 23:02:32 crc kubenswrapper[4791]: I1210 23:02:32.815162 4791 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/d1f6824d-b614-4d99-a538-2917ad1c4bb3-util\") pod \"af69698b82fdf008f5ff9e195cf808a654240e16b13dcd924b74994f83xxh8j\" (UID: \"d1f6824d-b614-4d99-a538-2917ad1c4bb3\") " pod="openshift-marketplace/af69698b82fdf008f5ff9e195cf808a654240e16b13dcd924b74994f83xxh8j" Dec 10 23:02:32 crc kubenswrapper[4791]: I1210 23:02:32.815651 4791 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/d1f6824d-b614-4d99-a538-2917ad1c4bb3-util\") pod \"af69698b82fdf008f5ff9e195cf808a654240e16b13dcd924b74994f83xxh8j\" (UID: \"d1f6824d-b614-4d99-a538-2917ad1c4bb3\") " pod="openshift-marketplace/af69698b82fdf008f5ff9e195cf808a654240e16b13dcd924b74994f83xxh8j" Dec 10 23:02:32 crc kubenswrapper[4791]: I1210 23:02:32.815917 4791 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/d1f6824d-b614-4d99-a538-2917ad1c4bb3-bundle\") pod \"af69698b82fdf008f5ff9e195cf808a654240e16b13dcd924b74994f83xxh8j\" (UID: \"d1f6824d-b614-4d99-a538-2917ad1c4bb3\") " pod="openshift-marketplace/af69698b82fdf008f5ff9e195cf808a654240e16b13dcd924b74994f83xxh8j" Dec 10 23:02:32 crc kubenswrapper[4791]: I1210 23:02:32.833568 4791 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-k9w6w\" (UniqueName: \"kubernetes.io/projected/d1f6824d-b614-4d99-a538-2917ad1c4bb3-kube-api-access-k9w6w\") pod \"af69698b82fdf008f5ff9e195cf808a654240e16b13dcd924b74994f83xxh8j\" (UID: \"d1f6824d-b614-4d99-a538-2917ad1c4bb3\") " pod="openshift-marketplace/af69698b82fdf008f5ff9e195cf808a654240e16b13dcd924b74994f83xxh8j" Dec 10 23:02:32 crc kubenswrapper[4791]: I1210 23:02:32.997808 4791 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/af69698b82fdf008f5ff9e195cf808a654240e16b13dcd924b74994f83xxh8j" Dec 10 23:02:33 crc kubenswrapper[4791]: I1210 23:02:33.441902 4791 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/af69698b82fdf008f5ff9e195cf808a654240e16b13dcd924b74994f83xxh8j"] Dec 10 23:02:34 crc kubenswrapper[4791]: I1210 23:02:34.383466 4791 generic.go:334] "Generic (PLEG): container finished" podID="d1f6824d-b614-4d99-a538-2917ad1c4bb3" containerID="7ff09cd640edbbe6f01426f32afd0357a83125c14a4cfba220902c130f2f15bb" exitCode=0 Dec 10 23:02:34 crc kubenswrapper[4791]: I1210 23:02:34.383590 4791 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/af69698b82fdf008f5ff9e195cf808a654240e16b13dcd924b74994f83xxh8j" event={"ID":"d1f6824d-b614-4d99-a538-2917ad1c4bb3","Type":"ContainerDied","Data":"7ff09cd640edbbe6f01426f32afd0357a83125c14a4cfba220902c130f2f15bb"} Dec 10 23:02:34 crc kubenswrapper[4791]: I1210 23:02:34.383993 4791 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/af69698b82fdf008f5ff9e195cf808a654240e16b13dcd924b74994f83xxh8j" event={"ID":"d1f6824d-b614-4d99-a538-2917ad1c4bb3","Type":"ContainerStarted","Data":"1269c7dc50f8a680e6f1067be6a2c5be260a8255ed29625825a28dbb1947a350"} Dec 10 23:02:35 crc kubenswrapper[4791]: I1210 23:02:35.031072 4791 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-operators-5f42v"] Dec 10 23:02:35 crc kubenswrapper[4791]: I1210 23:02:35.034686 4791 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-5f42v" Dec 10 23:02:35 crc kubenswrapper[4791]: I1210 23:02:35.041245 4791 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-5f42v"] Dec 10 23:02:35 crc kubenswrapper[4791]: I1210 23:02:35.145755 4791 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/8e8bd8f8-4aa4-4d7f-8c0a-a463ec584e50-utilities\") pod \"redhat-operators-5f42v\" (UID: \"8e8bd8f8-4aa4-4d7f-8c0a-a463ec584e50\") " pod="openshift-marketplace/redhat-operators-5f42v" Dec 10 23:02:35 crc kubenswrapper[4791]: I1210 23:02:35.145803 4791 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-p8z9h\" (UniqueName: \"kubernetes.io/projected/8e8bd8f8-4aa4-4d7f-8c0a-a463ec584e50-kube-api-access-p8z9h\") pod \"redhat-operators-5f42v\" (UID: \"8e8bd8f8-4aa4-4d7f-8c0a-a463ec584e50\") " pod="openshift-marketplace/redhat-operators-5f42v" Dec 10 23:02:35 crc kubenswrapper[4791]: I1210 23:02:35.145967 4791 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/8e8bd8f8-4aa4-4d7f-8c0a-a463ec584e50-catalog-content\") pod \"redhat-operators-5f42v\" (UID: \"8e8bd8f8-4aa4-4d7f-8c0a-a463ec584e50\") " pod="openshift-marketplace/redhat-operators-5f42v" Dec 10 23:02:35 crc kubenswrapper[4791]: I1210 23:02:35.247649 4791 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/8e8bd8f8-4aa4-4d7f-8c0a-a463ec584e50-utilities\") pod \"redhat-operators-5f42v\" (UID: \"8e8bd8f8-4aa4-4d7f-8c0a-a463ec584e50\") " pod="openshift-marketplace/redhat-operators-5f42v" Dec 10 23:02:35 crc kubenswrapper[4791]: I1210 23:02:35.247711 4791 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-p8z9h\" (UniqueName: \"kubernetes.io/projected/8e8bd8f8-4aa4-4d7f-8c0a-a463ec584e50-kube-api-access-p8z9h\") pod \"redhat-operators-5f42v\" (UID: \"8e8bd8f8-4aa4-4d7f-8c0a-a463ec584e50\") " pod="openshift-marketplace/redhat-operators-5f42v" Dec 10 23:02:35 crc kubenswrapper[4791]: I1210 23:02:35.247752 4791 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/8e8bd8f8-4aa4-4d7f-8c0a-a463ec584e50-catalog-content\") pod \"redhat-operators-5f42v\" (UID: \"8e8bd8f8-4aa4-4d7f-8c0a-a463ec584e50\") " pod="openshift-marketplace/redhat-operators-5f42v" Dec 10 23:02:35 crc kubenswrapper[4791]: I1210 23:02:35.248125 4791 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/8e8bd8f8-4aa4-4d7f-8c0a-a463ec584e50-utilities\") pod \"redhat-operators-5f42v\" (UID: \"8e8bd8f8-4aa4-4d7f-8c0a-a463ec584e50\") " pod="openshift-marketplace/redhat-operators-5f42v" Dec 10 23:02:35 crc kubenswrapper[4791]: I1210 23:02:35.248184 4791 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/8e8bd8f8-4aa4-4d7f-8c0a-a463ec584e50-catalog-content\") pod \"redhat-operators-5f42v\" (UID: \"8e8bd8f8-4aa4-4d7f-8c0a-a463ec584e50\") " pod="openshift-marketplace/redhat-operators-5f42v" Dec 10 23:02:35 crc kubenswrapper[4791]: I1210 23:02:35.276182 4791 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-p8z9h\" (UniqueName: \"kubernetes.io/projected/8e8bd8f8-4aa4-4d7f-8c0a-a463ec584e50-kube-api-access-p8z9h\") pod \"redhat-operators-5f42v\" (UID: \"8e8bd8f8-4aa4-4d7f-8c0a-a463ec584e50\") " pod="openshift-marketplace/redhat-operators-5f42v" Dec 10 23:02:35 crc kubenswrapper[4791]: I1210 23:02:35.361751 4791 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-5f42v" Dec 10 23:02:35 crc kubenswrapper[4791]: I1210 23:02:35.620193 4791 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-5f42v"] Dec 10 23:02:36 crc kubenswrapper[4791]: I1210 23:02:36.354602 4791 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-console/console-f9d7485db-zkfkv" podUID="1b61a973-db5e-4aa7-96e3-7fbdd9dd8cc0" containerName="console" containerID="cri-o://82877649b51625497004f259c4773b2cd94796ab144537ed6ed9dc6355b4a741" gracePeriod=15 Dec 10 23:02:36 crc kubenswrapper[4791]: I1210 23:02:36.397181 4791 generic.go:334] "Generic (PLEG): container finished" podID="d1f6824d-b614-4d99-a538-2917ad1c4bb3" containerID="1b6e596211abe16cc6537f3922350a9ebb9a192757525498cec029680218ccda" exitCode=0 Dec 10 23:02:36 crc kubenswrapper[4791]: I1210 23:02:36.397279 4791 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/af69698b82fdf008f5ff9e195cf808a654240e16b13dcd924b74994f83xxh8j" event={"ID":"d1f6824d-b614-4d99-a538-2917ad1c4bb3","Type":"ContainerDied","Data":"1b6e596211abe16cc6537f3922350a9ebb9a192757525498cec029680218ccda"} Dec 10 23:02:36 crc kubenswrapper[4791]: I1210 23:02:36.398500 4791 generic.go:334] "Generic (PLEG): container finished" podID="8e8bd8f8-4aa4-4d7f-8c0a-a463ec584e50" containerID="4c4864681123f42be2c04ec86b24f334e4a999dab31aff7e11a01a26b70ed561" exitCode=0 Dec 10 23:02:36 crc kubenswrapper[4791]: I1210 23:02:36.398538 4791 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-5f42v" event={"ID":"8e8bd8f8-4aa4-4d7f-8c0a-a463ec584e50","Type":"ContainerDied","Data":"4c4864681123f42be2c04ec86b24f334e4a999dab31aff7e11a01a26b70ed561"} Dec 10 23:02:36 crc kubenswrapper[4791]: I1210 23:02:36.398560 4791 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-5f42v" event={"ID":"8e8bd8f8-4aa4-4d7f-8c0a-a463ec584e50","Type":"ContainerStarted","Data":"d2a7c3f59813d771fd051db5cc1db664ce0ba8463be4b539b5b95efac7473f18"} Dec 10 23:02:36 crc kubenswrapper[4791]: I1210 23:02:36.766796 4791 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-console_console-f9d7485db-zkfkv_1b61a973-db5e-4aa7-96e3-7fbdd9dd8cc0/console/0.log" Dec 10 23:02:36 crc kubenswrapper[4791]: I1210 23:02:36.767075 4791 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-console/console-f9d7485db-zkfkv" Dec 10 23:02:36 crc kubenswrapper[4791]: I1210 23:02:36.869059 4791 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/1b61a973-db5e-4aa7-96e3-7fbdd9dd8cc0-service-ca\") pod \"1b61a973-db5e-4aa7-96e3-7fbdd9dd8cc0\" (UID: \"1b61a973-db5e-4aa7-96e3-7fbdd9dd8cc0\") " Dec 10 23:02:36 crc kubenswrapper[4791]: I1210 23:02:36.869133 4791 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-f2tld\" (UniqueName: \"kubernetes.io/projected/1b61a973-db5e-4aa7-96e3-7fbdd9dd8cc0-kube-api-access-f2tld\") pod \"1b61a973-db5e-4aa7-96e3-7fbdd9dd8cc0\" (UID: \"1b61a973-db5e-4aa7-96e3-7fbdd9dd8cc0\") " Dec 10 23:02:36 crc kubenswrapper[4791]: I1210 23:02:36.869178 4791 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"console-serving-cert\" (UniqueName: \"kubernetes.io/secret/1b61a973-db5e-4aa7-96e3-7fbdd9dd8cc0-console-serving-cert\") pod \"1b61a973-db5e-4aa7-96e3-7fbdd9dd8cc0\" (UID: \"1b61a973-db5e-4aa7-96e3-7fbdd9dd8cc0\") " Dec 10 23:02:36 crc kubenswrapper[4791]: I1210 23:02:36.869201 4791 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"oauth-serving-cert\" (UniqueName: \"kubernetes.io/configmap/1b61a973-db5e-4aa7-96e3-7fbdd9dd8cc0-oauth-serving-cert\") pod \"1b61a973-db5e-4aa7-96e3-7fbdd9dd8cc0\" (UID: \"1b61a973-db5e-4aa7-96e3-7fbdd9dd8cc0\") " Dec 10 23:02:36 crc kubenswrapper[4791]: I1210 23:02:36.869264 4791 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"console-oauth-config\" (UniqueName: \"kubernetes.io/secret/1b61a973-db5e-4aa7-96e3-7fbdd9dd8cc0-console-oauth-config\") pod \"1b61a973-db5e-4aa7-96e3-7fbdd9dd8cc0\" (UID: \"1b61a973-db5e-4aa7-96e3-7fbdd9dd8cc0\") " Dec 10 23:02:36 crc kubenswrapper[4791]: I1210 23:02:36.869286 4791 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"console-config\" (UniqueName: \"kubernetes.io/configmap/1b61a973-db5e-4aa7-96e3-7fbdd9dd8cc0-console-config\") pod \"1b61a973-db5e-4aa7-96e3-7fbdd9dd8cc0\" (UID: \"1b61a973-db5e-4aa7-96e3-7fbdd9dd8cc0\") " Dec 10 23:02:36 crc kubenswrapper[4791]: I1210 23:02:36.869303 4791 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/1b61a973-db5e-4aa7-96e3-7fbdd9dd8cc0-trusted-ca-bundle\") pod \"1b61a973-db5e-4aa7-96e3-7fbdd9dd8cc0\" (UID: \"1b61a973-db5e-4aa7-96e3-7fbdd9dd8cc0\") " Dec 10 23:02:36 crc kubenswrapper[4791]: I1210 23:02:36.870150 4791 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/1b61a973-db5e-4aa7-96e3-7fbdd9dd8cc0-console-config" (OuterVolumeSpecName: "console-config") pod "1b61a973-db5e-4aa7-96e3-7fbdd9dd8cc0" (UID: "1b61a973-db5e-4aa7-96e3-7fbdd9dd8cc0"). InnerVolumeSpecName "console-config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 10 23:02:36 crc kubenswrapper[4791]: I1210 23:02:36.870201 4791 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/1b61a973-db5e-4aa7-96e3-7fbdd9dd8cc0-trusted-ca-bundle" (OuterVolumeSpecName: "trusted-ca-bundle") pod "1b61a973-db5e-4aa7-96e3-7fbdd9dd8cc0" (UID: "1b61a973-db5e-4aa7-96e3-7fbdd9dd8cc0"). InnerVolumeSpecName "trusted-ca-bundle". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 10 23:02:36 crc kubenswrapper[4791]: I1210 23:02:36.870320 4791 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/1b61a973-db5e-4aa7-96e3-7fbdd9dd8cc0-oauth-serving-cert" (OuterVolumeSpecName: "oauth-serving-cert") pod "1b61a973-db5e-4aa7-96e3-7fbdd9dd8cc0" (UID: "1b61a973-db5e-4aa7-96e3-7fbdd9dd8cc0"). InnerVolumeSpecName "oauth-serving-cert". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 10 23:02:36 crc kubenswrapper[4791]: I1210 23:02:36.870539 4791 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/1b61a973-db5e-4aa7-96e3-7fbdd9dd8cc0-service-ca" (OuterVolumeSpecName: "service-ca") pod "1b61a973-db5e-4aa7-96e3-7fbdd9dd8cc0" (UID: "1b61a973-db5e-4aa7-96e3-7fbdd9dd8cc0"). InnerVolumeSpecName "service-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 10 23:02:36 crc kubenswrapper[4791]: I1210 23:02:36.870726 4791 reconciler_common.go:293] "Volume detached for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/1b61a973-db5e-4aa7-96e3-7fbdd9dd8cc0-service-ca\") on node \"crc\" DevicePath \"\"" Dec 10 23:02:36 crc kubenswrapper[4791]: I1210 23:02:36.870743 4791 reconciler_common.go:293] "Volume detached for volume \"oauth-serving-cert\" (UniqueName: \"kubernetes.io/configmap/1b61a973-db5e-4aa7-96e3-7fbdd9dd8cc0-oauth-serving-cert\") on node \"crc\" DevicePath \"\"" Dec 10 23:02:36 crc kubenswrapper[4791]: I1210 23:02:36.870757 4791 reconciler_common.go:293] "Volume detached for volume \"console-config\" (UniqueName: \"kubernetes.io/configmap/1b61a973-db5e-4aa7-96e3-7fbdd9dd8cc0-console-config\") on node \"crc\" DevicePath \"\"" Dec 10 23:02:36 crc kubenswrapper[4791]: I1210 23:02:36.870768 4791 reconciler_common.go:293] "Volume detached for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/1b61a973-db5e-4aa7-96e3-7fbdd9dd8cc0-trusted-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 10 23:02:36 crc kubenswrapper[4791]: I1210 23:02:36.874606 4791 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/1b61a973-db5e-4aa7-96e3-7fbdd9dd8cc0-console-serving-cert" (OuterVolumeSpecName: "console-serving-cert") pod "1b61a973-db5e-4aa7-96e3-7fbdd9dd8cc0" (UID: "1b61a973-db5e-4aa7-96e3-7fbdd9dd8cc0"). InnerVolumeSpecName "console-serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 10 23:02:36 crc kubenswrapper[4791]: I1210 23:02:36.874667 4791 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/1b61a973-db5e-4aa7-96e3-7fbdd9dd8cc0-kube-api-access-f2tld" (OuterVolumeSpecName: "kube-api-access-f2tld") pod "1b61a973-db5e-4aa7-96e3-7fbdd9dd8cc0" (UID: "1b61a973-db5e-4aa7-96e3-7fbdd9dd8cc0"). InnerVolumeSpecName "kube-api-access-f2tld". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 10 23:02:36 crc kubenswrapper[4791]: I1210 23:02:36.874824 4791 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/1b61a973-db5e-4aa7-96e3-7fbdd9dd8cc0-console-oauth-config" (OuterVolumeSpecName: "console-oauth-config") pod "1b61a973-db5e-4aa7-96e3-7fbdd9dd8cc0" (UID: "1b61a973-db5e-4aa7-96e3-7fbdd9dd8cc0"). InnerVolumeSpecName "console-oauth-config". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 10 23:02:36 crc kubenswrapper[4791]: I1210 23:02:36.972202 4791 reconciler_common.go:293] "Volume detached for volume \"console-oauth-config\" (UniqueName: \"kubernetes.io/secret/1b61a973-db5e-4aa7-96e3-7fbdd9dd8cc0-console-oauth-config\") on node \"crc\" DevicePath \"\"" Dec 10 23:02:36 crc kubenswrapper[4791]: I1210 23:02:36.972235 4791 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-f2tld\" (UniqueName: \"kubernetes.io/projected/1b61a973-db5e-4aa7-96e3-7fbdd9dd8cc0-kube-api-access-f2tld\") on node \"crc\" DevicePath \"\"" Dec 10 23:02:36 crc kubenswrapper[4791]: I1210 23:02:36.972245 4791 reconciler_common.go:293] "Volume detached for volume \"console-serving-cert\" (UniqueName: \"kubernetes.io/secret/1b61a973-db5e-4aa7-96e3-7fbdd9dd8cc0-console-serving-cert\") on node \"crc\" DevicePath \"\"" Dec 10 23:02:37 crc kubenswrapper[4791]: I1210 23:02:37.408903 4791 generic.go:334] "Generic (PLEG): container finished" podID="d1f6824d-b614-4d99-a538-2917ad1c4bb3" containerID="953f96467e1738aa59224aa40e43993aede19e8bf84cab707a375df69fba6556" exitCode=0 Dec 10 23:02:37 crc kubenswrapper[4791]: I1210 23:02:37.409121 4791 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/af69698b82fdf008f5ff9e195cf808a654240e16b13dcd924b74994f83xxh8j" event={"ID":"d1f6824d-b614-4d99-a538-2917ad1c4bb3","Type":"ContainerDied","Data":"953f96467e1738aa59224aa40e43993aede19e8bf84cab707a375df69fba6556"} Dec 10 23:02:37 crc kubenswrapper[4791]: I1210 23:02:37.411758 4791 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-5f42v" event={"ID":"8e8bd8f8-4aa4-4d7f-8c0a-a463ec584e50","Type":"ContainerStarted","Data":"25915819fdcc19309fe7285b7a21111a6a7920a7b4828600cd98564a38009d00"} Dec 10 23:02:37 crc kubenswrapper[4791]: I1210 23:02:37.414437 4791 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-console_console-f9d7485db-zkfkv_1b61a973-db5e-4aa7-96e3-7fbdd9dd8cc0/console/0.log" Dec 10 23:02:37 crc kubenswrapper[4791]: I1210 23:02:37.414491 4791 generic.go:334] "Generic (PLEG): container finished" podID="1b61a973-db5e-4aa7-96e3-7fbdd9dd8cc0" containerID="82877649b51625497004f259c4773b2cd94796ab144537ed6ed9dc6355b4a741" exitCode=2 Dec 10 23:02:37 crc kubenswrapper[4791]: I1210 23:02:37.414522 4791 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-console/console-f9d7485db-zkfkv" event={"ID":"1b61a973-db5e-4aa7-96e3-7fbdd9dd8cc0","Type":"ContainerDied","Data":"82877649b51625497004f259c4773b2cd94796ab144537ed6ed9dc6355b4a741"} Dec 10 23:02:37 crc kubenswrapper[4791]: I1210 23:02:37.414553 4791 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-console/console-f9d7485db-zkfkv" event={"ID":"1b61a973-db5e-4aa7-96e3-7fbdd9dd8cc0","Type":"ContainerDied","Data":"8be5e83a023b172a495e79f8fe719fc8e54ba3500d683e4fce82952ba90d0523"} Dec 10 23:02:37 crc kubenswrapper[4791]: I1210 23:02:37.414571 4791 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-console/console-f9d7485db-zkfkv" Dec 10 23:02:37 crc kubenswrapper[4791]: I1210 23:02:37.414573 4791 scope.go:117] "RemoveContainer" containerID="82877649b51625497004f259c4773b2cd94796ab144537ed6ed9dc6355b4a741" Dec 10 23:02:37 crc kubenswrapper[4791]: I1210 23:02:37.434671 4791 scope.go:117] "RemoveContainer" containerID="82877649b51625497004f259c4773b2cd94796ab144537ed6ed9dc6355b4a741" Dec 10 23:02:37 crc kubenswrapper[4791]: E1210 23:02:37.435770 4791 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"82877649b51625497004f259c4773b2cd94796ab144537ed6ed9dc6355b4a741\": container with ID starting with 82877649b51625497004f259c4773b2cd94796ab144537ed6ed9dc6355b4a741 not found: ID does not exist" containerID="82877649b51625497004f259c4773b2cd94796ab144537ed6ed9dc6355b4a741" Dec 10 23:02:37 crc kubenswrapper[4791]: I1210 23:02:37.435809 4791 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"82877649b51625497004f259c4773b2cd94796ab144537ed6ed9dc6355b4a741"} err="failed to get container status \"82877649b51625497004f259c4773b2cd94796ab144537ed6ed9dc6355b4a741\": rpc error: code = NotFound desc = could not find container \"82877649b51625497004f259c4773b2cd94796ab144537ed6ed9dc6355b4a741\": container with ID starting with 82877649b51625497004f259c4773b2cd94796ab144537ed6ed9dc6355b4a741 not found: ID does not exist" Dec 10 23:02:37 crc kubenswrapper[4791]: I1210 23:02:37.450972 4791 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-console/console-f9d7485db-zkfkv"] Dec 10 23:02:37 crc kubenswrapper[4791]: I1210 23:02:37.456560 4791 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-console/console-f9d7485db-zkfkv"] Dec 10 23:02:37 crc kubenswrapper[4791]: I1210 23:02:37.896456 4791 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="1b61a973-db5e-4aa7-96e3-7fbdd9dd8cc0" path="/var/lib/kubelet/pods/1b61a973-db5e-4aa7-96e3-7fbdd9dd8cc0/volumes" Dec 10 23:02:38 crc kubenswrapper[4791]: I1210 23:02:38.890868 4791 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/af69698b82fdf008f5ff9e195cf808a654240e16b13dcd924b74994f83xxh8j" Dec 10 23:02:38 crc kubenswrapper[4791]: I1210 23:02:38.995771 4791 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/d1f6824d-b614-4d99-a538-2917ad1c4bb3-util\") pod \"d1f6824d-b614-4d99-a538-2917ad1c4bb3\" (UID: \"d1f6824d-b614-4d99-a538-2917ad1c4bb3\") " Dec 10 23:02:38 crc kubenswrapper[4791]: I1210 23:02:38.995846 4791 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-k9w6w\" (UniqueName: \"kubernetes.io/projected/d1f6824d-b614-4d99-a538-2917ad1c4bb3-kube-api-access-k9w6w\") pod \"d1f6824d-b614-4d99-a538-2917ad1c4bb3\" (UID: \"d1f6824d-b614-4d99-a538-2917ad1c4bb3\") " Dec 10 23:02:38 crc kubenswrapper[4791]: I1210 23:02:38.995883 4791 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/d1f6824d-b614-4d99-a538-2917ad1c4bb3-bundle\") pod \"d1f6824d-b614-4d99-a538-2917ad1c4bb3\" (UID: \"d1f6824d-b614-4d99-a538-2917ad1c4bb3\") " Dec 10 23:02:38 crc kubenswrapper[4791]: I1210 23:02:38.997393 4791 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/d1f6824d-b614-4d99-a538-2917ad1c4bb3-bundle" (OuterVolumeSpecName: "bundle") pod "d1f6824d-b614-4d99-a538-2917ad1c4bb3" (UID: "d1f6824d-b614-4d99-a538-2917ad1c4bb3"). InnerVolumeSpecName "bundle". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 10 23:02:39 crc kubenswrapper[4791]: I1210 23:02:39.001031 4791 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/d1f6824d-b614-4d99-a538-2917ad1c4bb3-kube-api-access-k9w6w" (OuterVolumeSpecName: "kube-api-access-k9w6w") pod "d1f6824d-b614-4d99-a538-2917ad1c4bb3" (UID: "d1f6824d-b614-4d99-a538-2917ad1c4bb3"). InnerVolumeSpecName "kube-api-access-k9w6w". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 10 23:02:39 crc kubenswrapper[4791]: I1210 23:02:39.016309 4791 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/d1f6824d-b614-4d99-a538-2917ad1c4bb3-util" (OuterVolumeSpecName: "util") pod "d1f6824d-b614-4d99-a538-2917ad1c4bb3" (UID: "d1f6824d-b614-4d99-a538-2917ad1c4bb3"). InnerVolumeSpecName "util". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 10 23:02:39 crc kubenswrapper[4791]: I1210 23:02:39.097511 4791 reconciler_common.go:293] "Volume detached for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/d1f6824d-b614-4d99-a538-2917ad1c4bb3-util\") on node \"crc\" DevicePath \"\"" Dec 10 23:02:39 crc kubenswrapper[4791]: I1210 23:02:39.097542 4791 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-k9w6w\" (UniqueName: \"kubernetes.io/projected/d1f6824d-b614-4d99-a538-2917ad1c4bb3-kube-api-access-k9w6w\") on node \"crc\" DevicePath \"\"" Dec 10 23:02:39 crc kubenswrapper[4791]: I1210 23:02:39.097554 4791 reconciler_common.go:293] "Volume detached for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/d1f6824d-b614-4d99-a538-2917ad1c4bb3-bundle\") on node \"crc\" DevicePath \"\"" Dec 10 23:02:39 crc kubenswrapper[4791]: I1210 23:02:39.429905 4791 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/af69698b82fdf008f5ff9e195cf808a654240e16b13dcd924b74994f83xxh8j" event={"ID":"d1f6824d-b614-4d99-a538-2917ad1c4bb3","Type":"ContainerDied","Data":"1269c7dc50f8a680e6f1067be6a2c5be260a8255ed29625825a28dbb1947a350"} Dec 10 23:02:39 crc kubenswrapper[4791]: I1210 23:02:39.429953 4791 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="1269c7dc50f8a680e6f1067be6a2c5be260a8255ed29625825a28dbb1947a350" Dec 10 23:02:39 crc kubenswrapper[4791]: I1210 23:02:39.429961 4791 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/af69698b82fdf008f5ff9e195cf808a654240e16b13dcd924b74994f83xxh8j" Dec 10 23:02:40 crc kubenswrapper[4791]: I1210 23:02:40.493840 4791 generic.go:334] "Generic (PLEG): container finished" podID="8e8bd8f8-4aa4-4d7f-8c0a-a463ec584e50" containerID="25915819fdcc19309fe7285b7a21111a6a7920a7b4828600cd98564a38009d00" exitCode=0 Dec 10 23:02:40 crc kubenswrapper[4791]: I1210 23:02:40.493880 4791 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-5f42v" event={"ID":"8e8bd8f8-4aa4-4d7f-8c0a-a463ec584e50","Type":"ContainerDied","Data":"25915819fdcc19309fe7285b7a21111a6a7920a7b4828600cd98564a38009d00"} Dec 10 23:02:42 crc kubenswrapper[4791]: I1210 23:02:42.505865 4791 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-5f42v" event={"ID":"8e8bd8f8-4aa4-4d7f-8c0a-a463ec584e50","Type":"ContainerStarted","Data":"1e472052621dbee8bc1f69f15edefadbe91ec8b5bfeee098182af13632806572"} Dec 10 23:02:42 crc kubenswrapper[4791]: I1210 23:02:42.522803 4791 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-operators-5f42v" podStartSLOduration=1.849317645 podStartE2EDuration="7.522784567s" podCreationTimestamp="2025-12-10 23:02:35 +0000 UTC" firstStartedPulling="2025-12-10 23:02:36.400040208 +0000 UTC m=+790.829657821" lastFinishedPulling="2025-12-10 23:02:42.07350713 +0000 UTC m=+796.503124743" observedRunningTime="2025-12-10 23:02:42.519331587 +0000 UTC m=+796.948949200" watchObservedRunningTime="2025-12-10 23:02:42.522784567 +0000 UTC m=+796.952402180" Dec 10 23:02:45 crc kubenswrapper[4791]: I1210 23:02:45.362770 4791 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-operators-5f42v" Dec 10 23:02:45 crc kubenswrapper[4791]: I1210 23:02:45.363113 4791 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-operators-5f42v" Dec 10 23:02:46 crc kubenswrapper[4791]: I1210 23:02:46.411304 4791 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-marketplace/redhat-operators-5f42v" podUID="8e8bd8f8-4aa4-4d7f-8c0a-a463ec584e50" containerName="registry-server" probeResult="failure" output=< Dec 10 23:02:46 crc kubenswrapper[4791]: timeout: failed to connect service ":50051" within 1s Dec 10 23:02:46 crc kubenswrapper[4791]: > Dec 10 23:02:49 crc kubenswrapper[4791]: I1210 23:02:49.442059 4791 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["metallb-system/metallb-operator-controller-manager-867ddc9894-mhxh7"] Dec 10 23:02:49 crc kubenswrapper[4791]: E1210 23:02:49.442319 4791 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="d1f6824d-b614-4d99-a538-2917ad1c4bb3" containerName="pull" Dec 10 23:02:49 crc kubenswrapper[4791]: I1210 23:02:49.442349 4791 state_mem.go:107] "Deleted CPUSet assignment" podUID="d1f6824d-b614-4d99-a538-2917ad1c4bb3" containerName="pull" Dec 10 23:02:49 crc kubenswrapper[4791]: E1210 23:02:49.442369 4791 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="d1f6824d-b614-4d99-a538-2917ad1c4bb3" containerName="util" Dec 10 23:02:49 crc kubenswrapper[4791]: I1210 23:02:49.442379 4791 state_mem.go:107] "Deleted CPUSet assignment" podUID="d1f6824d-b614-4d99-a538-2917ad1c4bb3" containerName="util" Dec 10 23:02:49 crc kubenswrapper[4791]: E1210 23:02:49.442392 4791 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="1b61a973-db5e-4aa7-96e3-7fbdd9dd8cc0" containerName="console" Dec 10 23:02:49 crc kubenswrapper[4791]: I1210 23:02:49.442400 4791 state_mem.go:107] "Deleted CPUSet assignment" podUID="1b61a973-db5e-4aa7-96e3-7fbdd9dd8cc0" containerName="console" Dec 10 23:02:49 crc kubenswrapper[4791]: E1210 23:02:49.442420 4791 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="d1f6824d-b614-4d99-a538-2917ad1c4bb3" containerName="extract" Dec 10 23:02:49 crc kubenswrapper[4791]: I1210 23:02:49.442426 4791 state_mem.go:107] "Deleted CPUSet assignment" podUID="d1f6824d-b614-4d99-a538-2917ad1c4bb3" containerName="extract" Dec 10 23:02:49 crc kubenswrapper[4791]: I1210 23:02:49.442563 4791 memory_manager.go:354] "RemoveStaleState removing state" podUID="d1f6824d-b614-4d99-a538-2917ad1c4bb3" containerName="extract" Dec 10 23:02:49 crc kubenswrapper[4791]: I1210 23:02:49.442579 4791 memory_manager.go:354] "RemoveStaleState removing state" podUID="1b61a973-db5e-4aa7-96e3-7fbdd9dd8cc0" containerName="console" Dec 10 23:02:49 crc kubenswrapper[4791]: I1210 23:02:49.443045 4791 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="metallb-system/metallb-operator-controller-manager-867ddc9894-mhxh7" Dec 10 23:02:49 crc kubenswrapper[4791]: I1210 23:02:49.446143 4791 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"manager-account-dockercfg-95h5f" Dec 10 23:02:49 crc kubenswrapper[4791]: I1210 23:02:49.456559 4791 reflector.go:368] Caches populated for *v1.ConfigMap from object-"metallb-system"/"openshift-service-ca.crt" Dec 10 23:02:49 crc kubenswrapper[4791]: I1210 23:02:49.456873 4791 reflector.go:368] Caches populated for *v1.ConfigMap from object-"metallb-system"/"kube-root-ca.crt" Dec 10 23:02:49 crc kubenswrapper[4791]: I1210 23:02:49.456903 4791 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"metallb-operator-webhook-server-cert" Dec 10 23:02:49 crc kubenswrapper[4791]: I1210 23:02:49.460851 4791 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"metallb-operator-controller-manager-service-cert" Dec 10 23:02:49 crc kubenswrapper[4791]: I1210 23:02:49.477868 4791 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["metallb-system/metallb-operator-controller-manager-867ddc9894-mhxh7"] Dec 10 23:02:49 crc kubenswrapper[4791]: I1210 23:02:49.528222 4791 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/72a07cc5-20c1-4f60-b6d7-84dfe1f78be2-apiservice-cert\") pod \"metallb-operator-controller-manager-867ddc9894-mhxh7\" (UID: \"72a07cc5-20c1-4f60-b6d7-84dfe1f78be2\") " pod="metallb-system/metallb-operator-controller-manager-867ddc9894-mhxh7" Dec 10 23:02:49 crc kubenswrapper[4791]: I1210 23:02:49.528276 4791 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-vhbjr\" (UniqueName: \"kubernetes.io/projected/72a07cc5-20c1-4f60-b6d7-84dfe1f78be2-kube-api-access-vhbjr\") pod \"metallb-operator-controller-manager-867ddc9894-mhxh7\" (UID: \"72a07cc5-20c1-4f60-b6d7-84dfe1f78be2\") " pod="metallb-system/metallb-operator-controller-manager-867ddc9894-mhxh7" Dec 10 23:02:49 crc kubenswrapper[4791]: I1210 23:02:49.528312 4791 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/72a07cc5-20c1-4f60-b6d7-84dfe1f78be2-webhook-cert\") pod \"metallb-operator-controller-manager-867ddc9894-mhxh7\" (UID: \"72a07cc5-20c1-4f60-b6d7-84dfe1f78be2\") " pod="metallb-system/metallb-operator-controller-manager-867ddc9894-mhxh7" Dec 10 23:02:49 crc kubenswrapper[4791]: I1210 23:02:49.630039 4791 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/72a07cc5-20c1-4f60-b6d7-84dfe1f78be2-apiservice-cert\") pod \"metallb-operator-controller-manager-867ddc9894-mhxh7\" (UID: \"72a07cc5-20c1-4f60-b6d7-84dfe1f78be2\") " pod="metallb-system/metallb-operator-controller-manager-867ddc9894-mhxh7" Dec 10 23:02:49 crc kubenswrapper[4791]: I1210 23:02:49.630108 4791 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-vhbjr\" (UniqueName: \"kubernetes.io/projected/72a07cc5-20c1-4f60-b6d7-84dfe1f78be2-kube-api-access-vhbjr\") pod \"metallb-operator-controller-manager-867ddc9894-mhxh7\" (UID: \"72a07cc5-20c1-4f60-b6d7-84dfe1f78be2\") " pod="metallb-system/metallb-operator-controller-manager-867ddc9894-mhxh7" Dec 10 23:02:49 crc kubenswrapper[4791]: I1210 23:02:49.630149 4791 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/72a07cc5-20c1-4f60-b6d7-84dfe1f78be2-webhook-cert\") pod \"metallb-operator-controller-manager-867ddc9894-mhxh7\" (UID: \"72a07cc5-20c1-4f60-b6d7-84dfe1f78be2\") " pod="metallb-system/metallb-operator-controller-manager-867ddc9894-mhxh7" Dec 10 23:02:49 crc kubenswrapper[4791]: I1210 23:02:49.638104 4791 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/72a07cc5-20c1-4f60-b6d7-84dfe1f78be2-webhook-cert\") pod \"metallb-operator-controller-manager-867ddc9894-mhxh7\" (UID: \"72a07cc5-20c1-4f60-b6d7-84dfe1f78be2\") " pod="metallb-system/metallb-operator-controller-manager-867ddc9894-mhxh7" Dec 10 23:02:49 crc kubenswrapper[4791]: I1210 23:02:49.647712 4791 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/72a07cc5-20c1-4f60-b6d7-84dfe1f78be2-apiservice-cert\") pod \"metallb-operator-controller-manager-867ddc9894-mhxh7\" (UID: \"72a07cc5-20c1-4f60-b6d7-84dfe1f78be2\") " pod="metallb-system/metallb-operator-controller-manager-867ddc9894-mhxh7" Dec 10 23:02:49 crc kubenswrapper[4791]: I1210 23:02:49.650985 4791 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-vhbjr\" (UniqueName: \"kubernetes.io/projected/72a07cc5-20c1-4f60-b6d7-84dfe1f78be2-kube-api-access-vhbjr\") pod \"metallb-operator-controller-manager-867ddc9894-mhxh7\" (UID: \"72a07cc5-20c1-4f60-b6d7-84dfe1f78be2\") " pod="metallb-system/metallb-operator-controller-manager-867ddc9894-mhxh7" Dec 10 23:02:49 crc kubenswrapper[4791]: I1210 23:02:49.749535 4791 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["metallb-system/metallb-operator-webhook-server-568c9f7b75-dtmkk"] Dec 10 23:02:49 crc kubenswrapper[4791]: I1210 23:02:49.750210 4791 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="metallb-system/metallb-operator-webhook-server-568c9f7b75-dtmkk" Dec 10 23:02:49 crc kubenswrapper[4791]: I1210 23:02:49.755062 4791 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"metallb-operator-webhook-server-service-cert" Dec 10 23:02:49 crc kubenswrapper[4791]: I1210 23:02:49.755435 4791 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"controller-dockercfg-mf7vz" Dec 10 23:02:49 crc kubenswrapper[4791]: I1210 23:02:49.758274 4791 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"metallb-webhook-cert" Dec 10 23:02:49 crc kubenswrapper[4791]: I1210 23:02:49.767958 4791 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="metallb-system/metallb-operator-controller-manager-867ddc9894-mhxh7" Dec 10 23:02:49 crc kubenswrapper[4791]: I1210 23:02:49.768879 4791 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["metallb-system/metallb-operator-webhook-server-568c9f7b75-dtmkk"] Dec 10 23:02:49 crc kubenswrapper[4791]: I1210 23:02:49.832302 4791 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/f1ea99a7-677c-4d8b-89c6-3b24466632c2-webhook-cert\") pod \"metallb-operator-webhook-server-568c9f7b75-dtmkk\" (UID: \"f1ea99a7-677c-4d8b-89c6-3b24466632c2\") " pod="metallb-system/metallb-operator-webhook-server-568c9f7b75-dtmkk" Dec 10 23:02:49 crc kubenswrapper[4791]: I1210 23:02:49.832402 4791 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-rl9td\" (UniqueName: \"kubernetes.io/projected/f1ea99a7-677c-4d8b-89c6-3b24466632c2-kube-api-access-rl9td\") pod \"metallb-operator-webhook-server-568c9f7b75-dtmkk\" (UID: \"f1ea99a7-677c-4d8b-89c6-3b24466632c2\") " pod="metallb-system/metallb-operator-webhook-server-568c9f7b75-dtmkk" Dec 10 23:02:49 crc kubenswrapper[4791]: I1210 23:02:49.832444 4791 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/f1ea99a7-677c-4d8b-89c6-3b24466632c2-apiservice-cert\") pod \"metallb-operator-webhook-server-568c9f7b75-dtmkk\" (UID: \"f1ea99a7-677c-4d8b-89c6-3b24466632c2\") " pod="metallb-system/metallb-operator-webhook-server-568c9f7b75-dtmkk" Dec 10 23:02:49 crc kubenswrapper[4791]: I1210 23:02:49.934039 4791 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/f1ea99a7-677c-4d8b-89c6-3b24466632c2-webhook-cert\") pod \"metallb-operator-webhook-server-568c9f7b75-dtmkk\" (UID: \"f1ea99a7-677c-4d8b-89c6-3b24466632c2\") " pod="metallb-system/metallb-operator-webhook-server-568c9f7b75-dtmkk" Dec 10 23:02:49 crc kubenswrapper[4791]: I1210 23:02:49.934142 4791 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-rl9td\" (UniqueName: \"kubernetes.io/projected/f1ea99a7-677c-4d8b-89c6-3b24466632c2-kube-api-access-rl9td\") pod \"metallb-operator-webhook-server-568c9f7b75-dtmkk\" (UID: \"f1ea99a7-677c-4d8b-89c6-3b24466632c2\") " pod="metallb-system/metallb-operator-webhook-server-568c9f7b75-dtmkk" Dec 10 23:02:49 crc kubenswrapper[4791]: I1210 23:02:49.934183 4791 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/f1ea99a7-677c-4d8b-89c6-3b24466632c2-apiservice-cert\") pod \"metallb-operator-webhook-server-568c9f7b75-dtmkk\" (UID: \"f1ea99a7-677c-4d8b-89c6-3b24466632c2\") " pod="metallb-system/metallb-operator-webhook-server-568c9f7b75-dtmkk" Dec 10 23:02:49 crc kubenswrapper[4791]: I1210 23:02:49.937588 4791 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/f1ea99a7-677c-4d8b-89c6-3b24466632c2-webhook-cert\") pod \"metallb-operator-webhook-server-568c9f7b75-dtmkk\" (UID: \"f1ea99a7-677c-4d8b-89c6-3b24466632c2\") " pod="metallb-system/metallb-operator-webhook-server-568c9f7b75-dtmkk" Dec 10 23:02:49 crc kubenswrapper[4791]: I1210 23:02:49.937592 4791 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/f1ea99a7-677c-4d8b-89c6-3b24466632c2-apiservice-cert\") pod \"metallb-operator-webhook-server-568c9f7b75-dtmkk\" (UID: \"f1ea99a7-677c-4d8b-89c6-3b24466632c2\") " pod="metallb-system/metallb-operator-webhook-server-568c9f7b75-dtmkk" Dec 10 23:02:49 crc kubenswrapper[4791]: I1210 23:02:49.953516 4791 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-rl9td\" (UniqueName: \"kubernetes.io/projected/f1ea99a7-677c-4d8b-89c6-3b24466632c2-kube-api-access-rl9td\") pod \"metallb-operator-webhook-server-568c9f7b75-dtmkk\" (UID: \"f1ea99a7-677c-4d8b-89c6-3b24466632c2\") " pod="metallb-system/metallb-operator-webhook-server-568c9f7b75-dtmkk" Dec 10 23:02:50 crc kubenswrapper[4791]: I1210 23:02:50.069209 4791 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="metallb-system/metallb-operator-webhook-server-568c9f7b75-dtmkk" Dec 10 23:02:50 crc kubenswrapper[4791]: I1210 23:02:50.183390 4791 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["metallb-system/metallb-operator-controller-manager-867ddc9894-mhxh7"] Dec 10 23:02:50 crc kubenswrapper[4791]: I1210 23:02:50.333396 4791 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["metallb-system/metallb-operator-webhook-server-568c9f7b75-dtmkk"] Dec 10 23:02:50 crc kubenswrapper[4791]: W1210 23:02:50.340278 4791 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podf1ea99a7_677c_4d8b_89c6_3b24466632c2.slice/crio-924d63a9d6f3e895e02083d7f646d8e048c0f7fdd1f7f24f4cc2195f0aa221e8 WatchSource:0}: Error finding container 924d63a9d6f3e895e02083d7f646d8e048c0f7fdd1f7f24f4cc2195f0aa221e8: Status 404 returned error can't find the container with id 924d63a9d6f3e895e02083d7f646d8e048c0f7fdd1f7f24f4cc2195f0aa221e8 Dec 10 23:02:50 crc kubenswrapper[4791]: I1210 23:02:50.545928 4791 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/metallb-operator-controller-manager-867ddc9894-mhxh7" event={"ID":"72a07cc5-20c1-4f60-b6d7-84dfe1f78be2","Type":"ContainerStarted","Data":"6d19b06d2f04103762d8a671970c1511d0b1cc43cf65b0db0ed4359ff1f44931"} Dec 10 23:02:50 crc kubenswrapper[4791]: I1210 23:02:50.546851 4791 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/metallb-operator-webhook-server-568c9f7b75-dtmkk" event={"ID":"f1ea99a7-677c-4d8b-89c6-3b24466632c2","Type":"ContainerStarted","Data":"924d63a9d6f3e895e02083d7f646d8e048c0f7fdd1f7f24f4cc2195f0aa221e8"} Dec 10 23:02:55 crc kubenswrapper[4791]: I1210 23:02:55.038515 4791 patch_prober.go:28] interesting pod/machine-config-daemon-5rb5l container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 10 23:02:55 crc kubenswrapper[4791]: I1210 23:02:55.039087 4791 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-5rb5l" podUID="ba35653c-6e06-4cee-a4d6-137764090d18" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 10 23:02:55 crc kubenswrapper[4791]: I1210 23:02:55.039134 4791 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-5rb5l" Dec 10 23:02:55 crc kubenswrapper[4791]: I1210 23:02:55.039779 4791 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"cb32d752921f31f5faa2ab64f1cbd0e43f7e3f3e95a29f0f2fbc2ed42452d743"} pod="openshift-machine-config-operator/machine-config-daemon-5rb5l" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Dec 10 23:02:55 crc kubenswrapper[4791]: I1210 23:02:55.039833 4791 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-5rb5l" podUID="ba35653c-6e06-4cee-a4d6-137764090d18" containerName="machine-config-daemon" containerID="cri-o://cb32d752921f31f5faa2ab64f1cbd0e43f7e3f3e95a29f0f2fbc2ed42452d743" gracePeriod=600 Dec 10 23:02:55 crc kubenswrapper[4791]: I1210 23:02:55.417611 4791 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-operators-5f42v" Dec 10 23:02:55 crc kubenswrapper[4791]: I1210 23:02:55.505877 4791 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-operators-5f42v" Dec 10 23:02:55 crc kubenswrapper[4791]: I1210 23:02:55.589548 4791 generic.go:334] "Generic (PLEG): container finished" podID="ba35653c-6e06-4cee-a4d6-137764090d18" containerID="cb32d752921f31f5faa2ab64f1cbd0e43f7e3f3e95a29f0f2fbc2ed42452d743" exitCode=0 Dec 10 23:02:55 crc kubenswrapper[4791]: I1210 23:02:55.589618 4791 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-5rb5l" event={"ID":"ba35653c-6e06-4cee-a4d6-137764090d18","Type":"ContainerDied","Data":"cb32d752921f31f5faa2ab64f1cbd0e43f7e3f3e95a29f0f2fbc2ed42452d743"} Dec 10 23:02:55 crc kubenswrapper[4791]: I1210 23:02:55.589658 4791 scope.go:117] "RemoveContainer" containerID="c09d66dfdd1ff68cb29fe7936cd36ccb2e44990bd887b93547e1487b370efa4e" Dec 10 23:02:57 crc kubenswrapper[4791]: I1210 23:02:57.411804 4791 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-5f42v"] Dec 10 23:02:57 crc kubenswrapper[4791]: I1210 23:02:57.412837 4791 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-operators-5f42v" podUID="8e8bd8f8-4aa4-4d7f-8c0a-a463ec584e50" containerName="registry-server" containerID="cri-o://1e472052621dbee8bc1f69f15edefadbe91ec8b5bfeee098182af13632806572" gracePeriod=2 Dec 10 23:02:58 crc kubenswrapper[4791]: I1210 23:02:58.607402 4791 generic.go:334] "Generic (PLEG): container finished" podID="8e8bd8f8-4aa4-4d7f-8c0a-a463ec584e50" containerID="1e472052621dbee8bc1f69f15edefadbe91ec8b5bfeee098182af13632806572" exitCode=0 Dec 10 23:02:58 crc kubenswrapper[4791]: I1210 23:02:58.607463 4791 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-5f42v" event={"ID":"8e8bd8f8-4aa4-4d7f-8c0a-a463ec584e50","Type":"ContainerDied","Data":"1e472052621dbee8bc1f69f15edefadbe91ec8b5bfeee098182af13632806572"} Dec 10 23:02:59 crc kubenswrapper[4791]: I1210 23:02:59.379055 4791 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-5f42v" Dec 10 23:02:59 crc kubenswrapper[4791]: I1210 23:02:59.490199 4791 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/8e8bd8f8-4aa4-4d7f-8c0a-a463ec584e50-catalog-content\") pod \"8e8bd8f8-4aa4-4d7f-8c0a-a463ec584e50\" (UID: \"8e8bd8f8-4aa4-4d7f-8c0a-a463ec584e50\") " Dec 10 23:02:59 crc kubenswrapper[4791]: I1210 23:02:59.490663 4791 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/8e8bd8f8-4aa4-4d7f-8c0a-a463ec584e50-utilities\") pod \"8e8bd8f8-4aa4-4d7f-8c0a-a463ec584e50\" (UID: \"8e8bd8f8-4aa4-4d7f-8c0a-a463ec584e50\") " Dec 10 23:02:59 crc kubenswrapper[4791]: I1210 23:02:59.490705 4791 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-p8z9h\" (UniqueName: \"kubernetes.io/projected/8e8bd8f8-4aa4-4d7f-8c0a-a463ec584e50-kube-api-access-p8z9h\") pod \"8e8bd8f8-4aa4-4d7f-8c0a-a463ec584e50\" (UID: \"8e8bd8f8-4aa4-4d7f-8c0a-a463ec584e50\") " Dec 10 23:02:59 crc kubenswrapper[4791]: I1210 23:02:59.491636 4791 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/8e8bd8f8-4aa4-4d7f-8c0a-a463ec584e50-utilities" (OuterVolumeSpecName: "utilities") pod "8e8bd8f8-4aa4-4d7f-8c0a-a463ec584e50" (UID: "8e8bd8f8-4aa4-4d7f-8c0a-a463ec584e50"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 10 23:02:59 crc kubenswrapper[4791]: I1210 23:02:59.505645 4791 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/8e8bd8f8-4aa4-4d7f-8c0a-a463ec584e50-kube-api-access-p8z9h" (OuterVolumeSpecName: "kube-api-access-p8z9h") pod "8e8bd8f8-4aa4-4d7f-8c0a-a463ec584e50" (UID: "8e8bd8f8-4aa4-4d7f-8c0a-a463ec584e50"). InnerVolumeSpecName "kube-api-access-p8z9h". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 10 23:02:59 crc kubenswrapper[4791]: I1210 23:02:59.592180 4791 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/8e8bd8f8-4aa4-4d7f-8c0a-a463ec584e50-utilities\") on node \"crc\" DevicePath \"\"" Dec 10 23:02:59 crc kubenswrapper[4791]: I1210 23:02:59.592227 4791 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-p8z9h\" (UniqueName: \"kubernetes.io/projected/8e8bd8f8-4aa4-4d7f-8c0a-a463ec584e50-kube-api-access-p8z9h\") on node \"crc\" DevicePath \"\"" Dec 10 23:02:59 crc kubenswrapper[4791]: I1210 23:02:59.611744 4791 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/8e8bd8f8-4aa4-4d7f-8c0a-a463ec584e50-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "8e8bd8f8-4aa4-4d7f-8c0a-a463ec584e50" (UID: "8e8bd8f8-4aa4-4d7f-8c0a-a463ec584e50"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 10 23:02:59 crc kubenswrapper[4791]: I1210 23:02:59.617780 4791 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/metallb-operator-webhook-server-568c9f7b75-dtmkk" event={"ID":"f1ea99a7-677c-4d8b-89c6-3b24466632c2","Type":"ContainerStarted","Data":"380e70e0d674ac437ab47eec8e824d09a64f475cc368266e3b7526bbb43746e5"} Dec 10 23:02:59 crc kubenswrapper[4791]: I1210 23:02:59.617913 4791 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="metallb-system/metallb-operator-webhook-server-568c9f7b75-dtmkk" Dec 10 23:02:59 crc kubenswrapper[4791]: I1210 23:02:59.621684 4791 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/metallb-operator-controller-manager-867ddc9894-mhxh7" event={"ID":"72a07cc5-20c1-4f60-b6d7-84dfe1f78be2","Type":"ContainerStarted","Data":"96c7fd090fbd0ee2a8374400262b84cd1418e1a3b1e21b494a67387e87b0c3a6"} Dec 10 23:02:59 crc kubenswrapper[4791]: I1210 23:02:59.621763 4791 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="metallb-system/metallb-operator-controller-manager-867ddc9894-mhxh7" Dec 10 23:02:59 crc kubenswrapper[4791]: I1210 23:02:59.626028 4791 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-5rb5l" event={"ID":"ba35653c-6e06-4cee-a4d6-137764090d18","Type":"ContainerStarted","Data":"6bffb04499032491e4580b048fb3dc3c71ca5a93b7a98e3a62cb913d0da4b5c3"} Dec 10 23:02:59 crc kubenswrapper[4791]: I1210 23:02:59.628149 4791 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-5f42v" event={"ID":"8e8bd8f8-4aa4-4d7f-8c0a-a463ec584e50","Type":"ContainerDied","Data":"d2a7c3f59813d771fd051db5cc1db664ce0ba8463be4b539b5b95efac7473f18"} Dec 10 23:02:59 crc kubenswrapper[4791]: I1210 23:02:59.628202 4791 scope.go:117] "RemoveContainer" containerID="1e472052621dbee8bc1f69f15edefadbe91ec8b5bfeee098182af13632806572" Dec 10 23:02:59 crc kubenswrapper[4791]: I1210 23:02:59.628201 4791 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-5f42v" Dec 10 23:02:59 crc kubenswrapper[4791]: I1210 23:02:59.642749 4791 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="metallb-system/metallb-operator-webhook-server-568c9f7b75-dtmkk" podStartSLOduration=1.763294481 podStartE2EDuration="10.642731962s" podCreationTimestamp="2025-12-10 23:02:49 +0000 UTC" firstStartedPulling="2025-12-10 23:02:50.342775344 +0000 UTC m=+804.772392957" lastFinishedPulling="2025-12-10 23:02:59.222212825 +0000 UTC m=+813.651830438" observedRunningTime="2025-12-10 23:02:59.636130581 +0000 UTC m=+814.065748234" watchObservedRunningTime="2025-12-10 23:02:59.642731962 +0000 UTC m=+814.072349575" Dec 10 23:02:59 crc kubenswrapper[4791]: I1210 23:02:59.646859 4791 scope.go:117] "RemoveContainer" containerID="25915819fdcc19309fe7285b7a21111a6a7920a7b4828600cd98564a38009d00" Dec 10 23:02:59 crc kubenswrapper[4791]: I1210 23:02:59.659191 4791 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="metallb-system/metallb-operator-controller-manager-867ddc9894-mhxh7" podStartSLOduration=1.658771622 podStartE2EDuration="10.659170337s" podCreationTimestamp="2025-12-10 23:02:49 +0000 UTC" firstStartedPulling="2025-12-10 23:02:50.205572771 +0000 UTC m=+804.635190384" lastFinishedPulling="2025-12-10 23:02:59.205971486 +0000 UTC m=+813.635589099" observedRunningTime="2025-12-10 23:02:59.655283934 +0000 UTC m=+814.084901547" watchObservedRunningTime="2025-12-10 23:02:59.659170337 +0000 UTC m=+814.088787950" Dec 10 23:02:59 crc kubenswrapper[4791]: I1210 23:02:59.692915 4791 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/8e8bd8f8-4aa4-4d7f-8c0a-a463ec584e50-catalog-content\") on node \"crc\" DevicePath \"\"" Dec 10 23:02:59 crc kubenswrapper[4791]: I1210 23:02:59.703143 4791 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-5f42v"] Dec 10 23:02:59 crc kubenswrapper[4791]: I1210 23:02:59.706437 4791 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-operators-5f42v"] Dec 10 23:02:59 crc kubenswrapper[4791]: I1210 23:02:59.707873 4791 scope.go:117] "RemoveContainer" containerID="4c4864681123f42be2c04ec86b24f334e4a999dab31aff7e11a01a26b70ed561" Dec 10 23:02:59 crc kubenswrapper[4791]: I1210 23:02:59.893104 4791 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="8e8bd8f8-4aa4-4d7f-8c0a-a463ec584e50" path="/var/lib/kubelet/pods/8e8bd8f8-4aa4-4d7f-8c0a-a463ec584e50/volumes" Dec 10 23:03:10 crc kubenswrapper[4791]: I1210 23:03:10.076751 4791 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="metallb-system/metallb-operator-webhook-server-568c9f7b75-dtmkk" Dec 10 23:03:29 crc kubenswrapper[4791]: I1210 23:03:29.771364 4791 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="metallb-system/metallb-operator-controller-manager-867ddc9894-mhxh7" Dec 10 23:03:30 crc kubenswrapper[4791]: I1210 23:03:30.524724 4791 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["metallb-system/frr-k8s-webhook-server-7fcb986d4-whw8m"] Dec 10 23:03:30 crc kubenswrapper[4791]: E1210 23:03:30.525304 4791 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="8e8bd8f8-4aa4-4d7f-8c0a-a463ec584e50" containerName="extract-utilities" Dec 10 23:03:30 crc kubenswrapper[4791]: I1210 23:03:30.525320 4791 state_mem.go:107] "Deleted CPUSet assignment" podUID="8e8bd8f8-4aa4-4d7f-8c0a-a463ec584e50" containerName="extract-utilities" Dec 10 23:03:30 crc kubenswrapper[4791]: E1210 23:03:30.525353 4791 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="8e8bd8f8-4aa4-4d7f-8c0a-a463ec584e50" containerName="registry-server" Dec 10 23:03:30 crc kubenswrapper[4791]: I1210 23:03:30.525364 4791 state_mem.go:107] "Deleted CPUSet assignment" podUID="8e8bd8f8-4aa4-4d7f-8c0a-a463ec584e50" containerName="registry-server" Dec 10 23:03:30 crc kubenswrapper[4791]: E1210 23:03:30.525374 4791 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="8e8bd8f8-4aa4-4d7f-8c0a-a463ec584e50" containerName="extract-content" Dec 10 23:03:30 crc kubenswrapper[4791]: I1210 23:03:30.525381 4791 state_mem.go:107] "Deleted CPUSet assignment" podUID="8e8bd8f8-4aa4-4d7f-8c0a-a463ec584e50" containerName="extract-content" Dec 10 23:03:30 crc kubenswrapper[4791]: I1210 23:03:30.525514 4791 memory_manager.go:354] "RemoveStaleState removing state" podUID="8e8bd8f8-4aa4-4d7f-8c0a-a463ec584e50" containerName="registry-server" Dec 10 23:03:30 crc kubenswrapper[4791]: I1210 23:03:30.526024 4791 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="metallb-system/frr-k8s-webhook-server-7fcb986d4-whw8m" Dec 10 23:03:30 crc kubenswrapper[4791]: I1210 23:03:30.529738 4791 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"frr-k8s-webhook-server-cert" Dec 10 23:03:30 crc kubenswrapper[4791]: I1210 23:03:30.534011 4791 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["metallb-system/frr-k8s-gqlpm"] Dec 10 23:03:30 crc kubenswrapper[4791]: I1210 23:03:30.536685 4791 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="metallb-system/frr-k8s-gqlpm" Dec 10 23:03:30 crc kubenswrapper[4791]: I1210 23:03:30.537510 4791 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"frr-k8s-daemon-dockercfg-wdzlq" Dec 10 23:03:30 crc kubenswrapper[4791]: I1210 23:03:30.540162 4791 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"frr-k8s-certs-secret" Dec 10 23:03:30 crc kubenswrapper[4791]: I1210 23:03:30.543691 4791 reflector.go:368] Caches populated for *v1.ConfigMap from object-"metallb-system"/"frr-startup" Dec 10 23:03:30 crc kubenswrapper[4791]: I1210 23:03:30.545663 4791 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["metallb-system/frr-k8s-webhook-server-7fcb986d4-whw8m"] Dec 10 23:03:30 crc kubenswrapper[4791]: I1210 23:03:30.643672 4791 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["metallb-system/speaker-jfggv"] Dec 10 23:03:30 crc kubenswrapper[4791]: I1210 23:03:30.644540 4791 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="metallb-system/speaker-jfggv" Dec 10 23:03:30 crc kubenswrapper[4791]: I1210 23:03:30.646758 4791 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"speaker-certs-secret" Dec 10 23:03:30 crc kubenswrapper[4791]: I1210 23:03:30.646886 4791 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"speaker-dockercfg-5ctcg" Dec 10 23:03:30 crc kubenswrapper[4791]: I1210 23:03:30.648414 4791 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"metallb-memberlist" Dec 10 23:03:30 crc kubenswrapper[4791]: I1210 23:03:30.652902 4791 reflector.go:368] Caches populated for *v1.ConfigMap from object-"metallb-system"/"metallb-excludel2" Dec 10 23:03:30 crc kubenswrapper[4791]: I1210 23:03:30.681531 4791 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics\" (UniqueName: \"kubernetes.io/empty-dir/a68714bc-7787-46c8-ba8f-21f7af6d9d67-metrics\") pod \"frr-k8s-gqlpm\" (UID: \"a68714bc-7787-46c8-ba8f-21f7af6d9d67\") " pod="metallb-system/frr-k8s-gqlpm" Dec 10 23:03:30 crc kubenswrapper[4791]: I1210 23:03:30.681600 4791 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metallb-excludel2\" (UniqueName: \"kubernetes.io/configmap/10e04a8d-1cac-4c7a-9b14-be8db393651a-metallb-excludel2\") pod \"speaker-jfggv\" (UID: \"10e04a8d-1cac-4c7a-9b14-be8db393651a\") " pod="metallb-system/speaker-jfggv" Dec 10 23:03:30 crc kubenswrapper[4791]: I1210 23:03:30.681632 4791 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/a68714bc-7787-46c8-ba8f-21f7af6d9d67-metrics-certs\") pod \"frr-k8s-gqlpm\" (UID: \"a68714bc-7787-46c8-ba8f-21f7af6d9d67\") " pod="metallb-system/frr-k8s-gqlpm" Dec 10 23:03:30 crc kubenswrapper[4791]: I1210 23:03:30.681730 4791 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/d1dc382e-929c-44ce-b5c3-73410e626d0c-cert\") pod \"frr-k8s-webhook-server-7fcb986d4-whw8m\" (UID: \"d1dc382e-929c-44ce-b5c3-73410e626d0c\") " pod="metallb-system/frr-k8s-webhook-server-7fcb986d4-whw8m" Dec 10 23:03:30 crc kubenswrapper[4791]: I1210 23:03:30.681783 4791 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"frr-sockets\" (UniqueName: \"kubernetes.io/empty-dir/a68714bc-7787-46c8-ba8f-21f7af6d9d67-frr-sockets\") pod \"frr-k8s-gqlpm\" (UID: \"a68714bc-7787-46c8-ba8f-21f7af6d9d67\") " pod="metallb-system/frr-k8s-gqlpm" Dec 10 23:03:30 crc kubenswrapper[4791]: I1210 23:03:30.681814 4791 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-b7bf9\" (UniqueName: \"kubernetes.io/projected/d1dc382e-929c-44ce-b5c3-73410e626d0c-kube-api-access-b7bf9\") pod \"frr-k8s-webhook-server-7fcb986d4-whw8m\" (UID: \"d1dc382e-929c-44ce-b5c3-73410e626d0c\") " pod="metallb-system/frr-k8s-webhook-server-7fcb986d4-whw8m" Dec 10 23:03:30 crc kubenswrapper[4791]: I1210 23:03:30.681907 4791 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-vfp2x\" (UniqueName: \"kubernetes.io/projected/10e04a8d-1cac-4c7a-9b14-be8db393651a-kube-api-access-vfp2x\") pod \"speaker-jfggv\" (UID: \"10e04a8d-1cac-4c7a-9b14-be8db393651a\") " pod="metallb-system/speaker-jfggv" Dec 10 23:03:30 crc kubenswrapper[4791]: I1210 23:03:30.681932 4791 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"memberlist\" (UniqueName: \"kubernetes.io/secret/10e04a8d-1cac-4c7a-9b14-be8db393651a-memberlist\") pod \"speaker-jfggv\" (UID: \"10e04a8d-1cac-4c7a-9b14-be8db393651a\") " pod="metallb-system/speaker-jfggv" Dec 10 23:03:30 crc kubenswrapper[4791]: I1210 23:03:30.681998 4791 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"frr-conf\" (UniqueName: \"kubernetes.io/empty-dir/a68714bc-7787-46c8-ba8f-21f7af6d9d67-frr-conf\") pod \"frr-k8s-gqlpm\" (UID: \"a68714bc-7787-46c8-ba8f-21f7af6d9d67\") " pod="metallb-system/frr-k8s-gqlpm" Dec 10 23:03:30 crc kubenswrapper[4791]: I1210 23:03:30.682036 4791 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/10e04a8d-1cac-4c7a-9b14-be8db393651a-metrics-certs\") pod \"speaker-jfggv\" (UID: \"10e04a8d-1cac-4c7a-9b14-be8db393651a\") " pod="metallb-system/speaker-jfggv" Dec 10 23:03:30 crc kubenswrapper[4791]: I1210 23:03:30.682056 4791 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"reloader\" (UniqueName: \"kubernetes.io/empty-dir/a68714bc-7787-46c8-ba8f-21f7af6d9d67-reloader\") pod \"frr-k8s-gqlpm\" (UID: \"a68714bc-7787-46c8-ba8f-21f7af6d9d67\") " pod="metallb-system/frr-k8s-gqlpm" Dec 10 23:03:30 crc kubenswrapper[4791]: I1210 23:03:30.682084 4791 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-p8v2x\" (UniqueName: \"kubernetes.io/projected/a68714bc-7787-46c8-ba8f-21f7af6d9d67-kube-api-access-p8v2x\") pod \"frr-k8s-gqlpm\" (UID: \"a68714bc-7787-46c8-ba8f-21f7af6d9d67\") " pod="metallb-system/frr-k8s-gqlpm" Dec 10 23:03:30 crc kubenswrapper[4791]: I1210 23:03:30.682104 4791 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"frr-startup\" (UniqueName: \"kubernetes.io/configmap/a68714bc-7787-46c8-ba8f-21f7af6d9d67-frr-startup\") pod \"frr-k8s-gqlpm\" (UID: \"a68714bc-7787-46c8-ba8f-21f7af6d9d67\") " pod="metallb-system/frr-k8s-gqlpm" Dec 10 23:03:30 crc kubenswrapper[4791]: I1210 23:03:30.684244 4791 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["metallb-system/controller-f8648f98b-q9cdb"] Dec 10 23:03:30 crc kubenswrapper[4791]: I1210 23:03:30.685118 4791 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="metallb-system/controller-f8648f98b-q9cdb" Dec 10 23:03:30 crc kubenswrapper[4791]: I1210 23:03:30.689568 4791 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"controller-certs-secret" Dec 10 23:03:30 crc kubenswrapper[4791]: I1210 23:03:30.702884 4791 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["metallb-system/controller-f8648f98b-q9cdb"] Dec 10 23:03:30 crc kubenswrapper[4791]: I1210 23:03:30.783457 4791 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-vfp2x\" (UniqueName: \"kubernetes.io/projected/10e04a8d-1cac-4c7a-9b14-be8db393651a-kube-api-access-vfp2x\") pod \"speaker-jfggv\" (UID: \"10e04a8d-1cac-4c7a-9b14-be8db393651a\") " pod="metallb-system/speaker-jfggv" Dec 10 23:03:30 crc kubenswrapper[4791]: I1210 23:03:30.783528 4791 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"memberlist\" (UniqueName: \"kubernetes.io/secret/10e04a8d-1cac-4c7a-9b14-be8db393651a-memberlist\") pod \"speaker-jfggv\" (UID: \"10e04a8d-1cac-4c7a-9b14-be8db393651a\") " pod="metallb-system/speaker-jfggv" Dec 10 23:03:30 crc kubenswrapper[4791]: I1210 23:03:30.783565 4791 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"frr-conf\" (UniqueName: \"kubernetes.io/empty-dir/a68714bc-7787-46c8-ba8f-21f7af6d9d67-frr-conf\") pod \"frr-k8s-gqlpm\" (UID: \"a68714bc-7787-46c8-ba8f-21f7af6d9d67\") " pod="metallb-system/frr-k8s-gqlpm" Dec 10 23:03:30 crc kubenswrapper[4791]: I1210 23:03:30.783607 4791 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/10e04a8d-1cac-4c7a-9b14-be8db393651a-metrics-certs\") pod \"speaker-jfggv\" (UID: \"10e04a8d-1cac-4c7a-9b14-be8db393651a\") " pod="metallb-system/speaker-jfggv" Dec 10 23:03:30 crc kubenswrapper[4791]: I1210 23:03:30.783635 4791 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"reloader\" (UniqueName: \"kubernetes.io/empty-dir/a68714bc-7787-46c8-ba8f-21f7af6d9d67-reloader\") pod \"frr-k8s-gqlpm\" (UID: \"a68714bc-7787-46c8-ba8f-21f7af6d9d67\") " pod="metallb-system/frr-k8s-gqlpm" Dec 10 23:03:30 crc kubenswrapper[4791]: I1210 23:03:30.783686 4791 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-gv6x4\" (UniqueName: \"kubernetes.io/projected/e52e894c-94e4-4958-b261-c5c373c29040-kube-api-access-gv6x4\") pod \"controller-f8648f98b-q9cdb\" (UID: \"e52e894c-94e4-4958-b261-c5c373c29040\") " pod="metallb-system/controller-f8648f98b-q9cdb" Dec 10 23:03:30 crc kubenswrapper[4791]: E1210 23:03:30.783698 4791 secret.go:188] Couldn't get secret metallb-system/metallb-memberlist: secret "metallb-memberlist" not found Dec 10 23:03:30 crc kubenswrapper[4791]: I1210 23:03:30.783715 4791 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-p8v2x\" (UniqueName: \"kubernetes.io/projected/a68714bc-7787-46c8-ba8f-21f7af6d9d67-kube-api-access-p8v2x\") pod \"frr-k8s-gqlpm\" (UID: \"a68714bc-7787-46c8-ba8f-21f7af6d9d67\") " pod="metallb-system/frr-k8s-gqlpm" Dec 10 23:03:30 crc kubenswrapper[4791]: E1210 23:03:30.783778 4791 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/10e04a8d-1cac-4c7a-9b14-be8db393651a-memberlist podName:10e04a8d-1cac-4c7a-9b14-be8db393651a nodeName:}" failed. No retries permitted until 2025-12-10 23:03:31.283759837 +0000 UTC m=+845.713377450 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "memberlist" (UniqueName: "kubernetes.io/secret/10e04a8d-1cac-4c7a-9b14-be8db393651a-memberlist") pod "speaker-jfggv" (UID: "10e04a8d-1cac-4c7a-9b14-be8db393651a") : secret "metallb-memberlist" not found Dec 10 23:03:30 crc kubenswrapper[4791]: E1210 23:03:30.783785 4791 secret.go:188] Couldn't get secret metallb-system/speaker-certs-secret: secret "speaker-certs-secret" not found Dec 10 23:03:30 crc kubenswrapper[4791]: E1210 23:03:30.783883 4791 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/10e04a8d-1cac-4c7a-9b14-be8db393651a-metrics-certs podName:10e04a8d-1cac-4c7a-9b14-be8db393651a nodeName:}" failed. No retries permitted until 2025-12-10 23:03:31.28387622 +0000 UTC m=+845.713493833 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "metrics-certs" (UniqueName: "kubernetes.io/secret/10e04a8d-1cac-4c7a-9b14-be8db393651a-metrics-certs") pod "speaker-jfggv" (UID: "10e04a8d-1cac-4c7a-9b14-be8db393651a") : secret "speaker-certs-secret" not found Dec 10 23:03:30 crc kubenswrapper[4791]: I1210 23:03:30.783948 4791 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"frr-startup\" (UniqueName: \"kubernetes.io/configmap/a68714bc-7787-46c8-ba8f-21f7af6d9d67-frr-startup\") pod \"frr-k8s-gqlpm\" (UID: \"a68714bc-7787-46c8-ba8f-21f7af6d9d67\") " pod="metallb-system/frr-k8s-gqlpm" Dec 10 23:03:30 crc kubenswrapper[4791]: I1210 23:03:30.784068 4791 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/e52e894c-94e4-4958-b261-c5c373c29040-metrics-certs\") pod \"controller-f8648f98b-q9cdb\" (UID: \"e52e894c-94e4-4958-b261-c5c373c29040\") " pod="metallb-system/controller-f8648f98b-q9cdb" Dec 10 23:03:30 crc kubenswrapper[4791]: I1210 23:03:30.784135 4791 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics\" (UniqueName: \"kubernetes.io/empty-dir/a68714bc-7787-46c8-ba8f-21f7af6d9d67-metrics\") pod \"frr-k8s-gqlpm\" (UID: \"a68714bc-7787-46c8-ba8f-21f7af6d9d67\") " pod="metallb-system/frr-k8s-gqlpm" Dec 10 23:03:30 crc kubenswrapper[4791]: I1210 23:03:30.784239 4791 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metallb-excludel2\" (UniqueName: \"kubernetes.io/configmap/10e04a8d-1cac-4c7a-9b14-be8db393651a-metallb-excludel2\") pod \"speaker-jfggv\" (UID: \"10e04a8d-1cac-4c7a-9b14-be8db393651a\") " pod="metallb-system/speaker-jfggv" Dec 10 23:03:30 crc kubenswrapper[4791]: I1210 23:03:30.784259 4791 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"frr-conf\" (UniqueName: \"kubernetes.io/empty-dir/a68714bc-7787-46c8-ba8f-21f7af6d9d67-frr-conf\") pod \"frr-k8s-gqlpm\" (UID: \"a68714bc-7787-46c8-ba8f-21f7af6d9d67\") " pod="metallb-system/frr-k8s-gqlpm" Dec 10 23:03:30 crc kubenswrapper[4791]: I1210 23:03:30.784282 4791 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/a68714bc-7787-46c8-ba8f-21f7af6d9d67-metrics-certs\") pod \"frr-k8s-gqlpm\" (UID: \"a68714bc-7787-46c8-ba8f-21f7af6d9d67\") " pod="metallb-system/frr-k8s-gqlpm" Dec 10 23:03:30 crc kubenswrapper[4791]: I1210 23:03:30.784269 4791 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"reloader\" (UniqueName: \"kubernetes.io/empty-dir/a68714bc-7787-46c8-ba8f-21f7af6d9d67-reloader\") pod \"frr-k8s-gqlpm\" (UID: \"a68714bc-7787-46c8-ba8f-21f7af6d9d67\") " pod="metallb-system/frr-k8s-gqlpm" Dec 10 23:03:30 crc kubenswrapper[4791]: I1210 23:03:30.784316 4791 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/d1dc382e-929c-44ce-b5c3-73410e626d0c-cert\") pod \"frr-k8s-webhook-server-7fcb986d4-whw8m\" (UID: \"d1dc382e-929c-44ce-b5c3-73410e626d0c\") " pod="metallb-system/frr-k8s-webhook-server-7fcb986d4-whw8m" Dec 10 23:03:30 crc kubenswrapper[4791]: I1210 23:03:30.784468 4791 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics\" (UniqueName: \"kubernetes.io/empty-dir/a68714bc-7787-46c8-ba8f-21f7af6d9d67-metrics\") pod \"frr-k8s-gqlpm\" (UID: \"a68714bc-7787-46c8-ba8f-21f7af6d9d67\") " pod="metallb-system/frr-k8s-gqlpm" Dec 10 23:03:30 crc kubenswrapper[4791]: I1210 23:03:30.784493 4791 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"frr-sockets\" (UniqueName: \"kubernetes.io/empty-dir/a68714bc-7787-46c8-ba8f-21f7af6d9d67-frr-sockets\") pod \"frr-k8s-gqlpm\" (UID: \"a68714bc-7787-46c8-ba8f-21f7af6d9d67\") " pod="metallb-system/frr-k8s-gqlpm" Dec 10 23:03:30 crc kubenswrapper[4791]: I1210 23:03:30.784547 4791 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-b7bf9\" (UniqueName: \"kubernetes.io/projected/d1dc382e-929c-44ce-b5c3-73410e626d0c-kube-api-access-b7bf9\") pod \"frr-k8s-webhook-server-7fcb986d4-whw8m\" (UID: \"d1dc382e-929c-44ce-b5c3-73410e626d0c\") " pod="metallb-system/frr-k8s-webhook-server-7fcb986d4-whw8m" Dec 10 23:03:30 crc kubenswrapper[4791]: I1210 23:03:30.784606 4791 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/e52e894c-94e4-4958-b261-c5c373c29040-cert\") pod \"controller-f8648f98b-q9cdb\" (UID: \"e52e894c-94e4-4958-b261-c5c373c29040\") " pod="metallb-system/controller-f8648f98b-q9cdb" Dec 10 23:03:30 crc kubenswrapper[4791]: I1210 23:03:30.784916 4791 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"frr-sockets\" (UniqueName: \"kubernetes.io/empty-dir/a68714bc-7787-46c8-ba8f-21f7af6d9d67-frr-sockets\") pod \"frr-k8s-gqlpm\" (UID: \"a68714bc-7787-46c8-ba8f-21f7af6d9d67\") " pod="metallb-system/frr-k8s-gqlpm" Dec 10 23:03:30 crc kubenswrapper[4791]: I1210 23:03:30.785008 4791 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metallb-excludel2\" (UniqueName: \"kubernetes.io/configmap/10e04a8d-1cac-4c7a-9b14-be8db393651a-metallb-excludel2\") pod \"speaker-jfggv\" (UID: \"10e04a8d-1cac-4c7a-9b14-be8db393651a\") " pod="metallb-system/speaker-jfggv" Dec 10 23:03:30 crc kubenswrapper[4791]: I1210 23:03:30.785026 4791 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"frr-startup\" (UniqueName: \"kubernetes.io/configmap/a68714bc-7787-46c8-ba8f-21f7af6d9d67-frr-startup\") pod \"frr-k8s-gqlpm\" (UID: \"a68714bc-7787-46c8-ba8f-21f7af6d9d67\") " pod="metallb-system/frr-k8s-gqlpm" Dec 10 23:03:30 crc kubenswrapper[4791]: I1210 23:03:30.791128 4791 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/a68714bc-7787-46c8-ba8f-21f7af6d9d67-metrics-certs\") pod \"frr-k8s-gqlpm\" (UID: \"a68714bc-7787-46c8-ba8f-21f7af6d9d67\") " pod="metallb-system/frr-k8s-gqlpm" Dec 10 23:03:30 crc kubenswrapper[4791]: I1210 23:03:30.804210 4791 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cert\" (UniqueName: \"kubernetes.io/secret/d1dc382e-929c-44ce-b5c3-73410e626d0c-cert\") pod \"frr-k8s-webhook-server-7fcb986d4-whw8m\" (UID: \"d1dc382e-929c-44ce-b5c3-73410e626d0c\") " pod="metallb-system/frr-k8s-webhook-server-7fcb986d4-whw8m" Dec 10 23:03:30 crc kubenswrapper[4791]: I1210 23:03:30.811391 4791 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-b7bf9\" (UniqueName: \"kubernetes.io/projected/d1dc382e-929c-44ce-b5c3-73410e626d0c-kube-api-access-b7bf9\") pod \"frr-k8s-webhook-server-7fcb986d4-whw8m\" (UID: \"d1dc382e-929c-44ce-b5c3-73410e626d0c\") " pod="metallb-system/frr-k8s-webhook-server-7fcb986d4-whw8m" Dec 10 23:03:30 crc kubenswrapper[4791]: I1210 23:03:30.812008 4791 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-vfp2x\" (UniqueName: \"kubernetes.io/projected/10e04a8d-1cac-4c7a-9b14-be8db393651a-kube-api-access-vfp2x\") pod \"speaker-jfggv\" (UID: \"10e04a8d-1cac-4c7a-9b14-be8db393651a\") " pod="metallb-system/speaker-jfggv" Dec 10 23:03:30 crc kubenswrapper[4791]: I1210 23:03:30.814664 4791 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-p8v2x\" (UniqueName: \"kubernetes.io/projected/a68714bc-7787-46c8-ba8f-21f7af6d9d67-kube-api-access-p8v2x\") pod \"frr-k8s-gqlpm\" (UID: \"a68714bc-7787-46c8-ba8f-21f7af6d9d67\") " pod="metallb-system/frr-k8s-gqlpm" Dec 10 23:03:30 crc kubenswrapper[4791]: I1210 23:03:30.843847 4791 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="metallb-system/frr-k8s-webhook-server-7fcb986d4-whw8m" Dec 10 23:03:30 crc kubenswrapper[4791]: I1210 23:03:30.854085 4791 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="metallb-system/frr-k8s-gqlpm" Dec 10 23:03:30 crc kubenswrapper[4791]: I1210 23:03:30.885520 4791 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/e52e894c-94e4-4958-b261-c5c373c29040-cert\") pod \"controller-f8648f98b-q9cdb\" (UID: \"e52e894c-94e4-4958-b261-c5c373c29040\") " pod="metallb-system/controller-f8648f98b-q9cdb" Dec 10 23:03:30 crc kubenswrapper[4791]: I1210 23:03:30.885606 4791 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-gv6x4\" (UniqueName: \"kubernetes.io/projected/e52e894c-94e4-4958-b261-c5c373c29040-kube-api-access-gv6x4\") pod \"controller-f8648f98b-q9cdb\" (UID: \"e52e894c-94e4-4958-b261-c5c373c29040\") " pod="metallb-system/controller-f8648f98b-q9cdb" Dec 10 23:03:30 crc kubenswrapper[4791]: I1210 23:03:30.885655 4791 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/e52e894c-94e4-4958-b261-c5c373c29040-metrics-certs\") pod \"controller-f8648f98b-q9cdb\" (UID: \"e52e894c-94e4-4958-b261-c5c373c29040\") " pod="metallb-system/controller-f8648f98b-q9cdb" Dec 10 23:03:30 crc kubenswrapper[4791]: I1210 23:03:30.888505 4791 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"metallb-webhook-cert" Dec 10 23:03:30 crc kubenswrapper[4791]: I1210 23:03:30.891051 4791 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/e52e894c-94e4-4958-b261-c5c373c29040-metrics-certs\") pod \"controller-f8648f98b-q9cdb\" (UID: \"e52e894c-94e4-4958-b261-c5c373c29040\") " pod="metallb-system/controller-f8648f98b-q9cdb" Dec 10 23:03:30 crc kubenswrapper[4791]: I1210 23:03:30.901254 4791 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cert\" (UniqueName: \"kubernetes.io/secret/e52e894c-94e4-4958-b261-c5c373c29040-cert\") pod \"controller-f8648f98b-q9cdb\" (UID: \"e52e894c-94e4-4958-b261-c5c373c29040\") " pod="metallb-system/controller-f8648f98b-q9cdb" Dec 10 23:03:30 crc kubenswrapper[4791]: I1210 23:03:30.908433 4791 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-gv6x4\" (UniqueName: \"kubernetes.io/projected/e52e894c-94e4-4958-b261-c5c373c29040-kube-api-access-gv6x4\") pod \"controller-f8648f98b-q9cdb\" (UID: \"e52e894c-94e4-4958-b261-c5c373c29040\") " pod="metallb-system/controller-f8648f98b-q9cdb" Dec 10 23:03:31 crc kubenswrapper[4791]: I1210 23:03:30.999229 4791 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="metallb-system/controller-f8648f98b-q9cdb" Dec 10 23:03:31 crc kubenswrapper[4791]: I1210 23:03:31.208703 4791 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["metallb-system/controller-f8648f98b-q9cdb"] Dec 10 23:03:31 crc kubenswrapper[4791]: W1210 23:03:31.216656 4791 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pode52e894c_94e4_4958_b261_c5c373c29040.slice/crio-567120788cbb648b90b5d04176c7e62c6f9d0ddb93e4f34ea5cd10988227e658 WatchSource:0}: Error finding container 567120788cbb648b90b5d04176c7e62c6f9d0ddb93e4f34ea5cd10988227e658: Status 404 returned error can't find the container with id 567120788cbb648b90b5d04176c7e62c6f9d0ddb93e4f34ea5cd10988227e658 Dec 10 23:03:31 crc kubenswrapper[4791]: I1210 23:03:31.258816 4791 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["metallb-system/frr-k8s-webhook-server-7fcb986d4-whw8m"] Dec 10 23:03:31 crc kubenswrapper[4791]: I1210 23:03:31.289966 4791 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"memberlist\" (UniqueName: \"kubernetes.io/secret/10e04a8d-1cac-4c7a-9b14-be8db393651a-memberlist\") pod \"speaker-jfggv\" (UID: \"10e04a8d-1cac-4c7a-9b14-be8db393651a\") " pod="metallb-system/speaker-jfggv" Dec 10 23:03:31 crc kubenswrapper[4791]: I1210 23:03:31.290012 4791 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/10e04a8d-1cac-4c7a-9b14-be8db393651a-metrics-certs\") pod \"speaker-jfggv\" (UID: \"10e04a8d-1cac-4c7a-9b14-be8db393651a\") " pod="metallb-system/speaker-jfggv" Dec 10 23:03:31 crc kubenswrapper[4791]: E1210 23:03:31.290273 4791 secret.go:188] Couldn't get secret metallb-system/metallb-memberlist: secret "metallb-memberlist" not found Dec 10 23:03:31 crc kubenswrapper[4791]: E1210 23:03:31.290370 4791 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/10e04a8d-1cac-4c7a-9b14-be8db393651a-memberlist podName:10e04a8d-1cac-4c7a-9b14-be8db393651a nodeName:}" failed. No retries permitted until 2025-12-10 23:03:32.290317159 +0000 UTC m=+846.719934772 (durationBeforeRetry 1s). Error: MountVolume.SetUp failed for volume "memberlist" (UniqueName: "kubernetes.io/secret/10e04a8d-1cac-4c7a-9b14-be8db393651a-memberlist") pod "speaker-jfggv" (UID: "10e04a8d-1cac-4c7a-9b14-be8db393651a") : secret "metallb-memberlist" not found Dec 10 23:03:31 crc kubenswrapper[4791]: I1210 23:03:31.296531 4791 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/10e04a8d-1cac-4c7a-9b14-be8db393651a-metrics-certs\") pod \"speaker-jfggv\" (UID: \"10e04a8d-1cac-4c7a-9b14-be8db393651a\") " pod="metallb-system/speaker-jfggv" Dec 10 23:03:31 crc kubenswrapper[4791]: I1210 23:03:31.804044 4791 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/frr-k8s-webhook-server-7fcb986d4-whw8m" event={"ID":"d1dc382e-929c-44ce-b5c3-73410e626d0c","Type":"ContainerStarted","Data":"fc2de19401d6cb36741bd1aaca464c727a67b943b35ee3f16275c4513d7796d3"} Dec 10 23:03:31 crc kubenswrapper[4791]: I1210 23:03:31.805101 4791 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/frr-k8s-gqlpm" event={"ID":"a68714bc-7787-46c8-ba8f-21f7af6d9d67","Type":"ContainerStarted","Data":"6075a28cbc5188192cfc4f57f66eb6d6a3c53c6ba61081d4cd12ba4b39f9cf02"} Dec 10 23:03:31 crc kubenswrapper[4791]: I1210 23:03:31.805985 4791 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/controller-f8648f98b-q9cdb" event={"ID":"e52e894c-94e4-4958-b261-c5c373c29040","Type":"ContainerStarted","Data":"8674303d2c79fc39bb41b0506f308864e0503fc802ee2f837a44e0ab4ae3e9a3"} Dec 10 23:03:31 crc kubenswrapper[4791]: I1210 23:03:31.806047 4791 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/controller-f8648f98b-q9cdb" event={"ID":"e52e894c-94e4-4958-b261-c5c373c29040","Type":"ContainerStarted","Data":"ab5d5602a2b0e7c1c0460b803d96b09c2d5bff8ccc673d7d9e99af13b44db399"} Dec 10 23:03:31 crc kubenswrapper[4791]: I1210 23:03:31.806061 4791 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/controller-f8648f98b-q9cdb" event={"ID":"e52e894c-94e4-4958-b261-c5c373c29040","Type":"ContainerStarted","Data":"567120788cbb648b90b5d04176c7e62c6f9d0ddb93e4f34ea5cd10988227e658"} Dec 10 23:03:31 crc kubenswrapper[4791]: I1210 23:03:31.806298 4791 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="metallb-system/controller-f8648f98b-q9cdb" Dec 10 23:03:31 crc kubenswrapper[4791]: I1210 23:03:31.822603 4791 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="metallb-system/controller-f8648f98b-q9cdb" podStartSLOduration=1.822583863 podStartE2EDuration="1.822583863s" podCreationTimestamp="2025-12-10 23:03:30 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-10 23:03:31.819125754 +0000 UTC m=+846.248743367" watchObservedRunningTime="2025-12-10 23:03:31.822583863 +0000 UTC m=+846.252201476" Dec 10 23:03:32 crc kubenswrapper[4791]: I1210 23:03:32.305394 4791 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"memberlist\" (UniqueName: \"kubernetes.io/secret/10e04a8d-1cac-4c7a-9b14-be8db393651a-memberlist\") pod \"speaker-jfggv\" (UID: \"10e04a8d-1cac-4c7a-9b14-be8db393651a\") " pod="metallb-system/speaker-jfggv" Dec 10 23:03:32 crc kubenswrapper[4791]: I1210 23:03:32.310268 4791 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"memberlist\" (UniqueName: \"kubernetes.io/secret/10e04a8d-1cac-4c7a-9b14-be8db393651a-memberlist\") pod \"speaker-jfggv\" (UID: \"10e04a8d-1cac-4c7a-9b14-be8db393651a\") " pod="metallb-system/speaker-jfggv" Dec 10 23:03:32 crc kubenswrapper[4791]: I1210 23:03:32.460107 4791 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="metallb-system/speaker-jfggv" Dec 10 23:03:32 crc kubenswrapper[4791]: W1210 23:03:32.481785 4791 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod10e04a8d_1cac_4c7a_9b14_be8db393651a.slice/crio-b8a5d72983ce114ac419b973537a232bb31a9669c224e0ecdf3f45b648f09fbd WatchSource:0}: Error finding container b8a5d72983ce114ac419b973537a232bb31a9669c224e0ecdf3f45b648f09fbd: Status 404 returned error can't find the container with id b8a5d72983ce114ac419b973537a232bb31a9669c224e0ecdf3f45b648f09fbd Dec 10 23:03:32 crc kubenswrapper[4791]: I1210 23:03:32.815233 4791 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/speaker-jfggv" event={"ID":"10e04a8d-1cac-4c7a-9b14-be8db393651a","Type":"ContainerStarted","Data":"80a6846ebd38d758b8f69e569ff751ba71f38dd74e73d0c5afc76c30905ff238"} Dec 10 23:03:32 crc kubenswrapper[4791]: I1210 23:03:32.815277 4791 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/speaker-jfggv" event={"ID":"10e04a8d-1cac-4c7a-9b14-be8db393651a","Type":"ContainerStarted","Data":"b8a5d72983ce114ac419b973537a232bb31a9669c224e0ecdf3f45b648f09fbd"} Dec 10 23:03:33 crc kubenswrapper[4791]: I1210 23:03:33.834328 4791 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/speaker-jfggv" event={"ID":"10e04a8d-1cac-4c7a-9b14-be8db393651a","Type":"ContainerStarted","Data":"17a0ee94a9956249510ad5d642bd8184687ff0f19ccc2e412ab486bbcc381ad4"} Dec 10 23:03:33 crc kubenswrapper[4791]: I1210 23:03:33.834655 4791 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="metallb-system/speaker-jfggv" Dec 10 23:03:33 crc kubenswrapper[4791]: I1210 23:03:33.867929 4791 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="metallb-system/speaker-jfggv" podStartSLOduration=3.867909522 podStartE2EDuration="3.867909522s" podCreationTimestamp="2025-12-10 23:03:30 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-10 23:03:33.862093634 +0000 UTC m=+848.291711297" watchObservedRunningTime="2025-12-10 23:03:33.867909522 +0000 UTC m=+848.297527135" Dec 10 23:03:41 crc kubenswrapper[4791]: I1210 23:03:41.003650 4791 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="metallb-system/controller-f8648f98b-q9cdb" Dec 10 23:03:42 crc kubenswrapper[4791]: I1210 23:03:42.464195 4791 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="metallb-system/speaker-jfggv" Dec 10 23:03:42 crc kubenswrapper[4791]: I1210 23:03:42.897030 4791 generic.go:334] "Generic (PLEG): container finished" podID="a68714bc-7787-46c8-ba8f-21f7af6d9d67" containerID="369ca4f3b8acce88d0b68dc755a4b9eb18a90bf0568cee3f2c25a21c232ac17a" exitCode=0 Dec 10 23:03:42 crc kubenswrapper[4791]: I1210 23:03:42.897095 4791 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/frr-k8s-gqlpm" event={"ID":"a68714bc-7787-46c8-ba8f-21f7af6d9d67","Type":"ContainerDied","Data":"369ca4f3b8acce88d0b68dc755a4b9eb18a90bf0568cee3f2c25a21c232ac17a"} Dec 10 23:03:42 crc kubenswrapper[4791]: I1210 23:03:42.899181 4791 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/frr-k8s-webhook-server-7fcb986d4-whw8m" event={"ID":"d1dc382e-929c-44ce-b5c3-73410e626d0c","Type":"ContainerStarted","Data":"74d514bafd08a206e4f2320ca98253ff6803efa97c6800ddbf2c97a4005c917b"} Dec 10 23:03:42 crc kubenswrapper[4791]: I1210 23:03:42.899356 4791 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="metallb-system/frr-k8s-webhook-server-7fcb986d4-whw8m" Dec 10 23:03:42 crc kubenswrapper[4791]: I1210 23:03:42.937776 4791 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="metallb-system/frr-k8s-webhook-server-7fcb986d4-whw8m" podStartSLOduration=1.782781694 podStartE2EDuration="12.937758812s" podCreationTimestamp="2025-12-10 23:03:30 +0000 UTC" firstStartedPulling="2025-12-10 23:03:31.274569044 +0000 UTC m=+845.704186657" lastFinishedPulling="2025-12-10 23:03:42.429546152 +0000 UTC m=+856.859163775" observedRunningTime="2025-12-10 23:03:42.937110403 +0000 UTC m=+857.366728016" watchObservedRunningTime="2025-12-10 23:03:42.937758812 +0000 UTC m=+857.367376425" Dec 10 23:03:43 crc kubenswrapper[4791]: I1210 23:03:43.907688 4791 generic.go:334] "Generic (PLEG): container finished" podID="a68714bc-7787-46c8-ba8f-21f7af6d9d67" containerID="d99218049a64b8f1dc33eedfe259b07ed9862ed8a356f6ee0a26ed0dab7834c5" exitCode=0 Dec 10 23:03:43 crc kubenswrapper[4791]: I1210 23:03:43.907795 4791 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/frr-k8s-gqlpm" event={"ID":"a68714bc-7787-46c8-ba8f-21f7af6d9d67","Type":"ContainerDied","Data":"d99218049a64b8f1dc33eedfe259b07ed9862ed8a356f6ee0a26ed0dab7834c5"} Dec 10 23:03:44 crc kubenswrapper[4791]: I1210 23:03:44.917033 4791 generic.go:334] "Generic (PLEG): container finished" podID="a68714bc-7787-46c8-ba8f-21f7af6d9d67" containerID="bb9d2de691310e566887c1ccfcb9a7ba8a740ec67cc24cd9b89892c7f51f438a" exitCode=0 Dec 10 23:03:44 crc kubenswrapper[4791]: I1210 23:03:44.917078 4791 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/frr-k8s-gqlpm" event={"ID":"a68714bc-7787-46c8-ba8f-21f7af6d9d67","Type":"ContainerDied","Data":"bb9d2de691310e566887c1ccfcb9a7ba8a740ec67cc24cd9b89892c7f51f438a"} Dec 10 23:03:45 crc kubenswrapper[4791]: I1210 23:03:45.276947 4791 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/openstack-operator-index-fk8v7"] Dec 10 23:03:45 crc kubenswrapper[4791]: I1210 23:03:45.279046 4791 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/openstack-operator-index-fk8v7" Dec 10 23:03:45 crc kubenswrapper[4791]: I1210 23:03:45.291924 4791 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"openstack-operator-index-dockercfg-hpm45" Dec 10 23:03:45 crc kubenswrapper[4791]: I1210 23:03:45.292469 4791 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack-operators"/"openshift-service-ca.crt" Dec 10 23:03:45 crc kubenswrapper[4791]: I1210 23:03:45.294614 4791 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack-operators"/"kube-root-ca.crt" Dec 10 23:03:45 crc kubenswrapper[4791]: I1210 23:03:45.300170 4791 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/openstack-operator-index-fk8v7"] Dec 10 23:03:45 crc kubenswrapper[4791]: I1210 23:03:45.340293 4791 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-744dv\" (UniqueName: \"kubernetes.io/projected/ba6e8e36-4a00-49a9-b66b-ca3a1802f56a-kube-api-access-744dv\") pod \"openstack-operator-index-fk8v7\" (UID: \"ba6e8e36-4a00-49a9-b66b-ca3a1802f56a\") " pod="openstack-operators/openstack-operator-index-fk8v7" Dec 10 23:03:45 crc kubenswrapper[4791]: I1210 23:03:45.442437 4791 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-744dv\" (UniqueName: \"kubernetes.io/projected/ba6e8e36-4a00-49a9-b66b-ca3a1802f56a-kube-api-access-744dv\") pod \"openstack-operator-index-fk8v7\" (UID: \"ba6e8e36-4a00-49a9-b66b-ca3a1802f56a\") " pod="openstack-operators/openstack-operator-index-fk8v7" Dec 10 23:03:45 crc kubenswrapper[4791]: I1210 23:03:45.461700 4791 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-744dv\" (UniqueName: \"kubernetes.io/projected/ba6e8e36-4a00-49a9-b66b-ca3a1802f56a-kube-api-access-744dv\") pod \"openstack-operator-index-fk8v7\" (UID: \"ba6e8e36-4a00-49a9-b66b-ca3a1802f56a\") " pod="openstack-operators/openstack-operator-index-fk8v7" Dec 10 23:03:45 crc kubenswrapper[4791]: I1210 23:03:45.609866 4791 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/openstack-operator-index-fk8v7" Dec 10 23:03:45 crc kubenswrapper[4791]: I1210 23:03:45.864194 4791 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/openstack-operator-index-fk8v7"] Dec 10 23:03:45 crc kubenswrapper[4791]: W1210 23:03:45.873322 4791 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podba6e8e36_4a00_49a9_b66b_ca3a1802f56a.slice/crio-27ea0162278ccdb4cf1baab434b0e4b227447abc31dec6ad6dd010544f441712 WatchSource:0}: Error finding container 27ea0162278ccdb4cf1baab434b0e4b227447abc31dec6ad6dd010544f441712: Status 404 returned error can't find the container with id 27ea0162278ccdb4cf1baab434b0e4b227447abc31dec6ad6dd010544f441712 Dec 10 23:03:45 crc kubenswrapper[4791]: I1210 23:03:45.925441 4791 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/openstack-operator-index-fk8v7" event={"ID":"ba6e8e36-4a00-49a9-b66b-ca3a1802f56a","Type":"ContainerStarted","Data":"27ea0162278ccdb4cf1baab434b0e4b227447abc31dec6ad6dd010544f441712"} Dec 10 23:03:45 crc kubenswrapper[4791]: I1210 23:03:45.929217 4791 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/frr-k8s-gqlpm" event={"ID":"a68714bc-7787-46c8-ba8f-21f7af6d9d67","Type":"ContainerStarted","Data":"663612343fcd0a6c871a0b10c139471f006a94e1d275440a8c6008eb0e0fda13"} Dec 10 23:03:45 crc kubenswrapper[4791]: I1210 23:03:45.929261 4791 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/frr-k8s-gqlpm" event={"ID":"a68714bc-7787-46c8-ba8f-21f7af6d9d67","Type":"ContainerStarted","Data":"197dade81d1ab3c8b45625cf7bedb23f6a51d862471ed210293f53733c5ed9f0"} Dec 10 23:03:46 crc kubenswrapper[4791]: I1210 23:03:46.959399 4791 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/frr-k8s-gqlpm" event={"ID":"a68714bc-7787-46c8-ba8f-21f7af6d9d67","Type":"ContainerStarted","Data":"a443e85261a7ac1ca0c29bba48aa720a8da0ab0f5629ca74459648a7aa7e6ae9"} Dec 10 23:03:46 crc kubenswrapper[4791]: I1210 23:03:46.959921 4791 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/frr-k8s-gqlpm" event={"ID":"a68714bc-7787-46c8-ba8f-21f7af6d9d67","Type":"ContainerStarted","Data":"131643c839a9f31d55dcf3c61ba43942aa9e0c35157a6947c39428418638ca53"} Dec 10 23:03:46 crc kubenswrapper[4791]: I1210 23:03:46.959932 4791 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/frr-k8s-gqlpm" event={"ID":"a68714bc-7787-46c8-ba8f-21f7af6d9d67","Type":"ContainerStarted","Data":"fb95a3f501e5925970b4a65f9c774c18655a4e80a7cbdaa83af6c4d8ca0e06b8"} Dec 10 23:03:47 crc kubenswrapper[4791]: I1210 23:03:47.972979 4791 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/frr-k8s-gqlpm" event={"ID":"a68714bc-7787-46c8-ba8f-21f7af6d9d67","Type":"ContainerStarted","Data":"58b97ec5b9e567e1380425d7e10b656d58b414f746b297e5e6a65e9c85949b87"} Dec 10 23:03:47 crc kubenswrapper[4791]: I1210 23:03:47.973377 4791 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="metallb-system/frr-k8s-gqlpm" Dec 10 23:03:48 crc kubenswrapper[4791]: I1210 23:03:48.019882 4791 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="metallb-system/frr-k8s-gqlpm" podStartSLOduration=6.601796551 podStartE2EDuration="18.019863038s" podCreationTimestamp="2025-12-10 23:03:30 +0000 UTC" firstStartedPulling="2025-12-10 23:03:30.987936835 +0000 UTC m=+845.417554448" lastFinishedPulling="2025-12-10 23:03:42.406003312 +0000 UTC m=+856.835620935" observedRunningTime="2025-12-10 23:03:48.013460863 +0000 UTC m=+862.443078486" watchObservedRunningTime="2025-12-10 23:03:48.019863038 +0000 UTC m=+862.449480651" Dec 10 23:03:49 crc kubenswrapper[4791]: I1210 23:03:49.251931 4791 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack-operators/openstack-operator-index-fk8v7"] Dec 10 23:03:50 crc kubenswrapper[4791]: I1210 23:03:50.061048 4791 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/openstack-operator-index-zg7wn"] Dec 10 23:03:50 crc kubenswrapper[4791]: I1210 23:03:50.061940 4791 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/openstack-operator-index-zg7wn" Dec 10 23:03:50 crc kubenswrapper[4791]: I1210 23:03:50.081193 4791 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/openstack-operator-index-zg7wn"] Dec 10 23:03:50 crc kubenswrapper[4791]: I1210 23:03:50.106539 4791 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-k4sjb\" (UniqueName: \"kubernetes.io/projected/f96dcde6-8c38-41a4-b1bd-b1c3007ff451-kube-api-access-k4sjb\") pod \"openstack-operator-index-zg7wn\" (UID: \"f96dcde6-8c38-41a4-b1bd-b1c3007ff451\") " pod="openstack-operators/openstack-operator-index-zg7wn" Dec 10 23:03:50 crc kubenswrapper[4791]: I1210 23:03:50.207785 4791 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-k4sjb\" (UniqueName: \"kubernetes.io/projected/f96dcde6-8c38-41a4-b1bd-b1c3007ff451-kube-api-access-k4sjb\") pod \"openstack-operator-index-zg7wn\" (UID: \"f96dcde6-8c38-41a4-b1bd-b1c3007ff451\") " pod="openstack-operators/openstack-operator-index-zg7wn" Dec 10 23:03:50 crc kubenswrapper[4791]: I1210 23:03:50.234934 4791 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-k4sjb\" (UniqueName: \"kubernetes.io/projected/f96dcde6-8c38-41a4-b1bd-b1c3007ff451-kube-api-access-k4sjb\") pod \"openstack-operator-index-zg7wn\" (UID: \"f96dcde6-8c38-41a4-b1bd-b1c3007ff451\") " pod="openstack-operators/openstack-operator-index-zg7wn" Dec 10 23:03:50 crc kubenswrapper[4791]: I1210 23:03:50.415144 4791 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/openstack-operator-index-zg7wn" Dec 10 23:03:50 crc kubenswrapper[4791]: I1210 23:03:50.472518 4791 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/community-operators-8s8cs"] Dec 10 23:03:50 crc kubenswrapper[4791]: I1210 23:03:50.480099 4791 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-8s8cs" Dec 10 23:03:50 crc kubenswrapper[4791]: I1210 23:03:50.495363 4791 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-8s8cs"] Dec 10 23:03:50 crc kubenswrapper[4791]: I1210 23:03:50.612139 4791 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-g7w9c\" (UniqueName: \"kubernetes.io/projected/08b6de5f-b1b6-4fdc-90c2-422a913bd692-kube-api-access-g7w9c\") pod \"community-operators-8s8cs\" (UID: \"08b6de5f-b1b6-4fdc-90c2-422a913bd692\") " pod="openshift-marketplace/community-operators-8s8cs" Dec 10 23:03:50 crc kubenswrapper[4791]: I1210 23:03:50.612652 4791 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/08b6de5f-b1b6-4fdc-90c2-422a913bd692-catalog-content\") pod \"community-operators-8s8cs\" (UID: \"08b6de5f-b1b6-4fdc-90c2-422a913bd692\") " pod="openshift-marketplace/community-operators-8s8cs" Dec 10 23:03:50 crc kubenswrapper[4791]: I1210 23:03:50.612686 4791 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/08b6de5f-b1b6-4fdc-90c2-422a913bd692-utilities\") pod \"community-operators-8s8cs\" (UID: \"08b6de5f-b1b6-4fdc-90c2-422a913bd692\") " pod="openshift-marketplace/community-operators-8s8cs" Dec 10 23:03:50 crc kubenswrapper[4791]: I1210 23:03:50.714641 4791 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/08b6de5f-b1b6-4fdc-90c2-422a913bd692-catalog-content\") pod \"community-operators-8s8cs\" (UID: \"08b6de5f-b1b6-4fdc-90c2-422a913bd692\") " pod="openshift-marketplace/community-operators-8s8cs" Dec 10 23:03:50 crc kubenswrapper[4791]: I1210 23:03:50.714911 4791 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/08b6de5f-b1b6-4fdc-90c2-422a913bd692-utilities\") pod \"community-operators-8s8cs\" (UID: \"08b6de5f-b1b6-4fdc-90c2-422a913bd692\") " pod="openshift-marketplace/community-operators-8s8cs" Dec 10 23:03:50 crc kubenswrapper[4791]: I1210 23:03:50.715024 4791 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-g7w9c\" (UniqueName: \"kubernetes.io/projected/08b6de5f-b1b6-4fdc-90c2-422a913bd692-kube-api-access-g7w9c\") pod \"community-operators-8s8cs\" (UID: \"08b6de5f-b1b6-4fdc-90c2-422a913bd692\") " pod="openshift-marketplace/community-operators-8s8cs" Dec 10 23:03:50 crc kubenswrapper[4791]: I1210 23:03:50.715729 4791 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/08b6de5f-b1b6-4fdc-90c2-422a913bd692-catalog-content\") pod \"community-operators-8s8cs\" (UID: \"08b6de5f-b1b6-4fdc-90c2-422a913bd692\") " pod="openshift-marketplace/community-operators-8s8cs" Dec 10 23:03:50 crc kubenswrapper[4791]: I1210 23:03:50.715740 4791 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/08b6de5f-b1b6-4fdc-90c2-422a913bd692-utilities\") pod \"community-operators-8s8cs\" (UID: \"08b6de5f-b1b6-4fdc-90c2-422a913bd692\") " pod="openshift-marketplace/community-operators-8s8cs" Dec 10 23:03:50 crc kubenswrapper[4791]: I1210 23:03:50.735235 4791 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-g7w9c\" (UniqueName: \"kubernetes.io/projected/08b6de5f-b1b6-4fdc-90c2-422a913bd692-kube-api-access-g7w9c\") pod \"community-operators-8s8cs\" (UID: \"08b6de5f-b1b6-4fdc-90c2-422a913bd692\") " pod="openshift-marketplace/community-operators-8s8cs" Dec 10 23:03:50 crc kubenswrapper[4791]: I1210 23:03:50.820301 4791 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-8s8cs" Dec 10 23:03:50 crc kubenswrapper[4791]: I1210 23:03:50.855329 4791 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="metallb-system/frr-k8s-gqlpm" Dec 10 23:03:50 crc kubenswrapper[4791]: I1210 23:03:50.913116 4791 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/openstack-operator-index-zg7wn"] Dec 10 23:03:50 crc kubenswrapper[4791]: I1210 23:03:50.942907 4791 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="metallb-system/frr-k8s-gqlpm" Dec 10 23:03:50 crc kubenswrapper[4791]: I1210 23:03:50.992951 4791 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/openstack-operator-index-fk8v7" event={"ID":"ba6e8e36-4a00-49a9-b66b-ca3a1802f56a","Type":"ContainerStarted","Data":"fbeb722c72a83e246bf0fc0b8101530a96d848905627d4b261b4f9ae3b3504ad"} Dec 10 23:03:50 crc kubenswrapper[4791]: I1210 23:03:50.993089 4791 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack-operators/openstack-operator-index-fk8v7" podUID="ba6e8e36-4a00-49a9-b66b-ca3a1802f56a" containerName="registry-server" containerID="cri-o://fbeb722c72a83e246bf0fc0b8101530a96d848905627d4b261b4f9ae3b3504ad" gracePeriod=2 Dec 10 23:03:50 crc kubenswrapper[4791]: I1210 23:03:50.994390 4791 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/openstack-operator-index-zg7wn" event={"ID":"f96dcde6-8c38-41a4-b1bd-b1c3007ff451","Type":"ContainerStarted","Data":"f6e7e90557b717ba0b6251ebd4af845aeecc7c134afbaa164da41ad7c95aa66b"} Dec 10 23:03:51 crc kubenswrapper[4791]: I1210 23:03:51.010681 4791 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/openstack-operator-index-fk8v7" podStartSLOduration=2.919099648 podStartE2EDuration="6.010659688s" podCreationTimestamp="2025-12-10 23:03:45 +0000 UTC" firstStartedPulling="2025-12-10 23:03:45.87657058 +0000 UTC m=+860.306188193" lastFinishedPulling="2025-12-10 23:03:48.96813062 +0000 UTC m=+863.397748233" observedRunningTime="2025-12-10 23:03:51.009961268 +0000 UTC m=+865.439578881" watchObservedRunningTime="2025-12-10 23:03:51.010659688 +0000 UTC m=+865.440277291" Dec 10 23:03:51 crc kubenswrapper[4791]: I1210 23:03:51.115092 4791 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-8s8cs"] Dec 10 23:03:51 crc kubenswrapper[4791]: I1210 23:03:51.386453 4791 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack-operators/openstack-operator-index-fk8v7" Dec 10 23:03:51 crc kubenswrapper[4791]: I1210 23:03:51.548003 4791 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-744dv\" (UniqueName: \"kubernetes.io/projected/ba6e8e36-4a00-49a9-b66b-ca3a1802f56a-kube-api-access-744dv\") pod \"ba6e8e36-4a00-49a9-b66b-ca3a1802f56a\" (UID: \"ba6e8e36-4a00-49a9-b66b-ca3a1802f56a\") " Dec 10 23:03:51 crc kubenswrapper[4791]: I1210 23:03:51.556881 4791 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/ba6e8e36-4a00-49a9-b66b-ca3a1802f56a-kube-api-access-744dv" (OuterVolumeSpecName: "kube-api-access-744dv") pod "ba6e8e36-4a00-49a9-b66b-ca3a1802f56a" (UID: "ba6e8e36-4a00-49a9-b66b-ca3a1802f56a"). InnerVolumeSpecName "kube-api-access-744dv". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 10 23:03:51 crc kubenswrapper[4791]: I1210 23:03:51.650456 4791 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-744dv\" (UniqueName: \"kubernetes.io/projected/ba6e8e36-4a00-49a9-b66b-ca3a1802f56a-kube-api-access-744dv\") on node \"crc\" DevicePath \"\"" Dec 10 23:03:52 crc kubenswrapper[4791]: I1210 23:03:52.004462 4791 generic.go:334] "Generic (PLEG): container finished" podID="08b6de5f-b1b6-4fdc-90c2-422a913bd692" containerID="5cdaa81f058684afa6f6754ac22ce77f292479dec9d91eabebf5d99f53186c71" exitCode=0 Dec 10 23:03:52 crc kubenswrapper[4791]: I1210 23:03:52.004573 4791 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-8s8cs" event={"ID":"08b6de5f-b1b6-4fdc-90c2-422a913bd692","Type":"ContainerDied","Data":"5cdaa81f058684afa6f6754ac22ce77f292479dec9d91eabebf5d99f53186c71"} Dec 10 23:03:52 crc kubenswrapper[4791]: I1210 23:03:52.004649 4791 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-8s8cs" event={"ID":"08b6de5f-b1b6-4fdc-90c2-422a913bd692","Type":"ContainerStarted","Data":"e3d719bfc2fabd865199479bb723fdd8413e7c5a1f9d98e2a6a32da259dfcc21"} Dec 10 23:03:52 crc kubenswrapper[4791]: I1210 23:03:52.009296 4791 generic.go:334] "Generic (PLEG): container finished" podID="ba6e8e36-4a00-49a9-b66b-ca3a1802f56a" containerID="fbeb722c72a83e246bf0fc0b8101530a96d848905627d4b261b4f9ae3b3504ad" exitCode=0 Dec 10 23:03:52 crc kubenswrapper[4791]: I1210 23:03:52.009368 4791 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/openstack-operator-index-fk8v7" event={"ID":"ba6e8e36-4a00-49a9-b66b-ca3a1802f56a","Type":"ContainerDied","Data":"fbeb722c72a83e246bf0fc0b8101530a96d848905627d4b261b4f9ae3b3504ad"} Dec 10 23:03:52 crc kubenswrapper[4791]: I1210 23:03:52.009440 4791 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/openstack-operator-index-fk8v7" event={"ID":"ba6e8e36-4a00-49a9-b66b-ca3a1802f56a","Type":"ContainerDied","Data":"27ea0162278ccdb4cf1baab434b0e4b227447abc31dec6ad6dd010544f441712"} Dec 10 23:03:52 crc kubenswrapper[4791]: I1210 23:03:52.009459 4791 scope.go:117] "RemoveContainer" containerID="fbeb722c72a83e246bf0fc0b8101530a96d848905627d4b261b4f9ae3b3504ad" Dec 10 23:03:52 crc kubenswrapper[4791]: I1210 23:03:52.009417 4791 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack-operators/openstack-operator-index-fk8v7" Dec 10 23:03:52 crc kubenswrapper[4791]: I1210 23:03:52.011535 4791 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/openstack-operator-index-zg7wn" event={"ID":"f96dcde6-8c38-41a4-b1bd-b1c3007ff451","Type":"ContainerStarted","Data":"df877846eeb2e80cda08de0fc6acb52c1b79b943dd6e6a1f2f86201891f2c3ae"} Dec 10 23:03:52 crc kubenswrapper[4791]: I1210 23:03:52.028124 4791 scope.go:117] "RemoveContainer" containerID="fbeb722c72a83e246bf0fc0b8101530a96d848905627d4b261b4f9ae3b3504ad" Dec 10 23:03:52 crc kubenswrapper[4791]: E1210 23:03:52.028705 4791 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"fbeb722c72a83e246bf0fc0b8101530a96d848905627d4b261b4f9ae3b3504ad\": container with ID starting with fbeb722c72a83e246bf0fc0b8101530a96d848905627d4b261b4f9ae3b3504ad not found: ID does not exist" containerID="fbeb722c72a83e246bf0fc0b8101530a96d848905627d4b261b4f9ae3b3504ad" Dec 10 23:03:52 crc kubenswrapper[4791]: I1210 23:03:52.028751 4791 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"fbeb722c72a83e246bf0fc0b8101530a96d848905627d4b261b4f9ae3b3504ad"} err="failed to get container status \"fbeb722c72a83e246bf0fc0b8101530a96d848905627d4b261b4f9ae3b3504ad\": rpc error: code = NotFound desc = could not find container \"fbeb722c72a83e246bf0fc0b8101530a96d848905627d4b261b4f9ae3b3504ad\": container with ID starting with fbeb722c72a83e246bf0fc0b8101530a96d848905627d4b261b4f9ae3b3504ad not found: ID does not exist" Dec 10 23:03:52 crc kubenswrapper[4791]: I1210 23:03:52.035878 4791 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack-operators/openstack-operator-index-fk8v7"] Dec 10 23:03:52 crc kubenswrapper[4791]: I1210 23:03:52.043865 4791 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack-operators/openstack-operator-index-fk8v7"] Dec 10 23:03:52 crc kubenswrapper[4791]: I1210 23:03:52.049821 4791 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/openstack-operator-index-zg7wn" podStartSLOduration=1.959940808 podStartE2EDuration="2.049801224s" podCreationTimestamp="2025-12-10 23:03:50 +0000 UTC" firstStartedPulling="2025-12-10 23:03:50.96121801 +0000 UTC m=+865.390835633" lastFinishedPulling="2025-12-10 23:03:51.051078436 +0000 UTC m=+865.480696049" observedRunningTime="2025-12-10 23:03:52.048144596 +0000 UTC m=+866.477762209" watchObservedRunningTime="2025-12-10 23:03:52.049801224 +0000 UTC m=+866.479418837" Dec 10 23:03:53 crc kubenswrapper[4791]: I1210 23:03:53.897769 4791 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="ba6e8e36-4a00-49a9-b66b-ca3a1802f56a" path="/var/lib/kubelet/pods/ba6e8e36-4a00-49a9-b66b-ca3a1802f56a/volumes" Dec 10 23:03:54 crc kubenswrapper[4791]: I1210 23:03:54.032797 4791 generic.go:334] "Generic (PLEG): container finished" podID="08b6de5f-b1b6-4fdc-90c2-422a913bd692" containerID="b91feecbb4259e2b6ecd1d2216e6801fa16b60d7d0c8bd7df6eb9f5d6ae6cfff" exitCode=0 Dec 10 23:03:54 crc kubenswrapper[4791]: I1210 23:03:54.032862 4791 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-8s8cs" event={"ID":"08b6de5f-b1b6-4fdc-90c2-422a913bd692","Type":"ContainerDied","Data":"b91feecbb4259e2b6ecd1d2216e6801fa16b60d7d0c8bd7df6eb9f5d6ae6cfff"} Dec 10 23:03:57 crc kubenswrapper[4791]: I1210 23:03:57.055190 4791 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-8s8cs" event={"ID":"08b6de5f-b1b6-4fdc-90c2-422a913bd692","Type":"ContainerStarted","Data":"89ce3c3679b6aaa0760a5b2fce4fcb9658c096c0518af5f4c656cae6344335db"} Dec 10 23:03:57 crc kubenswrapper[4791]: I1210 23:03:57.079602 4791 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/community-operators-8s8cs" podStartSLOduration=2.703581449 podStartE2EDuration="7.079586001s" podCreationTimestamp="2025-12-10 23:03:50 +0000 UTC" firstStartedPulling="2025-12-10 23:03:52.007576964 +0000 UTC m=+866.437194577" lastFinishedPulling="2025-12-10 23:03:56.383581516 +0000 UTC m=+870.813199129" observedRunningTime="2025-12-10 23:03:57.077729687 +0000 UTC m=+871.507347300" watchObservedRunningTime="2025-12-10 23:03:57.079586001 +0000 UTC m=+871.509203614" Dec 10 23:04:00 crc kubenswrapper[4791]: I1210 23:04:00.415647 4791 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/openstack-operator-index-zg7wn" Dec 10 23:04:00 crc kubenswrapper[4791]: I1210 23:04:00.416039 4791 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack-operators/openstack-operator-index-zg7wn" Dec 10 23:04:00 crc kubenswrapper[4791]: I1210 23:04:00.459896 4791 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack-operators/openstack-operator-index-zg7wn" Dec 10 23:04:00 crc kubenswrapper[4791]: I1210 23:04:00.820872 4791 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/community-operators-8s8cs" Dec 10 23:04:00 crc kubenswrapper[4791]: I1210 23:04:00.821137 4791 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/community-operators-8s8cs" Dec 10 23:04:00 crc kubenswrapper[4791]: I1210 23:04:00.853283 4791 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="metallb-system/frr-k8s-webhook-server-7fcb986d4-whw8m" Dec 10 23:04:00 crc kubenswrapper[4791]: I1210 23:04:00.857107 4791 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="metallb-system/frr-k8s-gqlpm" Dec 10 23:04:00 crc kubenswrapper[4791]: I1210 23:04:00.879058 4791 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/community-operators-8s8cs" Dec 10 23:04:01 crc kubenswrapper[4791]: I1210 23:04:01.110668 4791 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/openstack-operator-index-zg7wn" Dec 10 23:04:01 crc kubenswrapper[4791]: I1210 23:04:01.150988 4791 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/community-operators-8s8cs" Dec 10 23:04:01 crc kubenswrapper[4791]: I1210 23:04:01.855241 4791 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-8s8cs"] Dec 10 23:04:03 crc kubenswrapper[4791]: I1210 23:04:03.091725 4791 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/community-operators-8s8cs" podUID="08b6de5f-b1b6-4fdc-90c2-422a913bd692" containerName="registry-server" containerID="cri-o://89ce3c3679b6aaa0760a5b2fce4fcb9658c096c0518af5f4c656cae6344335db" gracePeriod=2 Dec 10 23:04:04 crc kubenswrapper[4791]: I1210 23:04:04.100152 4791 generic.go:334] "Generic (PLEG): container finished" podID="08b6de5f-b1b6-4fdc-90c2-422a913bd692" containerID="89ce3c3679b6aaa0760a5b2fce4fcb9658c096c0518af5f4c656cae6344335db" exitCode=0 Dec 10 23:04:04 crc kubenswrapper[4791]: I1210 23:04:04.100230 4791 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-8s8cs" event={"ID":"08b6de5f-b1b6-4fdc-90c2-422a913bd692","Type":"ContainerDied","Data":"89ce3c3679b6aaa0760a5b2fce4fcb9658c096c0518af5f4c656cae6344335db"} Dec 10 23:04:04 crc kubenswrapper[4791]: I1210 23:04:04.272084 4791 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-8s8cs" Dec 10 23:04:04 crc kubenswrapper[4791]: I1210 23:04:04.328969 4791 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/08b6de5f-b1b6-4fdc-90c2-422a913bd692-utilities\") pod \"08b6de5f-b1b6-4fdc-90c2-422a913bd692\" (UID: \"08b6de5f-b1b6-4fdc-90c2-422a913bd692\") " Dec 10 23:04:04 crc kubenswrapper[4791]: I1210 23:04:04.329050 4791 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/08b6de5f-b1b6-4fdc-90c2-422a913bd692-catalog-content\") pod \"08b6de5f-b1b6-4fdc-90c2-422a913bd692\" (UID: \"08b6de5f-b1b6-4fdc-90c2-422a913bd692\") " Dec 10 23:04:04 crc kubenswrapper[4791]: I1210 23:04:04.329175 4791 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-g7w9c\" (UniqueName: \"kubernetes.io/projected/08b6de5f-b1b6-4fdc-90c2-422a913bd692-kube-api-access-g7w9c\") pod \"08b6de5f-b1b6-4fdc-90c2-422a913bd692\" (UID: \"08b6de5f-b1b6-4fdc-90c2-422a913bd692\") " Dec 10 23:04:04 crc kubenswrapper[4791]: I1210 23:04:04.330617 4791 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/08b6de5f-b1b6-4fdc-90c2-422a913bd692-utilities" (OuterVolumeSpecName: "utilities") pod "08b6de5f-b1b6-4fdc-90c2-422a913bd692" (UID: "08b6de5f-b1b6-4fdc-90c2-422a913bd692"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 10 23:04:04 crc kubenswrapper[4791]: I1210 23:04:04.341896 4791 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/08b6de5f-b1b6-4fdc-90c2-422a913bd692-kube-api-access-g7w9c" (OuterVolumeSpecName: "kube-api-access-g7w9c") pod "08b6de5f-b1b6-4fdc-90c2-422a913bd692" (UID: "08b6de5f-b1b6-4fdc-90c2-422a913bd692"). InnerVolumeSpecName "kube-api-access-g7w9c". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 10 23:04:04 crc kubenswrapper[4791]: I1210 23:04:04.387395 4791 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/08b6de5f-b1b6-4fdc-90c2-422a913bd692-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "08b6de5f-b1b6-4fdc-90c2-422a913bd692" (UID: "08b6de5f-b1b6-4fdc-90c2-422a913bd692"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 10 23:04:04 crc kubenswrapper[4791]: I1210 23:04:04.430678 4791 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-g7w9c\" (UniqueName: \"kubernetes.io/projected/08b6de5f-b1b6-4fdc-90c2-422a913bd692-kube-api-access-g7w9c\") on node \"crc\" DevicePath \"\"" Dec 10 23:04:04 crc kubenswrapper[4791]: I1210 23:04:04.430743 4791 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/08b6de5f-b1b6-4fdc-90c2-422a913bd692-utilities\") on node \"crc\" DevicePath \"\"" Dec 10 23:04:04 crc kubenswrapper[4791]: I1210 23:04:04.430765 4791 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/08b6de5f-b1b6-4fdc-90c2-422a913bd692-catalog-content\") on node \"crc\" DevicePath \"\"" Dec 10 23:04:05 crc kubenswrapper[4791]: I1210 23:04:05.110410 4791 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-8s8cs" event={"ID":"08b6de5f-b1b6-4fdc-90c2-422a913bd692","Type":"ContainerDied","Data":"e3d719bfc2fabd865199479bb723fdd8413e7c5a1f9d98e2a6a32da259dfcc21"} Dec 10 23:04:05 crc kubenswrapper[4791]: I1210 23:04:05.110489 4791 scope.go:117] "RemoveContainer" containerID="89ce3c3679b6aaa0760a5b2fce4fcb9658c096c0518af5f4c656cae6344335db" Dec 10 23:04:05 crc kubenswrapper[4791]: I1210 23:04:05.110509 4791 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-8s8cs" Dec 10 23:04:05 crc kubenswrapper[4791]: I1210 23:04:05.132837 4791 scope.go:117] "RemoveContainer" containerID="b91feecbb4259e2b6ecd1d2216e6801fa16b60d7d0c8bd7df6eb9f5d6ae6cfff" Dec 10 23:04:05 crc kubenswrapper[4791]: I1210 23:04:05.155299 4791 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-8s8cs"] Dec 10 23:04:05 crc kubenswrapper[4791]: I1210 23:04:05.159674 4791 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/community-operators-8s8cs"] Dec 10 23:04:05 crc kubenswrapper[4791]: I1210 23:04:05.161870 4791 scope.go:117] "RemoveContainer" containerID="5cdaa81f058684afa6f6754ac22ce77f292479dec9d91eabebf5d99f53186c71" Dec 10 23:04:05 crc kubenswrapper[4791]: I1210 23:04:05.894423 4791 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="08b6de5f-b1b6-4fdc-90c2-422a913bd692" path="/var/lib/kubelet/pods/08b6de5f-b1b6-4fdc-90c2-422a913bd692/volumes" Dec 10 23:04:07 crc kubenswrapper[4791]: I1210 23:04:07.320301 4791 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/50d664eb6e76f12a148f821b8e085abb03568bf5cb6cd29276e1382c05lclqs"] Dec 10 23:04:07 crc kubenswrapper[4791]: E1210 23:04:07.320895 4791 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="ba6e8e36-4a00-49a9-b66b-ca3a1802f56a" containerName="registry-server" Dec 10 23:04:07 crc kubenswrapper[4791]: I1210 23:04:07.320914 4791 state_mem.go:107] "Deleted CPUSet assignment" podUID="ba6e8e36-4a00-49a9-b66b-ca3a1802f56a" containerName="registry-server" Dec 10 23:04:07 crc kubenswrapper[4791]: E1210 23:04:07.320942 4791 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="08b6de5f-b1b6-4fdc-90c2-422a913bd692" containerName="extract-utilities" Dec 10 23:04:07 crc kubenswrapper[4791]: I1210 23:04:07.320950 4791 state_mem.go:107] "Deleted CPUSet assignment" podUID="08b6de5f-b1b6-4fdc-90c2-422a913bd692" containerName="extract-utilities" Dec 10 23:04:07 crc kubenswrapper[4791]: E1210 23:04:07.320962 4791 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="08b6de5f-b1b6-4fdc-90c2-422a913bd692" containerName="registry-server" Dec 10 23:04:07 crc kubenswrapper[4791]: I1210 23:04:07.320969 4791 state_mem.go:107] "Deleted CPUSet assignment" podUID="08b6de5f-b1b6-4fdc-90c2-422a913bd692" containerName="registry-server" Dec 10 23:04:07 crc kubenswrapper[4791]: E1210 23:04:07.320980 4791 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="08b6de5f-b1b6-4fdc-90c2-422a913bd692" containerName="extract-content" Dec 10 23:04:07 crc kubenswrapper[4791]: I1210 23:04:07.320987 4791 state_mem.go:107] "Deleted CPUSet assignment" podUID="08b6de5f-b1b6-4fdc-90c2-422a913bd692" containerName="extract-content" Dec 10 23:04:07 crc kubenswrapper[4791]: I1210 23:04:07.321109 4791 memory_manager.go:354] "RemoveStaleState removing state" podUID="08b6de5f-b1b6-4fdc-90c2-422a913bd692" containerName="registry-server" Dec 10 23:04:07 crc kubenswrapper[4791]: I1210 23:04:07.321128 4791 memory_manager.go:354] "RemoveStaleState removing state" podUID="ba6e8e36-4a00-49a9-b66b-ca3a1802f56a" containerName="registry-server" Dec 10 23:04:07 crc kubenswrapper[4791]: I1210 23:04:07.322078 4791 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/50d664eb6e76f12a148f821b8e085abb03568bf5cb6cd29276e1382c05lclqs" Dec 10 23:04:07 crc kubenswrapper[4791]: I1210 23:04:07.324046 4791 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"default-dockercfg-n44zc" Dec 10 23:04:07 crc kubenswrapper[4791]: I1210 23:04:07.346778 4791 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/50d664eb6e76f12a148f821b8e085abb03568bf5cb6cd29276e1382c05lclqs"] Dec 10 23:04:07 crc kubenswrapper[4791]: I1210 23:04:07.372411 4791 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/c76a21e1-575e-42e2-8372-e79986d26dbb-util\") pod \"50d664eb6e76f12a148f821b8e085abb03568bf5cb6cd29276e1382c05lclqs\" (UID: \"c76a21e1-575e-42e2-8372-e79986d26dbb\") " pod="openstack-operators/50d664eb6e76f12a148f821b8e085abb03568bf5cb6cd29276e1382c05lclqs" Dec 10 23:04:07 crc kubenswrapper[4791]: I1210 23:04:07.372923 4791 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/c76a21e1-575e-42e2-8372-e79986d26dbb-bundle\") pod \"50d664eb6e76f12a148f821b8e085abb03568bf5cb6cd29276e1382c05lclqs\" (UID: \"c76a21e1-575e-42e2-8372-e79986d26dbb\") " pod="openstack-operators/50d664eb6e76f12a148f821b8e085abb03568bf5cb6cd29276e1382c05lclqs" Dec 10 23:04:07 crc kubenswrapper[4791]: I1210 23:04:07.373133 4791 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-964mm\" (UniqueName: \"kubernetes.io/projected/c76a21e1-575e-42e2-8372-e79986d26dbb-kube-api-access-964mm\") pod \"50d664eb6e76f12a148f821b8e085abb03568bf5cb6cd29276e1382c05lclqs\" (UID: \"c76a21e1-575e-42e2-8372-e79986d26dbb\") " pod="openstack-operators/50d664eb6e76f12a148f821b8e085abb03568bf5cb6cd29276e1382c05lclqs" Dec 10 23:04:07 crc kubenswrapper[4791]: I1210 23:04:07.475010 4791 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/c76a21e1-575e-42e2-8372-e79986d26dbb-util\") pod \"50d664eb6e76f12a148f821b8e085abb03568bf5cb6cd29276e1382c05lclqs\" (UID: \"c76a21e1-575e-42e2-8372-e79986d26dbb\") " pod="openstack-operators/50d664eb6e76f12a148f821b8e085abb03568bf5cb6cd29276e1382c05lclqs" Dec 10 23:04:07 crc kubenswrapper[4791]: I1210 23:04:07.475515 4791 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/c76a21e1-575e-42e2-8372-e79986d26dbb-bundle\") pod \"50d664eb6e76f12a148f821b8e085abb03568bf5cb6cd29276e1382c05lclqs\" (UID: \"c76a21e1-575e-42e2-8372-e79986d26dbb\") " pod="openstack-operators/50d664eb6e76f12a148f821b8e085abb03568bf5cb6cd29276e1382c05lclqs" Dec 10 23:04:07 crc kubenswrapper[4791]: I1210 23:04:07.475694 4791 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-964mm\" (UniqueName: \"kubernetes.io/projected/c76a21e1-575e-42e2-8372-e79986d26dbb-kube-api-access-964mm\") pod \"50d664eb6e76f12a148f821b8e085abb03568bf5cb6cd29276e1382c05lclqs\" (UID: \"c76a21e1-575e-42e2-8372-e79986d26dbb\") " pod="openstack-operators/50d664eb6e76f12a148f821b8e085abb03568bf5cb6cd29276e1382c05lclqs" Dec 10 23:04:07 crc kubenswrapper[4791]: I1210 23:04:07.475912 4791 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/c76a21e1-575e-42e2-8372-e79986d26dbb-bundle\") pod \"50d664eb6e76f12a148f821b8e085abb03568bf5cb6cd29276e1382c05lclqs\" (UID: \"c76a21e1-575e-42e2-8372-e79986d26dbb\") " pod="openstack-operators/50d664eb6e76f12a148f821b8e085abb03568bf5cb6cd29276e1382c05lclqs" Dec 10 23:04:07 crc kubenswrapper[4791]: I1210 23:04:07.476026 4791 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/c76a21e1-575e-42e2-8372-e79986d26dbb-util\") pod \"50d664eb6e76f12a148f821b8e085abb03568bf5cb6cd29276e1382c05lclqs\" (UID: \"c76a21e1-575e-42e2-8372-e79986d26dbb\") " pod="openstack-operators/50d664eb6e76f12a148f821b8e085abb03568bf5cb6cd29276e1382c05lclqs" Dec 10 23:04:07 crc kubenswrapper[4791]: I1210 23:04:07.495827 4791 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-964mm\" (UniqueName: \"kubernetes.io/projected/c76a21e1-575e-42e2-8372-e79986d26dbb-kube-api-access-964mm\") pod \"50d664eb6e76f12a148f821b8e085abb03568bf5cb6cd29276e1382c05lclqs\" (UID: \"c76a21e1-575e-42e2-8372-e79986d26dbb\") " pod="openstack-operators/50d664eb6e76f12a148f821b8e085abb03568bf5cb6cd29276e1382c05lclqs" Dec 10 23:04:07 crc kubenswrapper[4791]: I1210 23:04:07.636629 4791 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/50d664eb6e76f12a148f821b8e085abb03568bf5cb6cd29276e1382c05lclqs" Dec 10 23:04:08 crc kubenswrapper[4791]: I1210 23:04:08.071095 4791 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/50d664eb6e76f12a148f821b8e085abb03568bf5cb6cd29276e1382c05lclqs"] Dec 10 23:04:08 crc kubenswrapper[4791]: W1210 23:04:08.075126 4791 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podc76a21e1_575e_42e2_8372_e79986d26dbb.slice/crio-c07a2befadab0fb7d3c1539ed4a8d23d6fa53291514908a3fd48ef115e9f99ff WatchSource:0}: Error finding container c07a2befadab0fb7d3c1539ed4a8d23d6fa53291514908a3fd48ef115e9f99ff: Status 404 returned error can't find the container with id c07a2befadab0fb7d3c1539ed4a8d23d6fa53291514908a3fd48ef115e9f99ff Dec 10 23:04:08 crc kubenswrapper[4791]: I1210 23:04:08.141000 4791 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/50d664eb6e76f12a148f821b8e085abb03568bf5cb6cd29276e1382c05lclqs" event={"ID":"c76a21e1-575e-42e2-8372-e79986d26dbb","Type":"ContainerStarted","Data":"c07a2befadab0fb7d3c1539ed4a8d23d6fa53291514908a3fd48ef115e9f99ff"} Dec 10 23:04:16 crc kubenswrapper[4791]: I1210 23:04:11.165109 4791 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/50d664eb6e76f12a148f821b8e085abb03568bf5cb6cd29276e1382c05lclqs" event={"ID":"c76a21e1-575e-42e2-8372-e79986d26dbb","Type":"ContainerStarted","Data":"b3831a59607083de94321c6f2fa2cb4b9652d7dc65671ebf302668f70970c5f4"} Dec 10 23:04:16 crc kubenswrapper[4791]: I1210 23:04:14.060668 4791 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-marketplace-8d275"] Dec 10 23:04:16 crc kubenswrapper[4791]: I1210 23:04:14.062361 4791 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-8d275" Dec 10 23:04:16 crc kubenswrapper[4791]: I1210 23:04:14.079655 4791 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-8d275"] Dec 10 23:04:16 crc kubenswrapper[4791]: I1210 23:04:14.171813 4791 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/2d2fb71e-03b5-49f1-989f-04abd05025ed-catalog-content\") pod \"redhat-marketplace-8d275\" (UID: \"2d2fb71e-03b5-49f1-989f-04abd05025ed\") " pod="openshift-marketplace/redhat-marketplace-8d275" Dec 10 23:04:16 crc kubenswrapper[4791]: I1210 23:04:14.172024 4791 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-rph6l\" (UniqueName: \"kubernetes.io/projected/2d2fb71e-03b5-49f1-989f-04abd05025ed-kube-api-access-rph6l\") pod \"redhat-marketplace-8d275\" (UID: \"2d2fb71e-03b5-49f1-989f-04abd05025ed\") " pod="openshift-marketplace/redhat-marketplace-8d275" Dec 10 23:04:16 crc kubenswrapper[4791]: I1210 23:04:14.172070 4791 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/2d2fb71e-03b5-49f1-989f-04abd05025ed-utilities\") pod \"redhat-marketplace-8d275\" (UID: \"2d2fb71e-03b5-49f1-989f-04abd05025ed\") " pod="openshift-marketplace/redhat-marketplace-8d275" Dec 10 23:04:16 crc kubenswrapper[4791]: I1210 23:04:14.273724 4791 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/2d2fb71e-03b5-49f1-989f-04abd05025ed-catalog-content\") pod \"redhat-marketplace-8d275\" (UID: \"2d2fb71e-03b5-49f1-989f-04abd05025ed\") " pod="openshift-marketplace/redhat-marketplace-8d275" Dec 10 23:04:16 crc kubenswrapper[4791]: I1210 23:04:14.273896 4791 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-rph6l\" (UniqueName: \"kubernetes.io/projected/2d2fb71e-03b5-49f1-989f-04abd05025ed-kube-api-access-rph6l\") pod \"redhat-marketplace-8d275\" (UID: \"2d2fb71e-03b5-49f1-989f-04abd05025ed\") " pod="openshift-marketplace/redhat-marketplace-8d275" Dec 10 23:04:16 crc kubenswrapper[4791]: I1210 23:04:14.273937 4791 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/2d2fb71e-03b5-49f1-989f-04abd05025ed-utilities\") pod \"redhat-marketplace-8d275\" (UID: \"2d2fb71e-03b5-49f1-989f-04abd05025ed\") " pod="openshift-marketplace/redhat-marketplace-8d275" Dec 10 23:04:16 crc kubenswrapper[4791]: I1210 23:04:14.274489 4791 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/2d2fb71e-03b5-49f1-989f-04abd05025ed-utilities\") pod \"redhat-marketplace-8d275\" (UID: \"2d2fb71e-03b5-49f1-989f-04abd05025ed\") " pod="openshift-marketplace/redhat-marketplace-8d275" Dec 10 23:04:16 crc kubenswrapper[4791]: I1210 23:04:14.275153 4791 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/2d2fb71e-03b5-49f1-989f-04abd05025ed-catalog-content\") pod \"redhat-marketplace-8d275\" (UID: \"2d2fb71e-03b5-49f1-989f-04abd05025ed\") " pod="openshift-marketplace/redhat-marketplace-8d275" Dec 10 23:04:16 crc kubenswrapper[4791]: I1210 23:04:14.303187 4791 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-rph6l\" (UniqueName: \"kubernetes.io/projected/2d2fb71e-03b5-49f1-989f-04abd05025ed-kube-api-access-rph6l\") pod \"redhat-marketplace-8d275\" (UID: \"2d2fb71e-03b5-49f1-989f-04abd05025ed\") " pod="openshift-marketplace/redhat-marketplace-8d275" Dec 10 23:04:16 crc kubenswrapper[4791]: I1210 23:04:14.376257 4791 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-8d275" Dec 10 23:04:16 crc kubenswrapper[4791]: I1210 23:04:16.199752 4791 generic.go:334] "Generic (PLEG): container finished" podID="c76a21e1-575e-42e2-8372-e79986d26dbb" containerID="b3831a59607083de94321c6f2fa2cb4b9652d7dc65671ebf302668f70970c5f4" exitCode=0 Dec 10 23:04:16 crc kubenswrapper[4791]: I1210 23:04:16.199844 4791 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/50d664eb6e76f12a148f821b8e085abb03568bf5cb6cd29276e1382c05lclqs" event={"ID":"c76a21e1-575e-42e2-8372-e79986d26dbb","Type":"ContainerDied","Data":"b3831a59607083de94321c6f2fa2cb4b9652d7dc65671ebf302668f70970c5f4"} Dec 10 23:04:16 crc kubenswrapper[4791]: I1210 23:04:16.866008 4791 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/certified-operators-dv5kn"] Dec 10 23:04:16 crc kubenswrapper[4791]: I1210 23:04:16.868702 4791 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-dv5kn" Dec 10 23:04:16 crc kubenswrapper[4791]: I1210 23:04:16.874141 4791 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-dv5kn"] Dec 10 23:04:17 crc kubenswrapper[4791]: I1210 23:04:17.010582 4791 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/385c4850-9b08-4531-af6a-4f2ec3662165-catalog-content\") pod \"certified-operators-dv5kn\" (UID: \"385c4850-9b08-4531-af6a-4f2ec3662165\") " pod="openshift-marketplace/certified-operators-dv5kn" Dec 10 23:04:17 crc kubenswrapper[4791]: I1210 23:04:17.010632 4791 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-fr6ld\" (UniqueName: \"kubernetes.io/projected/385c4850-9b08-4531-af6a-4f2ec3662165-kube-api-access-fr6ld\") pod \"certified-operators-dv5kn\" (UID: \"385c4850-9b08-4531-af6a-4f2ec3662165\") " pod="openshift-marketplace/certified-operators-dv5kn" Dec 10 23:04:17 crc kubenswrapper[4791]: I1210 23:04:17.010747 4791 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/385c4850-9b08-4531-af6a-4f2ec3662165-utilities\") pod \"certified-operators-dv5kn\" (UID: \"385c4850-9b08-4531-af6a-4f2ec3662165\") " pod="openshift-marketplace/certified-operators-dv5kn" Dec 10 23:04:17 crc kubenswrapper[4791]: I1210 23:04:17.112544 4791 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/385c4850-9b08-4531-af6a-4f2ec3662165-catalog-content\") pod \"certified-operators-dv5kn\" (UID: \"385c4850-9b08-4531-af6a-4f2ec3662165\") " pod="openshift-marketplace/certified-operators-dv5kn" Dec 10 23:04:17 crc kubenswrapper[4791]: I1210 23:04:17.112888 4791 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-fr6ld\" (UniqueName: \"kubernetes.io/projected/385c4850-9b08-4531-af6a-4f2ec3662165-kube-api-access-fr6ld\") pod \"certified-operators-dv5kn\" (UID: \"385c4850-9b08-4531-af6a-4f2ec3662165\") " pod="openshift-marketplace/certified-operators-dv5kn" Dec 10 23:04:17 crc kubenswrapper[4791]: I1210 23:04:17.113051 4791 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/385c4850-9b08-4531-af6a-4f2ec3662165-utilities\") pod \"certified-operators-dv5kn\" (UID: \"385c4850-9b08-4531-af6a-4f2ec3662165\") " pod="openshift-marketplace/certified-operators-dv5kn" Dec 10 23:04:17 crc kubenswrapper[4791]: I1210 23:04:17.113187 4791 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/385c4850-9b08-4531-af6a-4f2ec3662165-catalog-content\") pod \"certified-operators-dv5kn\" (UID: \"385c4850-9b08-4531-af6a-4f2ec3662165\") " pod="openshift-marketplace/certified-operators-dv5kn" Dec 10 23:04:17 crc kubenswrapper[4791]: I1210 23:04:17.113494 4791 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/385c4850-9b08-4531-af6a-4f2ec3662165-utilities\") pod \"certified-operators-dv5kn\" (UID: \"385c4850-9b08-4531-af6a-4f2ec3662165\") " pod="openshift-marketplace/certified-operators-dv5kn" Dec 10 23:04:17 crc kubenswrapper[4791]: I1210 23:04:17.130639 4791 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-fr6ld\" (UniqueName: \"kubernetes.io/projected/385c4850-9b08-4531-af6a-4f2ec3662165-kube-api-access-fr6ld\") pod \"certified-operators-dv5kn\" (UID: \"385c4850-9b08-4531-af6a-4f2ec3662165\") " pod="openshift-marketplace/certified-operators-dv5kn" Dec 10 23:04:17 crc kubenswrapper[4791]: I1210 23:04:17.167332 4791 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-8d275"] Dec 10 23:04:17 crc kubenswrapper[4791]: I1210 23:04:17.191373 4791 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-dv5kn" Dec 10 23:04:17 crc kubenswrapper[4791]: I1210 23:04:17.225133 4791 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-8d275" event={"ID":"2d2fb71e-03b5-49f1-989f-04abd05025ed","Type":"ContainerStarted","Data":"ab64e4b53de80ac89a1209a312351f65c694d8c878e74d8773e0689c67ce69e3"} Dec 10 23:04:17 crc kubenswrapper[4791]: I1210 23:04:17.458535 4791 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-dv5kn"] Dec 10 23:04:17 crc kubenswrapper[4791]: W1210 23:04:17.909156 4791 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod385c4850_9b08_4531_af6a_4f2ec3662165.slice/crio-2e566a80f948467a2a15904dc502b50a91380874bf6aaead971aec03b93ccaeb WatchSource:0}: Error finding container 2e566a80f948467a2a15904dc502b50a91380874bf6aaead971aec03b93ccaeb: Status 404 returned error can't find the container with id 2e566a80f948467a2a15904dc502b50a91380874bf6aaead971aec03b93ccaeb Dec 10 23:04:18 crc kubenswrapper[4791]: I1210 23:04:18.231719 4791 generic.go:334] "Generic (PLEG): container finished" podID="385c4850-9b08-4531-af6a-4f2ec3662165" containerID="81c9a2ed6acf321c299c94ac0b832db857870faee381baaa320c63752b74d8f6" exitCode=0 Dec 10 23:04:18 crc kubenswrapper[4791]: I1210 23:04:18.231756 4791 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-dv5kn" event={"ID":"385c4850-9b08-4531-af6a-4f2ec3662165","Type":"ContainerDied","Data":"81c9a2ed6acf321c299c94ac0b832db857870faee381baaa320c63752b74d8f6"} Dec 10 23:04:18 crc kubenswrapper[4791]: I1210 23:04:18.231793 4791 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-dv5kn" event={"ID":"385c4850-9b08-4531-af6a-4f2ec3662165","Type":"ContainerStarted","Data":"2e566a80f948467a2a15904dc502b50a91380874bf6aaead971aec03b93ccaeb"} Dec 10 23:04:18 crc kubenswrapper[4791]: I1210 23:04:18.233570 4791 generic.go:334] "Generic (PLEG): container finished" podID="2d2fb71e-03b5-49f1-989f-04abd05025ed" containerID="0ec718af0284ac70e99400b053ca758b3cfd95f87a5c1ca0605e9cefc2693011" exitCode=0 Dec 10 23:04:18 crc kubenswrapper[4791]: I1210 23:04:18.233620 4791 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-8d275" event={"ID":"2d2fb71e-03b5-49f1-989f-04abd05025ed","Type":"ContainerDied","Data":"0ec718af0284ac70e99400b053ca758b3cfd95f87a5c1ca0605e9cefc2693011"} Dec 10 23:04:18 crc kubenswrapper[4791]: I1210 23:04:18.238053 4791 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/50d664eb6e76f12a148f821b8e085abb03568bf5cb6cd29276e1382c05lclqs" event={"ID":"c76a21e1-575e-42e2-8372-e79986d26dbb","Type":"ContainerStarted","Data":"c4ac2bead06f7cc1a446e3ab9657f0752ef83995753c73afb0192178dd9c0e0a"} Dec 10 23:04:19 crc kubenswrapper[4791]: I1210 23:04:19.245221 4791 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-dv5kn" event={"ID":"385c4850-9b08-4531-af6a-4f2ec3662165","Type":"ContainerStarted","Data":"6b8c1f2ea7591d9890aacec0ab79fad7fb59b7b73a2cf120d3c182fbcbda9143"} Dec 10 23:04:19 crc kubenswrapper[4791]: I1210 23:04:19.247104 4791 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-8d275" event={"ID":"2d2fb71e-03b5-49f1-989f-04abd05025ed","Type":"ContainerStarted","Data":"6acb3b7203ea21b48a5a52715a95bc46a7cee69ef68469e7816c2a24a29b0370"} Dec 10 23:04:19 crc kubenswrapper[4791]: I1210 23:04:19.249118 4791 generic.go:334] "Generic (PLEG): container finished" podID="c76a21e1-575e-42e2-8372-e79986d26dbb" containerID="c4ac2bead06f7cc1a446e3ab9657f0752ef83995753c73afb0192178dd9c0e0a" exitCode=0 Dec 10 23:04:19 crc kubenswrapper[4791]: I1210 23:04:19.249164 4791 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/50d664eb6e76f12a148f821b8e085abb03568bf5cb6cd29276e1382c05lclqs" event={"ID":"c76a21e1-575e-42e2-8372-e79986d26dbb","Type":"ContainerDied","Data":"c4ac2bead06f7cc1a446e3ab9657f0752ef83995753c73afb0192178dd9c0e0a"} Dec 10 23:04:20 crc kubenswrapper[4791]: I1210 23:04:20.257604 4791 generic.go:334] "Generic (PLEG): container finished" podID="385c4850-9b08-4531-af6a-4f2ec3662165" containerID="6b8c1f2ea7591d9890aacec0ab79fad7fb59b7b73a2cf120d3c182fbcbda9143" exitCode=0 Dec 10 23:04:20 crc kubenswrapper[4791]: I1210 23:04:20.257666 4791 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-dv5kn" event={"ID":"385c4850-9b08-4531-af6a-4f2ec3662165","Type":"ContainerDied","Data":"6b8c1f2ea7591d9890aacec0ab79fad7fb59b7b73a2cf120d3c182fbcbda9143"} Dec 10 23:04:20 crc kubenswrapper[4791]: I1210 23:04:20.259732 4791 generic.go:334] "Generic (PLEG): container finished" podID="2d2fb71e-03b5-49f1-989f-04abd05025ed" containerID="6acb3b7203ea21b48a5a52715a95bc46a7cee69ef68469e7816c2a24a29b0370" exitCode=0 Dec 10 23:04:20 crc kubenswrapper[4791]: I1210 23:04:20.259819 4791 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-8d275" event={"ID":"2d2fb71e-03b5-49f1-989f-04abd05025ed","Type":"ContainerDied","Data":"6acb3b7203ea21b48a5a52715a95bc46a7cee69ef68469e7816c2a24a29b0370"} Dec 10 23:04:20 crc kubenswrapper[4791]: I1210 23:04:20.262027 4791 generic.go:334] "Generic (PLEG): container finished" podID="c76a21e1-575e-42e2-8372-e79986d26dbb" containerID="67329632dfad6f482e474635452a16cf1177ffdd82d5c12f0e52e526fd42d5c7" exitCode=0 Dec 10 23:04:20 crc kubenswrapper[4791]: I1210 23:04:20.262078 4791 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/50d664eb6e76f12a148f821b8e085abb03568bf5cb6cd29276e1382c05lclqs" event={"ID":"c76a21e1-575e-42e2-8372-e79986d26dbb","Type":"ContainerDied","Data":"67329632dfad6f482e474635452a16cf1177ffdd82d5c12f0e52e526fd42d5c7"} Dec 10 23:04:21 crc kubenswrapper[4791]: I1210 23:04:21.268734 4791 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-8d275" event={"ID":"2d2fb71e-03b5-49f1-989f-04abd05025ed","Type":"ContainerStarted","Data":"dd74ffe8d86cfeebfb98270cf33328e090328a5b6a11013e60eb39fb74a5dc81"} Dec 10 23:04:21 crc kubenswrapper[4791]: I1210 23:04:21.290600 4791 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-marketplace-8d275" podStartSLOduration=4.542660895 podStartE2EDuration="7.290580467s" podCreationTimestamp="2025-12-10 23:04:14 +0000 UTC" firstStartedPulling="2025-12-10 23:04:18.235959046 +0000 UTC m=+892.665576659" lastFinishedPulling="2025-12-10 23:04:20.983878628 +0000 UTC m=+895.413496231" observedRunningTime="2025-12-10 23:04:21.283898124 +0000 UTC m=+895.713515737" watchObservedRunningTime="2025-12-10 23:04:21.290580467 +0000 UTC m=+895.720198080" Dec 10 23:04:21 crc kubenswrapper[4791]: I1210 23:04:21.638701 4791 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack-operators/50d664eb6e76f12a148f821b8e085abb03568bf5cb6cd29276e1382c05lclqs" Dec 10 23:04:21 crc kubenswrapper[4791]: I1210 23:04:21.669238 4791 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/c76a21e1-575e-42e2-8372-e79986d26dbb-bundle\") pod \"c76a21e1-575e-42e2-8372-e79986d26dbb\" (UID: \"c76a21e1-575e-42e2-8372-e79986d26dbb\") " Dec 10 23:04:21 crc kubenswrapper[4791]: I1210 23:04:21.669399 4791 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-964mm\" (UniqueName: \"kubernetes.io/projected/c76a21e1-575e-42e2-8372-e79986d26dbb-kube-api-access-964mm\") pod \"c76a21e1-575e-42e2-8372-e79986d26dbb\" (UID: \"c76a21e1-575e-42e2-8372-e79986d26dbb\") " Dec 10 23:04:21 crc kubenswrapper[4791]: I1210 23:04:21.669438 4791 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/c76a21e1-575e-42e2-8372-e79986d26dbb-util\") pod \"c76a21e1-575e-42e2-8372-e79986d26dbb\" (UID: \"c76a21e1-575e-42e2-8372-e79986d26dbb\") " Dec 10 23:04:21 crc kubenswrapper[4791]: I1210 23:04:21.670908 4791 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/c76a21e1-575e-42e2-8372-e79986d26dbb-bundle" (OuterVolumeSpecName: "bundle") pod "c76a21e1-575e-42e2-8372-e79986d26dbb" (UID: "c76a21e1-575e-42e2-8372-e79986d26dbb"). InnerVolumeSpecName "bundle". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 10 23:04:21 crc kubenswrapper[4791]: I1210 23:04:21.679574 4791 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/c76a21e1-575e-42e2-8372-e79986d26dbb-kube-api-access-964mm" (OuterVolumeSpecName: "kube-api-access-964mm") pod "c76a21e1-575e-42e2-8372-e79986d26dbb" (UID: "c76a21e1-575e-42e2-8372-e79986d26dbb"). InnerVolumeSpecName "kube-api-access-964mm". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 10 23:04:21 crc kubenswrapper[4791]: I1210 23:04:21.680120 4791 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/c76a21e1-575e-42e2-8372-e79986d26dbb-util" (OuterVolumeSpecName: "util") pod "c76a21e1-575e-42e2-8372-e79986d26dbb" (UID: "c76a21e1-575e-42e2-8372-e79986d26dbb"). InnerVolumeSpecName "util". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 10 23:04:21 crc kubenswrapper[4791]: I1210 23:04:21.770498 4791 reconciler_common.go:293] "Volume detached for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/c76a21e1-575e-42e2-8372-e79986d26dbb-bundle\") on node \"crc\" DevicePath \"\"" Dec 10 23:04:21 crc kubenswrapper[4791]: I1210 23:04:21.770532 4791 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-964mm\" (UniqueName: \"kubernetes.io/projected/c76a21e1-575e-42e2-8372-e79986d26dbb-kube-api-access-964mm\") on node \"crc\" DevicePath \"\"" Dec 10 23:04:21 crc kubenswrapper[4791]: I1210 23:04:21.770543 4791 reconciler_common.go:293] "Volume detached for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/c76a21e1-575e-42e2-8372-e79986d26dbb-util\") on node \"crc\" DevicePath \"\"" Dec 10 23:04:22 crc kubenswrapper[4791]: I1210 23:04:22.276485 4791 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack-operators/50d664eb6e76f12a148f821b8e085abb03568bf5cb6cd29276e1382c05lclqs" Dec 10 23:04:22 crc kubenswrapper[4791]: I1210 23:04:22.276475 4791 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/50d664eb6e76f12a148f821b8e085abb03568bf5cb6cd29276e1382c05lclqs" event={"ID":"c76a21e1-575e-42e2-8372-e79986d26dbb","Type":"ContainerDied","Data":"c07a2befadab0fb7d3c1539ed4a8d23d6fa53291514908a3fd48ef115e9f99ff"} Dec 10 23:04:22 crc kubenswrapper[4791]: I1210 23:04:22.276589 4791 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="c07a2befadab0fb7d3c1539ed4a8d23d6fa53291514908a3fd48ef115e9f99ff" Dec 10 23:04:22 crc kubenswrapper[4791]: I1210 23:04:22.279952 4791 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-dv5kn" event={"ID":"385c4850-9b08-4531-af6a-4f2ec3662165","Type":"ContainerStarted","Data":"8e49eaec7ca0642fb7e5ea9e5e7b880f82bbe27050168091cd5c379e0766c1b5"} Dec 10 23:04:22 crc kubenswrapper[4791]: I1210 23:04:22.302151 4791 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/certified-operators-dv5kn" podStartSLOduration=2.911086917 podStartE2EDuration="6.302134246s" podCreationTimestamp="2025-12-10 23:04:16 +0000 UTC" firstStartedPulling="2025-12-10 23:04:18.232857676 +0000 UTC m=+892.662475289" lastFinishedPulling="2025-12-10 23:04:21.623905005 +0000 UTC m=+896.053522618" observedRunningTime="2025-12-10 23:04:22.297951175 +0000 UTC m=+896.727568808" watchObservedRunningTime="2025-12-10 23:04:22.302134246 +0000 UTC m=+896.731751859" Dec 10 23:04:23 crc kubenswrapper[4791]: I1210 23:04:23.586013 4791 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/openstack-operator-controller-operator-6cc5d54456-w47nt"] Dec 10 23:04:23 crc kubenswrapper[4791]: E1210 23:04:23.586610 4791 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="c76a21e1-575e-42e2-8372-e79986d26dbb" containerName="pull" Dec 10 23:04:23 crc kubenswrapper[4791]: I1210 23:04:23.586625 4791 state_mem.go:107] "Deleted CPUSet assignment" podUID="c76a21e1-575e-42e2-8372-e79986d26dbb" containerName="pull" Dec 10 23:04:23 crc kubenswrapper[4791]: E1210 23:04:23.586638 4791 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="c76a21e1-575e-42e2-8372-e79986d26dbb" containerName="extract" Dec 10 23:04:23 crc kubenswrapper[4791]: I1210 23:04:23.586647 4791 state_mem.go:107] "Deleted CPUSet assignment" podUID="c76a21e1-575e-42e2-8372-e79986d26dbb" containerName="extract" Dec 10 23:04:23 crc kubenswrapper[4791]: E1210 23:04:23.586656 4791 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="c76a21e1-575e-42e2-8372-e79986d26dbb" containerName="util" Dec 10 23:04:23 crc kubenswrapper[4791]: I1210 23:04:23.586664 4791 state_mem.go:107] "Deleted CPUSet assignment" podUID="c76a21e1-575e-42e2-8372-e79986d26dbb" containerName="util" Dec 10 23:04:23 crc kubenswrapper[4791]: I1210 23:04:23.586806 4791 memory_manager.go:354] "RemoveStaleState removing state" podUID="c76a21e1-575e-42e2-8372-e79986d26dbb" containerName="extract" Dec 10 23:04:23 crc kubenswrapper[4791]: I1210 23:04:23.587326 4791 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/openstack-operator-controller-operator-6cc5d54456-w47nt" Dec 10 23:04:23 crc kubenswrapper[4791]: I1210 23:04:23.589194 4791 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"openstack-operator-controller-operator-dockercfg-fjfrp" Dec 10 23:04:23 crc kubenswrapper[4791]: I1210 23:04:23.607740 4791 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/openstack-operator-controller-operator-6cc5d54456-w47nt"] Dec 10 23:04:23 crc kubenswrapper[4791]: I1210 23:04:23.695447 4791 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-xvf9x\" (UniqueName: \"kubernetes.io/projected/35509dee-f21b-4ab9-95ae-b93d2a4d93d2-kube-api-access-xvf9x\") pod \"openstack-operator-controller-operator-6cc5d54456-w47nt\" (UID: \"35509dee-f21b-4ab9-95ae-b93d2a4d93d2\") " pod="openstack-operators/openstack-operator-controller-operator-6cc5d54456-w47nt" Dec 10 23:04:23 crc kubenswrapper[4791]: I1210 23:04:23.796239 4791 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-xvf9x\" (UniqueName: \"kubernetes.io/projected/35509dee-f21b-4ab9-95ae-b93d2a4d93d2-kube-api-access-xvf9x\") pod \"openstack-operator-controller-operator-6cc5d54456-w47nt\" (UID: \"35509dee-f21b-4ab9-95ae-b93d2a4d93d2\") " pod="openstack-operators/openstack-operator-controller-operator-6cc5d54456-w47nt" Dec 10 23:04:23 crc kubenswrapper[4791]: I1210 23:04:23.819986 4791 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-xvf9x\" (UniqueName: \"kubernetes.io/projected/35509dee-f21b-4ab9-95ae-b93d2a4d93d2-kube-api-access-xvf9x\") pod \"openstack-operator-controller-operator-6cc5d54456-w47nt\" (UID: \"35509dee-f21b-4ab9-95ae-b93d2a4d93d2\") " pod="openstack-operators/openstack-operator-controller-operator-6cc5d54456-w47nt" Dec 10 23:04:23 crc kubenswrapper[4791]: I1210 23:04:23.904408 4791 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/openstack-operator-controller-operator-6cc5d54456-w47nt" Dec 10 23:04:24 crc kubenswrapper[4791]: I1210 23:04:24.262553 4791 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/openstack-operator-controller-operator-6cc5d54456-w47nt"] Dec 10 23:04:24 crc kubenswrapper[4791]: I1210 23:04:24.305679 4791 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/openstack-operator-controller-operator-6cc5d54456-w47nt" event={"ID":"35509dee-f21b-4ab9-95ae-b93d2a4d93d2","Type":"ContainerStarted","Data":"1cba7f8448e3c03526ac52f433309e298e671164e2b2ec8d7ad52196c1011516"} Dec 10 23:04:24 crc kubenswrapper[4791]: I1210 23:04:24.377464 4791 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-marketplace-8d275" Dec 10 23:04:24 crc kubenswrapper[4791]: I1210 23:04:24.378179 4791 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-marketplace-8d275" Dec 10 23:04:24 crc kubenswrapper[4791]: I1210 23:04:24.428572 4791 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-marketplace-8d275" Dec 10 23:04:26 crc kubenswrapper[4791]: I1210 23:04:26.418124 4791 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-marketplace-8d275" Dec 10 23:04:27 crc kubenswrapper[4791]: I1210 23:04:27.192615 4791 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/certified-operators-dv5kn" Dec 10 23:04:27 crc kubenswrapper[4791]: I1210 23:04:27.193280 4791 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/certified-operators-dv5kn" Dec 10 23:04:27 crc kubenswrapper[4791]: I1210 23:04:27.249378 4791 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/certified-operators-dv5kn" Dec 10 23:04:27 crc kubenswrapper[4791]: I1210 23:04:27.371942 4791 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/certified-operators-dv5kn" Dec 10 23:04:27 crc kubenswrapper[4791]: I1210 23:04:27.651492 4791 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-8d275"] Dec 10 23:04:28 crc kubenswrapper[4791]: I1210 23:04:28.327503 4791 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-marketplace-8d275" podUID="2d2fb71e-03b5-49f1-989f-04abd05025ed" containerName="registry-server" containerID="cri-o://dd74ffe8d86cfeebfb98270cf33328e090328a5b6a11013e60eb39fb74a5dc81" gracePeriod=2 Dec 10 23:04:29 crc kubenswrapper[4791]: I1210 23:04:29.335324 4791 generic.go:334] "Generic (PLEG): container finished" podID="2d2fb71e-03b5-49f1-989f-04abd05025ed" containerID="dd74ffe8d86cfeebfb98270cf33328e090328a5b6a11013e60eb39fb74a5dc81" exitCode=0 Dec 10 23:04:29 crc kubenswrapper[4791]: I1210 23:04:29.335650 4791 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-8d275" event={"ID":"2d2fb71e-03b5-49f1-989f-04abd05025ed","Type":"ContainerDied","Data":"dd74ffe8d86cfeebfb98270cf33328e090328a5b6a11013e60eb39fb74a5dc81"} Dec 10 23:04:30 crc kubenswrapper[4791]: I1210 23:04:30.855174 4791 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-dv5kn"] Dec 10 23:04:30 crc kubenswrapper[4791]: I1210 23:04:30.855598 4791 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/certified-operators-dv5kn" podUID="385c4850-9b08-4531-af6a-4f2ec3662165" containerName="registry-server" containerID="cri-o://8e49eaec7ca0642fb7e5ea9e5e7b880f82bbe27050168091cd5c379e0766c1b5" gracePeriod=2 Dec 10 23:04:32 crc kubenswrapper[4791]: I1210 23:04:32.360958 4791 generic.go:334] "Generic (PLEG): container finished" podID="385c4850-9b08-4531-af6a-4f2ec3662165" containerID="8e49eaec7ca0642fb7e5ea9e5e7b880f82bbe27050168091cd5c379e0766c1b5" exitCode=0 Dec 10 23:04:32 crc kubenswrapper[4791]: I1210 23:04:32.361081 4791 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-dv5kn" event={"ID":"385c4850-9b08-4531-af6a-4f2ec3662165","Type":"ContainerDied","Data":"8e49eaec7ca0642fb7e5ea9e5e7b880f82bbe27050168091cd5c379e0766c1b5"} Dec 10 23:04:33 crc kubenswrapper[4791]: I1210 23:04:33.879007 4791 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-8d275" Dec 10 23:04:33 crc kubenswrapper[4791]: I1210 23:04:33.938662 4791 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-dv5kn" Dec 10 23:04:33 crc kubenswrapper[4791]: I1210 23:04:33.994440 4791 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/2d2fb71e-03b5-49f1-989f-04abd05025ed-catalog-content\") pod \"2d2fb71e-03b5-49f1-989f-04abd05025ed\" (UID: \"2d2fb71e-03b5-49f1-989f-04abd05025ed\") " Dec 10 23:04:33 crc kubenswrapper[4791]: I1210 23:04:33.994510 4791 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/2d2fb71e-03b5-49f1-989f-04abd05025ed-utilities\") pod \"2d2fb71e-03b5-49f1-989f-04abd05025ed\" (UID: \"2d2fb71e-03b5-49f1-989f-04abd05025ed\") " Dec 10 23:04:33 crc kubenswrapper[4791]: I1210 23:04:33.994537 4791 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-rph6l\" (UniqueName: \"kubernetes.io/projected/2d2fb71e-03b5-49f1-989f-04abd05025ed-kube-api-access-rph6l\") pod \"2d2fb71e-03b5-49f1-989f-04abd05025ed\" (UID: \"2d2fb71e-03b5-49f1-989f-04abd05025ed\") " Dec 10 23:04:33 crc kubenswrapper[4791]: I1210 23:04:33.995279 4791 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/2d2fb71e-03b5-49f1-989f-04abd05025ed-utilities" (OuterVolumeSpecName: "utilities") pod "2d2fb71e-03b5-49f1-989f-04abd05025ed" (UID: "2d2fb71e-03b5-49f1-989f-04abd05025ed"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 10 23:04:34 crc kubenswrapper[4791]: I1210 23:04:34.017251 4791 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/2d2fb71e-03b5-49f1-989f-04abd05025ed-kube-api-access-rph6l" (OuterVolumeSpecName: "kube-api-access-rph6l") pod "2d2fb71e-03b5-49f1-989f-04abd05025ed" (UID: "2d2fb71e-03b5-49f1-989f-04abd05025ed"). InnerVolumeSpecName "kube-api-access-rph6l". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 10 23:04:34 crc kubenswrapper[4791]: I1210 23:04:34.030567 4791 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/2d2fb71e-03b5-49f1-989f-04abd05025ed-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "2d2fb71e-03b5-49f1-989f-04abd05025ed" (UID: "2d2fb71e-03b5-49f1-989f-04abd05025ed"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 10 23:04:34 crc kubenswrapper[4791]: I1210 23:04:34.095842 4791 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/385c4850-9b08-4531-af6a-4f2ec3662165-utilities\") pod \"385c4850-9b08-4531-af6a-4f2ec3662165\" (UID: \"385c4850-9b08-4531-af6a-4f2ec3662165\") " Dec 10 23:04:34 crc kubenswrapper[4791]: I1210 23:04:34.095979 4791 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-fr6ld\" (UniqueName: \"kubernetes.io/projected/385c4850-9b08-4531-af6a-4f2ec3662165-kube-api-access-fr6ld\") pod \"385c4850-9b08-4531-af6a-4f2ec3662165\" (UID: \"385c4850-9b08-4531-af6a-4f2ec3662165\") " Dec 10 23:04:34 crc kubenswrapper[4791]: I1210 23:04:34.096001 4791 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/385c4850-9b08-4531-af6a-4f2ec3662165-catalog-content\") pod \"385c4850-9b08-4531-af6a-4f2ec3662165\" (UID: \"385c4850-9b08-4531-af6a-4f2ec3662165\") " Dec 10 23:04:34 crc kubenswrapper[4791]: I1210 23:04:34.096272 4791 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/2d2fb71e-03b5-49f1-989f-04abd05025ed-catalog-content\") on node \"crc\" DevicePath \"\"" Dec 10 23:04:34 crc kubenswrapper[4791]: I1210 23:04:34.096296 4791 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/2d2fb71e-03b5-49f1-989f-04abd05025ed-utilities\") on node \"crc\" DevicePath \"\"" Dec 10 23:04:34 crc kubenswrapper[4791]: I1210 23:04:34.096305 4791 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-rph6l\" (UniqueName: \"kubernetes.io/projected/2d2fb71e-03b5-49f1-989f-04abd05025ed-kube-api-access-rph6l\") on node \"crc\" DevicePath \"\"" Dec 10 23:04:34 crc kubenswrapper[4791]: I1210 23:04:34.097079 4791 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/385c4850-9b08-4531-af6a-4f2ec3662165-utilities" (OuterVolumeSpecName: "utilities") pod "385c4850-9b08-4531-af6a-4f2ec3662165" (UID: "385c4850-9b08-4531-af6a-4f2ec3662165"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 10 23:04:34 crc kubenswrapper[4791]: I1210 23:04:34.099832 4791 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/385c4850-9b08-4531-af6a-4f2ec3662165-kube-api-access-fr6ld" (OuterVolumeSpecName: "kube-api-access-fr6ld") pod "385c4850-9b08-4531-af6a-4f2ec3662165" (UID: "385c4850-9b08-4531-af6a-4f2ec3662165"). InnerVolumeSpecName "kube-api-access-fr6ld". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 10 23:04:34 crc kubenswrapper[4791]: I1210 23:04:34.145678 4791 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/385c4850-9b08-4531-af6a-4f2ec3662165-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "385c4850-9b08-4531-af6a-4f2ec3662165" (UID: "385c4850-9b08-4531-af6a-4f2ec3662165"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 10 23:04:34 crc kubenswrapper[4791]: I1210 23:04:34.197662 4791 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/385c4850-9b08-4531-af6a-4f2ec3662165-utilities\") on node \"crc\" DevicePath \"\"" Dec 10 23:04:34 crc kubenswrapper[4791]: I1210 23:04:34.197710 4791 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-fr6ld\" (UniqueName: \"kubernetes.io/projected/385c4850-9b08-4531-af6a-4f2ec3662165-kube-api-access-fr6ld\") on node \"crc\" DevicePath \"\"" Dec 10 23:04:34 crc kubenswrapper[4791]: I1210 23:04:34.197722 4791 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/385c4850-9b08-4531-af6a-4f2ec3662165-catalog-content\") on node \"crc\" DevicePath \"\"" Dec 10 23:04:34 crc kubenswrapper[4791]: I1210 23:04:34.375183 4791 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/openstack-operator-controller-operator-6cc5d54456-w47nt" event={"ID":"35509dee-f21b-4ab9-95ae-b93d2a4d93d2","Type":"ContainerStarted","Data":"f82930acd5d908181b82a0b81dda2fa0bc99006a54b1333dcbe631fe3a2f97cc"} Dec 10 23:04:34 crc kubenswrapper[4791]: I1210 23:04:34.375649 4791 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/openstack-operator-controller-operator-6cc5d54456-w47nt" Dec 10 23:04:34 crc kubenswrapper[4791]: I1210 23:04:34.382440 4791 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-dv5kn" event={"ID":"385c4850-9b08-4531-af6a-4f2ec3662165","Type":"ContainerDied","Data":"2e566a80f948467a2a15904dc502b50a91380874bf6aaead971aec03b93ccaeb"} Dec 10 23:04:34 crc kubenswrapper[4791]: I1210 23:04:34.382453 4791 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-dv5kn" Dec 10 23:04:34 crc kubenswrapper[4791]: I1210 23:04:34.383079 4791 scope.go:117] "RemoveContainer" containerID="8e49eaec7ca0642fb7e5ea9e5e7b880f82bbe27050168091cd5c379e0766c1b5" Dec 10 23:04:34 crc kubenswrapper[4791]: I1210 23:04:34.385557 4791 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-8d275" event={"ID":"2d2fb71e-03b5-49f1-989f-04abd05025ed","Type":"ContainerDied","Data":"ab64e4b53de80ac89a1209a312351f65c694d8c878e74d8773e0689c67ce69e3"} Dec 10 23:04:34 crc kubenswrapper[4791]: I1210 23:04:34.385628 4791 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-8d275" Dec 10 23:04:34 crc kubenswrapper[4791]: I1210 23:04:34.406375 4791 scope.go:117] "RemoveContainer" containerID="6b8c1f2ea7591d9890aacec0ab79fad7fb59b7b73a2cf120d3c182fbcbda9143" Dec 10 23:04:34 crc kubenswrapper[4791]: I1210 23:04:34.417527 4791 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/openstack-operator-controller-operator-6cc5d54456-w47nt" podStartSLOduration=1.970825558 podStartE2EDuration="11.417503422s" podCreationTimestamp="2025-12-10 23:04:23 +0000 UTC" firstStartedPulling="2025-12-10 23:04:24.269602216 +0000 UTC m=+898.699219819" lastFinishedPulling="2025-12-10 23:04:33.71628007 +0000 UTC m=+908.145897683" observedRunningTime="2025-12-10 23:04:34.412053402 +0000 UTC m=+908.841671065" watchObservedRunningTime="2025-12-10 23:04:34.417503422 +0000 UTC m=+908.847121065" Dec 10 23:04:34 crc kubenswrapper[4791]: I1210 23:04:34.433536 4791 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-dv5kn"] Dec 10 23:04:34 crc kubenswrapper[4791]: I1210 23:04:34.440095 4791 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/certified-operators-dv5kn"] Dec 10 23:04:34 crc kubenswrapper[4791]: I1210 23:04:34.449820 4791 scope.go:117] "RemoveContainer" containerID="81c9a2ed6acf321c299c94ac0b832db857870faee381baaa320c63752b74d8f6" Dec 10 23:04:34 crc kubenswrapper[4791]: I1210 23:04:34.451904 4791 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-8d275"] Dec 10 23:04:34 crc kubenswrapper[4791]: I1210 23:04:34.456994 4791 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-marketplace-8d275"] Dec 10 23:04:34 crc kubenswrapper[4791]: I1210 23:04:34.462762 4791 scope.go:117] "RemoveContainer" containerID="dd74ffe8d86cfeebfb98270cf33328e090328a5b6a11013e60eb39fb74a5dc81" Dec 10 23:04:34 crc kubenswrapper[4791]: I1210 23:04:34.481838 4791 scope.go:117] "RemoveContainer" containerID="6acb3b7203ea21b48a5a52715a95bc46a7cee69ef68469e7816c2a24a29b0370" Dec 10 23:04:34 crc kubenswrapper[4791]: I1210 23:04:34.497615 4791 scope.go:117] "RemoveContainer" containerID="0ec718af0284ac70e99400b053ca758b3cfd95f87a5c1ca0605e9cefc2693011" Dec 10 23:04:35 crc kubenswrapper[4791]: I1210 23:04:35.894184 4791 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="2d2fb71e-03b5-49f1-989f-04abd05025ed" path="/var/lib/kubelet/pods/2d2fb71e-03b5-49f1-989f-04abd05025ed/volumes" Dec 10 23:04:35 crc kubenswrapper[4791]: I1210 23:04:35.895458 4791 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="385c4850-9b08-4531-af6a-4f2ec3662165" path="/var/lib/kubelet/pods/385c4850-9b08-4531-af6a-4f2ec3662165/volumes" Dec 10 23:04:43 crc kubenswrapper[4791]: I1210 23:04:43.907501 4791 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/openstack-operator-controller-operator-6cc5d54456-w47nt" Dec 10 23:05:07 crc kubenswrapper[4791]: I1210 23:05:07.068602 4791 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/barbican-operator-controller-manager-7d9dfd778-4ndx5"] Dec 10 23:05:07 crc kubenswrapper[4791]: E1210 23:05:07.069515 4791 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="2d2fb71e-03b5-49f1-989f-04abd05025ed" containerName="extract-utilities" Dec 10 23:05:07 crc kubenswrapper[4791]: I1210 23:05:07.069533 4791 state_mem.go:107] "Deleted CPUSet assignment" podUID="2d2fb71e-03b5-49f1-989f-04abd05025ed" containerName="extract-utilities" Dec 10 23:05:07 crc kubenswrapper[4791]: E1210 23:05:07.069550 4791 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="385c4850-9b08-4531-af6a-4f2ec3662165" containerName="extract-content" Dec 10 23:05:07 crc kubenswrapper[4791]: I1210 23:05:07.069558 4791 state_mem.go:107] "Deleted CPUSet assignment" podUID="385c4850-9b08-4531-af6a-4f2ec3662165" containerName="extract-content" Dec 10 23:05:07 crc kubenswrapper[4791]: E1210 23:05:07.069572 4791 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="2d2fb71e-03b5-49f1-989f-04abd05025ed" containerName="extract-content" Dec 10 23:05:07 crc kubenswrapper[4791]: I1210 23:05:07.069580 4791 state_mem.go:107] "Deleted CPUSet assignment" podUID="2d2fb71e-03b5-49f1-989f-04abd05025ed" containerName="extract-content" Dec 10 23:05:07 crc kubenswrapper[4791]: E1210 23:05:07.069592 4791 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="385c4850-9b08-4531-af6a-4f2ec3662165" containerName="registry-server" Dec 10 23:05:07 crc kubenswrapper[4791]: I1210 23:05:07.069600 4791 state_mem.go:107] "Deleted CPUSet assignment" podUID="385c4850-9b08-4531-af6a-4f2ec3662165" containerName="registry-server" Dec 10 23:05:07 crc kubenswrapper[4791]: E1210 23:05:07.069616 4791 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="385c4850-9b08-4531-af6a-4f2ec3662165" containerName="extract-utilities" Dec 10 23:05:07 crc kubenswrapper[4791]: I1210 23:05:07.069623 4791 state_mem.go:107] "Deleted CPUSet assignment" podUID="385c4850-9b08-4531-af6a-4f2ec3662165" containerName="extract-utilities" Dec 10 23:05:07 crc kubenswrapper[4791]: E1210 23:05:07.069632 4791 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="2d2fb71e-03b5-49f1-989f-04abd05025ed" containerName="registry-server" Dec 10 23:05:07 crc kubenswrapper[4791]: I1210 23:05:07.069639 4791 state_mem.go:107] "Deleted CPUSet assignment" podUID="2d2fb71e-03b5-49f1-989f-04abd05025ed" containerName="registry-server" Dec 10 23:05:07 crc kubenswrapper[4791]: I1210 23:05:07.069792 4791 memory_manager.go:354] "RemoveStaleState removing state" podUID="385c4850-9b08-4531-af6a-4f2ec3662165" containerName="registry-server" Dec 10 23:05:07 crc kubenswrapper[4791]: I1210 23:05:07.069804 4791 memory_manager.go:354] "RemoveStaleState removing state" podUID="2d2fb71e-03b5-49f1-989f-04abd05025ed" containerName="registry-server" Dec 10 23:05:07 crc kubenswrapper[4791]: I1210 23:05:07.070561 4791 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/barbican-operator-controller-manager-7d9dfd778-4ndx5" Dec 10 23:05:07 crc kubenswrapper[4791]: I1210 23:05:07.072824 4791 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"barbican-operator-controller-manager-dockercfg-vx7wd" Dec 10 23:05:07 crc kubenswrapper[4791]: I1210 23:05:07.079512 4791 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/cinder-operator-controller-manager-6c677c69b-6jmn2"] Dec 10 23:05:07 crc kubenswrapper[4791]: I1210 23:05:07.082916 4791 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/cinder-operator-controller-manager-6c677c69b-6jmn2" Dec 10 23:05:07 crc kubenswrapper[4791]: I1210 23:05:07.084969 4791 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"cinder-operator-controller-manager-dockercfg-4z4tk" Dec 10 23:05:07 crc kubenswrapper[4791]: I1210 23:05:07.121399 4791 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/designate-operator-controller-manager-697fb699cf-vg956"] Dec 10 23:05:07 crc kubenswrapper[4791]: I1210 23:05:07.149166 4791 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/designate-operator-controller-manager-697fb699cf-vg956" Dec 10 23:05:07 crc kubenswrapper[4791]: I1210 23:05:07.150835 4791 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"designate-operator-controller-manager-dockercfg-wvtpp" Dec 10 23:05:07 crc kubenswrapper[4791]: I1210 23:05:07.158121 4791 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/cinder-operator-controller-manager-6c677c69b-6jmn2"] Dec 10 23:05:07 crc kubenswrapper[4791]: I1210 23:05:07.181550 4791 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/designate-operator-controller-manager-697fb699cf-vg956"] Dec 10 23:05:07 crc kubenswrapper[4791]: I1210 23:05:07.182923 4791 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/glance-operator-controller-manager-5697bb5779-cvbqc"] Dec 10 23:05:07 crc kubenswrapper[4791]: I1210 23:05:07.185381 4791 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/glance-operator-controller-manager-5697bb5779-cvbqc" Dec 10 23:05:07 crc kubenswrapper[4791]: I1210 23:05:07.187077 4791 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-x88kg\" (UniqueName: \"kubernetes.io/projected/dc042f4f-d336-484e-8c9d-60658c82f84a-kube-api-access-x88kg\") pod \"barbican-operator-controller-manager-7d9dfd778-4ndx5\" (UID: \"dc042f4f-d336-484e-8c9d-60658c82f84a\") " pod="openstack-operators/barbican-operator-controller-manager-7d9dfd778-4ndx5" Dec 10 23:05:07 crc kubenswrapper[4791]: I1210 23:05:07.187132 4791 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-857zc\" (UniqueName: \"kubernetes.io/projected/48cd2bae-61e0-446c-ac2a-48e70bff5187-kube-api-access-857zc\") pod \"cinder-operator-controller-manager-6c677c69b-6jmn2\" (UID: \"48cd2bae-61e0-446c-ac2a-48e70bff5187\") " pod="openstack-operators/cinder-operator-controller-manager-6c677c69b-6jmn2" Dec 10 23:05:07 crc kubenswrapper[4791]: I1210 23:05:07.187195 4791 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-pwzfm\" (UniqueName: \"kubernetes.io/projected/19bf0bfe-4e8b-47a4-a919-4227b3e47bb3-kube-api-access-pwzfm\") pod \"designate-operator-controller-manager-697fb699cf-vg956\" (UID: \"19bf0bfe-4e8b-47a4-a919-4227b3e47bb3\") " pod="openstack-operators/designate-operator-controller-manager-697fb699cf-vg956" Dec 10 23:05:07 crc kubenswrapper[4791]: I1210 23:05:07.187543 4791 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"glance-operator-controller-manager-dockercfg-mqb45" Dec 10 23:05:07 crc kubenswrapper[4791]: I1210 23:05:07.192546 4791 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/heat-operator-controller-manager-5f64f6f8bb-fx67c"] Dec 10 23:05:07 crc kubenswrapper[4791]: I1210 23:05:07.193590 4791 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/heat-operator-controller-manager-5f64f6f8bb-fx67c" Dec 10 23:05:07 crc kubenswrapper[4791]: I1210 23:05:07.200657 4791 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"heat-operator-controller-manager-dockercfg-ttmtj" Dec 10 23:05:07 crc kubenswrapper[4791]: I1210 23:05:07.239983 4791 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/glance-operator-controller-manager-5697bb5779-cvbqc"] Dec 10 23:05:07 crc kubenswrapper[4791]: I1210 23:05:07.253630 4791 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/barbican-operator-controller-manager-7d9dfd778-4ndx5"] Dec 10 23:05:07 crc kubenswrapper[4791]: I1210 23:05:07.263268 4791 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/heat-operator-controller-manager-5f64f6f8bb-fx67c"] Dec 10 23:05:07 crc kubenswrapper[4791]: I1210 23:05:07.275636 4791 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/horizon-operator-controller-manager-68c6d99b8f-2lq47"] Dec 10 23:05:07 crc kubenswrapper[4791]: I1210 23:05:07.276778 4791 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/horizon-operator-controller-manager-68c6d99b8f-2lq47" Dec 10 23:05:07 crc kubenswrapper[4791]: I1210 23:05:07.281757 4791 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"horizon-operator-controller-manager-dockercfg-d2jmt" Dec 10 23:05:07 crc kubenswrapper[4791]: I1210 23:05:07.293442 4791 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-857zc\" (UniqueName: \"kubernetes.io/projected/48cd2bae-61e0-446c-ac2a-48e70bff5187-kube-api-access-857zc\") pod \"cinder-operator-controller-manager-6c677c69b-6jmn2\" (UID: \"48cd2bae-61e0-446c-ac2a-48e70bff5187\") " pod="openstack-operators/cinder-operator-controller-manager-6c677c69b-6jmn2" Dec 10 23:05:07 crc kubenswrapper[4791]: I1210 23:05:07.293497 4791 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-jqxwc\" (UniqueName: \"kubernetes.io/projected/f0b16aa2-e60e-4f0e-9679-9afc0a5ae027-kube-api-access-jqxwc\") pod \"glance-operator-controller-manager-5697bb5779-cvbqc\" (UID: \"f0b16aa2-e60e-4f0e-9679-9afc0a5ae027\") " pod="openstack-operators/glance-operator-controller-manager-5697bb5779-cvbqc" Dec 10 23:05:07 crc kubenswrapper[4791]: I1210 23:05:07.293564 4791 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-pwzfm\" (UniqueName: \"kubernetes.io/projected/19bf0bfe-4e8b-47a4-a919-4227b3e47bb3-kube-api-access-pwzfm\") pod \"designate-operator-controller-manager-697fb699cf-vg956\" (UID: \"19bf0bfe-4e8b-47a4-a919-4227b3e47bb3\") " pod="openstack-operators/designate-operator-controller-manager-697fb699cf-vg956" Dec 10 23:05:07 crc kubenswrapper[4791]: I1210 23:05:07.293609 4791 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-x88kg\" (UniqueName: \"kubernetes.io/projected/dc042f4f-d336-484e-8c9d-60658c82f84a-kube-api-access-x88kg\") pod \"barbican-operator-controller-manager-7d9dfd778-4ndx5\" (UID: \"dc042f4f-d336-484e-8c9d-60658c82f84a\") " pod="openstack-operators/barbican-operator-controller-manager-7d9dfd778-4ndx5" Dec 10 23:05:07 crc kubenswrapper[4791]: I1210 23:05:07.293647 4791 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-h75zs\" (UniqueName: \"kubernetes.io/projected/d181199e-5cd2-408b-b167-059b34491a15-kube-api-access-h75zs\") pod \"heat-operator-controller-manager-5f64f6f8bb-fx67c\" (UID: \"d181199e-5cd2-408b-b167-059b34491a15\") " pod="openstack-operators/heat-operator-controller-manager-5f64f6f8bb-fx67c" Dec 10 23:05:07 crc kubenswrapper[4791]: I1210 23:05:07.300007 4791 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/horizon-operator-controller-manager-68c6d99b8f-2lq47"] Dec 10 23:05:07 crc kubenswrapper[4791]: I1210 23:05:07.313554 4791 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/infra-operator-controller-manager-78d48bff9d-mmczn"] Dec 10 23:05:07 crc kubenswrapper[4791]: I1210 23:05:07.314807 4791 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/infra-operator-controller-manager-78d48bff9d-mmczn" Dec 10 23:05:07 crc kubenswrapper[4791]: I1210 23:05:07.319811 4791 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"infra-operator-webhook-server-cert" Dec 10 23:05:07 crc kubenswrapper[4791]: I1210 23:05:07.320180 4791 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"infra-operator-controller-manager-dockercfg-q8lh9" Dec 10 23:05:07 crc kubenswrapper[4791]: I1210 23:05:07.323547 4791 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/infra-operator-controller-manager-78d48bff9d-mmczn"] Dec 10 23:05:07 crc kubenswrapper[4791]: I1210 23:05:07.326001 4791 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/ironic-operator-controller-manager-967d97867-r6jdv"] Dec 10 23:05:07 crc kubenswrapper[4791]: I1210 23:05:07.327260 4791 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/ironic-operator-controller-manager-967d97867-r6jdv" Dec 10 23:05:07 crc kubenswrapper[4791]: I1210 23:05:07.329079 4791 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-x88kg\" (UniqueName: \"kubernetes.io/projected/dc042f4f-d336-484e-8c9d-60658c82f84a-kube-api-access-x88kg\") pod \"barbican-operator-controller-manager-7d9dfd778-4ndx5\" (UID: \"dc042f4f-d336-484e-8c9d-60658c82f84a\") " pod="openstack-operators/barbican-operator-controller-manager-7d9dfd778-4ndx5" Dec 10 23:05:07 crc kubenswrapper[4791]: I1210 23:05:07.332612 4791 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"ironic-operator-controller-manager-dockercfg-k5lp6" Dec 10 23:05:07 crc kubenswrapper[4791]: I1210 23:05:07.335117 4791 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-pwzfm\" (UniqueName: \"kubernetes.io/projected/19bf0bfe-4e8b-47a4-a919-4227b3e47bb3-kube-api-access-pwzfm\") pod \"designate-operator-controller-manager-697fb699cf-vg956\" (UID: \"19bf0bfe-4e8b-47a4-a919-4227b3e47bb3\") " pod="openstack-operators/designate-operator-controller-manager-697fb699cf-vg956" Dec 10 23:05:07 crc kubenswrapper[4791]: I1210 23:05:07.351773 4791 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-857zc\" (UniqueName: \"kubernetes.io/projected/48cd2bae-61e0-446c-ac2a-48e70bff5187-kube-api-access-857zc\") pod \"cinder-operator-controller-manager-6c677c69b-6jmn2\" (UID: \"48cd2bae-61e0-446c-ac2a-48e70bff5187\") " pod="openstack-operators/cinder-operator-controller-manager-6c677c69b-6jmn2" Dec 10 23:05:07 crc kubenswrapper[4791]: I1210 23:05:07.356579 4791 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/ironic-operator-controller-manager-967d97867-r6jdv"] Dec 10 23:05:07 crc kubenswrapper[4791]: I1210 23:05:07.393405 4791 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/keystone-operator-controller-manager-7765d96ddf-2wsxl"] Dec 10 23:05:07 crc kubenswrapper[4791]: I1210 23:05:07.394723 4791 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/keystone-operator-controller-manager-7765d96ddf-2wsxl" Dec 10 23:05:07 crc kubenswrapper[4791]: I1210 23:05:07.396774 4791 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-kp7nv\" (UniqueName: \"kubernetes.io/projected/cb5fb221-76b0-48b4-ad27-7d0c51581ec4-kube-api-access-kp7nv\") pod \"ironic-operator-controller-manager-967d97867-r6jdv\" (UID: \"cb5fb221-76b0-48b4-ad27-7d0c51581ec4\") " pod="openstack-operators/ironic-operator-controller-manager-967d97867-r6jdv" Dec 10 23:05:07 crc kubenswrapper[4791]: I1210 23:05:07.396867 4791 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-h75zs\" (UniqueName: \"kubernetes.io/projected/d181199e-5cd2-408b-b167-059b34491a15-kube-api-access-h75zs\") pod \"heat-operator-controller-manager-5f64f6f8bb-fx67c\" (UID: \"d181199e-5cd2-408b-b167-059b34491a15\") " pod="openstack-operators/heat-operator-controller-manager-5f64f6f8bb-fx67c" Dec 10 23:05:07 crc kubenswrapper[4791]: I1210 23:05:07.396889 4791 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-fbkm9\" (UniqueName: \"kubernetes.io/projected/e3fdc812-fbd3-4d35-8555-d55e9390ebd0-kube-api-access-fbkm9\") pod \"horizon-operator-controller-manager-68c6d99b8f-2lq47\" (UID: \"e3fdc812-fbd3-4d35-8555-d55e9390ebd0\") " pod="openstack-operators/horizon-operator-controller-manager-68c6d99b8f-2lq47" Dec 10 23:05:07 crc kubenswrapper[4791]: I1210 23:05:07.396920 4791 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-jqxwc\" (UniqueName: \"kubernetes.io/projected/f0b16aa2-e60e-4f0e-9679-9afc0a5ae027-kube-api-access-jqxwc\") pod \"glance-operator-controller-manager-5697bb5779-cvbqc\" (UID: \"f0b16aa2-e60e-4f0e-9679-9afc0a5ae027\") " pod="openstack-operators/glance-operator-controller-manager-5697bb5779-cvbqc" Dec 10 23:05:07 crc kubenswrapper[4791]: I1210 23:05:07.396941 4791 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/416b455d-b397-4aad-baf1-88e880619eb5-cert\") pod \"infra-operator-controller-manager-78d48bff9d-mmczn\" (UID: \"416b455d-b397-4aad-baf1-88e880619eb5\") " pod="openstack-operators/infra-operator-controller-manager-78d48bff9d-mmczn" Dec 10 23:05:07 crc kubenswrapper[4791]: I1210 23:05:07.396963 4791 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-t2ssb\" (UniqueName: \"kubernetes.io/projected/416b455d-b397-4aad-baf1-88e880619eb5-kube-api-access-t2ssb\") pod \"infra-operator-controller-manager-78d48bff9d-mmczn\" (UID: \"416b455d-b397-4aad-baf1-88e880619eb5\") " pod="openstack-operators/infra-operator-controller-manager-78d48bff9d-mmczn" Dec 10 23:05:07 crc kubenswrapper[4791]: I1210 23:05:07.400876 4791 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"keystone-operator-controller-manager-dockercfg-7778t" Dec 10 23:05:07 crc kubenswrapper[4791]: I1210 23:05:07.426036 4791 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-h75zs\" (UniqueName: \"kubernetes.io/projected/d181199e-5cd2-408b-b167-059b34491a15-kube-api-access-h75zs\") pod \"heat-operator-controller-manager-5f64f6f8bb-fx67c\" (UID: \"d181199e-5cd2-408b-b167-059b34491a15\") " pod="openstack-operators/heat-operator-controller-manager-5f64f6f8bb-fx67c" Dec 10 23:05:07 crc kubenswrapper[4791]: I1210 23:05:07.428961 4791 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-jqxwc\" (UniqueName: \"kubernetes.io/projected/f0b16aa2-e60e-4f0e-9679-9afc0a5ae027-kube-api-access-jqxwc\") pod \"glance-operator-controller-manager-5697bb5779-cvbqc\" (UID: \"f0b16aa2-e60e-4f0e-9679-9afc0a5ae027\") " pod="openstack-operators/glance-operator-controller-manager-5697bb5779-cvbqc" Dec 10 23:05:07 crc kubenswrapper[4791]: I1210 23:05:07.441398 4791 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/manila-operator-controller-manager-5b5fd79c9c-tsbtp"] Dec 10 23:05:07 crc kubenswrapper[4791]: I1210 23:05:07.442795 4791 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/manila-operator-controller-manager-5b5fd79c9c-tsbtp" Dec 10 23:05:07 crc kubenswrapper[4791]: I1210 23:05:07.442883 4791 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/keystone-operator-controller-manager-7765d96ddf-2wsxl"] Dec 10 23:05:07 crc kubenswrapper[4791]: I1210 23:05:07.445997 4791 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"manila-operator-controller-manager-dockercfg-4vz9l" Dec 10 23:05:07 crc kubenswrapper[4791]: I1210 23:05:07.452899 4791 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/mariadb-operator-controller-manager-79c8c4686c-2wmm4"] Dec 10 23:05:07 crc kubenswrapper[4791]: I1210 23:05:07.453739 4791 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/barbican-operator-controller-manager-7d9dfd778-4ndx5" Dec 10 23:05:07 crc kubenswrapper[4791]: I1210 23:05:07.467702 4791 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/mariadb-operator-controller-manager-79c8c4686c-2wmm4" Dec 10 23:05:07 crc kubenswrapper[4791]: I1210 23:05:07.469447 4791 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"mariadb-operator-controller-manager-dockercfg-q77nk" Dec 10 23:05:07 crc kubenswrapper[4791]: I1210 23:05:07.481980 4791 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/cinder-operator-controller-manager-6c677c69b-6jmn2" Dec 10 23:05:07 crc kubenswrapper[4791]: I1210 23:05:07.489473 4791 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/manila-operator-controller-manager-5b5fd79c9c-tsbtp"] Dec 10 23:05:07 crc kubenswrapper[4791]: I1210 23:05:07.498853 4791 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/416b455d-b397-4aad-baf1-88e880619eb5-cert\") pod \"infra-operator-controller-manager-78d48bff9d-mmczn\" (UID: \"416b455d-b397-4aad-baf1-88e880619eb5\") " pod="openstack-operators/infra-operator-controller-manager-78d48bff9d-mmczn" Dec 10 23:05:07 crc kubenswrapper[4791]: I1210 23:05:07.498901 4791 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-t2ssb\" (UniqueName: \"kubernetes.io/projected/416b455d-b397-4aad-baf1-88e880619eb5-kube-api-access-t2ssb\") pod \"infra-operator-controller-manager-78d48bff9d-mmczn\" (UID: \"416b455d-b397-4aad-baf1-88e880619eb5\") " pod="openstack-operators/infra-operator-controller-manager-78d48bff9d-mmczn" Dec 10 23:05:07 crc kubenswrapper[4791]: I1210 23:05:07.498933 4791 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-kp7nv\" (UniqueName: \"kubernetes.io/projected/cb5fb221-76b0-48b4-ad27-7d0c51581ec4-kube-api-access-kp7nv\") pod \"ironic-operator-controller-manager-967d97867-r6jdv\" (UID: \"cb5fb221-76b0-48b4-ad27-7d0c51581ec4\") " pod="openstack-operators/ironic-operator-controller-manager-967d97867-r6jdv" Dec 10 23:05:07 crc kubenswrapper[4791]: I1210 23:05:07.499002 4791 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-fbkm9\" (UniqueName: \"kubernetes.io/projected/e3fdc812-fbd3-4d35-8555-d55e9390ebd0-kube-api-access-fbkm9\") pod \"horizon-operator-controller-manager-68c6d99b8f-2lq47\" (UID: \"e3fdc812-fbd3-4d35-8555-d55e9390ebd0\") " pod="openstack-operators/horizon-operator-controller-manager-68c6d99b8f-2lq47" Dec 10 23:05:07 crc kubenswrapper[4791]: I1210 23:05:07.499044 4791 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-988zl\" (UniqueName: \"kubernetes.io/projected/311d4e7c-a0c7-47f6-8deb-7262068e8beb-kube-api-access-988zl\") pod \"keystone-operator-controller-manager-7765d96ddf-2wsxl\" (UID: \"311d4e7c-a0c7-47f6-8deb-7262068e8beb\") " pod="openstack-operators/keystone-operator-controller-manager-7765d96ddf-2wsxl" Dec 10 23:05:07 crc kubenswrapper[4791]: E1210 23:05:07.499050 4791 secret.go:188] Couldn't get secret openstack-operators/infra-operator-webhook-server-cert: secret "infra-operator-webhook-server-cert" not found Dec 10 23:05:07 crc kubenswrapper[4791]: E1210 23:05:07.499134 4791 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/416b455d-b397-4aad-baf1-88e880619eb5-cert podName:416b455d-b397-4aad-baf1-88e880619eb5 nodeName:}" failed. No retries permitted until 2025-12-10 23:05:07.999110613 +0000 UTC m=+942.428728336 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "cert" (UniqueName: "kubernetes.io/secret/416b455d-b397-4aad-baf1-88e880619eb5-cert") pod "infra-operator-controller-manager-78d48bff9d-mmczn" (UID: "416b455d-b397-4aad-baf1-88e880619eb5") : secret "infra-operator-webhook-server-cert" not found Dec 10 23:05:07 crc kubenswrapper[4791]: I1210 23:05:07.499066 4791 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-kmv99\" (UniqueName: \"kubernetes.io/projected/12b75f38-fb6a-4020-95b9-29c7dd872849-kube-api-access-kmv99\") pod \"manila-operator-controller-manager-5b5fd79c9c-tsbtp\" (UID: \"12b75f38-fb6a-4020-95b9-29c7dd872849\") " pod="openstack-operators/manila-operator-controller-manager-5b5fd79c9c-tsbtp" Dec 10 23:05:07 crc kubenswrapper[4791]: I1210 23:05:07.502578 4791 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/mariadb-operator-controller-manager-79c8c4686c-2wmm4"] Dec 10 23:05:07 crc kubenswrapper[4791]: I1210 23:05:07.522109 4791 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-kp7nv\" (UniqueName: \"kubernetes.io/projected/cb5fb221-76b0-48b4-ad27-7d0c51581ec4-kube-api-access-kp7nv\") pod \"ironic-operator-controller-manager-967d97867-r6jdv\" (UID: \"cb5fb221-76b0-48b4-ad27-7d0c51581ec4\") " pod="openstack-operators/ironic-operator-controller-manager-967d97867-r6jdv" Dec 10 23:05:07 crc kubenswrapper[4791]: I1210 23:05:07.525523 4791 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/neutron-operator-controller-manager-5fdfd5b6b5-g8f76"] Dec 10 23:05:07 crc kubenswrapper[4791]: I1210 23:05:07.526558 4791 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-fbkm9\" (UniqueName: \"kubernetes.io/projected/e3fdc812-fbd3-4d35-8555-d55e9390ebd0-kube-api-access-fbkm9\") pod \"horizon-operator-controller-manager-68c6d99b8f-2lq47\" (UID: \"e3fdc812-fbd3-4d35-8555-d55e9390ebd0\") " pod="openstack-operators/horizon-operator-controller-manager-68c6d99b8f-2lq47" Dec 10 23:05:07 crc kubenswrapper[4791]: I1210 23:05:07.528055 4791 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/neutron-operator-controller-manager-5fdfd5b6b5-g8f76"] Dec 10 23:05:07 crc kubenswrapper[4791]: I1210 23:05:07.528078 4791 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/nova-operator-controller-manager-697bc559fc-2lkcl"] Dec 10 23:05:07 crc kubenswrapper[4791]: I1210 23:05:07.528994 4791 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/neutron-operator-controller-manager-5fdfd5b6b5-g8f76" Dec 10 23:05:07 crc kubenswrapper[4791]: I1210 23:05:07.535019 4791 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"neutron-operator-controller-manager-dockercfg-h2zzs" Dec 10 23:05:07 crc kubenswrapper[4791]: I1210 23:05:07.538655 4791 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/nova-operator-controller-manager-697bc559fc-2lkcl" Dec 10 23:05:07 crc kubenswrapper[4791]: I1210 23:05:07.538818 4791 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-t2ssb\" (UniqueName: \"kubernetes.io/projected/416b455d-b397-4aad-baf1-88e880619eb5-kube-api-access-t2ssb\") pod \"infra-operator-controller-manager-78d48bff9d-mmczn\" (UID: \"416b455d-b397-4aad-baf1-88e880619eb5\") " pod="openstack-operators/infra-operator-controller-manager-78d48bff9d-mmczn" Dec 10 23:05:07 crc kubenswrapper[4791]: I1210 23:05:07.543629 4791 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"nova-operator-controller-manager-dockercfg-2lxwn" Dec 10 23:05:07 crc kubenswrapper[4791]: I1210 23:05:07.552209 4791 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/designate-operator-controller-manager-697fb699cf-vg956" Dec 10 23:05:07 crc kubenswrapper[4791]: I1210 23:05:07.561626 4791 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/glance-operator-controller-manager-5697bb5779-cvbqc" Dec 10 23:05:07 crc kubenswrapper[4791]: I1210 23:05:07.563252 4791 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/octavia-operator-controller-manager-998648c74-sjp6q"] Dec 10 23:05:07 crc kubenswrapper[4791]: I1210 23:05:07.564110 4791 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/octavia-operator-controller-manager-998648c74-sjp6q" Dec 10 23:05:07 crc kubenswrapper[4791]: I1210 23:05:07.577172 4791 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"octavia-operator-controller-manager-dockercfg-s277g" Dec 10 23:05:07 crc kubenswrapper[4791]: I1210 23:05:07.577854 4791 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/heat-operator-controller-manager-5f64f6f8bb-fx67c" Dec 10 23:05:07 crc kubenswrapper[4791]: I1210 23:05:07.602390 4791 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-988zl\" (UniqueName: \"kubernetes.io/projected/311d4e7c-a0c7-47f6-8deb-7262068e8beb-kube-api-access-988zl\") pod \"keystone-operator-controller-manager-7765d96ddf-2wsxl\" (UID: \"311d4e7c-a0c7-47f6-8deb-7262068e8beb\") " pod="openstack-operators/keystone-operator-controller-manager-7765d96ddf-2wsxl" Dec 10 23:05:07 crc kubenswrapper[4791]: I1210 23:05:07.602437 4791 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-2w8tm\" (UniqueName: \"kubernetes.io/projected/590c13cc-efa0-431a-9f2f-65fb97c1bd3c-kube-api-access-2w8tm\") pod \"mariadb-operator-controller-manager-79c8c4686c-2wmm4\" (UID: \"590c13cc-efa0-431a-9f2f-65fb97c1bd3c\") " pod="openstack-operators/mariadb-operator-controller-manager-79c8c4686c-2wmm4" Dec 10 23:05:07 crc kubenswrapper[4791]: I1210 23:05:07.602460 4791 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-kmv99\" (UniqueName: \"kubernetes.io/projected/12b75f38-fb6a-4020-95b9-29c7dd872849-kube-api-access-kmv99\") pod \"manila-operator-controller-manager-5b5fd79c9c-tsbtp\" (UID: \"12b75f38-fb6a-4020-95b9-29c7dd872849\") " pod="openstack-operators/manila-operator-controller-manager-5b5fd79c9c-tsbtp" Dec 10 23:05:07 crc kubenswrapper[4791]: I1210 23:05:07.602562 4791 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-qff2w\" (UniqueName: \"kubernetes.io/projected/3b2af4ca-48b4-463f-b0aa-c5cfd9099d67-kube-api-access-qff2w\") pod \"nova-operator-controller-manager-697bc559fc-2lkcl\" (UID: \"3b2af4ca-48b4-463f-b0aa-c5cfd9099d67\") " pod="openstack-operators/nova-operator-controller-manager-697bc559fc-2lkcl" Dec 10 23:05:07 crc kubenswrapper[4791]: I1210 23:05:07.666022 4791 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-kgzzj\" (UniqueName: \"kubernetes.io/projected/86bbb7b2-5f55-46a2-b426-013ba183977f-kube-api-access-kgzzj\") pod \"neutron-operator-controller-manager-5fdfd5b6b5-g8f76\" (UID: \"86bbb7b2-5f55-46a2-b426-013ba183977f\") " pod="openstack-operators/neutron-operator-controller-manager-5fdfd5b6b5-g8f76" Dec 10 23:05:07 crc kubenswrapper[4791]: I1210 23:05:07.666805 4791 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/horizon-operator-controller-manager-68c6d99b8f-2lq47" Dec 10 23:05:07 crc kubenswrapper[4791]: I1210 23:05:07.673891 4791 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-988zl\" (UniqueName: \"kubernetes.io/projected/311d4e7c-a0c7-47f6-8deb-7262068e8beb-kube-api-access-988zl\") pod \"keystone-operator-controller-manager-7765d96ddf-2wsxl\" (UID: \"311d4e7c-a0c7-47f6-8deb-7262068e8beb\") " pod="openstack-operators/keystone-operator-controller-manager-7765d96ddf-2wsxl" Dec 10 23:05:07 crc kubenswrapper[4791]: I1210 23:05:07.676728 4791 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-kmv99\" (UniqueName: \"kubernetes.io/projected/12b75f38-fb6a-4020-95b9-29c7dd872849-kube-api-access-kmv99\") pod \"manila-operator-controller-manager-5b5fd79c9c-tsbtp\" (UID: \"12b75f38-fb6a-4020-95b9-29c7dd872849\") " pod="openstack-operators/manila-operator-controller-manager-5b5fd79c9c-tsbtp" Dec 10 23:05:07 crc kubenswrapper[4791]: I1210 23:05:07.716718 4791 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/ironic-operator-controller-manager-967d97867-r6jdv" Dec 10 23:05:07 crc kubenswrapper[4791]: I1210 23:05:07.731145 4791 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/nova-operator-controller-manager-697bc559fc-2lkcl"] Dec 10 23:05:07 crc kubenswrapper[4791]: I1210 23:05:07.750789 4791 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/octavia-operator-controller-manager-998648c74-sjp6q"] Dec 10 23:05:07 crc kubenswrapper[4791]: I1210 23:05:07.765200 4791 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/openstack-baremetal-operator-controller-manager-84b575879fv4t28"] Dec 10 23:05:07 crc kubenswrapper[4791]: I1210 23:05:07.768060 4791 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-p8ljk\" (UniqueName: \"kubernetes.io/projected/23a95e98-8ae3-4ac4-945d-0cae5af5d1ac-kube-api-access-p8ljk\") pod \"octavia-operator-controller-manager-998648c74-sjp6q\" (UID: \"23a95e98-8ae3-4ac4-945d-0cae5af5d1ac\") " pod="openstack-operators/octavia-operator-controller-manager-998648c74-sjp6q" Dec 10 23:05:07 crc kubenswrapper[4791]: I1210 23:05:07.768106 4791 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-qff2w\" (UniqueName: \"kubernetes.io/projected/3b2af4ca-48b4-463f-b0aa-c5cfd9099d67-kube-api-access-qff2w\") pod \"nova-operator-controller-manager-697bc559fc-2lkcl\" (UID: \"3b2af4ca-48b4-463f-b0aa-c5cfd9099d67\") " pod="openstack-operators/nova-operator-controller-manager-697bc559fc-2lkcl" Dec 10 23:05:07 crc kubenswrapper[4791]: I1210 23:05:07.768149 4791 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-kgzzj\" (UniqueName: \"kubernetes.io/projected/86bbb7b2-5f55-46a2-b426-013ba183977f-kube-api-access-kgzzj\") pod \"neutron-operator-controller-manager-5fdfd5b6b5-g8f76\" (UID: \"86bbb7b2-5f55-46a2-b426-013ba183977f\") " pod="openstack-operators/neutron-operator-controller-manager-5fdfd5b6b5-g8f76" Dec 10 23:05:07 crc kubenswrapper[4791]: I1210 23:05:07.768176 4791 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-2w8tm\" (UniqueName: \"kubernetes.io/projected/590c13cc-efa0-431a-9f2f-65fb97c1bd3c-kube-api-access-2w8tm\") pod \"mariadb-operator-controller-manager-79c8c4686c-2wmm4\" (UID: \"590c13cc-efa0-431a-9f2f-65fb97c1bd3c\") " pod="openstack-operators/mariadb-operator-controller-manager-79c8c4686c-2wmm4" Dec 10 23:05:07 crc kubenswrapper[4791]: I1210 23:05:07.769789 4791 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/openstack-baremetal-operator-controller-manager-84b575879fv4t28" Dec 10 23:05:07 crc kubenswrapper[4791]: I1210 23:05:07.775483 4791 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/keystone-operator-controller-manager-7765d96ddf-2wsxl" Dec 10 23:05:07 crc kubenswrapper[4791]: I1210 23:05:07.782971 4791 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/ovn-operator-controller-manager-b6456fdb6-6rddm"] Dec 10 23:05:07 crc kubenswrapper[4791]: I1210 23:05:07.784415 4791 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"openstack-baremetal-operator-controller-manager-dockercfg-7792b" Dec 10 23:05:07 crc kubenswrapper[4791]: I1210 23:05:07.785591 4791 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/ovn-operator-controller-manager-b6456fdb6-6rddm" Dec 10 23:05:07 crc kubenswrapper[4791]: I1210 23:05:07.787681 4791 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"openstack-baremetal-operator-webhook-server-cert" Dec 10 23:05:07 crc kubenswrapper[4791]: I1210 23:05:07.790087 4791 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"ovn-operator-controller-manager-dockercfg-848dc" Dec 10 23:05:07 crc kubenswrapper[4791]: I1210 23:05:07.800752 4791 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-qff2w\" (UniqueName: \"kubernetes.io/projected/3b2af4ca-48b4-463f-b0aa-c5cfd9099d67-kube-api-access-qff2w\") pod \"nova-operator-controller-manager-697bc559fc-2lkcl\" (UID: \"3b2af4ca-48b4-463f-b0aa-c5cfd9099d67\") " pod="openstack-operators/nova-operator-controller-manager-697bc559fc-2lkcl" Dec 10 23:05:07 crc kubenswrapper[4791]: I1210 23:05:07.824180 4791 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-2w8tm\" (UniqueName: \"kubernetes.io/projected/590c13cc-efa0-431a-9f2f-65fb97c1bd3c-kube-api-access-2w8tm\") pod \"mariadb-operator-controller-manager-79c8c4686c-2wmm4\" (UID: \"590c13cc-efa0-431a-9f2f-65fb97c1bd3c\") " pod="openstack-operators/mariadb-operator-controller-manager-79c8c4686c-2wmm4" Dec 10 23:05:07 crc kubenswrapper[4791]: I1210 23:05:07.825170 4791 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/openstack-baremetal-operator-controller-manager-84b575879fv4t28"] Dec 10 23:05:07 crc kubenswrapper[4791]: I1210 23:05:07.825496 4791 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/manila-operator-controller-manager-5b5fd79c9c-tsbtp" Dec 10 23:05:07 crc kubenswrapper[4791]: I1210 23:05:07.828059 4791 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-kgzzj\" (UniqueName: \"kubernetes.io/projected/86bbb7b2-5f55-46a2-b426-013ba183977f-kube-api-access-kgzzj\") pod \"neutron-operator-controller-manager-5fdfd5b6b5-g8f76\" (UID: \"86bbb7b2-5f55-46a2-b426-013ba183977f\") " pod="openstack-operators/neutron-operator-controller-manager-5fdfd5b6b5-g8f76" Dec 10 23:05:07 crc kubenswrapper[4791]: I1210 23:05:07.839102 4791 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/ovn-operator-controller-manager-b6456fdb6-6rddm"] Dec 10 23:05:07 crc kubenswrapper[4791]: I1210 23:05:07.840354 4791 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/mariadb-operator-controller-manager-79c8c4686c-2wmm4" Dec 10 23:05:07 crc kubenswrapper[4791]: I1210 23:05:07.850694 4791 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/placement-operator-controller-manager-78f8948974-bhgvx"] Dec 10 23:05:07 crc kubenswrapper[4791]: I1210 23:05:07.856880 4791 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/placement-operator-controller-manager-78f8948974-bhgvx" Dec 10 23:05:07 crc kubenswrapper[4791]: I1210 23:05:07.864864 4791 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"placement-operator-controller-manager-dockercfg-xwcdg" Dec 10 23:05:07 crc kubenswrapper[4791]: I1210 23:05:07.864878 4791 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/neutron-operator-controller-manager-5fdfd5b6b5-g8f76" Dec 10 23:05:07 crc kubenswrapper[4791]: I1210 23:05:07.865093 4791 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/swift-operator-controller-manager-9d58d64bc-8vwnp"] Dec 10 23:05:07 crc kubenswrapper[4791]: I1210 23:05:07.866200 4791 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/swift-operator-controller-manager-9d58d64bc-8vwnp" Dec 10 23:05:07 crc kubenswrapper[4791]: I1210 23:05:07.869637 4791 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-p8ljk\" (UniqueName: \"kubernetes.io/projected/23a95e98-8ae3-4ac4-945d-0cae5af5d1ac-kube-api-access-p8ljk\") pod \"octavia-operator-controller-manager-998648c74-sjp6q\" (UID: \"23a95e98-8ae3-4ac4-945d-0cae5af5d1ac\") " pod="openstack-operators/octavia-operator-controller-manager-998648c74-sjp6q" Dec 10 23:05:07 crc kubenswrapper[4791]: I1210 23:05:07.873122 4791 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"swift-operator-controller-manager-dockercfg-4rn7b" Dec 10 23:05:07 crc kubenswrapper[4791]: I1210 23:05:07.883453 4791 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/placement-operator-controller-manager-78f8948974-bhgvx"] Dec 10 23:05:07 crc kubenswrapper[4791]: I1210 23:05:07.887568 4791 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-p8ljk\" (UniqueName: \"kubernetes.io/projected/23a95e98-8ae3-4ac4-945d-0cae5af5d1ac-kube-api-access-p8ljk\") pod \"octavia-operator-controller-manager-998648c74-sjp6q\" (UID: \"23a95e98-8ae3-4ac4-945d-0cae5af5d1ac\") " pod="openstack-operators/octavia-operator-controller-manager-998648c74-sjp6q" Dec 10 23:05:07 crc kubenswrapper[4791]: I1210 23:05:07.894671 4791 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/nova-operator-controller-manager-697bc559fc-2lkcl" Dec 10 23:05:07 crc kubenswrapper[4791]: I1210 23:05:07.946106 4791 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/swift-operator-controller-manager-9d58d64bc-8vwnp"] Dec 10 23:05:07 crc kubenswrapper[4791]: I1210 23:05:07.946149 4791 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/telemetry-operator-controller-manager-58d5ff84df-t6tnb"] Dec 10 23:05:07 crc kubenswrapper[4791]: I1210 23:05:07.947260 4791 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/telemetry-operator-controller-manager-58d5ff84df-t6tnb"] Dec 10 23:05:07 crc kubenswrapper[4791]: I1210 23:05:07.947280 4791 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/test-operator-controller-manager-5854674fcc-rw7ct"] Dec 10 23:05:07 crc kubenswrapper[4791]: I1210 23:05:07.948322 4791 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/telemetry-operator-controller-manager-58d5ff84df-t6tnb" Dec 10 23:05:07 crc kubenswrapper[4791]: I1210 23:05:07.950622 4791 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"telemetry-operator-controller-manager-dockercfg-z74dv" Dec 10 23:05:07 crc kubenswrapper[4791]: I1210 23:05:07.950899 4791 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/test-operator-controller-manager-5854674fcc-rw7ct"] Dec 10 23:05:07 crc kubenswrapper[4791]: I1210 23:05:07.951010 4791 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/test-operator-controller-manager-5854674fcc-rw7ct" Dec 10 23:05:07 crc kubenswrapper[4791]: I1210 23:05:07.954554 4791 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"test-operator-controller-manager-dockercfg-nr74q" Dec 10 23:05:07 crc kubenswrapper[4791]: I1210 23:05:07.974368 4791 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/d5907af9-fe3f-4da0-ba70-993d36b25746-cert\") pod \"openstack-baremetal-operator-controller-manager-84b575879fv4t28\" (UID: \"d5907af9-fe3f-4da0-ba70-993d36b25746\") " pod="openstack-operators/openstack-baremetal-operator-controller-manager-84b575879fv4t28" Dec 10 23:05:07 crc kubenswrapper[4791]: I1210 23:05:07.975452 4791 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-7p5x5\" (UniqueName: \"kubernetes.io/projected/b67ef71c-85c5-41ca-bf93-d1ca8a2fd007-kube-api-access-7p5x5\") pod \"placement-operator-controller-manager-78f8948974-bhgvx\" (UID: \"b67ef71c-85c5-41ca-bf93-d1ca8a2fd007\") " pod="openstack-operators/placement-operator-controller-manager-78f8948974-bhgvx" Dec 10 23:05:07 crc kubenswrapper[4791]: I1210 23:05:07.975532 4791 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-gplnr\" (UniqueName: \"kubernetes.io/projected/87c40827-309c-4c68-ae46-bbe449a9e66c-kube-api-access-gplnr\") pod \"ovn-operator-controller-manager-b6456fdb6-6rddm\" (UID: \"87c40827-309c-4c68-ae46-bbe449a9e66c\") " pod="openstack-operators/ovn-operator-controller-manager-b6456fdb6-6rddm" Dec 10 23:05:07 crc kubenswrapper[4791]: I1210 23:05:07.975597 4791 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-mk2hf\" (UniqueName: \"kubernetes.io/projected/bda40eaa-446c-475e-8f2f-76f76caa0b85-kube-api-access-mk2hf\") pod \"test-operator-controller-manager-5854674fcc-rw7ct\" (UID: \"bda40eaa-446c-475e-8f2f-76f76caa0b85\") " pod="openstack-operators/test-operator-controller-manager-5854674fcc-rw7ct" Dec 10 23:05:07 crc kubenswrapper[4791]: I1210 23:05:07.975621 4791 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-s2pvb\" (UniqueName: \"kubernetes.io/projected/4cf9497a-4193-4d34-a68a-48085af6e8b5-kube-api-access-s2pvb\") pod \"swift-operator-controller-manager-9d58d64bc-8vwnp\" (UID: \"4cf9497a-4193-4d34-a68a-48085af6e8b5\") " pod="openstack-operators/swift-operator-controller-manager-9d58d64bc-8vwnp" Dec 10 23:05:07 crc kubenswrapper[4791]: I1210 23:05:07.975662 4791 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-7kld4\" (UniqueName: \"kubernetes.io/projected/2418f025-d37e-4316-b804-9ab53c326c82-kube-api-access-7kld4\") pod \"telemetry-operator-controller-manager-58d5ff84df-t6tnb\" (UID: \"2418f025-d37e-4316-b804-9ab53c326c82\") " pod="openstack-operators/telemetry-operator-controller-manager-58d5ff84df-t6tnb" Dec 10 23:05:07 crc kubenswrapper[4791]: I1210 23:05:07.975703 4791 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-nppmq\" (UniqueName: \"kubernetes.io/projected/d5907af9-fe3f-4da0-ba70-993d36b25746-kube-api-access-nppmq\") pod \"openstack-baremetal-operator-controller-manager-84b575879fv4t28\" (UID: \"d5907af9-fe3f-4da0-ba70-993d36b25746\") " pod="openstack-operators/openstack-baremetal-operator-controller-manager-84b575879fv4t28" Dec 10 23:05:07 crc kubenswrapper[4791]: I1210 23:05:07.980271 4791 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/watcher-operator-controller-manager-75944c9b7-7cdb7"] Dec 10 23:05:07 crc kubenswrapper[4791]: I1210 23:05:07.981601 4791 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/watcher-operator-controller-manager-75944c9b7-7cdb7" Dec 10 23:05:07 crc kubenswrapper[4791]: I1210 23:05:07.984873 4791 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/watcher-operator-controller-manager-75944c9b7-7cdb7"] Dec 10 23:05:07 crc kubenswrapper[4791]: I1210 23:05:07.988409 4791 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"watcher-operator-controller-manager-dockercfg-4xvbp" Dec 10 23:05:08 crc kubenswrapper[4791]: I1210 23:05:08.000391 4791 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/openstack-operator-controller-manager-686dfd865c-cft5f"] Dec 10 23:05:08 crc kubenswrapper[4791]: I1210 23:05:08.001975 4791 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/openstack-operator-controller-manager-686dfd865c-cft5f" Dec 10 23:05:08 crc kubenswrapper[4791]: I1210 23:05:08.005757 4791 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"webhook-server-cert" Dec 10 23:05:08 crc kubenswrapper[4791]: I1210 23:05:08.006067 4791 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"metrics-server-cert" Dec 10 23:05:08 crc kubenswrapper[4791]: I1210 23:05:08.006215 4791 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"openstack-operator-controller-manager-dockercfg-5vcct" Dec 10 23:05:08 crc kubenswrapper[4791]: I1210 23:05:08.018684 4791 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/octavia-operator-controller-manager-998648c74-sjp6q" Dec 10 23:05:08 crc kubenswrapper[4791]: I1210 23:05:08.021265 4791 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/openstack-operator-controller-manager-686dfd865c-cft5f"] Dec 10 23:05:08 crc kubenswrapper[4791]: I1210 23:05:08.049717 4791 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/rabbitmq-cluster-operator-manager-668c99d594-2bgc5"] Dec 10 23:05:08 crc kubenswrapper[4791]: I1210 23:05:08.050631 4791 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/rabbitmq-cluster-operator-manager-668c99d594-2bgc5" Dec 10 23:05:08 crc kubenswrapper[4791]: I1210 23:05:08.057231 4791 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"rabbitmq-cluster-operator-controller-manager-dockercfg-2qx7k" Dec 10 23:05:08 crc kubenswrapper[4791]: I1210 23:05:08.057420 4791 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/rabbitmq-cluster-operator-manager-668c99d594-2bgc5"] Dec 10 23:05:08 crc kubenswrapper[4791]: I1210 23:05:08.077164 4791 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-7p5x5\" (UniqueName: \"kubernetes.io/projected/b67ef71c-85c5-41ca-bf93-d1ca8a2fd007-kube-api-access-7p5x5\") pod \"placement-operator-controller-manager-78f8948974-bhgvx\" (UID: \"b67ef71c-85c5-41ca-bf93-d1ca8a2fd007\") " pod="openstack-operators/placement-operator-controller-manager-78f8948974-bhgvx" Dec 10 23:05:08 crc kubenswrapper[4791]: I1210 23:05:08.077213 4791 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-gplnr\" (UniqueName: \"kubernetes.io/projected/87c40827-309c-4c68-ae46-bbe449a9e66c-kube-api-access-gplnr\") pod \"ovn-operator-controller-manager-b6456fdb6-6rddm\" (UID: \"87c40827-309c-4c68-ae46-bbe449a9e66c\") " pod="openstack-operators/ovn-operator-controller-manager-b6456fdb6-6rddm" Dec 10 23:05:08 crc kubenswrapper[4791]: I1210 23:05:08.077243 4791 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-vsmg4\" (UniqueName: \"kubernetes.io/projected/21e66f28-5ed9-475e-aa7d-a105ad1f2f9e-kube-api-access-vsmg4\") pod \"watcher-operator-controller-manager-75944c9b7-7cdb7\" (UID: \"21e66f28-5ed9-475e-aa7d-a105ad1f2f9e\") " pod="openstack-operators/watcher-operator-controller-manager-75944c9b7-7cdb7" Dec 10 23:05:08 crc kubenswrapper[4791]: I1210 23:05:08.077260 4791 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-7sjtw\" (UniqueName: \"kubernetes.io/projected/b3ae5e53-9077-4dd0-996d-2417f28b4736-kube-api-access-7sjtw\") pod \"rabbitmq-cluster-operator-manager-668c99d594-2bgc5\" (UID: \"b3ae5e53-9077-4dd0-996d-2417f28b4736\") " pod="openstack-operators/rabbitmq-cluster-operator-manager-668c99d594-2bgc5" Dec 10 23:05:08 crc kubenswrapper[4791]: I1210 23:05:08.077284 4791 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-mk2hf\" (UniqueName: \"kubernetes.io/projected/bda40eaa-446c-475e-8f2f-76f76caa0b85-kube-api-access-mk2hf\") pod \"test-operator-controller-manager-5854674fcc-rw7ct\" (UID: \"bda40eaa-446c-475e-8f2f-76f76caa0b85\") " pod="openstack-operators/test-operator-controller-manager-5854674fcc-rw7ct" Dec 10 23:05:08 crc kubenswrapper[4791]: I1210 23:05:08.077302 4791 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-s2pvb\" (UniqueName: \"kubernetes.io/projected/4cf9497a-4193-4d34-a68a-48085af6e8b5-kube-api-access-s2pvb\") pod \"swift-operator-controller-manager-9d58d64bc-8vwnp\" (UID: \"4cf9497a-4193-4d34-a68a-48085af6e8b5\") " pod="openstack-operators/swift-operator-controller-manager-9d58d64bc-8vwnp" Dec 10 23:05:08 crc kubenswrapper[4791]: I1210 23:05:08.077329 4791 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-7kld4\" (UniqueName: \"kubernetes.io/projected/2418f025-d37e-4316-b804-9ab53c326c82-kube-api-access-7kld4\") pod \"telemetry-operator-controller-manager-58d5ff84df-t6tnb\" (UID: \"2418f025-d37e-4316-b804-9ab53c326c82\") " pod="openstack-operators/telemetry-operator-controller-manager-58d5ff84df-t6tnb" Dec 10 23:05:08 crc kubenswrapper[4791]: I1210 23:05:08.077377 4791 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"webhook-certs\" (UniqueName: \"kubernetes.io/secret/e2c4ecea-2852-4e39-93fb-aee7cbe31aa4-webhook-certs\") pod \"openstack-operator-controller-manager-686dfd865c-cft5f\" (UID: \"e2c4ecea-2852-4e39-93fb-aee7cbe31aa4\") " pod="openstack-operators/openstack-operator-controller-manager-686dfd865c-cft5f" Dec 10 23:05:08 crc kubenswrapper[4791]: I1210 23:05:08.077395 4791 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-nppmq\" (UniqueName: \"kubernetes.io/projected/d5907af9-fe3f-4da0-ba70-993d36b25746-kube-api-access-nppmq\") pod \"openstack-baremetal-operator-controller-manager-84b575879fv4t28\" (UID: \"d5907af9-fe3f-4da0-ba70-993d36b25746\") " pod="openstack-operators/openstack-baremetal-operator-controller-manager-84b575879fv4t28" Dec 10 23:05:08 crc kubenswrapper[4791]: I1210 23:05:08.077422 4791 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/d5907af9-fe3f-4da0-ba70-993d36b25746-cert\") pod \"openstack-baremetal-operator-controller-manager-84b575879fv4t28\" (UID: \"d5907af9-fe3f-4da0-ba70-993d36b25746\") " pod="openstack-operators/openstack-baremetal-operator-controller-manager-84b575879fv4t28" Dec 10 23:05:08 crc kubenswrapper[4791]: I1210 23:05:08.077438 4791 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/e2c4ecea-2852-4e39-93fb-aee7cbe31aa4-metrics-certs\") pod \"openstack-operator-controller-manager-686dfd865c-cft5f\" (UID: \"e2c4ecea-2852-4e39-93fb-aee7cbe31aa4\") " pod="openstack-operators/openstack-operator-controller-manager-686dfd865c-cft5f" Dec 10 23:05:08 crc kubenswrapper[4791]: I1210 23:05:08.077460 4791 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-xdbmk\" (UniqueName: \"kubernetes.io/projected/e2c4ecea-2852-4e39-93fb-aee7cbe31aa4-kube-api-access-xdbmk\") pod \"openstack-operator-controller-manager-686dfd865c-cft5f\" (UID: \"e2c4ecea-2852-4e39-93fb-aee7cbe31aa4\") " pod="openstack-operators/openstack-operator-controller-manager-686dfd865c-cft5f" Dec 10 23:05:08 crc kubenswrapper[4791]: I1210 23:05:08.077487 4791 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/416b455d-b397-4aad-baf1-88e880619eb5-cert\") pod \"infra-operator-controller-manager-78d48bff9d-mmczn\" (UID: \"416b455d-b397-4aad-baf1-88e880619eb5\") " pod="openstack-operators/infra-operator-controller-manager-78d48bff9d-mmczn" Dec 10 23:05:08 crc kubenswrapper[4791]: E1210 23:05:08.077590 4791 secret.go:188] Couldn't get secret openstack-operators/infra-operator-webhook-server-cert: secret "infra-operator-webhook-server-cert" not found Dec 10 23:05:08 crc kubenswrapper[4791]: E1210 23:05:08.077627 4791 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/416b455d-b397-4aad-baf1-88e880619eb5-cert podName:416b455d-b397-4aad-baf1-88e880619eb5 nodeName:}" failed. No retries permitted until 2025-12-10 23:05:09.077613537 +0000 UTC m=+943.507231150 (durationBeforeRetry 1s). Error: MountVolume.SetUp failed for volume "cert" (UniqueName: "kubernetes.io/secret/416b455d-b397-4aad-baf1-88e880619eb5-cert") pod "infra-operator-controller-manager-78d48bff9d-mmczn" (UID: "416b455d-b397-4aad-baf1-88e880619eb5") : secret "infra-operator-webhook-server-cert" not found Dec 10 23:05:08 crc kubenswrapper[4791]: E1210 23:05:08.079517 4791 secret.go:188] Couldn't get secret openstack-operators/openstack-baremetal-operator-webhook-server-cert: secret "openstack-baremetal-operator-webhook-server-cert" not found Dec 10 23:05:08 crc kubenswrapper[4791]: E1210 23:05:08.079559 4791 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/d5907af9-fe3f-4da0-ba70-993d36b25746-cert podName:d5907af9-fe3f-4da0-ba70-993d36b25746 nodeName:}" failed. No retries permitted until 2025-12-10 23:05:08.579547463 +0000 UTC m=+943.009165076 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "cert" (UniqueName: "kubernetes.io/secret/d5907af9-fe3f-4da0-ba70-993d36b25746-cert") pod "openstack-baremetal-operator-controller-manager-84b575879fv4t28" (UID: "d5907af9-fe3f-4da0-ba70-993d36b25746") : secret "openstack-baremetal-operator-webhook-server-cert" not found Dec 10 23:05:08 crc kubenswrapper[4791]: I1210 23:05:08.110546 4791 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-7p5x5\" (UniqueName: \"kubernetes.io/projected/b67ef71c-85c5-41ca-bf93-d1ca8a2fd007-kube-api-access-7p5x5\") pod \"placement-operator-controller-manager-78f8948974-bhgvx\" (UID: \"b67ef71c-85c5-41ca-bf93-d1ca8a2fd007\") " pod="openstack-operators/placement-operator-controller-manager-78f8948974-bhgvx" Dec 10 23:05:08 crc kubenswrapper[4791]: I1210 23:05:08.111506 4791 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-7kld4\" (UniqueName: \"kubernetes.io/projected/2418f025-d37e-4316-b804-9ab53c326c82-kube-api-access-7kld4\") pod \"telemetry-operator-controller-manager-58d5ff84df-t6tnb\" (UID: \"2418f025-d37e-4316-b804-9ab53c326c82\") " pod="openstack-operators/telemetry-operator-controller-manager-58d5ff84df-t6tnb" Dec 10 23:05:08 crc kubenswrapper[4791]: I1210 23:05:08.111788 4791 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-mk2hf\" (UniqueName: \"kubernetes.io/projected/bda40eaa-446c-475e-8f2f-76f76caa0b85-kube-api-access-mk2hf\") pod \"test-operator-controller-manager-5854674fcc-rw7ct\" (UID: \"bda40eaa-446c-475e-8f2f-76f76caa0b85\") " pod="openstack-operators/test-operator-controller-manager-5854674fcc-rw7ct" Dec 10 23:05:08 crc kubenswrapper[4791]: I1210 23:05:08.112674 4791 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-gplnr\" (UniqueName: \"kubernetes.io/projected/87c40827-309c-4c68-ae46-bbe449a9e66c-kube-api-access-gplnr\") pod \"ovn-operator-controller-manager-b6456fdb6-6rddm\" (UID: \"87c40827-309c-4c68-ae46-bbe449a9e66c\") " pod="openstack-operators/ovn-operator-controller-manager-b6456fdb6-6rddm" Dec 10 23:05:08 crc kubenswrapper[4791]: I1210 23:05:08.113785 4791 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-s2pvb\" (UniqueName: \"kubernetes.io/projected/4cf9497a-4193-4d34-a68a-48085af6e8b5-kube-api-access-s2pvb\") pod \"swift-operator-controller-manager-9d58d64bc-8vwnp\" (UID: \"4cf9497a-4193-4d34-a68a-48085af6e8b5\") " pod="openstack-operators/swift-operator-controller-manager-9d58d64bc-8vwnp" Dec 10 23:05:08 crc kubenswrapper[4791]: I1210 23:05:08.114819 4791 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-nppmq\" (UniqueName: \"kubernetes.io/projected/d5907af9-fe3f-4da0-ba70-993d36b25746-kube-api-access-nppmq\") pod \"openstack-baremetal-operator-controller-manager-84b575879fv4t28\" (UID: \"d5907af9-fe3f-4da0-ba70-993d36b25746\") " pod="openstack-operators/openstack-baremetal-operator-controller-manager-84b575879fv4t28" Dec 10 23:05:08 crc kubenswrapper[4791]: I1210 23:05:08.142509 4791 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/telemetry-operator-controller-manager-58d5ff84df-t6tnb" Dec 10 23:05:08 crc kubenswrapper[4791]: I1210 23:05:08.151854 4791 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/test-operator-controller-manager-5854674fcc-rw7ct" Dec 10 23:05:08 crc kubenswrapper[4791]: I1210 23:05:08.260056 4791 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-vsmg4\" (UniqueName: \"kubernetes.io/projected/21e66f28-5ed9-475e-aa7d-a105ad1f2f9e-kube-api-access-vsmg4\") pod \"watcher-operator-controller-manager-75944c9b7-7cdb7\" (UID: \"21e66f28-5ed9-475e-aa7d-a105ad1f2f9e\") " pod="openstack-operators/watcher-operator-controller-manager-75944c9b7-7cdb7" Dec 10 23:05:08 crc kubenswrapper[4791]: I1210 23:05:08.260116 4791 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-7sjtw\" (UniqueName: \"kubernetes.io/projected/b3ae5e53-9077-4dd0-996d-2417f28b4736-kube-api-access-7sjtw\") pod \"rabbitmq-cluster-operator-manager-668c99d594-2bgc5\" (UID: \"b3ae5e53-9077-4dd0-996d-2417f28b4736\") " pod="openstack-operators/rabbitmq-cluster-operator-manager-668c99d594-2bgc5" Dec 10 23:05:08 crc kubenswrapper[4791]: I1210 23:05:08.260231 4791 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"webhook-certs\" (UniqueName: \"kubernetes.io/secret/e2c4ecea-2852-4e39-93fb-aee7cbe31aa4-webhook-certs\") pod \"openstack-operator-controller-manager-686dfd865c-cft5f\" (UID: \"e2c4ecea-2852-4e39-93fb-aee7cbe31aa4\") " pod="openstack-operators/openstack-operator-controller-manager-686dfd865c-cft5f" Dec 10 23:05:08 crc kubenswrapper[4791]: I1210 23:05:08.260300 4791 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/e2c4ecea-2852-4e39-93fb-aee7cbe31aa4-metrics-certs\") pod \"openstack-operator-controller-manager-686dfd865c-cft5f\" (UID: \"e2c4ecea-2852-4e39-93fb-aee7cbe31aa4\") " pod="openstack-operators/openstack-operator-controller-manager-686dfd865c-cft5f" Dec 10 23:05:08 crc kubenswrapper[4791]: E1210 23:05:08.260432 4791 secret.go:188] Couldn't get secret openstack-operators/metrics-server-cert: secret "metrics-server-cert" not found Dec 10 23:05:08 crc kubenswrapper[4791]: E1210 23:05:08.260669 4791 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/e2c4ecea-2852-4e39-93fb-aee7cbe31aa4-metrics-certs podName:e2c4ecea-2852-4e39-93fb-aee7cbe31aa4 nodeName:}" failed. No retries permitted until 2025-12-10 23:05:08.760464332 +0000 UTC m=+943.190081945 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "metrics-certs" (UniqueName: "kubernetes.io/secret/e2c4ecea-2852-4e39-93fb-aee7cbe31aa4-metrics-certs") pod "openstack-operator-controller-manager-686dfd865c-cft5f" (UID: "e2c4ecea-2852-4e39-93fb-aee7cbe31aa4") : secret "metrics-server-cert" not found Dec 10 23:05:08 crc kubenswrapper[4791]: I1210 23:05:08.260698 4791 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-xdbmk\" (UniqueName: \"kubernetes.io/projected/e2c4ecea-2852-4e39-93fb-aee7cbe31aa4-kube-api-access-xdbmk\") pod \"openstack-operator-controller-manager-686dfd865c-cft5f\" (UID: \"e2c4ecea-2852-4e39-93fb-aee7cbe31aa4\") " pod="openstack-operators/openstack-operator-controller-manager-686dfd865c-cft5f" Dec 10 23:05:08 crc kubenswrapper[4791]: E1210 23:05:08.260838 4791 secret.go:188] Couldn't get secret openstack-operators/webhook-server-cert: secret "webhook-server-cert" not found Dec 10 23:05:08 crc kubenswrapper[4791]: E1210 23:05:08.260891 4791 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/e2c4ecea-2852-4e39-93fb-aee7cbe31aa4-webhook-certs podName:e2c4ecea-2852-4e39-93fb-aee7cbe31aa4 nodeName:}" failed. No retries permitted until 2025-12-10 23:05:08.760875695 +0000 UTC m=+943.190493308 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "webhook-certs" (UniqueName: "kubernetes.io/secret/e2c4ecea-2852-4e39-93fb-aee7cbe31aa4-webhook-certs") pod "openstack-operator-controller-manager-686dfd865c-cft5f" (UID: "e2c4ecea-2852-4e39-93fb-aee7cbe31aa4") : secret "webhook-server-cert" not found Dec 10 23:05:08 crc kubenswrapper[4791]: I1210 23:05:08.264099 4791 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/ovn-operator-controller-manager-b6456fdb6-6rddm" Dec 10 23:05:08 crc kubenswrapper[4791]: I1210 23:05:08.282836 4791 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-vsmg4\" (UniqueName: \"kubernetes.io/projected/21e66f28-5ed9-475e-aa7d-a105ad1f2f9e-kube-api-access-vsmg4\") pod \"watcher-operator-controller-manager-75944c9b7-7cdb7\" (UID: \"21e66f28-5ed9-475e-aa7d-a105ad1f2f9e\") " pod="openstack-operators/watcher-operator-controller-manager-75944c9b7-7cdb7" Dec 10 23:05:08 crc kubenswrapper[4791]: I1210 23:05:08.289103 4791 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-xdbmk\" (UniqueName: \"kubernetes.io/projected/e2c4ecea-2852-4e39-93fb-aee7cbe31aa4-kube-api-access-xdbmk\") pod \"openstack-operator-controller-manager-686dfd865c-cft5f\" (UID: \"e2c4ecea-2852-4e39-93fb-aee7cbe31aa4\") " pod="openstack-operators/openstack-operator-controller-manager-686dfd865c-cft5f" Dec 10 23:05:08 crc kubenswrapper[4791]: I1210 23:05:08.289539 4791 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/placement-operator-controller-manager-78f8948974-bhgvx" Dec 10 23:05:08 crc kubenswrapper[4791]: I1210 23:05:08.289923 4791 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-7sjtw\" (UniqueName: \"kubernetes.io/projected/b3ae5e53-9077-4dd0-996d-2417f28b4736-kube-api-access-7sjtw\") pod \"rabbitmq-cluster-operator-manager-668c99d594-2bgc5\" (UID: \"b3ae5e53-9077-4dd0-996d-2417f28b4736\") " pod="openstack-operators/rabbitmq-cluster-operator-manager-668c99d594-2bgc5" Dec 10 23:05:08 crc kubenswrapper[4791]: I1210 23:05:08.320608 4791 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/rabbitmq-cluster-operator-manager-668c99d594-2bgc5" Dec 10 23:05:08 crc kubenswrapper[4791]: I1210 23:05:08.327798 4791 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/swift-operator-controller-manager-9d58d64bc-8vwnp" Dec 10 23:05:08 crc kubenswrapper[4791]: I1210 23:05:08.479853 4791 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/watcher-operator-controller-manager-75944c9b7-7cdb7" Dec 10 23:05:08 crc kubenswrapper[4791]: I1210 23:05:08.480260 4791 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/barbican-operator-controller-manager-7d9dfd778-4ndx5"] Dec 10 23:05:08 crc kubenswrapper[4791]: I1210 23:05:08.668562 4791 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/d5907af9-fe3f-4da0-ba70-993d36b25746-cert\") pod \"openstack-baremetal-operator-controller-manager-84b575879fv4t28\" (UID: \"d5907af9-fe3f-4da0-ba70-993d36b25746\") " pod="openstack-operators/openstack-baremetal-operator-controller-manager-84b575879fv4t28" Dec 10 23:05:08 crc kubenswrapper[4791]: E1210 23:05:08.670475 4791 secret.go:188] Couldn't get secret openstack-operators/openstack-baremetal-operator-webhook-server-cert: secret "openstack-baremetal-operator-webhook-server-cert" not found Dec 10 23:05:08 crc kubenswrapper[4791]: E1210 23:05:08.671717 4791 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/d5907af9-fe3f-4da0-ba70-993d36b25746-cert podName:d5907af9-fe3f-4da0-ba70-993d36b25746 nodeName:}" failed. No retries permitted until 2025-12-10 23:05:09.671693806 +0000 UTC m=+944.101311419 (durationBeforeRetry 1s). Error: MountVolume.SetUp failed for volume "cert" (UniqueName: "kubernetes.io/secret/d5907af9-fe3f-4da0-ba70-993d36b25746-cert") pod "openstack-baremetal-operator-controller-manager-84b575879fv4t28" (UID: "d5907af9-fe3f-4da0-ba70-993d36b25746") : secret "openstack-baremetal-operator-webhook-server-cert" not found Dec 10 23:05:08 crc kubenswrapper[4791]: I1210 23:05:08.688098 4791 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/heat-operator-controller-manager-5f64f6f8bb-fx67c"] Dec 10 23:05:08 crc kubenswrapper[4791]: I1210 23:05:08.707593 4791 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/cinder-operator-controller-manager-6c677c69b-6jmn2"] Dec 10 23:05:08 crc kubenswrapper[4791]: W1210 23:05:08.740699 4791 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podd181199e_5cd2_408b_b167_059b34491a15.slice/crio-399cb5eea5c5c384068b27a04f38109888eeae12250deeaf1441dabbe6f1e96d WatchSource:0}: Error finding container 399cb5eea5c5c384068b27a04f38109888eeae12250deeaf1441dabbe6f1e96d: Status 404 returned error can't find the container with id 399cb5eea5c5c384068b27a04f38109888eeae12250deeaf1441dabbe6f1e96d Dec 10 23:05:08 crc kubenswrapper[4791]: I1210 23:05:08.770040 4791 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"webhook-certs\" (UniqueName: \"kubernetes.io/secret/e2c4ecea-2852-4e39-93fb-aee7cbe31aa4-webhook-certs\") pod \"openstack-operator-controller-manager-686dfd865c-cft5f\" (UID: \"e2c4ecea-2852-4e39-93fb-aee7cbe31aa4\") " pod="openstack-operators/openstack-operator-controller-manager-686dfd865c-cft5f" Dec 10 23:05:08 crc kubenswrapper[4791]: I1210 23:05:08.770137 4791 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/e2c4ecea-2852-4e39-93fb-aee7cbe31aa4-metrics-certs\") pod \"openstack-operator-controller-manager-686dfd865c-cft5f\" (UID: \"e2c4ecea-2852-4e39-93fb-aee7cbe31aa4\") " pod="openstack-operators/openstack-operator-controller-manager-686dfd865c-cft5f" Dec 10 23:05:08 crc kubenswrapper[4791]: E1210 23:05:08.770285 4791 secret.go:188] Couldn't get secret openstack-operators/metrics-server-cert: secret "metrics-server-cert" not found Dec 10 23:05:08 crc kubenswrapper[4791]: E1210 23:05:08.770292 4791 secret.go:188] Couldn't get secret openstack-operators/webhook-server-cert: secret "webhook-server-cert" not found Dec 10 23:05:08 crc kubenswrapper[4791]: E1210 23:05:08.770353 4791 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/e2c4ecea-2852-4e39-93fb-aee7cbe31aa4-metrics-certs podName:e2c4ecea-2852-4e39-93fb-aee7cbe31aa4 nodeName:}" failed. No retries permitted until 2025-12-10 23:05:09.770320969 +0000 UTC m=+944.199938582 (durationBeforeRetry 1s). Error: MountVolume.SetUp failed for volume "metrics-certs" (UniqueName: "kubernetes.io/secret/e2c4ecea-2852-4e39-93fb-aee7cbe31aa4-metrics-certs") pod "openstack-operator-controller-manager-686dfd865c-cft5f" (UID: "e2c4ecea-2852-4e39-93fb-aee7cbe31aa4") : secret "metrics-server-cert" not found Dec 10 23:05:08 crc kubenswrapper[4791]: E1210 23:05:08.787815 4791 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/e2c4ecea-2852-4e39-93fb-aee7cbe31aa4-webhook-certs podName:e2c4ecea-2852-4e39-93fb-aee7cbe31aa4 nodeName:}" failed. No retries permitted until 2025-12-10 23:05:09.787772829 +0000 UTC m=+944.217390442 (durationBeforeRetry 1s). Error: MountVolume.SetUp failed for volume "webhook-certs" (UniqueName: "kubernetes.io/secret/e2c4ecea-2852-4e39-93fb-aee7cbe31aa4-webhook-certs") pod "openstack-operator-controller-manager-686dfd865c-cft5f" (UID: "e2c4ecea-2852-4e39-93fb-aee7cbe31aa4") : secret "webhook-server-cert" not found Dec 10 23:05:08 crc kubenswrapper[4791]: I1210 23:05:08.800740 4791 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/heat-operator-controller-manager-5f64f6f8bb-fx67c" event={"ID":"d181199e-5cd2-408b-b167-059b34491a15","Type":"ContainerStarted","Data":"399cb5eea5c5c384068b27a04f38109888eeae12250deeaf1441dabbe6f1e96d"} Dec 10 23:05:08 crc kubenswrapper[4791]: I1210 23:05:08.803418 4791 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/cinder-operator-controller-manager-6c677c69b-6jmn2" event={"ID":"48cd2bae-61e0-446c-ac2a-48e70bff5187","Type":"ContainerStarted","Data":"412488e84009e78e0967dfb6cd650e655145d62f92a9304d74cfb7f724e3cde3"} Dec 10 23:05:08 crc kubenswrapper[4791]: I1210 23:05:08.811422 4791 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/barbican-operator-controller-manager-7d9dfd778-4ndx5" event={"ID":"dc042f4f-d336-484e-8c9d-60658c82f84a","Type":"ContainerStarted","Data":"060ca0e01739ee6f8f806b7035b631f0100cfa3768eb134d1d6754bec6ad1214"} Dec 10 23:05:08 crc kubenswrapper[4791]: I1210 23:05:08.847987 4791 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/designate-operator-controller-manager-697fb699cf-vg956"] Dec 10 23:05:08 crc kubenswrapper[4791]: W1210 23:05:08.849824 4791 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod19bf0bfe_4e8b_47a4_a919_4227b3e47bb3.slice/crio-6b74b19daeff9eee867efa33ec405543774e3bfc614ae2b08ce6702e719fd7de WatchSource:0}: Error finding container 6b74b19daeff9eee867efa33ec405543774e3bfc614ae2b08ce6702e719fd7de: Status 404 returned error can't find the container with id 6b74b19daeff9eee867efa33ec405543774e3bfc614ae2b08ce6702e719fd7de Dec 10 23:05:09 crc kubenswrapper[4791]: I1210 23:05:09.078720 4791 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/416b455d-b397-4aad-baf1-88e880619eb5-cert\") pod \"infra-operator-controller-manager-78d48bff9d-mmczn\" (UID: \"416b455d-b397-4aad-baf1-88e880619eb5\") " pod="openstack-operators/infra-operator-controller-manager-78d48bff9d-mmczn" Dec 10 23:05:09 crc kubenswrapper[4791]: E1210 23:05:09.079117 4791 secret.go:188] Couldn't get secret openstack-operators/infra-operator-webhook-server-cert: secret "infra-operator-webhook-server-cert" not found Dec 10 23:05:09 crc kubenswrapper[4791]: E1210 23:05:09.079217 4791 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/416b455d-b397-4aad-baf1-88e880619eb5-cert podName:416b455d-b397-4aad-baf1-88e880619eb5 nodeName:}" failed. No retries permitted until 2025-12-10 23:05:11.079194139 +0000 UTC m=+945.508811762 (durationBeforeRetry 2s). Error: MountVolume.SetUp failed for volume "cert" (UniqueName: "kubernetes.io/secret/416b455d-b397-4aad-baf1-88e880619eb5-cert") pod "infra-operator-controller-manager-78d48bff9d-mmczn" (UID: "416b455d-b397-4aad-baf1-88e880619eb5") : secret "infra-operator-webhook-server-cert" not found Dec 10 23:05:09 crc kubenswrapper[4791]: I1210 23:05:09.386975 4791 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/test-operator-controller-manager-5854674fcc-rw7ct"] Dec 10 23:05:09 crc kubenswrapper[4791]: I1210 23:05:09.424553 4791 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/ironic-operator-controller-manager-967d97867-r6jdv"] Dec 10 23:05:09 crc kubenswrapper[4791]: I1210 23:05:09.441092 4791 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/manila-operator-controller-manager-5b5fd79c9c-tsbtp"] Dec 10 23:05:09 crc kubenswrapper[4791]: W1210 23:05:09.449953 4791 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod4cf9497a_4193_4d34_a68a_48085af6e8b5.slice/crio-1ccb5c61702064b875151444c40160c34c2d4d5a04fe4515a9a72b114a540c42 WatchSource:0}: Error finding container 1ccb5c61702064b875151444c40160c34c2d4d5a04fe4515a9a72b114a540c42: Status 404 returned error can't find the container with id 1ccb5c61702064b875151444c40160c34c2d4d5a04fe4515a9a72b114a540c42 Dec 10 23:05:09 crc kubenswrapper[4791]: I1210 23:05:09.451212 4791 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/horizon-operator-controller-manager-68c6d99b8f-2lq47"] Dec 10 23:05:09 crc kubenswrapper[4791]: W1210 23:05:09.455637 4791 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podb67ef71c_85c5_41ca_bf93_d1ca8a2fd007.slice/crio-89196ca78d764e19337eeb5c14cd87fc7273374f1e7a8e5e1a93d1ce07e89ed8 WatchSource:0}: Error finding container 89196ca78d764e19337eeb5c14cd87fc7273374f1e7a8e5e1a93d1ce07e89ed8: Status 404 returned error can't find the container with id 89196ca78d764e19337eeb5c14cd87fc7273374f1e7a8e5e1a93d1ce07e89ed8 Dec 10 23:05:09 crc kubenswrapper[4791]: I1210 23:05:09.458534 4791 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/glance-operator-controller-manager-5697bb5779-cvbqc"] Dec 10 23:05:09 crc kubenswrapper[4791]: E1210 23:05:09.465875 4791 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:manager,Image:quay.io/openstack-k8s-operators/mariadb-operator@sha256:424da951f13f1fbe9083215dc9f5088f90676dd813f01fdf3c1a8639b61cbaad,Command:[/manager],Args:[--leader-elect --health-probe-bind-address=:8081 --metrics-bind-address=127.0.0.1:8080],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:LEASE_DURATION,Value:30,ValueFrom:nil,},EnvVar{Name:RENEW_DEADLINE,Value:20,ValueFrom:nil,},EnvVar{Name:RETRY_PERIOD,Value:5,ValueFrom:nil,},EnvVar{Name:ENABLE_WEBHOOKS,Value:false,ValueFrom:nil,},EnvVar{Name:METRICS_CERTS,Value:false,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{cpu: {{500 -3} {} 500m DecimalSI},memory: {{536870912 0} {} BinarySI},},Requests:ResourceList{cpu: {{10 -3} {} 10m DecimalSI},memory: {{268435456 0} {} BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kube-api-access-2w8tm,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/healthz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:15,TimeoutSeconds:1,PeriodSeconds:20,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},ReadinessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/readyz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:5,TimeoutSeconds:1,PeriodSeconds:10,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod mariadb-operator-controller-manager-79c8c4686c-2wmm4_openstack-operators(590c13cc-efa0-431a-9f2f-65fb97c1bd3c): ErrImagePull: pull QPS exceeded" logger="UnhandledError" Dec 10 23:05:09 crc kubenswrapper[4791]: E1210 23:05:09.465975 4791 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:manager,Image:quay.io/openstack-k8s-operators/octavia-operator@sha256:d9a3694865a7d54ee96397add18c3898886e98d079aa20876a0f4de1fa7a7168,Command:[/manager],Args:[--leader-elect --health-probe-bind-address=:8081 --metrics-bind-address=127.0.0.1:8080],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:LEASE_DURATION,Value:30,ValueFrom:nil,},EnvVar{Name:RENEW_DEADLINE,Value:20,ValueFrom:nil,},EnvVar{Name:RETRY_PERIOD,Value:5,ValueFrom:nil,},EnvVar{Name:ENABLE_WEBHOOKS,Value:false,ValueFrom:nil,},EnvVar{Name:METRICS_CERTS,Value:false,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{cpu: {{500 -3} {} 500m DecimalSI},memory: {{536870912 0} {} BinarySI},},Requests:ResourceList{cpu: {{10 -3} {} 10m DecimalSI},memory: {{268435456 0} {} BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kube-api-access-p8ljk,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/healthz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:15,TimeoutSeconds:1,PeriodSeconds:20,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},ReadinessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/readyz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:5,TimeoutSeconds:1,PeriodSeconds:10,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod octavia-operator-controller-manager-998648c74-sjp6q_openstack-operators(23a95e98-8ae3-4ac4-945d-0cae5af5d1ac): ErrImagePull: pull QPS exceeded" logger="UnhandledError" Dec 10 23:05:09 crc kubenswrapper[4791]: I1210 23:05:09.468384 4791 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/nova-operator-controller-manager-697bc559fc-2lkcl"] Dec 10 23:05:09 crc kubenswrapper[4791]: E1210 23:05:09.468475 4791 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:manager,Image:quay.io/openstack-k8s-operators/telemetry-operator@sha256:f27e732ec1faee765461bf137d9be81278b2fa39675019a73622755e1e610b6f,Command:[/manager],Args:[--leader-elect --health-probe-bind-address=:8081 --metrics-bind-address=127.0.0.1:8080],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:LEASE_DURATION,Value:30,ValueFrom:nil,},EnvVar{Name:RENEW_DEADLINE,Value:20,ValueFrom:nil,},EnvVar{Name:RETRY_PERIOD,Value:5,ValueFrom:nil,},EnvVar{Name:ENABLE_WEBHOOKS,Value:false,ValueFrom:nil,},EnvVar{Name:METRICS_CERTS,Value:false,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{cpu: {{500 -3} {} 500m DecimalSI},memory: {{536870912 0} {} BinarySI},},Requests:ResourceList{cpu: {{10 -3} {} 10m DecimalSI},memory: {{268435456 0} {} BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kube-api-access-7kld4,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/healthz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:15,TimeoutSeconds:1,PeriodSeconds:20,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},ReadinessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/readyz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:5,TimeoutSeconds:1,PeriodSeconds:10,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod telemetry-operator-controller-manager-58d5ff84df-t6tnb_openstack-operators(2418f025-d37e-4316-b804-9ab53c326c82): ErrImagePull: pull QPS exceeded" logger="UnhandledError" Dec 10 23:05:09 crc kubenswrapper[4791]: E1210 23:05:09.468760 4791 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:kube-rbac-proxy,Image:quay.io/openstack-k8s-operators/kube-rbac-proxy:v0.16.0,Command:[],Args:[--secure-listen-address=0.0.0.0:8443 --upstream=http://127.0.0.1:8080/ --logtostderr=true --v=0],WorkingDir:,Ports:[]ContainerPort{ContainerPort{Name:https,HostPort:0,ContainerPort:8443,Protocol:TCP,HostIP:,},},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{cpu: {{500 -3} {} 500m DecimalSI},memory: {{134217728 0} {} BinarySI},},Requests:ResourceList{cpu: {{5 -3} {} 5m DecimalSI},memory: {{67108864 0} {} BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kube-api-access-2w8tm,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod mariadb-operator-controller-manager-79c8c4686c-2wmm4_openstack-operators(590c13cc-efa0-431a-9f2f-65fb97c1bd3c): ErrImagePull: pull QPS exceeded" logger="UnhandledError" Dec 10 23:05:09 crc kubenswrapper[4791]: E1210 23:05:09.470583 4791 pod_workers.go:1301] "Error syncing pod, skipping" err="[failed to \"StartContainer\" for \"manager\" with ErrImagePull: \"pull QPS exceeded\", failed to \"StartContainer\" for \"kube-rbac-proxy\" with ErrImagePull: \"pull QPS exceeded\"]" pod="openstack-operators/mariadb-operator-controller-manager-79c8c4686c-2wmm4" podUID="590c13cc-efa0-431a-9f2f-65fb97c1bd3c" Dec 10 23:05:09 crc kubenswrapper[4791]: E1210 23:05:09.471101 4791 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:kube-rbac-proxy,Image:quay.io/openstack-k8s-operators/kube-rbac-proxy:v0.16.0,Command:[],Args:[--secure-listen-address=0.0.0.0:8443 --upstream=http://127.0.0.1:8080/ --logtostderr=true --v=0],WorkingDir:,Ports:[]ContainerPort{ContainerPort{Name:https,HostPort:0,ContainerPort:8443,Protocol:TCP,HostIP:,},},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{cpu: {{500 -3} {} 500m DecimalSI},memory: {{134217728 0} {} BinarySI},},Requests:ResourceList{cpu: {{5 -3} {} 5m DecimalSI},memory: {{67108864 0} {} BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kube-api-access-p8ljk,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod octavia-operator-controller-manager-998648c74-sjp6q_openstack-operators(23a95e98-8ae3-4ac4-945d-0cae5af5d1ac): ErrImagePull: pull QPS exceeded" logger="UnhandledError" Dec 10 23:05:09 crc kubenswrapper[4791]: E1210 23:05:09.471455 4791 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:kube-rbac-proxy,Image:quay.io/openstack-k8s-operators/kube-rbac-proxy:v0.16.0,Command:[],Args:[--secure-listen-address=0.0.0.0:8443 --upstream=http://127.0.0.1:8080/ --logtostderr=true --v=0],WorkingDir:,Ports:[]ContainerPort{ContainerPort{Name:https,HostPort:0,ContainerPort:8443,Protocol:TCP,HostIP:,},},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{cpu: {{500 -3} {} 500m DecimalSI},memory: {{134217728 0} {} BinarySI},},Requests:ResourceList{cpu: {{5 -3} {} 5m DecimalSI},memory: {{67108864 0} {} BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kube-api-access-7kld4,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod telemetry-operator-controller-manager-58d5ff84df-t6tnb_openstack-operators(2418f025-d37e-4316-b804-9ab53c326c82): ErrImagePull: pull QPS exceeded" logger="UnhandledError" Dec 10 23:05:09 crc kubenswrapper[4791]: E1210 23:05:09.472290 4791 pod_workers.go:1301] "Error syncing pod, skipping" err="[failed to \"StartContainer\" for \"manager\" with ErrImagePull: \"pull QPS exceeded\", failed to \"StartContainer\" for \"kube-rbac-proxy\" with ErrImagePull: \"pull QPS exceeded\"]" pod="openstack-operators/octavia-operator-controller-manager-998648c74-sjp6q" podUID="23a95e98-8ae3-4ac4-945d-0cae5af5d1ac" Dec 10 23:05:09 crc kubenswrapper[4791]: E1210 23:05:09.472677 4791 pod_workers.go:1301] "Error syncing pod, skipping" err="[failed to \"StartContainer\" for \"manager\" with ErrImagePull: \"pull QPS exceeded\", failed to \"StartContainer\" for \"kube-rbac-proxy\" with ErrImagePull: \"pull QPS exceeded\"]" pod="openstack-operators/telemetry-operator-controller-manager-58d5ff84df-t6tnb" podUID="2418f025-d37e-4316-b804-9ab53c326c82" Dec 10 23:05:09 crc kubenswrapper[4791]: E1210 23:05:09.472785 4791 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:manager,Image:quay.io/openstack-k8s-operators/ovn-operator@sha256:635a4aef9d6f0b799e8ec91333dbb312160c001d05b3c63f614c124e0b67cb59,Command:[/manager],Args:[--leader-elect --health-probe-bind-address=:8081 --metrics-bind-address=127.0.0.1:8080],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:LEASE_DURATION,Value:30,ValueFrom:nil,},EnvVar{Name:RENEW_DEADLINE,Value:20,ValueFrom:nil,},EnvVar{Name:RETRY_PERIOD,Value:5,ValueFrom:nil,},EnvVar{Name:ENABLE_WEBHOOKS,Value:false,ValueFrom:nil,},EnvVar{Name:METRICS_CERTS,Value:false,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{cpu: {{500 -3} {} 500m DecimalSI},memory: {{536870912 0} {} BinarySI},},Requests:ResourceList{cpu: {{10 -3} {} 10m DecimalSI},memory: {{268435456 0} {} BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kube-api-access-gplnr,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/healthz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:15,TimeoutSeconds:1,PeriodSeconds:20,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},ReadinessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/readyz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:5,TimeoutSeconds:1,PeriodSeconds:10,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod ovn-operator-controller-manager-b6456fdb6-6rddm_openstack-operators(87c40827-309c-4c68-ae46-bbe449a9e66c): ErrImagePull: pull QPS exceeded" logger="UnhandledError" Dec 10 23:05:09 crc kubenswrapper[4791]: I1210 23:05:09.473312 4791 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/keystone-operator-controller-manager-7765d96ddf-2wsxl"] Dec 10 23:05:09 crc kubenswrapper[4791]: E1210 23:05:09.475051 4791 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:kube-rbac-proxy,Image:quay.io/openstack-k8s-operators/kube-rbac-proxy:v0.16.0,Command:[],Args:[--secure-listen-address=0.0.0.0:8443 --upstream=http://127.0.0.1:8080/ --logtostderr=true --v=0],WorkingDir:,Ports:[]ContainerPort{ContainerPort{Name:https,HostPort:0,ContainerPort:8443,Protocol:TCP,HostIP:,},},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{cpu: {{500 -3} {} 500m DecimalSI},memory: {{134217728 0} {} BinarySI},},Requests:ResourceList{cpu: {{5 -3} {} 5m DecimalSI},memory: {{67108864 0} {} BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kube-api-access-gplnr,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod ovn-operator-controller-manager-b6456fdb6-6rddm_openstack-operators(87c40827-309c-4c68-ae46-bbe449a9e66c): ErrImagePull: pull QPS exceeded" logger="UnhandledError" Dec 10 23:05:09 crc kubenswrapper[4791]: E1210 23:05:09.476227 4791 pod_workers.go:1301] "Error syncing pod, skipping" err="[failed to \"StartContainer\" for \"manager\" with ErrImagePull: \"pull QPS exceeded\", failed to \"StartContainer\" for \"kube-rbac-proxy\" with ErrImagePull: \"pull QPS exceeded\"]" pod="openstack-operators/ovn-operator-controller-manager-b6456fdb6-6rddm" podUID="87c40827-309c-4c68-ae46-bbe449a9e66c" Dec 10 23:05:09 crc kubenswrapper[4791]: I1210 23:05:09.479273 4791 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/swift-operator-controller-manager-9d58d64bc-8vwnp"] Dec 10 23:05:09 crc kubenswrapper[4791]: I1210 23:05:09.488298 4791 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/neutron-operator-controller-manager-5fdfd5b6b5-g8f76"] Dec 10 23:05:09 crc kubenswrapper[4791]: I1210 23:05:09.492314 4791 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/placement-operator-controller-manager-78f8948974-bhgvx"] Dec 10 23:05:09 crc kubenswrapper[4791]: I1210 23:05:09.497666 4791 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/octavia-operator-controller-manager-998648c74-sjp6q"] Dec 10 23:05:09 crc kubenswrapper[4791]: I1210 23:05:09.503548 4791 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/mariadb-operator-controller-manager-79c8c4686c-2wmm4"] Dec 10 23:05:09 crc kubenswrapper[4791]: I1210 23:05:09.508800 4791 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/ovn-operator-controller-manager-b6456fdb6-6rddm"] Dec 10 23:05:09 crc kubenswrapper[4791]: I1210 23:05:09.514051 4791 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/telemetry-operator-controller-manager-58d5ff84df-t6tnb"] Dec 10 23:05:09 crc kubenswrapper[4791]: I1210 23:05:09.641191 4791 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/watcher-operator-controller-manager-75944c9b7-7cdb7"] Dec 10 23:05:09 crc kubenswrapper[4791]: W1210 23:05:09.645144 4791 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod21e66f28_5ed9_475e_aa7d_a105ad1f2f9e.slice/crio-c5feed110a5b09384195588574b292bdeeb96bf87316a49276c673cefe6dbf96 WatchSource:0}: Error finding container c5feed110a5b09384195588574b292bdeeb96bf87316a49276c673cefe6dbf96: Status 404 returned error can't find the container with id c5feed110a5b09384195588574b292bdeeb96bf87316a49276c673cefe6dbf96 Dec 10 23:05:09 crc kubenswrapper[4791]: I1210 23:05:09.646906 4791 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/rabbitmq-cluster-operator-manager-668c99d594-2bgc5"] Dec 10 23:05:09 crc kubenswrapper[4791]: W1210 23:05:09.648706 4791 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podb3ae5e53_9077_4dd0_996d_2417f28b4736.slice/crio-6879828903d95ae73dd41b8115c5dc5abcdcfce68361d167b0a57085efa18db0 WatchSource:0}: Error finding container 6879828903d95ae73dd41b8115c5dc5abcdcfce68361d167b0a57085efa18db0: Status 404 returned error can't find the container with id 6879828903d95ae73dd41b8115c5dc5abcdcfce68361d167b0a57085efa18db0 Dec 10 23:05:09 crc kubenswrapper[4791]: E1210 23:05:09.652781 4791 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:operator,Image:quay.io/openstack-k8s-operators/rabbitmq-cluster-operator@sha256:893e66303c1b0bc1d00a299a3f0380bad55c8dc813c8a1c6a4aab379f5aa12a2,Command:[/manager],Args:[],WorkingDir:,Ports:[]ContainerPort{ContainerPort{Name:metrics,HostPort:0,ContainerPort:9782,Protocol:TCP,HostIP:,},},Env:[]EnvVar{EnvVar{Name:OPERATOR_NAMESPACE,Value:,ValueFrom:&EnvVarSource{FieldRef:&ObjectFieldSelector{APIVersion:v1,FieldPath:metadata.namespace,},ResourceFieldRef:nil,ConfigMapKeyRef:nil,SecretKeyRef:nil,},},EnvVar{Name:LEASE_DURATION,Value:30,ValueFrom:nil,},EnvVar{Name:RENEW_DEADLINE,Value:20,ValueFrom:nil,},EnvVar{Name:RETRY_PERIOD,Value:5,ValueFrom:nil,},EnvVar{Name:ENABLE_WEBHOOKS,Value:false,ValueFrom:nil,},EnvVar{Name:METRICS_CERTS,Value:false,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{cpu: {{200 -3} {} 200m DecimalSI},memory: {{524288000 0} {} 500Mi BinarySI},},Requests:ResourceList{cpu: {{5 -3} {} 5m DecimalSI},memory: {{67108864 0} {} BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kube-api-access-7sjtw,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000660000,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod rabbitmq-cluster-operator-manager-668c99d594-2bgc5_openstack-operators(b3ae5e53-9077-4dd0-996d-2417f28b4736): ErrImagePull: pull QPS exceeded" logger="UnhandledError" Dec 10 23:05:09 crc kubenswrapper[4791]: E1210 23:05:09.653981 4791 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"operator\" with ErrImagePull: \"pull QPS exceeded\"" pod="openstack-operators/rabbitmq-cluster-operator-manager-668c99d594-2bgc5" podUID="b3ae5e53-9077-4dd0-996d-2417f28b4736" Dec 10 23:05:09 crc kubenswrapper[4791]: I1210 23:05:09.691486 4791 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/d5907af9-fe3f-4da0-ba70-993d36b25746-cert\") pod \"openstack-baremetal-operator-controller-manager-84b575879fv4t28\" (UID: \"d5907af9-fe3f-4da0-ba70-993d36b25746\") " pod="openstack-operators/openstack-baremetal-operator-controller-manager-84b575879fv4t28" Dec 10 23:05:09 crc kubenswrapper[4791]: E1210 23:05:09.691670 4791 secret.go:188] Couldn't get secret openstack-operators/openstack-baremetal-operator-webhook-server-cert: secret "openstack-baremetal-operator-webhook-server-cert" not found Dec 10 23:05:09 crc kubenswrapper[4791]: E1210 23:05:09.691750 4791 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/d5907af9-fe3f-4da0-ba70-993d36b25746-cert podName:d5907af9-fe3f-4da0-ba70-993d36b25746 nodeName:}" failed. No retries permitted until 2025-12-10 23:05:11.691730687 +0000 UTC m=+946.121348300 (durationBeforeRetry 2s). Error: MountVolume.SetUp failed for volume "cert" (UniqueName: "kubernetes.io/secret/d5907af9-fe3f-4da0-ba70-993d36b25746-cert") pod "openstack-baremetal-operator-controller-manager-84b575879fv4t28" (UID: "d5907af9-fe3f-4da0-ba70-993d36b25746") : secret "openstack-baremetal-operator-webhook-server-cert" not found Dec 10 23:05:09 crc kubenswrapper[4791]: I1210 23:05:09.793210 4791 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"webhook-certs\" (UniqueName: \"kubernetes.io/secret/e2c4ecea-2852-4e39-93fb-aee7cbe31aa4-webhook-certs\") pod \"openstack-operator-controller-manager-686dfd865c-cft5f\" (UID: \"e2c4ecea-2852-4e39-93fb-aee7cbe31aa4\") " pod="openstack-operators/openstack-operator-controller-manager-686dfd865c-cft5f" Dec 10 23:05:09 crc kubenswrapper[4791]: I1210 23:05:09.793309 4791 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/e2c4ecea-2852-4e39-93fb-aee7cbe31aa4-metrics-certs\") pod \"openstack-operator-controller-manager-686dfd865c-cft5f\" (UID: \"e2c4ecea-2852-4e39-93fb-aee7cbe31aa4\") " pod="openstack-operators/openstack-operator-controller-manager-686dfd865c-cft5f" Dec 10 23:05:09 crc kubenswrapper[4791]: E1210 23:05:09.793520 4791 secret.go:188] Couldn't get secret openstack-operators/metrics-server-cert: secret "metrics-server-cert" not found Dec 10 23:05:09 crc kubenswrapper[4791]: E1210 23:05:09.793594 4791 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/e2c4ecea-2852-4e39-93fb-aee7cbe31aa4-metrics-certs podName:e2c4ecea-2852-4e39-93fb-aee7cbe31aa4 nodeName:}" failed. No retries permitted until 2025-12-10 23:05:11.793578475 +0000 UTC m=+946.223196098 (durationBeforeRetry 2s). Error: MountVolume.SetUp failed for volume "metrics-certs" (UniqueName: "kubernetes.io/secret/e2c4ecea-2852-4e39-93fb-aee7cbe31aa4-metrics-certs") pod "openstack-operator-controller-manager-686dfd865c-cft5f" (UID: "e2c4ecea-2852-4e39-93fb-aee7cbe31aa4") : secret "metrics-server-cert" not found Dec 10 23:05:09 crc kubenswrapper[4791]: E1210 23:05:09.793648 4791 secret.go:188] Couldn't get secret openstack-operators/webhook-server-cert: secret "webhook-server-cert" not found Dec 10 23:05:09 crc kubenswrapper[4791]: E1210 23:05:09.793674 4791 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/e2c4ecea-2852-4e39-93fb-aee7cbe31aa4-webhook-certs podName:e2c4ecea-2852-4e39-93fb-aee7cbe31aa4 nodeName:}" failed. No retries permitted until 2025-12-10 23:05:11.793665207 +0000 UTC m=+946.223282830 (durationBeforeRetry 2s). Error: MountVolume.SetUp failed for volume "webhook-certs" (UniqueName: "kubernetes.io/secret/e2c4ecea-2852-4e39-93fb-aee7cbe31aa4-webhook-certs") pod "openstack-operator-controller-manager-686dfd865c-cft5f" (UID: "e2c4ecea-2852-4e39-93fb-aee7cbe31aa4") : secret "webhook-server-cert" not found Dec 10 23:05:09 crc kubenswrapper[4791]: I1210 23:05:09.820776 4791 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/test-operator-controller-manager-5854674fcc-rw7ct" event={"ID":"bda40eaa-446c-475e-8f2f-76f76caa0b85","Type":"ContainerStarted","Data":"1da9d266d8455a7279fcfa9c1ce2e6726b5c46627dd796a670b573c42bd940fd"} Dec 10 23:05:09 crc kubenswrapper[4791]: I1210 23:05:09.824860 4791 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/rabbitmq-cluster-operator-manager-668c99d594-2bgc5" event={"ID":"b3ae5e53-9077-4dd0-996d-2417f28b4736","Type":"ContainerStarted","Data":"6879828903d95ae73dd41b8115c5dc5abcdcfce68361d167b0a57085efa18db0"} Dec 10 23:05:09 crc kubenswrapper[4791]: E1210 23:05:09.826109 4791 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"operator\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/rabbitmq-cluster-operator@sha256:893e66303c1b0bc1d00a299a3f0380bad55c8dc813c8a1c6a4aab379f5aa12a2\\\"\"" pod="openstack-operators/rabbitmq-cluster-operator-manager-668c99d594-2bgc5" podUID="b3ae5e53-9077-4dd0-996d-2417f28b4736" Dec 10 23:05:09 crc kubenswrapper[4791]: I1210 23:05:09.830187 4791 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/ironic-operator-controller-manager-967d97867-r6jdv" event={"ID":"cb5fb221-76b0-48b4-ad27-7d0c51581ec4","Type":"ContainerStarted","Data":"d4c1f3513a8f66aa27e95d58d2fad1de7e57774580159c65ba7381e973f5d70f"} Dec 10 23:05:09 crc kubenswrapper[4791]: I1210 23:05:09.831746 4791 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/telemetry-operator-controller-manager-58d5ff84df-t6tnb" event={"ID":"2418f025-d37e-4316-b804-9ab53c326c82","Type":"ContainerStarted","Data":"ae9606448723b723d2890a2a45ff5caf7fd55fbbc240a379f4884ef1e21333de"} Dec 10 23:05:09 crc kubenswrapper[4791]: I1210 23:05:09.832874 4791 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/watcher-operator-controller-manager-75944c9b7-7cdb7" event={"ID":"21e66f28-5ed9-475e-aa7d-a105ad1f2f9e","Type":"ContainerStarted","Data":"c5feed110a5b09384195588574b292bdeeb96bf87316a49276c673cefe6dbf96"} Dec 10 23:05:09 crc kubenswrapper[4791]: I1210 23:05:09.834235 4791 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/octavia-operator-controller-manager-998648c74-sjp6q" event={"ID":"23a95e98-8ae3-4ac4-945d-0cae5af5d1ac","Type":"ContainerStarted","Data":"53b7ef95db9e2d9b2b9ff3444ab5155ce6c932447839e9358cfe33aef7b56bec"} Dec 10 23:05:09 crc kubenswrapper[4791]: E1210 23:05:09.834898 4791 pod_workers.go:1301] "Error syncing pod, skipping" err="[failed to \"StartContainer\" for \"manager\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/telemetry-operator@sha256:f27e732ec1faee765461bf137d9be81278b2fa39675019a73622755e1e610b6f\\\"\", failed to \"StartContainer\" for \"kube-rbac-proxy\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/kube-rbac-proxy:v0.16.0\\\"\"]" pod="openstack-operators/telemetry-operator-controller-manager-58d5ff84df-t6tnb" podUID="2418f025-d37e-4316-b804-9ab53c326c82" Dec 10 23:05:09 crc kubenswrapper[4791]: I1210 23:05:09.835779 4791 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/keystone-operator-controller-manager-7765d96ddf-2wsxl" event={"ID":"311d4e7c-a0c7-47f6-8deb-7262068e8beb","Type":"ContainerStarted","Data":"18e2fd8d6b56af75d3e5341c841fb32850c7204a22af8b787afb6f03cc8e59ec"} Dec 10 23:05:09 crc kubenswrapper[4791]: E1210 23:05:09.836480 4791 pod_workers.go:1301] "Error syncing pod, skipping" err="[failed to \"StartContainer\" for \"manager\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/octavia-operator@sha256:d9a3694865a7d54ee96397add18c3898886e98d079aa20876a0f4de1fa7a7168\\\"\", failed to \"StartContainer\" for \"kube-rbac-proxy\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/kube-rbac-proxy:v0.16.0\\\"\"]" pod="openstack-operators/octavia-operator-controller-manager-998648c74-sjp6q" podUID="23a95e98-8ae3-4ac4-945d-0cae5af5d1ac" Dec 10 23:05:09 crc kubenswrapper[4791]: I1210 23:05:09.838111 4791 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/glance-operator-controller-manager-5697bb5779-cvbqc" event={"ID":"f0b16aa2-e60e-4f0e-9679-9afc0a5ae027","Type":"ContainerStarted","Data":"049e73adc710d2ded6581c0c73ef3b87dd5c92c8a32bff101315f137f09914ca"} Dec 10 23:05:09 crc kubenswrapper[4791]: I1210 23:05:09.851637 4791 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/placement-operator-controller-manager-78f8948974-bhgvx" event={"ID":"b67ef71c-85c5-41ca-bf93-d1ca8a2fd007","Type":"ContainerStarted","Data":"89196ca78d764e19337eeb5c14cd87fc7273374f1e7a8e5e1a93d1ce07e89ed8"} Dec 10 23:05:09 crc kubenswrapper[4791]: I1210 23:05:09.853895 4791 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/nova-operator-controller-manager-697bc559fc-2lkcl" event={"ID":"3b2af4ca-48b4-463f-b0aa-c5cfd9099d67","Type":"ContainerStarted","Data":"ad518bc11a6b8e6512cbd1b38b64a308b7e42d9d336f637dde61973d6660f9bd"} Dec 10 23:05:09 crc kubenswrapper[4791]: I1210 23:05:09.855477 4791 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/designate-operator-controller-manager-697fb699cf-vg956" event={"ID":"19bf0bfe-4e8b-47a4-a919-4227b3e47bb3","Type":"ContainerStarted","Data":"6b74b19daeff9eee867efa33ec405543774e3bfc614ae2b08ce6702e719fd7de"} Dec 10 23:05:09 crc kubenswrapper[4791]: I1210 23:05:09.856525 4791 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/ovn-operator-controller-manager-b6456fdb6-6rddm" event={"ID":"87c40827-309c-4c68-ae46-bbe449a9e66c","Type":"ContainerStarted","Data":"e6915d6501d6d8b7a6d7828c5a663656751a928f90efce28f9ab97982bb0c36a"} Dec 10 23:05:09 crc kubenswrapper[4791]: I1210 23:05:09.862490 4791 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/horizon-operator-controller-manager-68c6d99b8f-2lq47" event={"ID":"e3fdc812-fbd3-4d35-8555-d55e9390ebd0","Type":"ContainerStarted","Data":"cdb375d15192c73607b0ae64fd2f69d99f45177b4d782244a373020eb0444ce4"} Dec 10 23:05:09 crc kubenswrapper[4791]: E1210 23:05:09.862600 4791 pod_workers.go:1301] "Error syncing pod, skipping" err="[failed to \"StartContainer\" for \"manager\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/ovn-operator@sha256:635a4aef9d6f0b799e8ec91333dbb312160c001d05b3c63f614c124e0b67cb59\\\"\", failed to \"StartContainer\" for \"kube-rbac-proxy\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/kube-rbac-proxy:v0.16.0\\\"\"]" pod="openstack-operators/ovn-operator-controller-manager-b6456fdb6-6rddm" podUID="87c40827-309c-4c68-ae46-bbe449a9e66c" Dec 10 23:05:09 crc kubenswrapper[4791]: I1210 23:05:09.864367 4791 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/neutron-operator-controller-manager-5fdfd5b6b5-g8f76" event={"ID":"86bbb7b2-5f55-46a2-b426-013ba183977f","Type":"ContainerStarted","Data":"d279d88285b984b85594523584d3ce07d0118717bef801c315a773e111eeaac5"} Dec 10 23:05:09 crc kubenswrapper[4791]: I1210 23:05:09.865555 4791 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/manila-operator-controller-manager-5b5fd79c9c-tsbtp" event={"ID":"12b75f38-fb6a-4020-95b9-29c7dd872849","Type":"ContainerStarted","Data":"86383d168b0796a94b1c08d76141bc57f38111fa8b23efb28b5693d38533c9c0"} Dec 10 23:05:09 crc kubenswrapper[4791]: I1210 23:05:09.866725 4791 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/swift-operator-controller-manager-9d58d64bc-8vwnp" event={"ID":"4cf9497a-4193-4d34-a68a-48085af6e8b5","Type":"ContainerStarted","Data":"1ccb5c61702064b875151444c40160c34c2d4d5a04fe4515a9a72b114a540c42"} Dec 10 23:05:09 crc kubenswrapper[4791]: I1210 23:05:09.870584 4791 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/mariadb-operator-controller-manager-79c8c4686c-2wmm4" event={"ID":"590c13cc-efa0-431a-9f2f-65fb97c1bd3c","Type":"ContainerStarted","Data":"b359013c39cadb1598e37b48c0bb26e3cf1c07b99353d132ccc11db1aa5f65d5"} Dec 10 23:05:09 crc kubenswrapper[4791]: E1210 23:05:09.872695 4791 pod_workers.go:1301] "Error syncing pod, skipping" err="[failed to \"StartContainer\" for \"manager\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/mariadb-operator@sha256:424da951f13f1fbe9083215dc9f5088f90676dd813f01fdf3c1a8639b61cbaad\\\"\", failed to \"StartContainer\" for \"kube-rbac-proxy\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/kube-rbac-proxy:v0.16.0\\\"\"]" pod="openstack-operators/mariadb-operator-controller-manager-79c8c4686c-2wmm4" podUID="590c13cc-efa0-431a-9f2f-65fb97c1bd3c" Dec 10 23:05:10 crc kubenswrapper[4791]: E1210 23:05:10.878751 4791 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"operator\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/rabbitmq-cluster-operator@sha256:893e66303c1b0bc1d00a299a3f0380bad55c8dc813c8a1c6a4aab379f5aa12a2\\\"\"" pod="openstack-operators/rabbitmq-cluster-operator-manager-668c99d594-2bgc5" podUID="b3ae5e53-9077-4dd0-996d-2417f28b4736" Dec 10 23:05:10 crc kubenswrapper[4791]: E1210 23:05:10.879297 4791 pod_workers.go:1301] "Error syncing pod, skipping" err="[failed to \"StartContainer\" for \"manager\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/ovn-operator@sha256:635a4aef9d6f0b799e8ec91333dbb312160c001d05b3c63f614c124e0b67cb59\\\"\", failed to \"StartContainer\" for \"kube-rbac-proxy\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/kube-rbac-proxy:v0.16.0\\\"\"]" pod="openstack-operators/ovn-operator-controller-manager-b6456fdb6-6rddm" podUID="87c40827-309c-4c68-ae46-bbe449a9e66c" Dec 10 23:05:10 crc kubenswrapper[4791]: E1210 23:05:10.879790 4791 pod_workers.go:1301] "Error syncing pod, skipping" err="[failed to \"StartContainer\" for \"manager\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/telemetry-operator@sha256:f27e732ec1faee765461bf137d9be81278b2fa39675019a73622755e1e610b6f\\\"\", failed to \"StartContainer\" for \"kube-rbac-proxy\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/kube-rbac-proxy:v0.16.0\\\"\"]" pod="openstack-operators/telemetry-operator-controller-manager-58d5ff84df-t6tnb" podUID="2418f025-d37e-4316-b804-9ab53c326c82" Dec 10 23:05:10 crc kubenswrapper[4791]: E1210 23:05:10.880439 4791 pod_workers.go:1301] "Error syncing pod, skipping" err="[failed to \"StartContainer\" for \"manager\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/octavia-operator@sha256:d9a3694865a7d54ee96397add18c3898886e98d079aa20876a0f4de1fa7a7168\\\"\", failed to \"StartContainer\" for \"kube-rbac-proxy\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/kube-rbac-proxy:v0.16.0\\\"\"]" pod="openstack-operators/octavia-operator-controller-manager-998648c74-sjp6q" podUID="23a95e98-8ae3-4ac4-945d-0cae5af5d1ac" Dec 10 23:05:10 crc kubenswrapper[4791]: E1210 23:05:10.880573 4791 pod_workers.go:1301] "Error syncing pod, skipping" err="[failed to \"StartContainer\" for \"manager\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/mariadb-operator@sha256:424da951f13f1fbe9083215dc9f5088f90676dd813f01fdf3c1a8639b61cbaad\\\"\", failed to \"StartContainer\" for \"kube-rbac-proxy\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/kube-rbac-proxy:v0.16.0\\\"\"]" pod="openstack-operators/mariadb-operator-controller-manager-79c8c4686c-2wmm4" podUID="590c13cc-efa0-431a-9f2f-65fb97c1bd3c" Dec 10 23:05:11 crc kubenswrapper[4791]: I1210 23:05:11.122473 4791 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/416b455d-b397-4aad-baf1-88e880619eb5-cert\") pod \"infra-operator-controller-manager-78d48bff9d-mmczn\" (UID: \"416b455d-b397-4aad-baf1-88e880619eb5\") " pod="openstack-operators/infra-operator-controller-manager-78d48bff9d-mmczn" Dec 10 23:05:11 crc kubenswrapper[4791]: E1210 23:05:11.122658 4791 secret.go:188] Couldn't get secret openstack-operators/infra-operator-webhook-server-cert: secret "infra-operator-webhook-server-cert" not found Dec 10 23:05:11 crc kubenswrapper[4791]: E1210 23:05:11.122723 4791 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/416b455d-b397-4aad-baf1-88e880619eb5-cert podName:416b455d-b397-4aad-baf1-88e880619eb5 nodeName:}" failed. No retries permitted until 2025-12-10 23:05:15.122706484 +0000 UTC m=+949.552324097 (durationBeforeRetry 4s). Error: MountVolume.SetUp failed for volume "cert" (UniqueName: "kubernetes.io/secret/416b455d-b397-4aad-baf1-88e880619eb5-cert") pod "infra-operator-controller-manager-78d48bff9d-mmczn" (UID: "416b455d-b397-4aad-baf1-88e880619eb5") : secret "infra-operator-webhook-server-cert" not found Dec 10 23:05:11 crc kubenswrapper[4791]: I1210 23:05:11.740182 4791 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/d5907af9-fe3f-4da0-ba70-993d36b25746-cert\") pod \"openstack-baremetal-operator-controller-manager-84b575879fv4t28\" (UID: \"d5907af9-fe3f-4da0-ba70-993d36b25746\") " pod="openstack-operators/openstack-baremetal-operator-controller-manager-84b575879fv4t28" Dec 10 23:05:11 crc kubenswrapper[4791]: E1210 23:05:11.740386 4791 secret.go:188] Couldn't get secret openstack-operators/openstack-baremetal-operator-webhook-server-cert: secret "openstack-baremetal-operator-webhook-server-cert" not found Dec 10 23:05:11 crc kubenswrapper[4791]: E1210 23:05:11.740475 4791 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/d5907af9-fe3f-4da0-ba70-993d36b25746-cert podName:d5907af9-fe3f-4da0-ba70-993d36b25746 nodeName:}" failed. No retries permitted until 2025-12-10 23:05:15.740455475 +0000 UTC m=+950.170073088 (durationBeforeRetry 4s). Error: MountVolume.SetUp failed for volume "cert" (UniqueName: "kubernetes.io/secret/d5907af9-fe3f-4da0-ba70-993d36b25746-cert") pod "openstack-baremetal-operator-controller-manager-84b575879fv4t28" (UID: "d5907af9-fe3f-4da0-ba70-993d36b25746") : secret "openstack-baremetal-operator-webhook-server-cert" not found Dec 10 23:05:11 crc kubenswrapper[4791]: I1210 23:05:11.841527 4791 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"webhook-certs\" (UniqueName: \"kubernetes.io/secret/e2c4ecea-2852-4e39-93fb-aee7cbe31aa4-webhook-certs\") pod \"openstack-operator-controller-manager-686dfd865c-cft5f\" (UID: \"e2c4ecea-2852-4e39-93fb-aee7cbe31aa4\") " pod="openstack-operators/openstack-operator-controller-manager-686dfd865c-cft5f" Dec 10 23:05:11 crc kubenswrapper[4791]: I1210 23:05:11.841585 4791 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/e2c4ecea-2852-4e39-93fb-aee7cbe31aa4-metrics-certs\") pod \"openstack-operator-controller-manager-686dfd865c-cft5f\" (UID: \"e2c4ecea-2852-4e39-93fb-aee7cbe31aa4\") " pod="openstack-operators/openstack-operator-controller-manager-686dfd865c-cft5f" Dec 10 23:05:11 crc kubenswrapper[4791]: E1210 23:05:11.841717 4791 secret.go:188] Couldn't get secret openstack-operators/metrics-server-cert: secret "metrics-server-cert" not found Dec 10 23:05:11 crc kubenswrapper[4791]: E1210 23:05:11.841741 4791 secret.go:188] Couldn't get secret openstack-operators/webhook-server-cert: secret "webhook-server-cert" not found Dec 10 23:05:11 crc kubenswrapper[4791]: E1210 23:05:11.841762 4791 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/e2c4ecea-2852-4e39-93fb-aee7cbe31aa4-metrics-certs podName:e2c4ecea-2852-4e39-93fb-aee7cbe31aa4 nodeName:}" failed. No retries permitted until 2025-12-10 23:05:15.841748587 +0000 UTC m=+950.271366200 (durationBeforeRetry 4s). Error: MountVolume.SetUp failed for volume "metrics-certs" (UniqueName: "kubernetes.io/secret/e2c4ecea-2852-4e39-93fb-aee7cbe31aa4-metrics-certs") pod "openstack-operator-controller-manager-686dfd865c-cft5f" (UID: "e2c4ecea-2852-4e39-93fb-aee7cbe31aa4") : secret "metrics-server-cert" not found Dec 10 23:05:11 crc kubenswrapper[4791]: E1210 23:05:11.841829 4791 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/e2c4ecea-2852-4e39-93fb-aee7cbe31aa4-webhook-certs podName:e2c4ecea-2852-4e39-93fb-aee7cbe31aa4 nodeName:}" failed. No retries permitted until 2025-12-10 23:05:15.841816069 +0000 UTC m=+950.271433682 (durationBeforeRetry 4s). Error: MountVolume.SetUp failed for volume "webhook-certs" (UniqueName: "kubernetes.io/secret/e2c4ecea-2852-4e39-93fb-aee7cbe31aa4-webhook-certs") pod "openstack-operator-controller-manager-686dfd865c-cft5f" (UID: "e2c4ecea-2852-4e39-93fb-aee7cbe31aa4") : secret "webhook-server-cert" not found Dec 10 23:05:15 crc kubenswrapper[4791]: I1210 23:05:15.123592 4791 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/416b455d-b397-4aad-baf1-88e880619eb5-cert\") pod \"infra-operator-controller-manager-78d48bff9d-mmczn\" (UID: \"416b455d-b397-4aad-baf1-88e880619eb5\") " pod="openstack-operators/infra-operator-controller-manager-78d48bff9d-mmczn" Dec 10 23:05:15 crc kubenswrapper[4791]: E1210 23:05:15.123776 4791 secret.go:188] Couldn't get secret openstack-operators/infra-operator-webhook-server-cert: secret "infra-operator-webhook-server-cert" not found Dec 10 23:05:15 crc kubenswrapper[4791]: E1210 23:05:15.124094 4791 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/416b455d-b397-4aad-baf1-88e880619eb5-cert podName:416b455d-b397-4aad-baf1-88e880619eb5 nodeName:}" failed. No retries permitted until 2025-12-10 23:05:23.1240632 +0000 UTC m=+957.553680823 (durationBeforeRetry 8s). Error: MountVolume.SetUp failed for volume "cert" (UniqueName: "kubernetes.io/secret/416b455d-b397-4aad-baf1-88e880619eb5-cert") pod "infra-operator-controller-manager-78d48bff9d-mmczn" (UID: "416b455d-b397-4aad-baf1-88e880619eb5") : secret "infra-operator-webhook-server-cert" not found Dec 10 23:05:15 crc kubenswrapper[4791]: I1210 23:05:15.836188 4791 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/d5907af9-fe3f-4da0-ba70-993d36b25746-cert\") pod \"openstack-baremetal-operator-controller-manager-84b575879fv4t28\" (UID: \"d5907af9-fe3f-4da0-ba70-993d36b25746\") " pod="openstack-operators/openstack-baremetal-operator-controller-manager-84b575879fv4t28" Dec 10 23:05:15 crc kubenswrapper[4791]: E1210 23:05:15.836375 4791 secret.go:188] Couldn't get secret openstack-operators/openstack-baremetal-operator-webhook-server-cert: secret "openstack-baremetal-operator-webhook-server-cert" not found Dec 10 23:05:15 crc kubenswrapper[4791]: E1210 23:05:15.836455 4791 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/d5907af9-fe3f-4da0-ba70-993d36b25746-cert podName:d5907af9-fe3f-4da0-ba70-993d36b25746 nodeName:}" failed. No retries permitted until 2025-12-10 23:05:23.836434777 +0000 UTC m=+958.266052390 (durationBeforeRetry 8s). Error: MountVolume.SetUp failed for volume "cert" (UniqueName: "kubernetes.io/secret/d5907af9-fe3f-4da0-ba70-993d36b25746-cert") pod "openstack-baremetal-operator-controller-manager-84b575879fv4t28" (UID: "d5907af9-fe3f-4da0-ba70-993d36b25746") : secret "openstack-baremetal-operator-webhook-server-cert" not found Dec 10 23:05:15 crc kubenswrapper[4791]: I1210 23:05:15.937901 4791 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"webhook-certs\" (UniqueName: \"kubernetes.io/secret/e2c4ecea-2852-4e39-93fb-aee7cbe31aa4-webhook-certs\") pod \"openstack-operator-controller-manager-686dfd865c-cft5f\" (UID: \"e2c4ecea-2852-4e39-93fb-aee7cbe31aa4\") " pod="openstack-operators/openstack-operator-controller-manager-686dfd865c-cft5f" Dec 10 23:05:15 crc kubenswrapper[4791]: I1210 23:05:15.937997 4791 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/e2c4ecea-2852-4e39-93fb-aee7cbe31aa4-metrics-certs\") pod \"openstack-operator-controller-manager-686dfd865c-cft5f\" (UID: \"e2c4ecea-2852-4e39-93fb-aee7cbe31aa4\") " pod="openstack-operators/openstack-operator-controller-manager-686dfd865c-cft5f" Dec 10 23:05:15 crc kubenswrapper[4791]: E1210 23:05:15.938243 4791 secret.go:188] Couldn't get secret openstack-operators/metrics-server-cert: secret "metrics-server-cert" not found Dec 10 23:05:15 crc kubenswrapper[4791]: E1210 23:05:15.938243 4791 secret.go:188] Couldn't get secret openstack-operators/webhook-server-cert: secret "webhook-server-cert" not found Dec 10 23:05:15 crc kubenswrapper[4791]: E1210 23:05:15.938298 4791 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/e2c4ecea-2852-4e39-93fb-aee7cbe31aa4-metrics-certs podName:e2c4ecea-2852-4e39-93fb-aee7cbe31aa4 nodeName:}" failed. No retries permitted until 2025-12-10 23:05:23.938279025 +0000 UTC m=+958.367896638 (durationBeforeRetry 8s). Error: MountVolume.SetUp failed for volume "metrics-certs" (UniqueName: "kubernetes.io/secret/e2c4ecea-2852-4e39-93fb-aee7cbe31aa4-metrics-certs") pod "openstack-operator-controller-manager-686dfd865c-cft5f" (UID: "e2c4ecea-2852-4e39-93fb-aee7cbe31aa4") : secret "metrics-server-cert" not found Dec 10 23:05:15 crc kubenswrapper[4791]: E1210 23:05:15.938423 4791 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/e2c4ecea-2852-4e39-93fb-aee7cbe31aa4-webhook-certs podName:e2c4ecea-2852-4e39-93fb-aee7cbe31aa4 nodeName:}" failed. No retries permitted until 2025-12-10 23:05:23.938399378 +0000 UTC m=+958.368016991 (durationBeforeRetry 8s). Error: MountVolume.SetUp failed for volume "webhook-certs" (UniqueName: "kubernetes.io/secret/e2c4ecea-2852-4e39-93fb-aee7cbe31aa4-webhook-certs") pod "openstack-operator-controller-manager-686dfd865c-cft5f" (UID: "e2c4ecea-2852-4e39-93fb-aee7cbe31aa4") : secret "webhook-server-cert" not found Dec 10 23:05:21 crc kubenswrapper[4791]: E1210 23:05:21.382354 4791 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="quay.io/openstack-k8s-operators/designate-operator@sha256:900050d3501c0785b227db34b89883efe68247816e5c7427cacb74f8aa10605a" Dec 10 23:05:21 crc kubenswrapper[4791]: E1210 23:05:21.382838 4791 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:manager,Image:quay.io/openstack-k8s-operators/designate-operator@sha256:900050d3501c0785b227db34b89883efe68247816e5c7427cacb74f8aa10605a,Command:[/manager],Args:[--leader-elect --health-probe-bind-address=:8081 --metrics-bind-address=127.0.0.1:8080],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:LEASE_DURATION,Value:30,ValueFrom:nil,},EnvVar{Name:RENEW_DEADLINE,Value:20,ValueFrom:nil,},EnvVar{Name:RETRY_PERIOD,Value:5,ValueFrom:nil,},EnvVar{Name:ENABLE_WEBHOOKS,Value:false,ValueFrom:nil,},EnvVar{Name:METRICS_CERTS,Value:false,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{cpu: {{500 -3} {} 500m DecimalSI},memory: {{536870912 0} {} BinarySI},},Requests:ResourceList{cpu: {{10 -3} {} 10m DecimalSI},memory: {{268435456 0} {} BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kube-api-access-pwzfm,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/healthz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:15,TimeoutSeconds:1,PeriodSeconds:20,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},ReadinessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/readyz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:5,TimeoutSeconds:1,PeriodSeconds:10,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod designate-operator-controller-manager-697fb699cf-vg956_openstack-operators(19bf0bfe-4e8b-47a4-a919-4227b3e47bb3): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError" Dec 10 23:05:22 crc kubenswrapper[4791]: E1210 23:05:22.218887 4791 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="quay.io/openstack-k8s-operators/placement-operator@sha256:d29650b006da97eb9178fcc58f2eb9fead8c2b414fac18f86a3c3a1507488c4f" Dec 10 23:05:22 crc kubenswrapper[4791]: E1210 23:05:22.219133 4791 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:manager,Image:quay.io/openstack-k8s-operators/placement-operator@sha256:d29650b006da97eb9178fcc58f2eb9fead8c2b414fac18f86a3c3a1507488c4f,Command:[/manager],Args:[--leader-elect --health-probe-bind-address=:8081 --metrics-bind-address=127.0.0.1:8080],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:LEASE_DURATION,Value:30,ValueFrom:nil,},EnvVar{Name:RENEW_DEADLINE,Value:20,ValueFrom:nil,},EnvVar{Name:RETRY_PERIOD,Value:5,ValueFrom:nil,},EnvVar{Name:ENABLE_WEBHOOKS,Value:false,ValueFrom:nil,},EnvVar{Name:METRICS_CERTS,Value:false,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{cpu: {{500 -3} {} 500m DecimalSI},memory: {{536870912 0} {} BinarySI},},Requests:ResourceList{cpu: {{10 -3} {} 10m DecimalSI},memory: {{268435456 0} {} BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kube-api-access-7p5x5,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/healthz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:15,TimeoutSeconds:1,PeriodSeconds:20,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},ReadinessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/readyz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:5,TimeoutSeconds:1,PeriodSeconds:10,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod placement-operator-controller-manager-78f8948974-bhgvx_openstack-operators(b67ef71c-85c5-41ca-bf93-d1ca8a2fd007): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError" Dec 10 23:05:22 crc kubenswrapper[4791]: E1210 23:05:22.803863 4791 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="quay.io/openstack-k8s-operators/test-operator@sha256:101b3e007d8c9f2e183262d7712f986ad51256448099069bc14f1ea5f997ab94" Dec 10 23:05:22 crc kubenswrapper[4791]: E1210 23:05:22.804062 4791 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:manager,Image:quay.io/openstack-k8s-operators/test-operator@sha256:101b3e007d8c9f2e183262d7712f986ad51256448099069bc14f1ea5f997ab94,Command:[/manager],Args:[--leader-elect --health-probe-bind-address=:8081 --metrics-bind-address=127.0.0.1:8080],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:LEASE_DURATION,Value:30,ValueFrom:nil,},EnvVar{Name:RENEW_DEADLINE,Value:20,ValueFrom:nil,},EnvVar{Name:RETRY_PERIOD,Value:5,ValueFrom:nil,},EnvVar{Name:ENABLE_WEBHOOKS,Value:false,ValueFrom:nil,},EnvVar{Name:METRICS_CERTS,Value:false,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{cpu: {{500 -3} {} 500m DecimalSI},memory: {{536870912 0} {} BinarySI},},Requests:ResourceList{cpu: {{10 -3} {} 10m DecimalSI},memory: {{268435456 0} {} BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kube-api-access-mk2hf,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/healthz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:15,TimeoutSeconds:1,PeriodSeconds:20,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},ReadinessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/readyz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:5,TimeoutSeconds:1,PeriodSeconds:10,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod test-operator-controller-manager-5854674fcc-rw7ct_openstack-operators(bda40eaa-446c-475e-8f2f-76f76caa0b85): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError" Dec 10 23:05:23 crc kubenswrapper[4791]: I1210 23:05:23.177064 4791 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/416b455d-b397-4aad-baf1-88e880619eb5-cert\") pod \"infra-operator-controller-manager-78d48bff9d-mmczn\" (UID: \"416b455d-b397-4aad-baf1-88e880619eb5\") " pod="openstack-operators/infra-operator-controller-manager-78d48bff9d-mmczn" Dec 10 23:05:23 crc kubenswrapper[4791]: E1210 23:05:23.177197 4791 secret.go:188] Couldn't get secret openstack-operators/infra-operator-webhook-server-cert: secret "infra-operator-webhook-server-cert" not found Dec 10 23:05:23 crc kubenswrapper[4791]: E1210 23:05:23.177247 4791 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/416b455d-b397-4aad-baf1-88e880619eb5-cert podName:416b455d-b397-4aad-baf1-88e880619eb5 nodeName:}" failed. No retries permitted until 2025-12-10 23:05:39.177232997 +0000 UTC m=+973.606850610 (durationBeforeRetry 16s). Error: MountVolume.SetUp failed for volume "cert" (UniqueName: "kubernetes.io/secret/416b455d-b397-4aad-baf1-88e880619eb5-cert") pod "infra-operator-controller-manager-78d48bff9d-mmczn" (UID: "416b455d-b397-4aad-baf1-88e880619eb5") : secret "infra-operator-webhook-server-cert" not found Dec 10 23:05:23 crc kubenswrapper[4791]: I1210 23:05:23.923896 4791 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/d5907af9-fe3f-4da0-ba70-993d36b25746-cert\") pod \"openstack-baremetal-operator-controller-manager-84b575879fv4t28\" (UID: \"d5907af9-fe3f-4da0-ba70-993d36b25746\") " pod="openstack-operators/openstack-baremetal-operator-controller-manager-84b575879fv4t28" Dec 10 23:05:23 crc kubenswrapper[4791]: E1210 23:05:23.924057 4791 secret.go:188] Couldn't get secret openstack-operators/openstack-baremetal-operator-webhook-server-cert: secret "openstack-baremetal-operator-webhook-server-cert" not found Dec 10 23:05:23 crc kubenswrapper[4791]: E1210 23:05:23.924139 4791 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/d5907af9-fe3f-4da0-ba70-993d36b25746-cert podName:d5907af9-fe3f-4da0-ba70-993d36b25746 nodeName:}" failed. No retries permitted until 2025-12-10 23:05:39.924117772 +0000 UTC m=+974.353735425 (durationBeforeRetry 16s). Error: MountVolume.SetUp failed for volume "cert" (UniqueName: "kubernetes.io/secret/d5907af9-fe3f-4da0-ba70-993d36b25746-cert") pod "openstack-baremetal-operator-controller-manager-84b575879fv4t28" (UID: "d5907af9-fe3f-4da0-ba70-993d36b25746") : secret "openstack-baremetal-operator-webhook-server-cert" not found Dec 10 23:05:24 crc kubenswrapper[4791]: I1210 23:05:24.025402 4791 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"webhook-certs\" (UniqueName: \"kubernetes.io/secret/e2c4ecea-2852-4e39-93fb-aee7cbe31aa4-webhook-certs\") pod \"openstack-operator-controller-manager-686dfd865c-cft5f\" (UID: \"e2c4ecea-2852-4e39-93fb-aee7cbe31aa4\") " pod="openstack-operators/openstack-operator-controller-manager-686dfd865c-cft5f" Dec 10 23:05:24 crc kubenswrapper[4791]: I1210 23:05:24.025484 4791 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/e2c4ecea-2852-4e39-93fb-aee7cbe31aa4-metrics-certs\") pod \"openstack-operator-controller-manager-686dfd865c-cft5f\" (UID: \"e2c4ecea-2852-4e39-93fb-aee7cbe31aa4\") " pod="openstack-operators/openstack-operator-controller-manager-686dfd865c-cft5f" Dec 10 23:05:24 crc kubenswrapper[4791]: E1210 23:05:24.025609 4791 secret.go:188] Couldn't get secret openstack-operators/webhook-server-cert: secret "webhook-server-cert" not found Dec 10 23:05:24 crc kubenswrapper[4791]: E1210 23:05:24.025640 4791 secret.go:188] Couldn't get secret openstack-operators/metrics-server-cert: secret "metrics-server-cert" not found Dec 10 23:05:24 crc kubenswrapper[4791]: E1210 23:05:24.025664 4791 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/e2c4ecea-2852-4e39-93fb-aee7cbe31aa4-webhook-certs podName:e2c4ecea-2852-4e39-93fb-aee7cbe31aa4 nodeName:}" failed. No retries permitted until 2025-12-10 23:05:40.025647921 +0000 UTC m=+974.455265534 (durationBeforeRetry 16s). Error: MountVolume.SetUp failed for volume "webhook-certs" (UniqueName: "kubernetes.io/secret/e2c4ecea-2852-4e39-93fb-aee7cbe31aa4-webhook-certs") pod "openstack-operator-controller-manager-686dfd865c-cft5f" (UID: "e2c4ecea-2852-4e39-93fb-aee7cbe31aa4") : secret "webhook-server-cert" not found Dec 10 23:05:24 crc kubenswrapper[4791]: E1210 23:05:24.025707 4791 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/e2c4ecea-2852-4e39-93fb-aee7cbe31aa4-metrics-certs podName:e2c4ecea-2852-4e39-93fb-aee7cbe31aa4 nodeName:}" failed. No retries permitted until 2025-12-10 23:05:40.025689192 +0000 UTC m=+974.455306805 (durationBeforeRetry 16s). Error: MountVolume.SetUp failed for volume "metrics-certs" (UniqueName: "kubernetes.io/secret/e2c4ecea-2852-4e39-93fb-aee7cbe31aa4-metrics-certs") pod "openstack-operator-controller-manager-686dfd865c-cft5f" (UID: "e2c4ecea-2852-4e39-93fb-aee7cbe31aa4") : secret "metrics-server-cert" not found Dec 10 23:05:25 crc kubenswrapper[4791]: I1210 23:05:25.038194 4791 patch_prober.go:28] interesting pod/machine-config-daemon-5rb5l container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 10 23:05:25 crc kubenswrapper[4791]: I1210 23:05:25.038290 4791 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-5rb5l" podUID="ba35653c-6e06-4cee-a4d6-137764090d18" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 10 23:05:29 crc kubenswrapper[4791]: E1210 23:05:29.294509 4791 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="quay.io/openstack-k8s-operators/horizon-operator@sha256:9e847f4dbdea19ab997f32a02b3680a9bd966f9c705911645c3866a19fda9ea5" Dec 10 23:05:29 crc kubenswrapper[4791]: E1210 23:05:29.295518 4791 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:manager,Image:quay.io/openstack-k8s-operators/horizon-operator@sha256:9e847f4dbdea19ab997f32a02b3680a9bd966f9c705911645c3866a19fda9ea5,Command:[/manager],Args:[--leader-elect --health-probe-bind-address=:8081 --metrics-bind-address=127.0.0.1:8080],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:LEASE_DURATION,Value:30,ValueFrom:nil,},EnvVar{Name:RENEW_DEADLINE,Value:20,ValueFrom:nil,},EnvVar{Name:RETRY_PERIOD,Value:5,ValueFrom:nil,},EnvVar{Name:ENABLE_WEBHOOKS,Value:false,ValueFrom:nil,},EnvVar{Name:METRICS_CERTS,Value:false,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{cpu: {{500 -3} {} 500m DecimalSI},memory: {{536870912 0} {} BinarySI},},Requests:ResourceList{cpu: {{10 -3} {} 10m DecimalSI},memory: {{268435456 0} {} BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kube-api-access-fbkm9,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/healthz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:15,TimeoutSeconds:1,PeriodSeconds:20,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},ReadinessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/readyz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:5,TimeoutSeconds:1,PeriodSeconds:10,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod horizon-operator-controller-manager-68c6d99b8f-2lq47_openstack-operators(e3fdc812-fbd3-4d35-8555-d55e9390ebd0): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError" Dec 10 23:05:30 crc kubenswrapper[4791]: E1210 23:05:30.159162 4791 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="quay.io/openstack-k8s-operators/cinder-operator@sha256:981b6a8f95934a86c5f10ef6e198b07265aeba7f11cf84b9ccd13dfaf06f3ca3" Dec 10 23:05:30 crc kubenswrapper[4791]: E1210 23:05:30.159401 4791 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:manager,Image:quay.io/openstack-k8s-operators/cinder-operator@sha256:981b6a8f95934a86c5f10ef6e198b07265aeba7f11cf84b9ccd13dfaf06f3ca3,Command:[/manager],Args:[--leader-elect --health-probe-bind-address=:8081 --metrics-bind-address=127.0.0.1:8080],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:LEASE_DURATION,Value:30,ValueFrom:nil,},EnvVar{Name:RENEW_DEADLINE,Value:20,ValueFrom:nil,},EnvVar{Name:RETRY_PERIOD,Value:5,ValueFrom:nil,},EnvVar{Name:ENABLE_WEBHOOKS,Value:false,ValueFrom:nil,},EnvVar{Name:METRICS_CERTS,Value:false,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{cpu: {{500 -3} {} 500m DecimalSI},memory: {{536870912 0} {} BinarySI},},Requests:ResourceList{cpu: {{10 -3} {} 10m DecimalSI},memory: {{268435456 0} {} BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kube-api-access-857zc,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/healthz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:15,TimeoutSeconds:1,PeriodSeconds:20,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},ReadinessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/readyz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:5,TimeoutSeconds:1,PeriodSeconds:10,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod cinder-operator-controller-manager-6c677c69b-6jmn2_openstack-operators(48cd2bae-61e0-446c-ac2a-48e70bff5187): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError" Dec 10 23:05:31 crc kubenswrapper[4791]: E1210 23:05:31.100662 4791 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="quay.io/openstack-k8s-operators/glance-operator@sha256:5370dc4a8e776923eec00bb50cbdb2e390e9dde50be26bdc04a216bd2d6b5027" Dec 10 23:05:31 crc kubenswrapper[4791]: E1210 23:05:31.100873 4791 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:manager,Image:quay.io/openstack-k8s-operators/glance-operator@sha256:5370dc4a8e776923eec00bb50cbdb2e390e9dde50be26bdc04a216bd2d6b5027,Command:[/manager],Args:[--leader-elect --health-probe-bind-address=:8081 --metrics-bind-address=127.0.0.1:8080],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:LEASE_DURATION,Value:30,ValueFrom:nil,},EnvVar{Name:RENEW_DEADLINE,Value:20,ValueFrom:nil,},EnvVar{Name:RETRY_PERIOD,Value:5,ValueFrom:nil,},EnvVar{Name:ENABLE_WEBHOOKS,Value:false,ValueFrom:nil,},EnvVar{Name:METRICS_CERTS,Value:false,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{cpu: {{500 -3} {} 500m DecimalSI},memory: {{536870912 0} {} BinarySI},},Requests:ResourceList{cpu: {{10 -3} {} 10m DecimalSI},memory: {{268435456 0} {} BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kube-api-access-jqxwc,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/healthz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:15,TimeoutSeconds:1,PeriodSeconds:20,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},ReadinessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/readyz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:5,TimeoutSeconds:1,PeriodSeconds:10,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod glance-operator-controller-manager-5697bb5779-cvbqc_openstack-operators(f0b16aa2-e60e-4f0e-9679-9afc0a5ae027): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError" Dec 10 23:05:31 crc kubenswrapper[4791]: E1210 23:05:31.970145 4791 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="quay.io/openstack-k8s-operators/nova-operator@sha256:779f0cee6024d0fb8f259b036fe790e62aa5a3b0431ea9bf15a6e7d02e2e5670" Dec 10 23:05:31 crc kubenswrapper[4791]: E1210 23:05:31.971146 4791 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:manager,Image:quay.io/openstack-k8s-operators/nova-operator@sha256:779f0cee6024d0fb8f259b036fe790e62aa5a3b0431ea9bf15a6e7d02e2e5670,Command:[/manager],Args:[--leader-elect --health-probe-bind-address=:8081 --metrics-bind-address=127.0.0.1:8080],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:LEASE_DURATION,Value:30,ValueFrom:nil,},EnvVar{Name:RENEW_DEADLINE,Value:20,ValueFrom:nil,},EnvVar{Name:RETRY_PERIOD,Value:5,ValueFrom:nil,},EnvVar{Name:ENABLE_WEBHOOKS,Value:false,ValueFrom:nil,},EnvVar{Name:METRICS_CERTS,Value:false,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{cpu: {{500 -3} {} 500m DecimalSI},memory: {{536870912 0} {} BinarySI},},Requests:ResourceList{cpu: {{10 -3} {} 10m DecimalSI},memory: {{268435456 0} {} BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kube-api-access-qff2w,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/healthz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:15,TimeoutSeconds:1,PeriodSeconds:20,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},ReadinessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/readyz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:5,TimeoutSeconds:1,PeriodSeconds:10,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod nova-operator-controller-manager-697bc559fc-2lkcl_openstack-operators(3b2af4ca-48b4-463f-b0aa-c5cfd9099d67): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError" Dec 10 23:05:32 crc kubenswrapper[4791]: E1210 23:05:32.370873 4791 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="quay.io/openstack-k8s-operators/manila-operator@sha256:44126f9c6b1d2bf752ddf989e20a4fc4cc1c07723d4fcb78465ccb2f55da6b3a" Dec 10 23:05:32 crc kubenswrapper[4791]: E1210 23:05:32.371733 4791 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:manager,Image:quay.io/openstack-k8s-operators/manila-operator@sha256:44126f9c6b1d2bf752ddf989e20a4fc4cc1c07723d4fcb78465ccb2f55da6b3a,Command:[/manager],Args:[--leader-elect --health-probe-bind-address=:8081 --metrics-bind-address=127.0.0.1:8080],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:LEASE_DURATION,Value:30,ValueFrom:nil,},EnvVar{Name:RENEW_DEADLINE,Value:20,ValueFrom:nil,},EnvVar{Name:RETRY_PERIOD,Value:5,ValueFrom:nil,},EnvVar{Name:ENABLE_WEBHOOKS,Value:false,ValueFrom:nil,},EnvVar{Name:METRICS_CERTS,Value:false,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{cpu: {{500 -3} {} 500m DecimalSI},memory: {{536870912 0} {} BinarySI},},Requests:ResourceList{cpu: {{10 -3} {} 10m DecimalSI},memory: {{268435456 0} {} BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kube-api-access-kmv99,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/healthz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:15,TimeoutSeconds:1,PeriodSeconds:20,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},ReadinessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/readyz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:5,TimeoutSeconds:1,PeriodSeconds:10,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod manila-operator-controller-manager-5b5fd79c9c-tsbtp_openstack-operators(12b75f38-fb6a-4020-95b9-29c7dd872849): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError" Dec 10 23:05:32 crc kubenswrapper[4791]: E1210 23:05:32.853418 4791 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="quay.io/openstack-k8s-operators/neutron-operator@sha256:0b3fb69f35c151895d3dffd514974a9f9fe1c77c3bca69b78b81efb183cf4557" Dec 10 23:05:32 crc kubenswrapper[4791]: E1210 23:05:32.853626 4791 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:manager,Image:quay.io/openstack-k8s-operators/neutron-operator@sha256:0b3fb69f35c151895d3dffd514974a9f9fe1c77c3bca69b78b81efb183cf4557,Command:[/manager],Args:[--leader-elect --health-probe-bind-address=:8081 --metrics-bind-address=127.0.0.1:8080],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:LEASE_DURATION,Value:30,ValueFrom:nil,},EnvVar{Name:RENEW_DEADLINE,Value:20,ValueFrom:nil,},EnvVar{Name:RETRY_PERIOD,Value:5,ValueFrom:nil,},EnvVar{Name:ENABLE_WEBHOOKS,Value:false,ValueFrom:nil,},EnvVar{Name:METRICS_CERTS,Value:false,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{cpu: {{500 -3} {} 500m DecimalSI},memory: {{536870912 0} {} BinarySI},},Requests:ResourceList{cpu: {{10 -3} {} 10m DecimalSI},memory: {{268435456 0} {} BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kube-api-access-kgzzj,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/healthz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:15,TimeoutSeconds:1,PeriodSeconds:20,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},ReadinessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/readyz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:5,TimeoutSeconds:1,PeriodSeconds:10,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod neutron-operator-controller-manager-5fdfd5b6b5-g8f76_openstack-operators(86bbb7b2-5f55-46a2-b426-013ba183977f): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError" Dec 10 23:05:37 crc kubenswrapper[4791]: I1210 23:05:37.823646 4791 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Dec 10 23:05:38 crc kubenswrapper[4791]: I1210 23:05:38.171588 4791 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/heat-operator-controller-manager-5f64f6f8bb-fx67c" event={"ID":"d181199e-5cd2-408b-b167-059b34491a15","Type":"ContainerStarted","Data":"a5b437bae3f0df9e86cad6e84ef91a884d39c8609b5b7655bc74218d6de606c6"} Dec 10 23:05:38 crc kubenswrapper[4791]: I1210 23:05:38.173786 4791 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/keystone-operator-controller-manager-7765d96ddf-2wsxl" event={"ID":"311d4e7c-a0c7-47f6-8deb-7262068e8beb","Type":"ContainerStarted","Data":"1dc1b5621da28e75e031b5b39600ce3c2f565103d10c29d66e0c6c99d7220161"} Dec 10 23:05:38 crc kubenswrapper[4791]: I1210 23:05:38.175651 4791 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/barbican-operator-controller-manager-7d9dfd778-4ndx5" event={"ID":"dc042f4f-d336-484e-8c9d-60658c82f84a","Type":"ContainerStarted","Data":"15d10e5aa36926ebbba625a1aa53d70f2176e7230454a7b551cf1fc5c8e01f56"} Dec 10 23:05:38 crc kubenswrapper[4791]: I1210 23:05:38.184963 4791 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/ironic-operator-controller-manager-967d97867-r6jdv" event={"ID":"cb5fb221-76b0-48b4-ad27-7d0c51581ec4","Type":"ContainerStarted","Data":"dd361c5fc6f0a0b579f168eddd07e23aaea8bceefb235c6f6f1b702781fdccc7"} Dec 10 23:05:39 crc kubenswrapper[4791]: I1210 23:05:39.202155 4791 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/watcher-operator-controller-manager-75944c9b7-7cdb7" event={"ID":"21e66f28-5ed9-475e-aa7d-a105ad1f2f9e","Type":"ContainerStarted","Data":"89540d20aa5566c646a00f33fa1ea21f48fda72905037fb0178426923375a7fa"} Dec 10 23:05:39 crc kubenswrapper[4791]: I1210 23:05:39.205234 4791 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/swift-operator-controller-manager-9d58d64bc-8vwnp" event={"ID":"4cf9497a-4193-4d34-a68a-48085af6e8b5","Type":"ContainerStarted","Data":"6facbccd85f3900644bc53dfaf3818f30c7dba4451ef8ef15a89d7cbc37e4104"} Dec 10 23:05:39 crc kubenswrapper[4791]: I1210 23:05:39.206937 4791 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/ovn-operator-controller-manager-b6456fdb6-6rddm" event={"ID":"87c40827-309c-4c68-ae46-bbe449a9e66c","Type":"ContainerStarted","Data":"fbba46921cc909644327f028a14e666ce966d038e5de2f25540dd143e5d96b3f"} Dec 10 23:05:39 crc kubenswrapper[4791]: I1210 23:05:39.248111 4791 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/416b455d-b397-4aad-baf1-88e880619eb5-cert\") pod \"infra-operator-controller-manager-78d48bff9d-mmczn\" (UID: \"416b455d-b397-4aad-baf1-88e880619eb5\") " pod="openstack-operators/infra-operator-controller-manager-78d48bff9d-mmczn" Dec 10 23:05:39 crc kubenswrapper[4791]: I1210 23:05:39.254077 4791 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cert\" (UniqueName: \"kubernetes.io/secret/416b455d-b397-4aad-baf1-88e880619eb5-cert\") pod \"infra-operator-controller-manager-78d48bff9d-mmczn\" (UID: \"416b455d-b397-4aad-baf1-88e880619eb5\") " pod="openstack-operators/infra-operator-controller-manager-78d48bff9d-mmczn" Dec 10 23:05:39 crc kubenswrapper[4791]: I1210 23:05:39.496952 4791 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"infra-operator-controller-manager-dockercfg-q8lh9" Dec 10 23:05:39 crc kubenswrapper[4791]: I1210 23:05:39.506859 4791 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/infra-operator-controller-manager-78d48bff9d-mmczn" Dec 10 23:05:40 crc kubenswrapper[4791]: I1210 23:05:39.988328 4791 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/d5907af9-fe3f-4da0-ba70-993d36b25746-cert\") pod \"openstack-baremetal-operator-controller-manager-84b575879fv4t28\" (UID: \"d5907af9-fe3f-4da0-ba70-993d36b25746\") " pod="openstack-operators/openstack-baremetal-operator-controller-manager-84b575879fv4t28" Dec 10 23:05:40 crc kubenswrapper[4791]: I1210 23:05:40.008462 4791 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cert\" (UniqueName: \"kubernetes.io/secret/d5907af9-fe3f-4da0-ba70-993d36b25746-cert\") pod \"openstack-baremetal-operator-controller-manager-84b575879fv4t28\" (UID: \"d5907af9-fe3f-4da0-ba70-993d36b25746\") " pod="openstack-operators/openstack-baremetal-operator-controller-manager-84b575879fv4t28" Dec 10 23:05:40 crc kubenswrapper[4791]: I1210 23:05:40.049949 4791 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"openstack-baremetal-operator-controller-manager-dockercfg-7792b" Dec 10 23:05:40 crc kubenswrapper[4791]: I1210 23:05:40.058849 4791 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/openstack-baremetal-operator-controller-manager-84b575879fv4t28" Dec 10 23:05:40 crc kubenswrapper[4791]: I1210 23:05:40.102380 4791 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"webhook-certs\" (UniqueName: \"kubernetes.io/secret/e2c4ecea-2852-4e39-93fb-aee7cbe31aa4-webhook-certs\") pod \"openstack-operator-controller-manager-686dfd865c-cft5f\" (UID: \"e2c4ecea-2852-4e39-93fb-aee7cbe31aa4\") " pod="openstack-operators/openstack-operator-controller-manager-686dfd865c-cft5f" Dec 10 23:05:40 crc kubenswrapper[4791]: I1210 23:05:40.102434 4791 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/e2c4ecea-2852-4e39-93fb-aee7cbe31aa4-metrics-certs\") pod \"openstack-operator-controller-manager-686dfd865c-cft5f\" (UID: \"e2c4ecea-2852-4e39-93fb-aee7cbe31aa4\") " pod="openstack-operators/openstack-operator-controller-manager-686dfd865c-cft5f" Dec 10 23:05:40 crc kubenswrapper[4791]: I1210 23:05:40.127614 4791 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"webhook-certs\" (UniqueName: \"kubernetes.io/secret/e2c4ecea-2852-4e39-93fb-aee7cbe31aa4-webhook-certs\") pod \"openstack-operator-controller-manager-686dfd865c-cft5f\" (UID: \"e2c4ecea-2852-4e39-93fb-aee7cbe31aa4\") " pod="openstack-operators/openstack-operator-controller-manager-686dfd865c-cft5f" Dec 10 23:05:40 crc kubenswrapper[4791]: I1210 23:05:40.127650 4791 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/e2c4ecea-2852-4e39-93fb-aee7cbe31aa4-metrics-certs\") pod \"openstack-operator-controller-manager-686dfd865c-cft5f\" (UID: \"e2c4ecea-2852-4e39-93fb-aee7cbe31aa4\") " pod="openstack-operators/openstack-operator-controller-manager-686dfd865c-cft5f" Dec 10 23:05:40 crc kubenswrapper[4791]: I1210 23:05:40.186698 4791 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/infra-operator-controller-manager-78d48bff9d-mmczn"] Dec 10 23:05:40 crc kubenswrapper[4791]: I1210 23:05:40.218377 4791 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/rabbitmq-cluster-operator-manager-668c99d594-2bgc5" event={"ID":"b3ae5e53-9077-4dd0-996d-2417f28b4736","Type":"ContainerStarted","Data":"47fab4c50af7501b4403029740f0a9ee282b8049f45d00db9c6c16c12156199c"} Dec 10 23:05:40 crc kubenswrapper[4791]: I1210 23:05:40.220732 4791 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/infra-operator-controller-manager-78d48bff9d-mmczn" event={"ID":"416b455d-b397-4aad-baf1-88e880619eb5","Type":"ContainerStarted","Data":"b9d6bd384844691253eeb011d228de5b40c7333640dec9fc705f18a46821bb25"} Dec 10 23:05:40 crc kubenswrapper[4791]: I1210 23:05:40.223654 4791 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/mariadb-operator-controller-manager-79c8c4686c-2wmm4" event={"ID":"590c13cc-efa0-431a-9f2f-65fb97c1bd3c","Type":"ContainerStarted","Data":"6335f356d3e2d9f8fe39d3de27a3ff9a942f57161b135774b39c9136b40122c1"} Dec 10 23:05:40 crc kubenswrapper[4791]: I1210 23:05:40.224927 4791 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/telemetry-operator-controller-manager-58d5ff84df-t6tnb" event={"ID":"2418f025-d37e-4316-b804-9ab53c326c82","Type":"ContainerStarted","Data":"c765fcdcdfc26be881c28c2412df82115db9d809817dcbd758beefd9a9b22763"} Dec 10 23:05:40 crc kubenswrapper[4791]: I1210 23:05:40.226024 4791 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/octavia-operator-controller-manager-998648c74-sjp6q" event={"ID":"23a95e98-8ae3-4ac4-945d-0cae5af5d1ac","Type":"ContainerStarted","Data":"3fe283455c09ea4ead79dd62c4f8c4057998c4c093e2e98240cb862922638b77"} Dec 10 23:05:40 crc kubenswrapper[4791]: I1210 23:05:40.362101 4791 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"openstack-operator-controller-manager-dockercfg-5vcct" Dec 10 23:05:40 crc kubenswrapper[4791]: I1210 23:05:40.371496 4791 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/openstack-operator-controller-manager-686dfd865c-cft5f" Dec 10 23:05:40 crc kubenswrapper[4791]: E1210 23:05:40.767606 4791 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\"" pod="openstack-operators/placement-operator-controller-manager-78f8948974-bhgvx" podUID="b67ef71c-85c5-41ca-bf93-d1ca8a2fd007" Dec 10 23:05:40 crc kubenswrapper[4791]: E1210 23:05:40.801853 4791 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\"" pod="openstack-operators/nova-operator-controller-manager-697bc559fc-2lkcl" podUID="3b2af4ca-48b4-463f-b0aa-c5cfd9099d67" Dec 10 23:05:40 crc kubenswrapper[4791]: I1210 23:05:40.836516 4791 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/rabbitmq-cluster-operator-manager-668c99d594-2bgc5" podStartSLOduration=5.923350139 podStartE2EDuration="33.836498824s" podCreationTimestamp="2025-12-10 23:05:07 +0000 UTC" firstStartedPulling="2025-12-10 23:05:09.652658915 +0000 UTC m=+944.082276528" lastFinishedPulling="2025-12-10 23:05:37.5658076 +0000 UTC m=+971.995425213" observedRunningTime="2025-12-10 23:05:40.241763216 +0000 UTC m=+974.671380839" watchObservedRunningTime="2025-12-10 23:05:40.836498824 +0000 UTC m=+975.266116437" Dec 10 23:05:40 crc kubenswrapper[4791]: I1210 23:05:40.841253 4791 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/openstack-baremetal-operator-controller-manager-84b575879fv4t28"] Dec 10 23:05:41 crc kubenswrapper[4791]: E1210 23:05:41.050788 4791 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\"" pod="openstack-operators/designate-operator-controller-manager-697fb699cf-vg956" podUID="19bf0bfe-4e8b-47a4-a919-4227b3e47bb3" Dec 10 23:05:41 crc kubenswrapper[4791]: I1210 23:05:41.275109 4791 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/placement-operator-controller-manager-78f8948974-bhgvx" event={"ID":"b67ef71c-85c5-41ca-bf93-d1ca8a2fd007","Type":"ContainerStarted","Data":"d3f6ec63c13ca60731a8c4cc51fd971ff2dcb71bfc1b584776751efa5ec73bd1"} Dec 10 23:05:41 crc kubenswrapper[4791]: I1210 23:05:41.288316 4791 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/nova-operator-controller-manager-697bc559fc-2lkcl" event={"ID":"3b2af4ca-48b4-463f-b0aa-c5cfd9099d67","Type":"ContainerStarted","Data":"643ba1a97ebb50d07bc5aed2d1d9d1078d6aa228b9fcbcda435305062e1b827b"} Dec 10 23:05:41 crc kubenswrapper[4791]: E1210 23:05:41.295112 4791 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/nova-operator@sha256:779f0cee6024d0fb8f259b036fe790e62aa5a3b0431ea9bf15a6e7d02e2e5670\\\"\"" pod="openstack-operators/nova-operator-controller-manager-697bc559fc-2lkcl" podUID="3b2af4ca-48b4-463f-b0aa-c5cfd9099d67" Dec 10 23:05:41 crc kubenswrapper[4791]: I1210 23:05:41.298191 4791 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/openstack-baremetal-operator-controller-manager-84b575879fv4t28" event={"ID":"d5907af9-fe3f-4da0-ba70-993d36b25746","Type":"ContainerStarted","Data":"f792b552dbbfd9edc2b53a66c1a4622b8e34f4dfda835808d57da91165b91553"} Dec 10 23:05:41 crc kubenswrapper[4791]: I1210 23:05:41.320625 4791 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/designate-operator-controller-manager-697fb699cf-vg956" event={"ID":"19bf0bfe-4e8b-47a4-a919-4227b3e47bb3","Type":"ContainerStarted","Data":"fa404efa920366a57d43b0555d3256a30b171270c0206bf7bfaeaaa2ba70191f"} Dec 10 23:05:41 crc kubenswrapper[4791]: I1210 23:05:41.433413 4791 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/openstack-operator-controller-manager-686dfd865c-cft5f"] Dec 10 23:05:41 crc kubenswrapper[4791]: E1210 23:05:41.543735 4791 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\"" pod="openstack-operators/glance-operator-controller-manager-5697bb5779-cvbqc" podUID="f0b16aa2-e60e-4f0e-9679-9afc0a5ae027" Dec 10 23:05:41 crc kubenswrapper[4791]: E1210 23:05:41.570986 4791 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\"" pod="openstack-operators/neutron-operator-controller-manager-5fdfd5b6b5-g8f76" podUID="86bbb7b2-5f55-46a2-b426-013ba183977f" Dec 10 23:05:41 crc kubenswrapper[4791]: E1210 23:05:41.673417 4791 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\"" pod="openstack-operators/manila-operator-controller-manager-5b5fd79c9c-tsbtp" podUID="12b75f38-fb6a-4020-95b9-29c7dd872849" Dec 10 23:05:41 crc kubenswrapper[4791]: E1210 23:05:41.676660 4791 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\"" pod="openstack-operators/cinder-operator-controller-manager-6c677c69b-6jmn2" podUID="48cd2bae-61e0-446c-ac2a-48e70bff5187" Dec 10 23:05:41 crc kubenswrapper[4791]: E1210 23:05:41.717089 4791 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\"" pod="openstack-operators/horizon-operator-controller-manager-68c6d99b8f-2lq47" podUID="e3fdc812-fbd3-4d35-8555-d55e9390ebd0" Dec 10 23:05:41 crc kubenswrapper[4791]: E1210 23:05:41.839763 4791 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\"" pod="openstack-operators/test-operator-controller-manager-5854674fcc-rw7ct" podUID="bda40eaa-446c-475e-8f2f-76f76caa0b85" Dec 10 23:05:42 crc kubenswrapper[4791]: I1210 23:05:42.330220 4791 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/telemetry-operator-controller-manager-58d5ff84df-t6tnb" event={"ID":"2418f025-d37e-4316-b804-9ab53c326c82","Type":"ContainerStarted","Data":"9e97624027285334dbea9528b8629d6b4e32123bba52c3391819e392a8637741"} Dec 10 23:05:42 crc kubenswrapper[4791]: I1210 23:05:42.330411 4791 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/telemetry-operator-controller-manager-58d5ff84df-t6tnb" Dec 10 23:05:42 crc kubenswrapper[4791]: I1210 23:05:42.334680 4791 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/test-operator-controller-manager-5854674fcc-rw7ct" event={"ID":"bda40eaa-446c-475e-8f2f-76f76caa0b85","Type":"ContainerStarted","Data":"35bbe6b26f70831cf343a2004f6e6365e03b6959f8ef5e0c16cf63eeb06c9d34"} Dec 10 23:05:42 crc kubenswrapper[4791]: I1210 23:05:42.338198 4791 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/manila-operator-controller-manager-5b5fd79c9c-tsbtp" event={"ID":"12b75f38-fb6a-4020-95b9-29c7dd872849","Type":"ContainerStarted","Data":"afee316d1a33cde7c2a03b6f0331fa1cf43ef09674e8267e546ba61bc4de79e3"} Dec 10 23:05:42 crc kubenswrapper[4791]: E1210 23:05:42.341570 4791 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/manila-operator@sha256:44126f9c6b1d2bf752ddf989e20a4fc4cc1c07723d4fcb78465ccb2f55da6b3a\\\"\"" pod="openstack-operators/manila-operator-controller-manager-5b5fd79c9c-tsbtp" podUID="12b75f38-fb6a-4020-95b9-29c7dd872849" Dec 10 23:05:42 crc kubenswrapper[4791]: I1210 23:05:42.343618 4791 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/horizon-operator-controller-manager-68c6d99b8f-2lq47" event={"ID":"e3fdc812-fbd3-4d35-8555-d55e9390ebd0","Type":"ContainerStarted","Data":"19fd3ffb554c009071e1bea0bc50e840734e82ca3f10102daab5dec8b45460c0"} Dec 10 23:05:42 crc kubenswrapper[4791]: I1210 23:05:42.346242 4791 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/openstack-operator-controller-manager-686dfd865c-cft5f" event={"ID":"e2c4ecea-2852-4e39-93fb-aee7cbe31aa4","Type":"ContainerStarted","Data":"b1e8f7e418b39ecb625aba8321c1a97ea6df959f92d3625240f94f34901863e1"} Dec 10 23:05:42 crc kubenswrapper[4791]: I1210 23:05:42.346277 4791 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/openstack-operator-controller-manager-686dfd865c-cft5f" event={"ID":"e2c4ecea-2852-4e39-93fb-aee7cbe31aa4","Type":"ContainerStarted","Data":"88bfda76bf3995098e7f2e7cf083083088a0ccf4a4e0acf2614e22e8865cb722"} Dec 10 23:05:42 crc kubenswrapper[4791]: I1210 23:05:42.346309 4791 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/openstack-operator-controller-manager-686dfd865c-cft5f" Dec 10 23:05:42 crc kubenswrapper[4791]: I1210 23:05:42.349053 4791 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/ovn-operator-controller-manager-b6456fdb6-6rddm" event={"ID":"87c40827-309c-4c68-ae46-bbe449a9e66c","Type":"ContainerStarted","Data":"3f0d2199e26584e1b33afac8ec495d0eb811c81cad23f9d70eeaa3cc3703b9eb"} Dec 10 23:05:42 crc kubenswrapper[4791]: I1210 23:05:42.349161 4791 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/ovn-operator-controller-manager-b6456fdb6-6rddm" Dec 10 23:05:42 crc kubenswrapper[4791]: I1210 23:05:42.351703 4791 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/mariadb-operator-controller-manager-79c8c4686c-2wmm4" event={"ID":"590c13cc-efa0-431a-9f2f-65fb97c1bd3c","Type":"ContainerStarted","Data":"a8cc62a026370c9f9804d5ef1cffb04efc2c5823336efc365fe0d30b10673e1a"} Dec 10 23:05:42 crc kubenswrapper[4791]: I1210 23:05:42.351882 4791 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/mariadb-operator-controller-manager-79c8c4686c-2wmm4" Dec 10 23:05:42 crc kubenswrapper[4791]: I1210 23:05:42.354283 4791 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/neutron-operator-controller-manager-5fdfd5b6b5-g8f76" event={"ID":"86bbb7b2-5f55-46a2-b426-013ba183977f","Type":"ContainerStarted","Data":"ffe025f3bb3be53c0d6206e49fcfdaee3e96d0fe5af1251d3486204d0b567955"} Dec 10 23:05:42 crc kubenswrapper[4791]: E1210 23:05:42.355540 4791 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/neutron-operator@sha256:0b3fb69f35c151895d3dffd514974a9f9fe1c77c3bca69b78b81efb183cf4557\\\"\"" pod="openstack-operators/neutron-operator-controller-manager-5fdfd5b6b5-g8f76" podUID="86bbb7b2-5f55-46a2-b426-013ba183977f" Dec 10 23:05:42 crc kubenswrapper[4791]: I1210 23:05:42.357147 4791 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/keystone-operator-controller-manager-7765d96ddf-2wsxl" event={"ID":"311d4e7c-a0c7-47f6-8deb-7262068e8beb","Type":"ContainerStarted","Data":"171e0e5cf8566a49773c8610d3ad9add57de4235d8e429135c05cbb923389f07"} Dec 10 23:05:42 crc kubenswrapper[4791]: I1210 23:05:42.357689 4791 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/keystone-operator-controller-manager-7765d96ddf-2wsxl" Dec 10 23:05:42 crc kubenswrapper[4791]: I1210 23:05:42.359732 4791 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/keystone-operator-controller-manager-7765d96ddf-2wsxl" Dec 10 23:05:42 crc kubenswrapper[4791]: I1210 23:05:42.366576 4791 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/barbican-operator-controller-manager-7d9dfd778-4ndx5" event={"ID":"dc042f4f-d336-484e-8c9d-60658c82f84a","Type":"ContainerStarted","Data":"aafa6ea83196c5eafc08faa71e8b7d0f6df2307406396fc6a715ef657b0159af"} Dec 10 23:05:42 crc kubenswrapper[4791]: I1210 23:05:42.366759 4791 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/barbican-operator-controller-manager-7d9dfd778-4ndx5" Dec 10 23:05:42 crc kubenswrapper[4791]: I1210 23:05:42.367573 4791 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/telemetry-operator-controller-manager-58d5ff84df-t6tnb" podStartSLOduration=7.721940893 podStartE2EDuration="35.367546396s" podCreationTimestamp="2025-12-10 23:05:07 +0000 UTC" firstStartedPulling="2025-12-10 23:05:09.468270774 +0000 UTC m=+943.897888387" lastFinishedPulling="2025-12-10 23:05:37.113876277 +0000 UTC m=+971.543493890" observedRunningTime="2025-12-10 23:05:42.356576195 +0000 UTC m=+976.786193808" watchObservedRunningTime="2025-12-10 23:05:42.367546396 +0000 UTC m=+976.797164009" Dec 10 23:05:42 crc kubenswrapper[4791]: I1210 23:05:42.369260 4791 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/octavia-operator-controller-manager-998648c74-sjp6q" event={"ID":"23a95e98-8ae3-4ac4-945d-0cae5af5d1ac","Type":"ContainerStarted","Data":"bf74a6917a9891e9dc176f9fc0360b30915d45ad5e7f1879b4adfcf7c833e858"} Dec 10 23:05:42 crc kubenswrapper[4791]: I1210 23:05:42.369657 4791 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/octavia-operator-controller-manager-998648c74-sjp6q" Dec 10 23:05:42 crc kubenswrapper[4791]: I1210 23:05:42.370130 4791 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/barbican-operator-controller-manager-7d9dfd778-4ndx5" Dec 10 23:05:42 crc kubenswrapper[4791]: I1210 23:05:42.372581 4791 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/designate-operator-controller-manager-697fb699cf-vg956" event={"ID":"19bf0bfe-4e8b-47a4-a919-4227b3e47bb3","Type":"ContainerStarted","Data":"b4499dccd5e4265ffdfd2ba9152faab64df352d17adcf64a656c3bea0e7519bc"} Dec 10 23:05:42 crc kubenswrapper[4791]: I1210 23:05:42.373287 4791 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/designate-operator-controller-manager-697fb699cf-vg956" Dec 10 23:05:42 crc kubenswrapper[4791]: I1210 23:05:42.387570 4791 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/cinder-operator-controller-manager-6c677c69b-6jmn2" event={"ID":"48cd2bae-61e0-446c-ac2a-48e70bff5187","Type":"ContainerStarted","Data":"fc38a510ee021381b22d76ecfb34158e30a1cf9a2fb693ed7222eedfac72949f"} Dec 10 23:05:42 crc kubenswrapper[4791]: I1210 23:05:42.389614 4791 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/glance-operator-controller-manager-5697bb5779-cvbqc" event={"ID":"f0b16aa2-e60e-4f0e-9679-9afc0a5ae027","Type":"ContainerStarted","Data":"a710cd110463df49bf31bec4696a5ead9694c6a5441e545a907171546bab7699"} Dec 10 23:05:42 crc kubenswrapper[4791]: I1210 23:05:42.400107 4791 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/ironic-operator-controller-manager-967d97867-r6jdv" event={"ID":"cb5fb221-76b0-48b4-ad27-7d0c51581ec4","Type":"ContainerStarted","Data":"45a135d76927b46f8a1b0c672550397dbee14919054df6a215913fabf4d8e1e8"} Dec 10 23:05:42 crc kubenswrapper[4791]: I1210 23:05:42.401199 4791 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/ironic-operator-controller-manager-967d97867-r6jdv" Dec 10 23:05:42 crc kubenswrapper[4791]: I1210 23:05:42.404941 4791 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/ironic-operator-controller-manager-967d97867-r6jdv" Dec 10 23:05:42 crc kubenswrapper[4791]: I1210 23:05:42.417556 4791 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/heat-operator-controller-manager-5f64f6f8bb-fx67c" event={"ID":"d181199e-5cd2-408b-b167-059b34491a15","Type":"ContainerStarted","Data":"1c96f42a914a3565d8f0fbeef25ade2ad77ed7012e4be080dcb1fc7821afe6f5"} Dec 10 23:05:42 crc kubenswrapper[4791]: I1210 23:05:42.419016 4791 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/heat-operator-controller-manager-5f64f6f8bb-fx67c" Dec 10 23:05:42 crc kubenswrapper[4791]: I1210 23:05:42.431363 4791 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/heat-operator-controller-manager-5f64f6f8bb-fx67c" Dec 10 23:05:42 crc kubenswrapper[4791]: I1210 23:05:42.441776 4791 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/watcher-operator-controller-manager-75944c9b7-7cdb7" event={"ID":"21e66f28-5ed9-475e-aa7d-a105ad1f2f9e","Type":"ContainerStarted","Data":"c4d464cb7c2343184eccf6e44d5422077203d91c8e66129550f39d64e712cb92"} Dec 10 23:05:42 crc kubenswrapper[4791]: I1210 23:05:42.442148 4791 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/watcher-operator-controller-manager-75944c9b7-7cdb7" Dec 10 23:05:42 crc kubenswrapper[4791]: I1210 23:05:42.456663 4791 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/ovn-operator-controller-manager-b6456fdb6-6rddm" podStartSLOduration=4.766372481 podStartE2EDuration="35.45664113s" podCreationTimestamp="2025-12-10 23:05:07 +0000 UTC" firstStartedPulling="2025-12-10 23:05:09.472693173 +0000 UTC m=+943.902310776" lastFinishedPulling="2025-12-10 23:05:40.162961812 +0000 UTC m=+974.592579425" observedRunningTime="2025-12-10 23:05:42.452665204 +0000 UTC m=+976.882282817" watchObservedRunningTime="2025-12-10 23:05:42.45664113 +0000 UTC m=+976.886258743" Dec 10 23:05:42 crc kubenswrapper[4791]: I1210 23:05:42.473787 4791 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/swift-operator-controller-manager-9d58d64bc-8vwnp" event={"ID":"4cf9497a-4193-4d34-a68a-48085af6e8b5","Type":"ContainerStarted","Data":"1c75b87c698b0eaae3d215679f7996f70bd0ced9ac4efc8f02d5c9f9eae3c59d"} Dec 10 23:05:42 crc kubenswrapper[4791]: I1210 23:05:42.473838 4791 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/swift-operator-controller-manager-9d58d64bc-8vwnp" Dec 10 23:05:42 crc kubenswrapper[4791]: I1210 23:05:42.489600 4791 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/mariadb-operator-controller-manager-79c8c4686c-2wmm4" podStartSLOduration=7.814576881 podStartE2EDuration="35.489581154s" podCreationTimestamp="2025-12-10 23:05:07 +0000 UTC" firstStartedPulling="2025-12-10 23:05:09.465684618 +0000 UTC m=+943.895302231" lastFinishedPulling="2025-12-10 23:05:37.140688881 +0000 UTC m=+971.570306504" observedRunningTime="2025-12-10 23:05:42.484836385 +0000 UTC m=+976.914453998" watchObservedRunningTime="2025-12-10 23:05:42.489581154 +0000 UTC m=+976.919198767" Dec 10 23:05:42 crc kubenswrapper[4791]: I1210 23:05:42.795714 4791 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/openstack-operator-controller-manager-686dfd865c-cft5f" podStartSLOduration=35.795688233 podStartE2EDuration="35.795688233s" podCreationTimestamp="2025-12-10 23:05:07 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-10 23:05:42.782897959 +0000 UTC m=+977.212515572" watchObservedRunningTime="2025-12-10 23:05:42.795688233 +0000 UTC m=+977.225305846" Dec 10 23:05:42 crc kubenswrapper[4791]: I1210 23:05:42.852616 4791 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/designate-operator-controller-manager-697fb699cf-vg956" podStartSLOduration=2.686288529 podStartE2EDuration="35.852587917s" podCreationTimestamp="2025-12-10 23:05:07 +0000 UTC" firstStartedPulling="2025-12-10 23:05:08.85210707 +0000 UTC m=+943.281724713" lastFinishedPulling="2025-12-10 23:05:42.018406478 +0000 UTC m=+976.448024101" observedRunningTime="2025-12-10 23:05:42.819744596 +0000 UTC m=+977.249362229" watchObservedRunningTime="2025-12-10 23:05:42.852587917 +0000 UTC m=+977.282205530" Dec 10 23:05:42 crc kubenswrapper[4791]: I1210 23:05:42.878653 4791 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/watcher-operator-controller-manager-75944c9b7-7cdb7" podStartSLOduration=5.363197851 podStartE2EDuration="35.878629438s" podCreationTimestamp="2025-12-10 23:05:07 +0000 UTC" firstStartedPulling="2025-12-10 23:05:09.64737301 +0000 UTC m=+944.076990613" lastFinishedPulling="2025-12-10 23:05:40.162804587 +0000 UTC m=+974.592422200" observedRunningTime="2025-12-10 23:05:42.851726132 +0000 UTC m=+977.281343755" watchObservedRunningTime="2025-12-10 23:05:42.878629438 +0000 UTC m=+977.308247051" Dec 10 23:05:42 crc kubenswrapper[4791]: I1210 23:05:42.914617 4791 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/swift-operator-controller-manager-9d58d64bc-8vwnp" podStartSLOduration=4.60451362 podStartE2EDuration="35.91459484s" podCreationTimestamp="2025-12-10 23:05:07 +0000 UTC" firstStartedPulling="2025-12-10 23:05:09.453199933 +0000 UTC m=+943.882817546" lastFinishedPulling="2025-12-10 23:05:40.763281153 +0000 UTC m=+975.192898766" observedRunningTime="2025-12-10 23:05:42.914369313 +0000 UTC m=+977.343986936" watchObservedRunningTime="2025-12-10 23:05:42.91459484 +0000 UTC m=+977.344212453" Dec 10 23:05:42 crc kubenswrapper[4791]: I1210 23:05:42.917956 4791 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/octavia-operator-controller-manager-998648c74-sjp6q" podStartSLOduration=8.242127171 podStartE2EDuration="35.917943247s" podCreationTimestamp="2025-12-10 23:05:07 +0000 UTC" firstStartedPulling="2025-12-10 23:05:09.465833263 +0000 UTC m=+943.895450876" lastFinishedPulling="2025-12-10 23:05:37.141649339 +0000 UTC m=+971.571266952" observedRunningTime="2025-12-10 23:05:42.886704294 +0000 UTC m=+977.316321927" watchObservedRunningTime="2025-12-10 23:05:42.917943247 +0000 UTC m=+977.347560860" Dec 10 23:05:42 crc kubenswrapper[4791]: I1210 23:05:42.967791 4791 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/heat-operator-controller-manager-5f64f6f8bb-fx67c" podStartSLOduration=3.5501537450000002 podStartE2EDuration="35.967769224s" podCreationTimestamp="2025-12-10 23:05:07 +0000 UTC" firstStartedPulling="2025-12-10 23:05:08.742878227 +0000 UTC m=+943.172495840" lastFinishedPulling="2025-12-10 23:05:41.160493716 +0000 UTC m=+975.590111319" observedRunningTime="2025-12-10 23:05:42.966897769 +0000 UTC m=+977.396515392" watchObservedRunningTime="2025-12-10 23:05:42.967769224 +0000 UTC m=+977.397386837" Dec 10 23:05:43 crc kubenswrapper[4791]: I1210 23:05:43.003185 4791 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/keystone-operator-controller-manager-7765d96ddf-2wsxl" podStartSLOduration=5.441523231 podStartE2EDuration="36.003161659s" podCreationTimestamp="2025-12-10 23:05:07 +0000 UTC" firstStartedPulling="2025-12-10 23:05:09.447079294 +0000 UTC m=+943.876696907" lastFinishedPulling="2025-12-10 23:05:40.008717722 +0000 UTC m=+974.438335335" observedRunningTime="2025-12-10 23:05:43.001256843 +0000 UTC m=+977.430874466" watchObservedRunningTime="2025-12-10 23:05:43.003161659 +0000 UTC m=+977.432779282" Dec 10 23:05:43 crc kubenswrapper[4791]: I1210 23:05:43.075388 4791 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/ironic-operator-controller-manager-967d97867-r6jdv" podStartSLOduration=4.81624941 podStartE2EDuration="36.07536891s" podCreationTimestamp="2025-12-10 23:05:07 +0000 UTC" firstStartedPulling="2025-12-10 23:05:09.415446939 +0000 UTC m=+943.845064552" lastFinishedPulling="2025-12-10 23:05:40.674566429 +0000 UTC m=+975.104184052" observedRunningTime="2025-12-10 23:05:43.032996081 +0000 UTC m=+977.462613694" watchObservedRunningTime="2025-12-10 23:05:43.07536891 +0000 UTC m=+977.504986523" Dec 10 23:05:43 crc kubenswrapper[4791]: I1210 23:05:43.100004 4791 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/barbican-operator-controller-manager-7d9dfd778-4ndx5" podStartSLOduration=4.723702785 podStartE2EDuration="36.098530917s" podCreationTimestamp="2025-12-10 23:05:07 +0000 UTC" firstStartedPulling="2025-12-10 23:05:08.633085407 +0000 UTC m=+943.062703020" lastFinishedPulling="2025-12-10 23:05:40.007913539 +0000 UTC m=+974.437531152" observedRunningTime="2025-12-10 23:05:43.088295738 +0000 UTC m=+977.517913351" watchObservedRunningTime="2025-12-10 23:05:43.098530917 +0000 UTC m=+977.528148520" Dec 10 23:05:43 crc kubenswrapper[4791]: I1210 23:05:43.495722 4791 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/test-operator-controller-manager-5854674fcc-rw7ct" event={"ID":"bda40eaa-446c-475e-8f2f-76f76caa0b85","Type":"ContainerStarted","Data":"b76b624c2ee5b6a2e58571a05c7e69ead755be9c17a17e95201191f5995b2c02"} Dec 10 23:05:43 crc kubenswrapper[4791]: I1210 23:05:43.496035 4791 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/test-operator-controller-manager-5854674fcc-rw7ct" Dec 10 23:05:43 crc kubenswrapper[4791]: I1210 23:05:43.506253 4791 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/glance-operator-controller-manager-5697bb5779-cvbqc" event={"ID":"f0b16aa2-e60e-4f0e-9679-9afc0a5ae027","Type":"ContainerStarted","Data":"7460a1cbbfb596ac597883eee688dc5afc7f7c3c9c72792def0a98be9b172a94"} Dec 10 23:05:43 crc kubenswrapper[4791]: I1210 23:05:43.506666 4791 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/glance-operator-controller-manager-5697bb5779-cvbqc" Dec 10 23:05:43 crc kubenswrapper[4791]: I1210 23:05:43.513401 4791 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/nova-operator-controller-manager-697bc559fc-2lkcl" event={"ID":"3b2af4ca-48b4-463f-b0aa-c5cfd9099d67","Type":"ContainerStarted","Data":"d8cce902c05c8a1d1b2d670554c52b3c15af5637a95ffd47152f8e5e05b271ba"} Dec 10 23:05:43 crc kubenswrapper[4791]: I1210 23:05:43.513810 4791 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/nova-operator-controller-manager-697bc559fc-2lkcl" Dec 10 23:05:43 crc kubenswrapper[4791]: I1210 23:05:43.550693 4791 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/test-operator-controller-manager-5854674fcc-rw7ct" podStartSLOduration=2.942634702 podStartE2EDuration="36.550670656s" podCreationTimestamp="2025-12-10 23:05:07 +0000 UTC" firstStartedPulling="2025-12-10 23:05:09.391804768 +0000 UTC m=+943.821422371" lastFinishedPulling="2025-12-10 23:05:42.999840712 +0000 UTC m=+977.429458325" observedRunningTime="2025-12-10 23:05:43.532142085 +0000 UTC m=+977.961759698" watchObservedRunningTime="2025-12-10 23:05:43.550670656 +0000 UTC m=+977.980288269" Dec 10 23:05:43 crc kubenswrapper[4791]: I1210 23:05:43.600986 4791 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/ovn-operator-controller-manager-b6456fdb6-6rddm" Dec 10 23:05:43 crc kubenswrapper[4791]: I1210 23:05:43.601040 4791 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/watcher-operator-controller-manager-75944c9b7-7cdb7" Dec 10 23:05:43 crc kubenswrapper[4791]: I1210 23:05:43.602785 4791 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/swift-operator-controller-manager-9d58d64bc-8vwnp" Dec 10 23:05:43 crc kubenswrapper[4791]: I1210 23:05:43.608929 4791 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/glance-operator-controller-manager-5697bb5779-cvbqc" podStartSLOduration=3.151338094 podStartE2EDuration="36.608908269s" podCreationTimestamp="2025-12-10 23:05:07 +0000 UTC" firstStartedPulling="2025-12-10 23:05:09.433396824 +0000 UTC m=+943.863014437" lastFinishedPulling="2025-12-10 23:05:42.890966999 +0000 UTC m=+977.320584612" observedRunningTime="2025-12-10 23:05:43.606668944 +0000 UTC m=+978.036286557" watchObservedRunningTime="2025-12-10 23:05:43.608908269 +0000 UTC m=+978.038525882" Dec 10 23:05:43 crc kubenswrapper[4791]: I1210 23:05:43.627974 4791 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/nova-operator-controller-manager-697bc559fc-2lkcl" podStartSLOduration=2.781853661 podStartE2EDuration="36.627955836s" podCreationTimestamp="2025-12-10 23:05:07 +0000 UTC" firstStartedPulling="2025-12-10 23:05:09.438023279 +0000 UTC m=+943.867640892" lastFinishedPulling="2025-12-10 23:05:43.284125444 +0000 UTC m=+977.713743067" observedRunningTime="2025-12-10 23:05:43.620511888 +0000 UTC m=+978.050129491" watchObservedRunningTime="2025-12-10 23:05:43.627955836 +0000 UTC m=+978.057573449" Dec 10 23:05:44 crc kubenswrapper[4791]: I1210 23:05:44.530642 4791 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/neutron-operator-controller-manager-5fdfd5b6b5-g8f76" event={"ID":"86bbb7b2-5f55-46a2-b426-013ba183977f","Type":"ContainerStarted","Data":"4fc9ff83a67030c3d72dd53a5685066f2ba52f71b2783c23bacbeb8e03d56b92"} Dec 10 23:05:44 crc kubenswrapper[4791]: I1210 23:05:44.531251 4791 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/neutron-operator-controller-manager-5fdfd5b6b5-g8f76" Dec 10 23:05:44 crc kubenswrapper[4791]: I1210 23:05:44.533269 4791 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/placement-operator-controller-manager-78f8948974-bhgvx" event={"ID":"b67ef71c-85c5-41ca-bf93-d1ca8a2fd007","Type":"ContainerStarted","Data":"c3a24ed7e009d2d95153a3b256ebaecc05578ff0e3966244251c5ce9efd9382d"} Dec 10 23:05:44 crc kubenswrapper[4791]: I1210 23:05:44.533362 4791 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/placement-operator-controller-manager-78f8948974-bhgvx" Dec 10 23:05:44 crc kubenswrapper[4791]: I1210 23:05:44.536862 4791 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/manila-operator-controller-manager-5b5fd79c9c-tsbtp" event={"ID":"12b75f38-fb6a-4020-95b9-29c7dd872849","Type":"ContainerStarted","Data":"897192747da9da843e39edee0eb417ff66480cc5c190c58db1f275118e7255ee"} Dec 10 23:05:44 crc kubenswrapper[4791]: I1210 23:05:44.549267 4791 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/neutron-operator-controller-manager-5fdfd5b6b5-g8f76" podStartSLOduration=2.802352382 podStartE2EDuration="37.549245542s" podCreationTimestamp="2025-12-10 23:05:07 +0000 UTC" firstStartedPulling="2025-12-10 23:05:09.449831075 +0000 UTC m=+943.879448688" lastFinishedPulling="2025-12-10 23:05:44.196724235 +0000 UTC m=+978.626341848" observedRunningTime="2025-12-10 23:05:44.546599024 +0000 UTC m=+978.976216647" watchObservedRunningTime="2025-12-10 23:05:44.549245542 +0000 UTC m=+978.978863155" Dec 10 23:05:44 crc kubenswrapper[4791]: I1210 23:05:44.579308 4791 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/placement-operator-controller-manager-78f8948974-bhgvx" podStartSLOduration=3.666198948 podStartE2EDuration="37.57928564s" podCreationTimestamp="2025-12-10 23:05:07 +0000 UTC" firstStartedPulling="2025-12-10 23:05:09.460292721 +0000 UTC m=+943.889910344" lastFinishedPulling="2025-12-10 23:05:43.373379423 +0000 UTC m=+977.802997036" observedRunningTime="2025-12-10 23:05:44.561868621 +0000 UTC m=+978.991486244" watchObservedRunningTime="2025-12-10 23:05:44.57928564 +0000 UTC m=+979.008903253" Dec 10 23:05:44 crc kubenswrapper[4791]: I1210 23:05:44.613591 4791 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/manila-operator-controller-manager-5b5fd79c9c-tsbtp" podStartSLOduration=2.893086844 podStartE2EDuration="37.613563422s" podCreationTimestamp="2025-12-10 23:05:07 +0000 UTC" firstStartedPulling="2025-12-10 23:05:09.415889002 +0000 UTC m=+943.845506615" lastFinishedPulling="2025-12-10 23:05:44.13636558 +0000 UTC m=+978.565983193" observedRunningTime="2025-12-10 23:05:44.603043155 +0000 UTC m=+979.032660768" watchObservedRunningTime="2025-12-10 23:05:44.613563422 +0000 UTC m=+979.043181035" Dec 10 23:05:47 crc kubenswrapper[4791]: I1210 23:05:47.555470 4791 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/designate-operator-controller-manager-697fb699cf-vg956" Dec 10 23:05:47 crc kubenswrapper[4791]: I1210 23:05:47.567113 4791 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/openstack-baremetal-operator-controller-manager-84b575879fv4t28" event={"ID":"d5907af9-fe3f-4da0-ba70-993d36b25746","Type":"ContainerStarted","Data":"36b5decc3256a3f3c026b3bc9f99949f38ee080c858be06185b4637f228bfb47"} Dec 10 23:05:47 crc kubenswrapper[4791]: I1210 23:05:47.567220 4791 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/openstack-baremetal-operator-controller-manager-84b575879fv4t28" Dec 10 23:05:47 crc kubenswrapper[4791]: I1210 23:05:47.567238 4791 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/openstack-baremetal-operator-controller-manager-84b575879fv4t28" event={"ID":"d5907af9-fe3f-4da0-ba70-993d36b25746","Type":"ContainerStarted","Data":"78aeeb616d261a92a715b53a3fc5c682f75948ab48631bd14259da088386eb75"} Dec 10 23:05:47 crc kubenswrapper[4791]: I1210 23:05:47.570845 4791 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/infra-operator-controller-manager-78d48bff9d-mmczn" event={"ID":"416b455d-b397-4aad-baf1-88e880619eb5","Type":"ContainerStarted","Data":"4a4acf0d2f38f251d014a1fce03e24e4ed8323eaa7a62c864f775e3e88c88ead"} Dec 10 23:05:47 crc kubenswrapper[4791]: I1210 23:05:47.570928 4791 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/infra-operator-controller-manager-78d48bff9d-mmczn" event={"ID":"416b455d-b397-4aad-baf1-88e880619eb5","Type":"ContainerStarted","Data":"a94ddda8bc38f8fc9aca7a3a8f7efd1048480b7bf22b0997819a9ce54b449461"} Dec 10 23:05:47 crc kubenswrapper[4791]: I1210 23:05:47.571006 4791 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/infra-operator-controller-manager-78d48bff9d-mmczn" Dec 10 23:05:47 crc kubenswrapper[4791]: I1210 23:05:47.587799 4791 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/cinder-operator-controller-manager-6c677c69b-6jmn2" event={"ID":"48cd2bae-61e0-446c-ac2a-48e70bff5187","Type":"ContainerStarted","Data":"cbe17bc054b82b1e5c74de3d33902037908a1351d2d6fa9695d163d763bd1307"} Dec 10 23:05:47 crc kubenswrapper[4791]: I1210 23:05:47.588526 4791 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/cinder-operator-controller-manager-6c677c69b-6jmn2" Dec 10 23:05:47 crc kubenswrapper[4791]: I1210 23:05:47.590235 4791 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/horizon-operator-controller-manager-68c6d99b8f-2lq47" event={"ID":"e3fdc812-fbd3-4d35-8555-d55e9390ebd0","Type":"ContainerStarted","Data":"09431d2444414af4614d966e0735e12b95e769d76cff9094ef6913adc5a30430"} Dec 10 23:05:47 crc kubenswrapper[4791]: I1210 23:05:47.590502 4791 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/horizon-operator-controller-manager-68c6d99b8f-2lq47" Dec 10 23:05:47 crc kubenswrapper[4791]: I1210 23:05:47.625149 4791 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/infra-operator-controller-manager-78d48bff9d-mmczn" podStartSLOduration=34.501043803 podStartE2EDuration="40.62512852s" podCreationTimestamp="2025-12-10 23:05:07 +0000 UTC" firstStartedPulling="2025-12-10 23:05:40.334871638 +0000 UTC m=+974.764489251" lastFinishedPulling="2025-12-10 23:05:46.458956345 +0000 UTC m=+980.888573968" observedRunningTime="2025-12-10 23:05:47.617165947 +0000 UTC m=+982.046783560" watchObservedRunningTime="2025-12-10 23:05:47.62512852 +0000 UTC m=+982.054746133" Dec 10 23:05:47 crc kubenswrapper[4791]: I1210 23:05:47.664389 4791 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/cinder-operator-controller-manager-6c677c69b-6jmn2" podStartSLOduration=2.975001141 podStartE2EDuration="40.664319076s" podCreationTimestamp="2025-12-10 23:05:07 +0000 UTC" firstStartedPulling="2025-12-10 23:05:08.769405053 +0000 UTC m=+943.199022666" lastFinishedPulling="2025-12-10 23:05:46.458722988 +0000 UTC m=+980.888340601" observedRunningTime="2025-12-10 23:05:47.660842444 +0000 UTC m=+982.090460057" watchObservedRunningTime="2025-12-10 23:05:47.664319076 +0000 UTC m=+982.093936689" Dec 10 23:05:47 crc kubenswrapper[4791]: I1210 23:05:47.664912 4791 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/openstack-baremetal-operator-controller-manager-84b575879fv4t28" podStartSLOduration=35.143785485 podStartE2EDuration="40.664905953s" podCreationTimestamp="2025-12-10 23:05:07 +0000 UTC" firstStartedPulling="2025-12-10 23:05:40.947326924 +0000 UTC m=+975.376944537" lastFinishedPulling="2025-12-10 23:05:46.468447372 +0000 UTC m=+980.898065005" observedRunningTime="2025-12-10 23:05:47.644166656 +0000 UTC m=+982.073784269" watchObservedRunningTime="2025-12-10 23:05:47.664905953 +0000 UTC m=+982.094523566" Dec 10 23:05:47 crc kubenswrapper[4791]: I1210 23:05:47.681272 4791 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/horizon-operator-controller-manager-68c6d99b8f-2lq47" podStartSLOduration=3.667816515 podStartE2EDuration="40.681249441s" podCreationTimestamp="2025-12-10 23:05:07 +0000 UTC" firstStartedPulling="2025-12-10 23:05:09.447020992 +0000 UTC m=+943.876638595" lastFinishedPulling="2025-12-10 23:05:46.460453908 +0000 UTC m=+980.890071521" observedRunningTime="2025-12-10 23:05:47.67473152 +0000 UTC m=+982.104349143" watchObservedRunningTime="2025-12-10 23:05:47.681249441 +0000 UTC m=+982.110867064" Dec 10 23:05:47 crc kubenswrapper[4791]: I1210 23:05:47.825929 4791 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/manila-operator-controller-manager-5b5fd79c9c-tsbtp" Dec 10 23:05:47 crc kubenswrapper[4791]: I1210 23:05:47.845033 4791 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/mariadb-operator-controller-manager-79c8c4686c-2wmm4" Dec 10 23:05:48 crc kubenswrapper[4791]: I1210 23:05:48.022231 4791 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/octavia-operator-controller-manager-998648c74-sjp6q" Dec 10 23:05:48 crc kubenswrapper[4791]: I1210 23:05:48.146474 4791 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/telemetry-operator-controller-manager-58d5ff84df-t6tnb" Dec 10 23:05:48 crc kubenswrapper[4791]: I1210 23:05:48.156158 4791 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/test-operator-controller-manager-5854674fcc-rw7ct" Dec 10 23:05:48 crc kubenswrapper[4791]: I1210 23:05:48.301024 4791 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/placement-operator-controller-manager-78f8948974-bhgvx" Dec 10 23:05:50 crc kubenswrapper[4791]: I1210 23:05:50.392938 4791 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/openstack-operator-controller-manager-686dfd865c-cft5f" Dec 10 23:05:55 crc kubenswrapper[4791]: I1210 23:05:55.038634 4791 patch_prober.go:28] interesting pod/machine-config-daemon-5rb5l container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 10 23:05:55 crc kubenswrapper[4791]: I1210 23:05:55.038700 4791 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-5rb5l" podUID="ba35653c-6e06-4cee-a4d6-137764090d18" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 10 23:05:57 crc kubenswrapper[4791]: I1210 23:05:57.485748 4791 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/cinder-operator-controller-manager-6c677c69b-6jmn2" Dec 10 23:05:57 crc kubenswrapper[4791]: I1210 23:05:57.566059 4791 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/glance-operator-controller-manager-5697bb5779-cvbqc" Dec 10 23:05:57 crc kubenswrapper[4791]: I1210 23:05:57.672168 4791 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/horizon-operator-controller-manager-68c6d99b8f-2lq47" Dec 10 23:05:57 crc kubenswrapper[4791]: I1210 23:05:57.830907 4791 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/manila-operator-controller-manager-5b5fd79c9c-tsbtp" Dec 10 23:05:57 crc kubenswrapper[4791]: I1210 23:05:57.868847 4791 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/neutron-operator-controller-manager-5fdfd5b6b5-g8f76" Dec 10 23:05:57 crc kubenswrapper[4791]: I1210 23:05:57.900845 4791 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/nova-operator-controller-manager-697bc559fc-2lkcl" Dec 10 23:05:59 crc kubenswrapper[4791]: I1210 23:05:59.517791 4791 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/infra-operator-controller-manager-78d48bff9d-mmczn" Dec 10 23:06:00 crc kubenswrapper[4791]: I1210 23:06:00.067309 4791 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/openstack-baremetal-operator-controller-manager-84b575879fv4t28" Dec 10 23:06:19 crc kubenswrapper[4791]: I1210 23:06:19.209857 4791 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-675f4bcbfc-68bpm"] Dec 10 23:06:19 crc kubenswrapper[4791]: I1210 23:06:19.212310 4791 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-675f4bcbfc-68bpm" Dec 10 23:06:19 crc kubenswrapper[4791]: I1210 23:06:19.218235 4791 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-675f4bcbfc-68bpm"] Dec 10 23:06:19 crc kubenswrapper[4791]: I1210 23:06:19.219768 4791 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dnsmasq-dns-dockercfg-l6gv5" Dec 10 23:06:19 crc kubenswrapper[4791]: I1210 23:06:19.220001 4791 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"dns" Dec 10 23:06:19 crc kubenswrapper[4791]: I1210 23:06:19.220225 4791 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"kube-root-ca.crt" Dec 10 23:06:19 crc kubenswrapper[4791]: I1210 23:06:19.220657 4791 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openshift-service-ca.crt" Dec 10 23:06:19 crc kubenswrapper[4791]: I1210 23:06:19.275425 4791 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-78dd6ddcc-gg848"] Dec 10 23:06:19 crc kubenswrapper[4791]: I1210 23:06:19.276991 4791 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-78dd6ddcc-gg848" Dec 10 23:06:19 crc kubenswrapper[4791]: I1210 23:06:19.280399 4791 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"dns-svc" Dec 10 23:06:19 crc kubenswrapper[4791]: I1210 23:06:19.295986 4791 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-78dd6ddcc-gg848"] Dec 10 23:06:19 crc kubenswrapper[4791]: I1210 23:06:19.328899 4791 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-578pp\" (UniqueName: \"kubernetes.io/projected/6dd522dc-1554-4d27-a289-78b0650df8d1-kube-api-access-578pp\") pod \"dnsmasq-dns-675f4bcbfc-68bpm\" (UID: \"6dd522dc-1554-4d27-a289-78b0650df8d1\") " pod="openstack/dnsmasq-dns-675f4bcbfc-68bpm" Dec 10 23:06:19 crc kubenswrapper[4791]: I1210 23:06:19.329201 4791 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/6dd522dc-1554-4d27-a289-78b0650df8d1-config\") pod \"dnsmasq-dns-675f4bcbfc-68bpm\" (UID: \"6dd522dc-1554-4d27-a289-78b0650df8d1\") " pod="openstack/dnsmasq-dns-675f4bcbfc-68bpm" Dec 10 23:06:19 crc kubenswrapper[4791]: I1210 23:06:19.430844 4791 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/bb91eb46-389f-46d4-8fc3-a84266bb5aff-config\") pod \"dnsmasq-dns-78dd6ddcc-gg848\" (UID: \"bb91eb46-389f-46d4-8fc3-a84266bb5aff\") " pod="openstack/dnsmasq-dns-78dd6ddcc-gg848" Dec 10 23:06:19 crc kubenswrapper[4791]: I1210 23:06:19.430889 4791 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-578pp\" (UniqueName: \"kubernetes.io/projected/6dd522dc-1554-4d27-a289-78b0650df8d1-kube-api-access-578pp\") pod \"dnsmasq-dns-675f4bcbfc-68bpm\" (UID: \"6dd522dc-1554-4d27-a289-78b0650df8d1\") " pod="openstack/dnsmasq-dns-675f4bcbfc-68bpm" Dec 10 23:06:19 crc kubenswrapper[4791]: I1210 23:06:19.430917 4791 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/6dd522dc-1554-4d27-a289-78b0650df8d1-config\") pod \"dnsmasq-dns-675f4bcbfc-68bpm\" (UID: \"6dd522dc-1554-4d27-a289-78b0650df8d1\") " pod="openstack/dnsmasq-dns-675f4bcbfc-68bpm" Dec 10 23:06:19 crc kubenswrapper[4791]: I1210 23:06:19.430937 4791 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-fbfng\" (UniqueName: \"kubernetes.io/projected/bb91eb46-389f-46d4-8fc3-a84266bb5aff-kube-api-access-fbfng\") pod \"dnsmasq-dns-78dd6ddcc-gg848\" (UID: \"bb91eb46-389f-46d4-8fc3-a84266bb5aff\") " pod="openstack/dnsmasq-dns-78dd6ddcc-gg848" Dec 10 23:06:19 crc kubenswrapper[4791]: I1210 23:06:19.430969 4791 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/bb91eb46-389f-46d4-8fc3-a84266bb5aff-dns-svc\") pod \"dnsmasq-dns-78dd6ddcc-gg848\" (UID: \"bb91eb46-389f-46d4-8fc3-a84266bb5aff\") " pod="openstack/dnsmasq-dns-78dd6ddcc-gg848" Dec 10 23:06:19 crc kubenswrapper[4791]: I1210 23:06:19.432039 4791 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/6dd522dc-1554-4d27-a289-78b0650df8d1-config\") pod \"dnsmasq-dns-675f4bcbfc-68bpm\" (UID: \"6dd522dc-1554-4d27-a289-78b0650df8d1\") " pod="openstack/dnsmasq-dns-675f4bcbfc-68bpm" Dec 10 23:06:19 crc kubenswrapper[4791]: I1210 23:06:19.449838 4791 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-578pp\" (UniqueName: \"kubernetes.io/projected/6dd522dc-1554-4d27-a289-78b0650df8d1-kube-api-access-578pp\") pod \"dnsmasq-dns-675f4bcbfc-68bpm\" (UID: \"6dd522dc-1554-4d27-a289-78b0650df8d1\") " pod="openstack/dnsmasq-dns-675f4bcbfc-68bpm" Dec 10 23:06:19 crc kubenswrapper[4791]: I1210 23:06:19.532181 4791 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/bb91eb46-389f-46d4-8fc3-a84266bb5aff-config\") pod \"dnsmasq-dns-78dd6ddcc-gg848\" (UID: \"bb91eb46-389f-46d4-8fc3-a84266bb5aff\") " pod="openstack/dnsmasq-dns-78dd6ddcc-gg848" Dec 10 23:06:19 crc kubenswrapper[4791]: I1210 23:06:19.532249 4791 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-fbfng\" (UniqueName: \"kubernetes.io/projected/bb91eb46-389f-46d4-8fc3-a84266bb5aff-kube-api-access-fbfng\") pod \"dnsmasq-dns-78dd6ddcc-gg848\" (UID: \"bb91eb46-389f-46d4-8fc3-a84266bb5aff\") " pod="openstack/dnsmasq-dns-78dd6ddcc-gg848" Dec 10 23:06:19 crc kubenswrapper[4791]: I1210 23:06:19.532298 4791 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/bb91eb46-389f-46d4-8fc3-a84266bb5aff-dns-svc\") pod \"dnsmasq-dns-78dd6ddcc-gg848\" (UID: \"bb91eb46-389f-46d4-8fc3-a84266bb5aff\") " pod="openstack/dnsmasq-dns-78dd6ddcc-gg848" Dec 10 23:06:19 crc kubenswrapper[4791]: I1210 23:06:19.533033 4791 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/bb91eb46-389f-46d4-8fc3-a84266bb5aff-dns-svc\") pod \"dnsmasq-dns-78dd6ddcc-gg848\" (UID: \"bb91eb46-389f-46d4-8fc3-a84266bb5aff\") " pod="openstack/dnsmasq-dns-78dd6ddcc-gg848" Dec 10 23:06:19 crc kubenswrapper[4791]: I1210 23:06:19.533036 4791 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/bb91eb46-389f-46d4-8fc3-a84266bb5aff-config\") pod \"dnsmasq-dns-78dd6ddcc-gg848\" (UID: \"bb91eb46-389f-46d4-8fc3-a84266bb5aff\") " pod="openstack/dnsmasq-dns-78dd6ddcc-gg848" Dec 10 23:06:19 crc kubenswrapper[4791]: I1210 23:06:19.537960 4791 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-675f4bcbfc-68bpm" Dec 10 23:06:19 crc kubenswrapper[4791]: I1210 23:06:19.556057 4791 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-fbfng\" (UniqueName: \"kubernetes.io/projected/bb91eb46-389f-46d4-8fc3-a84266bb5aff-kube-api-access-fbfng\") pod \"dnsmasq-dns-78dd6ddcc-gg848\" (UID: \"bb91eb46-389f-46d4-8fc3-a84266bb5aff\") " pod="openstack/dnsmasq-dns-78dd6ddcc-gg848" Dec 10 23:06:19 crc kubenswrapper[4791]: I1210 23:06:19.604190 4791 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-78dd6ddcc-gg848" Dec 10 23:06:20 crc kubenswrapper[4791]: I1210 23:06:20.060172 4791 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-675f4bcbfc-68bpm"] Dec 10 23:06:20 crc kubenswrapper[4791]: I1210 23:06:20.113131 4791 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-78dd6ddcc-gg848"] Dec 10 23:06:20 crc kubenswrapper[4791]: W1210 23:06:20.118632 4791 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podbb91eb46_389f_46d4_8fc3_a84266bb5aff.slice/crio-c73c67e9a3cab9f4f79fce0e6c2a2b49d4fb828ea36b11a377c77404dc287ce7 WatchSource:0}: Error finding container c73c67e9a3cab9f4f79fce0e6c2a2b49d4fb828ea36b11a377c77404dc287ce7: Status 404 returned error can't find the container with id c73c67e9a3cab9f4f79fce0e6c2a2b49d4fb828ea36b11a377c77404dc287ce7 Dec 10 23:06:20 crc kubenswrapper[4791]: I1210 23:06:20.875641 4791 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-78dd6ddcc-gg848" event={"ID":"bb91eb46-389f-46d4-8fc3-a84266bb5aff","Type":"ContainerStarted","Data":"c73c67e9a3cab9f4f79fce0e6c2a2b49d4fb828ea36b11a377c77404dc287ce7"} Dec 10 23:06:20 crc kubenswrapper[4791]: I1210 23:06:20.877849 4791 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-675f4bcbfc-68bpm" event={"ID":"6dd522dc-1554-4d27-a289-78b0650df8d1","Type":"ContainerStarted","Data":"7338da6d30033fbd48e70f4fc83255e34b5b5143034535de719e8f6318120a52"} Dec 10 23:06:22 crc kubenswrapper[4791]: I1210 23:06:22.125773 4791 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-675f4bcbfc-68bpm"] Dec 10 23:06:22 crc kubenswrapper[4791]: I1210 23:06:22.147753 4791 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-5ccc8479f9-8w9hb"] Dec 10 23:06:22 crc kubenswrapper[4791]: I1210 23:06:22.150217 4791 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-5ccc8479f9-8w9hb" Dec 10 23:06:22 crc kubenswrapper[4791]: I1210 23:06:22.195557 4791 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-5ccc8479f9-8w9hb"] Dec 10 23:06:22 crc kubenswrapper[4791]: I1210 23:06:22.268827 4791 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/73f16e68-0124-4784-977f-72ea5d6d04bf-dns-svc\") pod \"dnsmasq-dns-5ccc8479f9-8w9hb\" (UID: \"73f16e68-0124-4784-977f-72ea5d6d04bf\") " pod="openstack/dnsmasq-dns-5ccc8479f9-8w9hb" Dec 10 23:06:22 crc kubenswrapper[4791]: I1210 23:06:22.268886 4791 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/73f16e68-0124-4784-977f-72ea5d6d04bf-config\") pod \"dnsmasq-dns-5ccc8479f9-8w9hb\" (UID: \"73f16e68-0124-4784-977f-72ea5d6d04bf\") " pod="openstack/dnsmasq-dns-5ccc8479f9-8w9hb" Dec 10 23:06:22 crc kubenswrapper[4791]: I1210 23:06:22.268917 4791 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-zn4rz\" (UniqueName: \"kubernetes.io/projected/73f16e68-0124-4784-977f-72ea5d6d04bf-kube-api-access-zn4rz\") pod \"dnsmasq-dns-5ccc8479f9-8w9hb\" (UID: \"73f16e68-0124-4784-977f-72ea5d6d04bf\") " pod="openstack/dnsmasq-dns-5ccc8479f9-8w9hb" Dec 10 23:06:22 crc kubenswrapper[4791]: I1210 23:06:22.369681 4791 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/73f16e68-0124-4784-977f-72ea5d6d04bf-dns-svc\") pod \"dnsmasq-dns-5ccc8479f9-8w9hb\" (UID: \"73f16e68-0124-4784-977f-72ea5d6d04bf\") " pod="openstack/dnsmasq-dns-5ccc8479f9-8w9hb" Dec 10 23:06:22 crc kubenswrapper[4791]: I1210 23:06:22.369720 4791 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/73f16e68-0124-4784-977f-72ea5d6d04bf-config\") pod \"dnsmasq-dns-5ccc8479f9-8w9hb\" (UID: \"73f16e68-0124-4784-977f-72ea5d6d04bf\") " pod="openstack/dnsmasq-dns-5ccc8479f9-8w9hb" Dec 10 23:06:22 crc kubenswrapper[4791]: I1210 23:06:22.369750 4791 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-zn4rz\" (UniqueName: \"kubernetes.io/projected/73f16e68-0124-4784-977f-72ea5d6d04bf-kube-api-access-zn4rz\") pod \"dnsmasq-dns-5ccc8479f9-8w9hb\" (UID: \"73f16e68-0124-4784-977f-72ea5d6d04bf\") " pod="openstack/dnsmasq-dns-5ccc8479f9-8w9hb" Dec 10 23:06:22 crc kubenswrapper[4791]: I1210 23:06:22.370659 4791 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/73f16e68-0124-4784-977f-72ea5d6d04bf-dns-svc\") pod \"dnsmasq-dns-5ccc8479f9-8w9hb\" (UID: \"73f16e68-0124-4784-977f-72ea5d6d04bf\") " pod="openstack/dnsmasq-dns-5ccc8479f9-8w9hb" Dec 10 23:06:22 crc kubenswrapper[4791]: I1210 23:06:22.370760 4791 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/73f16e68-0124-4784-977f-72ea5d6d04bf-config\") pod \"dnsmasq-dns-5ccc8479f9-8w9hb\" (UID: \"73f16e68-0124-4784-977f-72ea5d6d04bf\") " pod="openstack/dnsmasq-dns-5ccc8479f9-8w9hb" Dec 10 23:06:22 crc kubenswrapper[4791]: I1210 23:06:22.376013 4791 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-78dd6ddcc-gg848"] Dec 10 23:06:22 crc kubenswrapper[4791]: I1210 23:06:22.388403 4791 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-zn4rz\" (UniqueName: \"kubernetes.io/projected/73f16e68-0124-4784-977f-72ea5d6d04bf-kube-api-access-zn4rz\") pod \"dnsmasq-dns-5ccc8479f9-8w9hb\" (UID: \"73f16e68-0124-4784-977f-72ea5d6d04bf\") " pod="openstack/dnsmasq-dns-5ccc8479f9-8w9hb" Dec 10 23:06:22 crc kubenswrapper[4791]: I1210 23:06:22.475862 4791 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-5ccc8479f9-8w9hb" Dec 10 23:06:22 crc kubenswrapper[4791]: I1210 23:06:22.834694 4791 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-5ccc8479f9-8w9hb"] Dec 10 23:06:22 crc kubenswrapper[4791]: W1210 23:06:22.840471 4791 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod73f16e68_0124_4784_977f_72ea5d6d04bf.slice/crio-1cf836f70d0946a02cb65df3a7c27da984eb3de1f80c9e4d62a08bc0b144969b WatchSource:0}: Error finding container 1cf836f70d0946a02cb65df3a7c27da984eb3de1f80c9e4d62a08bc0b144969b: Status 404 returned error can't find the container with id 1cf836f70d0946a02cb65df3a7c27da984eb3de1f80c9e4d62a08bc0b144969b Dec 10 23:06:22 crc kubenswrapper[4791]: I1210 23:06:22.890505 4791 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-5ccc8479f9-8w9hb" event={"ID":"73f16e68-0124-4784-977f-72ea5d6d04bf","Type":"ContainerStarted","Data":"1cf836f70d0946a02cb65df3a7c27da984eb3de1f80c9e4d62a08bc0b144969b"} Dec 10 23:06:23 crc kubenswrapper[4791]: I1210 23:06:23.617128 4791 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-57d769cc4f-tnfbj"] Dec 10 23:06:23 crc kubenswrapper[4791]: I1210 23:06:23.618778 4791 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-57d769cc4f-tnfbj" Dec 10 23:06:23 crc kubenswrapper[4791]: I1210 23:06:23.625368 4791 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-57d769cc4f-tnfbj"] Dec 10 23:06:23 crc kubenswrapper[4791]: I1210 23:06:23.788083 4791 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-kwvlk\" (UniqueName: \"kubernetes.io/projected/b2c7ad15-ca5a-4042-9d55-31524249a851-kube-api-access-kwvlk\") pod \"dnsmasq-dns-57d769cc4f-tnfbj\" (UID: \"b2c7ad15-ca5a-4042-9d55-31524249a851\") " pod="openstack/dnsmasq-dns-57d769cc4f-tnfbj" Dec 10 23:06:23 crc kubenswrapper[4791]: I1210 23:06:23.788435 4791 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/b2c7ad15-ca5a-4042-9d55-31524249a851-config\") pod \"dnsmasq-dns-57d769cc4f-tnfbj\" (UID: \"b2c7ad15-ca5a-4042-9d55-31524249a851\") " pod="openstack/dnsmasq-dns-57d769cc4f-tnfbj" Dec 10 23:06:23 crc kubenswrapper[4791]: I1210 23:06:23.788570 4791 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/b2c7ad15-ca5a-4042-9d55-31524249a851-dns-svc\") pod \"dnsmasq-dns-57d769cc4f-tnfbj\" (UID: \"b2c7ad15-ca5a-4042-9d55-31524249a851\") " pod="openstack/dnsmasq-dns-57d769cc4f-tnfbj" Dec 10 23:06:23 crc kubenswrapper[4791]: I1210 23:06:23.892806 4791 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-kwvlk\" (UniqueName: \"kubernetes.io/projected/b2c7ad15-ca5a-4042-9d55-31524249a851-kube-api-access-kwvlk\") pod \"dnsmasq-dns-57d769cc4f-tnfbj\" (UID: \"b2c7ad15-ca5a-4042-9d55-31524249a851\") " pod="openstack/dnsmasq-dns-57d769cc4f-tnfbj" Dec 10 23:06:23 crc kubenswrapper[4791]: I1210 23:06:23.892884 4791 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/b2c7ad15-ca5a-4042-9d55-31524249a851-config\") pod \"dnsmasq-dns-57d769cc4f-tnfbj\" (UID: \"b2c7ad15-ca5a-4042-9d55-31524249a851\") " pod="openstack/dnsmasq-dns-57d769cc4f-tnfbj" Dec 10 23:06:23 crc kubenswrapper[4791]: I1210 23:06:23.893862 4791 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/b2c7ad15-ca5a-4042-9d55-31524249a851-dns-svc\") pod \"dnsmasq-dns-57d769cc4f-tnfbj\" (UID: \"b2c7ad15-ca5a-4042-9d55-31524249a851\") " pod="openstack/dnsmasq-dns-57d769cc4f-tnfbj" Dec 10 23:06:23 crc kubenswrapper[4791]: I1210 23:06:23.895037 4791 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/b2c7ad15-ca5a-4042-9d55-31524249a851-dns-svc\") pod \"dnsmasq-dns-57d769cc4f-tnfbj\" (UID: \"b2c7ad15-ca5a-4042-9d55-31524249a851\") " pod="openstack/dnsmasq-dns-57d769cc4f-tnfbj" Dec 10 23:06:23 crc kubenswrapper[4791]: I1210 23:06:23.895200 4791 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/b2c7ad15-ca5a-4042-9d55-31524249a851-config\") pod \"dnsmasq-dns-57d769cc4f-tnfbj\" (UID: \"b2c7ad15-ca5a-4042-9d55-31524249a851\") " pod="openstack/dnsmasq-dns-57d769cc4f-tnfbj" Dec 10 23:06:23 crc kubenswrapper[4791]: I1210 23:06:23.919023 4791 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-kwvlk\" (UniqueName: \"kubernetes.io/projected/b2c7ad15-ca5a-4042-9d55-31524249a851-kube-api-access-kwvlk\") pod \"dnsmasq-dns-57d769cc4f-tnfbj\" (UID: \"b2c7ad15-ca5a-4042-9d55-31524249a851\") " pod="openstack/dnsmasq-dns-57d769cc4f-tnfbj" Dec 10 23:06:23 crc kubenswrapper[4791]: I1210 23:06:23.943313 4791 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-57d769cc4f-tnfbj" Dec 10 23:06:24 crc kubenswrapper[4791]: I1210 23:06:24.621232 4791 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/rabbitmq-server-0"] Dec 10 23:06:24 crc kubenswrapper[4791]: I1210 23:06:24.623279 4791 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/rabbitmq-server-0" Dec 10 23:06:24 crc kubenswrapper[4791]: I1210 23:06:24.625121 4791 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"rabbitmq-erlang-cookie" Dec 10 23:06:24 crc kubenswrapper[4791]: I1210 23:06:24.626009 4791 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"rabbitmq-default-user" Dec 10 23:06:24 crc kubenswrapper[4791]: I1210 23:06:24.626355 4791 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"rabbitmq-server-conf" Dec 10 23:06:24 crc kubenswrapper[4791]: I1210 23:06:24.626364 4791 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-rabbitmq-svc" Dec 10 23:06:24 crc kubenswrapper[4791]: I1210 23:06:24.626458 4791 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"rabbitmq-server-dockercfg-568t8" Dec 10 23:06:24 crc kubenswrapper[4791]: I1210 23:06:24.628297 4791 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"rabbitmq-config-data" Dec 10 23:06:24 crc kubenswrapper[4791]: I1210 23:06:24.628463 4791 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"rabbitmq-plugins-conf" Dec 10 23:06:24 crc kubenswrapper[4791]: I1210 23:06:24.633305 4791 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/rabbitmq-cell1-server-0"] Dec 10 23:06:24 crc kubenswrapper[4791]: I1210 23:06:24.634642 4791 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/rabbitmq-cell1-server-0" Dec 10 23:06:24 crc kubenswrapper[4791]: I1210 23:06:24.641164 4791 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/rabbitmq-cell1-server-0"] Dec 10 23:06:24 crc kubenswrapper[4791]: I1210 23:06:24.642655 4791 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"rabbitmq-cell1-erlang-cookie" Dec 10 23:06:24 crc kubenswrapper[4791]: I1210 23:06:24.642703 4791 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"rabbitmq-cell1-server-conf" Dec 10 23:06:24 crc kubenswrapper[4791]: I1210 23:06:24.642716 4791 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"rabbitmq-cell1-server-dockercfg-scxq4" Dec 10 23:06:24 crc kubenswrapper[4791]: I1210 23:06:24.642810 4791 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"rabbitmq-cell1-plugins-conf" Dec 10 23:06:24 crc kubenswrapper[4791]: I1210 23:06:24.642880 4791 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"rabbitmq-cell1-default-user" Dec 10 23:06:24 crc kubenswrapper[4791]: I1210 23:06:24.643315 4791 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-rabbitmq-cell1-svc" Dec 10 23:06:24 crc kubenswrapper[4791]: I1210 23:06:24.644179 4791 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"rabbitmq-cell1-config-data" Dec 10 23:06:24 crc kubenswrapper[4791]: I1210 23:06:24.648184 4791 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/rabbitmq-server-0"] Dec 10 23:06:24 crc kubenswrapper[4791]: I1210 23:06:24.691950 4791 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-57d769cc4f-tnfbj"] Dec 10 23:06:24 crc kubenswrapper[4791]: I1210 23:06:24.810432 4791 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/47eb9200-b0bc-41a6-abb9-f6167bd69c66-server-conf\") pod \"rabbitmq-cell1-server-0\" (UID: \"47eb9200-b0bc-41a6-abb9-f6167bd69c66\") " pod="openstack/rabbitmq-cell1-server-0" Dec 10 23:06:24 crc kubenswrapper[4791]: I1210 23:06:24.810513 4791 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-tls\" (UniqueName: \"kubernetes.io/projected/9ce8dbeb-dce8-47ac-98f1-fe92f74aa3de-rabbitmq-tls\") pod \"rabbitmq-server-0\" (UID: \"9ce8dbeb-dce8-47ac-98f1-fe92f74aa3de\") " pod="openstack/rabbitmq-server-0" Dec 10 23:06:24 crc kubenswrapper[4791]: I1210 23:06:24.810534 4791 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/9ce8dbeb-dce8-47ac-98f1-fe92f74aa3de-config-data\") pod \"rabbitmq-server-0\" (UID: \"9ce8dbeb-dce8-47ac-98f1-fe92f74aa3de\") " pod="openstack/rabbitmq-server-0" Dec 10 23:06:24 crc kubenswrapper[4791]: I1210 23:06:24.810549 4791 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/47eb9200-b0bc-41a6-abb9-f6167bd69c66-rabbitmq-erlang-cookie\") pod \"rabbitmq-cell1-server-0\" (UID: \"47eb9200-b0bc-41a6-abb9-f6167bd69c66\") " pod="openstack/rabbitmq-cell1-server-0" Dec 10 23:06:24 crc kubenswrapper[4791]: I1210 23:06:24.810665 4791 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/47eb9200-b0bc-41a6-abb9-f6167bd69c66-plugins-conf\") pod \"rabbitmq-cell1-server-0\" (UID: \"47eb9200-b0bc-41a6-abb9-f6167bd69c66\") " pod="openstack/rabbitmq-cell1-server-0" Dec 10 23:06:24 crc kubenswrapper[4791]: I1210 23:06:24.810853 4791 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/9ce8dbeb-dce8-47ac-98f1-fe92f74aa3de-plugins-conf\") pod \"rabbitmq-server-0\" (UID: \"9ce8dbeb-dce8-47ac-98f1-fe92f74aa3de\") " pod="openstack/rabbitmq-server-0" Dec 10 23:06:24 crc kubenswrapper[4791]: I1210 23:06:24.810896 4791 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/9ce8dbeb-dce8-47ac-98f1-fe92f74aa3de-rabbitmq-erlang-cookie\") pod \"rabbitmq-server-0\" (UID: \"9ce8dbeb-dce8-47ac-98f1-fe92f74aa3de\") " pod="openstack/rabbitmq-server-0" Dec 10 23:06:24 crc kubenswrapper[4791]: I1210 23:06:24.811109 4791 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/47eb9200-b0bc-41a6-abb9-f6167bd69c66-erlang-cookie-secret\") pod \"rabbitmq-cell1-server-0\" (UID: \"47eb9200-b0bc-41a6-abb9-f6167bd69c66\") " pod="openstack/rabbitmq-cell1-server-0" Dec 10 23:06:24 crc kubenswrapper[4791]: I1210 23:06:24.811150 4791 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/9ce8dbeb-dce8-47ac-98f1-fe92f74aa3de-rabbitmq-confd\") pod \"rabbitmq-server-0\" (UID: \"9ce8dbeb-dce8-47ac-98f1-fe92f74aa3de\") " pod="openstack/rabbitmq-server-0" Dec 10 23:06:24 crc kubenswrapper[4791]: I1210 23:06:24.811172 4791 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/9ce8dbeb-dce8-47ac-98f1-fe92f74aa3de-server-conf\") pod \"rabbitmq-server-0\" (UID: \"9ce8dbeb-dce8-47ac-98f1-fe92f74aa3de\") " pod="openstack/rabbitmq-server-0" Dec 10 23:06:24 crc kubenswrapper[4791]: I1210 23:06:24.811212 4791 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-kmc8x\" (UniqueName: \"kubernetes.io/projected/47eb9200-b0bc-41a6-abb9-f6167bd69c66-kube-api-access-kmc8x\") pod \"rabbitmq-cell1-server-0\" (UID: \"47eb9200-b0bc-41a6-abb9-f6167bd69c66\") " pod="openstack/rabbitmq-cell1-server-0" Dec 10 23:06:24 crc kubenswrapper[4791]: I1210 23:06:24.811264 4791 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/47eb9200-b0bc-41a6-abb9-f6167bd69c66-pod-info\") pod \"rabbitmq-cell1-server-0\" (UID: \"47eb9200-b0bc-41a6-abb9-f6167bd69c66\") " pod="openstack/rabbitmq-cell1-server-0" Dec 10 23:06:24 crc kubenswrapper[4791]: I1210 23:06:24.811281 4791 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/47eb9200-b0bc-41a6-abb9-f6167bd69c66-rabbitmq-confd\") pod \"rabbitmq-cell1-server-0\" (UID: \"47eb9200-b0bc-41a6-abb9-f6167bd69c66\") " pod="openstack/rabbitmq-cell1-server-0" Dec 10 23:06:24 crc kubenswrapper[4791]: I1210 23:06:24.811323 4791 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage07-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage07-crc\") pod \"rabbitmq-server-0\" (UID: \"9ce8dbeb-dce8-47ac-98f1-fe92f74aa3de\") " pod="openstack/rabbitmq-server-0" Dec 10 23:06:24 crc kubenswrapper[4791]: I1210 23:06:24.811354 4791 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/47eb9200-b0bc-41a6-abb9-f6167bd69c66-config-data\") pod \"rabbitmq-cell1-server-0\" (UID: \"47eb9200-b0bc-41a6-abb9-f6167bd69c66\") " pod="openstack/rabbitmq-cell1-server-0" Dec 10 23:06:24 crc kubenswrapper[4791]: I1210 23:06:24.811369 4791 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-85rrv\" (UniqueName: \"kubernetes.io/projected/9ce8dbeb-dce8-47ac-98f1-fe92f74aa3de-kube-api-access-85rrv\") pod \"rabbitmq-server-0\" (UID: \"9ce8dbeb-dce8-47ac-98f1-fe92f74aa3de\") " pod="openstack/rabbitmq-server-0" Dec 10 23:06:24 crc kubenswrapper[4791]: I1210 23:06:24.811406 4791 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-tls\" (UniqueName: \"kubernetes.io/projected/47eb9200-b0bc-41a6-abb9-f6167bd69c66-rabbitmq-tls\") pod \"rabbitmq-cell1-server-0\" (UID: \"47eb9200-b0bc-41a6-abb9-f6167bd69c66\") " pod="openstack/rabbitmq-cell1-server-0" Dec 10 23:06:24 crc kubenswrapper[4791]: I1210 23:06:24.811422 4791 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage08-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage08-crc\") pod \"rabbitmq-cell1-server-0\" (UID: \"47eb9200-b0bc-41a6-abb9-f6167bd69c66\") " pod="openstack/rabbitmq-cell1-server-0" Dec 10 23:06:24 crc kubenswrapper[4791]: I1210 23:06:24.811439 4791 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/9ce8dbeb-dce8-47ac-98f1-fe92f74aa3de-pod-info\") pod \"rabbitmq-server-0\" (UID: \"9ce8dbeb-dce8-47ac-98f1-fe92f74aa3de\") " pod="openstack/rabbitmq-server-0" Dec 10 23:06:24 crc kubenswrapper[4791]: I1210 23:06:24.811458 4791 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/9ce8dbeb-dce8-47ac-98f1-fe92f74aa3de-rabbitmq-plugins\") pod \"rabbitmq-server-0\" (UID: \"9ce8dbeb-dce8-47ac-98f1-fe92f74aa3de\") " pod="openstack/rabbitmq-server-0" Dec 10 23:06:24 crc kubenswrapper[4791]: I1210 23:06:24.811472 4791 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/47eb9200-b0bc-41a6-abb9-f6167bd69c66-rabbitmq-plugins\") pod \"rabbitmq-cell1-server-0\" (UID: \"47eb9200-b0bc-41a6-abb9-f6167bd69c66\") " pod="openstack/rabbitmq-cell1-server-0" Dec 10 23:06:24 crc kubenswrapper[4791]: I1210 23:06:24.811516 4791 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/9ce8dbeb-dce8-47ac-98f1-fe92f74aa3de-erlang-cookie-secret\") pod \"rabbitmq-server-0\" (UID: \"9ce8dbeb-dce8-47ac-98f1-fe92f74aa3de\") " pod="openstack/rabbitmq-server-0" Dec 10 23:06:24 crc kubenswrapper[4791]: I1210 23:06:24.912539 4791 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-kmc8x\" (UniqueName: \"kubernetes.io/projected/47eb9200-b0bc-41a6-abb9-f6167bd69c66-kube-api-access-kmc8x\") pod \"rabbitmq-cell1-server-0\" (UID: \"47eb9200-b0bc-41a6-abb9-f6167bd69c66\") " pod="openstack/rabbitmq-cell1-server-0" Dec 10 23:06:24 crc kubenswrapper[4791]: I1210 23:06:24.912623 4791 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/47eb9200-b0bc-41a6-abb9-f6167bd69c66-pod-info\") pod \"rabbitmq-cell1-server-0\" (UID: \"47eb9200-b0bc-41a6-abb9-f6167bd69c66\") " pod="openstack/rabbitmq-cell1-server-0" Dec 10 23:06:24 crc kubenswrapper[4791]: I1210 23:06:24.912644 4791 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/47eb9200-b0bc-41a6-abb9-f6167bd69c66-rabbitmq-confd\") pod \"rabbitmq-cell1-server-0\" (UID: \"47eb9200-b0bc-41a6-abb9-f6167bd69c66\") " pod="openstack/rabbitmq-cell1-server-0" Dec 10 23:06:24 crc kubenswrapper[4791]: I1210 23:06:24.912670 4791 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage07-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage07-crc\") pod \"rabbitmq-server-0\" (UID: \"9ce8dbeb-dce8-47ac-98f1-fe92f74aa3de\") " pod="openstack/rabbitmq-server-0" Dec 10 23:06:24 crc kubenswrapper[4791]: I1210 23:06:24.912693 4791 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/47eb9200-b0bc-41a6-abb9-f6167bd69c66-config-data\") pod \"rabbitmq-cell1-server-0\" (UID: \"47eb9200-b0bc-41a6-abb9-f6167bd69c66\") " pod="openstack/rabbitmq-cell1-server-0" Dec 10 23:06:24 crc kubenswrapper[4791]: I1210 23:06:24.912713 4791 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-85rrv\" (UniqueName: \"kubernetes.io/projected/9ce8dbeb-dce8-47ac-98f1-fe92f74aa3de-kube-api-access-85rrv\") pod \"rabbitmq-server-0\" (UID: \"9ce8dbeb-dce8-47ac-98f1-fe92f74aa3de\") " pod="openstack/rabbitmq-server-0" Dec 10 23:06:24 crc kubenswrapper[4791]: I1210 23:06:24.912878 4791 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-tls\" (UniqueName: \"kubernetes.io/projected/47eb9200-b0bc-41a6-abb9-f6167bd69c66-rabbitmq-tls\") pod \"rabbitmq-cell1-server-0\" (UID: \"47eb9200-b0bc-41a6-abb9-f6167bd69c66\") " pod="openstack/rabbitmq-cell1-server-0" Dec 10 23:06:24 crc kubenswrapper[4791]: I1210 23:06:24.912948 4791 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage08-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage08-crc\") pod \"rabbitmq-cell1-server-0\" (UID: \"47eb9200-b0bc-41a6-abb9-f6167bd69c66\") " pod="openstack/rabbitmq-cell1-server-0" Dec 10 23:06:24 crc kubenswrapper[4791]: I1210 23:06:24.912987 4791 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/9ce8dbeb-dce8-47ac-98f1-fe92f74aa3de-pod-info\") pod \"rabbitmq-server-0\" (UID: \"9ce8dbeb-dce8-47ac-98f1-fe92f74aa3de\") " pod="openstack/rabbitmq-server-0" Dec 10 23:06:24 crc kubenswrapper[4791]: I1210 23:06:24.913006 4791 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/9ce8dbeb-dce8-47ac-98f1-fe92f74aa3de-rabbitmq-plugins\") pod \"rabbitmq-server-0\" (UID: \"9ce8dbeb-dce8-47ac-98f1-fe92f74aa3de\") " pod="openstack/rabbitmq-server-0" Dec 10 23:06:24 crc kubenswrapper[4791]: I1210 23:06:24.913024 4791 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/47eb9200-b0bc-41a6-abb9-f6167bd69c66-rabbitmq-plugins\") pod \"rabbitmq-cell1-server-0\" (UID: \"47eb9200-b0bc-41a6-abb9-f6167bd69c66\") " pod="openstack/rabbitmq-cell1-server-0" Dec 10 23:06:24 crc kubenswrapper[4791]: I1210 23:06:24.913060 4791 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/9ce8dbeb-dce8-47ac-98f1-fe92f74aa3de-erlang-cookie-secret\") pod \"rabbitmq-server-0\" (UID: \"9ce8dbeb-dce8-47ac-98f1-fe92f74aa3de\") " pod="openstack/rabbitmq-server-0" Dec 10 23:06:24 crc kubenswrapper[4791]: I1210 23:06:24.913079 4791 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/47eb9200-b0bc-41a6-abb9-f6167bd69c66-server-conf\") pod \"rabbitmq-cell1-server-0\" (UID: \"47eb9200-b0bc-41a6-abb9-f6167bd69c66\") " pod="openstack/rabbitmq-cell1-server-0" Dec 10 23:06:24 crc kubenswrapper[4791]: I1210 23:06:24.913116 4791 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-tls\" (UniqueName: \"kubernetes.io/projected/9ce8dbeb-dce8-47ac-98f1-fe92f74aa3de-rabbitmq-tls\") pod \"rabbitmq-server-0\" (UID: \"9ce8dbeb-dce8-47ac-98f1-fe92f74aa3de\") " pod="openstack/rabbitmq-server-0" Dec 10 23:06:24 crc kubenswrapper[4791]: I1210 23:06:24.913132 4791 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/47eb9200-b0bc-41a6-abb9-f6167bd69c66-rabbitmq-erlang-cookie\") pod \"rabbitmq-cell1-server-0\" (UID: \"47eb9200-b0bc-41a6-abb9-f6167bd69c66\") " pod="openstack/rabbitmq-cell1-server-0" Dec 10 23:06:24 crc kubenswrapper[4791]: I1210 23:06:24.913148 4791 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/9ce8dbeb-dce8-47ac-98f1-fe92f74aa3de-config-data\") pod \"rabbitmq-server-0\" (UID: \"9ce8dbeb-dce8-47ac-98f1-fe92f74aa3de\") " pod="openstack/rabbitmq-server-0" Dec 10 23:06:24 crc kubenswrapper[4791]: I1210 23:06:24.913265 4791 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/47eb9200-b0bc-41a6-abb9-f6167bd69c66-plugins-conf\") pod \"rabbitmq-cell1-server-0\" (UID: \"47eb9200-b0bc-41a6-abb9-f6167bd69c66\") " pod="openstack/rabbitmq-cell1-server-0" Dec 10 23:06:24 crc kubenswrapper[4791]: I1210 23:06:24.913306 4791 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/9ce8dbeb-dce8-47ac-98f1-fe92f74aa3de-plugins-conf\") pod \"rabbitmq-server-0\" (UID: \"9ce8dbeb-dce8-47ac-98f1-fe92f74aa3de\") " pod="openstack/rabbitmq-server-0" Dec 10 23:06:24 crc kubenswrapper[4791]: I1210 23:06:24.913374 4791 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/9ce8dbeb-dce8-47ac-98f1-fe92f74aa3de-rabbitmq-erlang-cookie\") pod \"rabbitmq-server-0\" (UID: \"9ce8dbeb-dce8-47ac-98f1-fe92f74aa3de\") " pod="openstack/rabbitmq-server-0" Dec 10 23:06:24 crc kubenswrapper[4791]: I1210 23:06:24.913428 4791 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/47eb9200-b0bc-41a6-abb9-f6167bd69c66-erlang-cookie-secret\") pod \"rabbitmq-cell1-server-0\" (UID: \"47eb9200-b0bc-41a6-abb9-f6167bd69c66\") " pod="openstack/rabbitmq-cell1-server-0" Dec 10 23:06:24 crc kubenswrapper[4791]: I1210 23:06:24.913446 4791 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/9ce8dbeb-dce8-47ac-98f1-fe92f74aa3de-rabbitmq-confd\") pod \"rabbitmq-server-0\" (UID: \"9ce8dbeb-dce8-47ac-98f1-fe92f74aa3de\") " pod="openstack/rabbitmq-server-0" Dec 10 23:06:24 crc kubenswrapper[4791]: I1210 23:06:24.913467 4791 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/9ce8dbeb-dce8-47ac-98f1-fe92f74aa3de-server-conf\") pod \"rabbitmq-server-0\" (UID: \"9ce8dbeb-dce8-47ac-98f1-fe92f74aa3de\") " pod="openstack/rabbitmq-server-0" Dec 10 23:06:24 crc kubenswrapper[4791]: I1210 23:06:24.913688 4791 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage07-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage07-crc\") pod \"rabbitmq-server-0\" (UID: \"9ce8dbeb-dce8-47ac-98f1-fe92f74aa3de\") device mount path \"/mnt/openstack/pv07\"" pod="openstack/rabbitmq-server-0" Dec 10 23:06:24 crc kubenswrapper[4791]: I1210 23:06:24.913688 4791 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage08-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage08-crc\") pod \"rabbitmq-cell1-server-0\" (UID: \"47eb9200-b0bc-41a6-abb9-f6167bd69c66\") device mount path \"/mnt/openstack/pv08\"" pod="openstack/rabbitmq-cell1-server-0" Dec 10 23:06:24 crc kubenswrapper[4791]: I1210 23:06:24.914817 4791 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/47eb9200-b0bc-41a6-abb9-f6167bd69c66-server-conf\") pod \"rabbitmq-cell1-server-0\" (UID: \"47eb9200-b0bc-41a6-abb9-f6167bd69c66\") " pod="openstack/rabbitmq-cell1-server-0" Dec 10 23:06:24 crc kubenswrapper[4791]: I1210 23:06:24.915135 4791 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/47eb9200-b0bc-41a6-abb9-f6167bd69c66-rabbitmq-plugins\") pod \"rabbitmq-cell1-server-0\" (UID: \"47eb9200-b0bc-41a6-abb9-f6167bd69c66\") " pod="openstack/rabbitmq-cell1-server-0" Dec 10 23:06:24 crc kubenswrapper[4791]: I1210 23:06:24.915383 4791 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/9ce8dbeb-dce8-47ac-98f1-fe92f74aa3de-rabbitmq-plugins\") pod \"rabbitmq-server-0\" (UID: \"9ce8dbeb-dce8-47ac-98f1-fe92f74aa3de\") " pod="openstack/rabbitmq-server-0" Dec 10 23:06:24 crc kubenswrapper[4791]: I1210 23:06:24.915848 4791 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/47eb9200-b0bc-41a6-abb9-f6167bd69c66-plugins-conf\") pod \"rabbitmq-cell1-server-0\" (UID: \"47eb9200-b0bc-41a6-abb9-f6167bd69c66\") " pod="openstack/rabbitmq-cell1-server-0" Dec 10 23:06:24 crc kubenswrapper[4791]: I1210 23:06:24.915878 4791 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/9ce8dbeb-dce8-47ac-98f1-fe92f74aa3de-plugins-conf\") pod \"rabbitmq-server-0\" (UID: \"9ce8dbeb-dce8-47ac-98f1-fe92f74aa3de\") " pod="openstack/rabbitmq-server-0" Dec 10 23:06:24 crc kubenswrapper[4791]: I1210 23:06:24.915941 4791 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/9ce8dbeb-dce8-47ac-98f1-fe92f74aa3de-rabbitmq-erlang-cookie\") pod \"rabbitmq-server-0\" (UID: \"9ce8dbeb-dce8-47ac-98f1-fe92f74aa3de\") " pod="openstack/rabbitmq-server-0" Dec 10 23:06:24 crc kubenswrapper[4791]: I1210 23:06:24.916458 4791 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/47eb9200-b0bc-41a6-abb9-f6167bd69c66-rabbitmq-erlang-cookie\") pod \"rabbitmq-cell1-server-0\" (UID: \"47eb9200-b0bc-41a6-abb9-f6167bd69c66\") " pod="openstack/rabbitmq-cell1-server-0" Dec 10 23:06:24 crc kubenswrapper[4791]: I1210 23:06:24.918081 4791 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/47eb9200-b0bc-41a6-abb9-f6167bd69c66-config-data\") pod \"rabbitmq-cell1-server-0\" (UID: \"47eb9200-b0bc-41a6-abb9-f6167bd69c66\") " pod="openstack/rabbitmq-cell1-server-0" Dec 10 23:06:24 crc kubenswrapper[4791]: I1210 23:06:24.922130 4791 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/9ce8dbeb-dce8-47ac-98f1-fe92f74aa3de-server-conf\") pod \"rabbitmq-server-0\" (UID: \"9ce8dbeb-dce8-47ac-98f1-fe92f74aa3de\") " pod="openstack/rabbitmq-server-0" Dec 10 23:06:24 crc kubenswrapper[4791]: I1210 23:06:24.926732 4791 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/9ce8dbeb-dce8-47ac-98f1-fe92f74aa3de-pod-info\") pod \"rabbitmq-server-0\" (UID: \"9ce8dbeb-dce8-47ac-98f1-fe92f74aa3de\") " pod="openstack/rabbitmq-server-0" Dec 10 23:06:24 crc kubenswrapper[4791]: I1210 23:06:24.926968 4791 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-tls\" (UniqueName: \"kubernetes.io/projected/9ce8dbeb-dce8-47ac-98f1-fe92f74aa3de-rabbitmq-tls\") pod \"rabbitmq-server-0\" (UID: \"9ce8dbeb-dce8-47ac-98f1-fe92f74aa3de\") " pod="openstack/rabbitmq-server-0" Dec 10 23:06:24 crc kubenswrapper[4791]: I1210 23:06:24.927187 4791 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/9ce8dbeb-dce8-47ac-98f1-fe92f74aa3de-erlang-cookie-secret\") pod \"rabbitmq-server-0\" (UID: \"9ce8dbeb-dce8-47ac-98f1-fe92f74aa3de\") " pod="openstack/rabbitmq-server-0" Dec 10 23:06:24 crc kubenswrapper[4791]: I1210 23:06:24.928234 4791 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/9ce8dbeb-dce8-47ac-98f1-fe92f74aa3de-config-data\") pod \"rabbitmq-server-0\" (UID: \"9ce8dbeb-dce8-47ac-98f1-fe92f74aa3de\") " pod="openstack/rabbitmq-server-0" Dec 10 23:06:24 crc kubenswrapper[4791]: I1210 23:06:24.928591 4791 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/47eb9200-b0bc-41a6-abb9-f6167bd69c66-erlang-cookie-secret\") pod \"rabbitmq-cell1-server-0\" (UID: \"47eb9200-b0bc-41a6-abb9-f6167bd69c66\") " pod="openstack/rabbitmq-cell1-server-0" Dec 10 23:06:24 crc kubenswrapper[4791]: I1210 23:06:24.930540 4791 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-tls\" (UniqueName: \"kubernetes.io/projected/47eb9200-b0bc-41a6-abb9-f6167bd69c66-rabbitmq-tls\") pod \"rabbitmq-cell1-server-0\" (UID: \"47eb9200-b0bc-41a6-abb9-f6167bd69c66\") " pod="openstack/rabbitmq-cell1-server-0" Dec 10 23:06:24 crc kubenswrapper[4791]: I1210 23:06:24.931406 4791 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/47eb9200-b0bc-41a6-abb9-f6167bd69c66-pod-info\") pod \"rabbitmq-cell1-server-0\" (UID: \"47eb9200-b0bc-41a6-abb9-f6167bd69c66\") " pod="openstack/rabbitmq-cell1-server-0" Dec 10 23:06:24 crc kubenswrapper[4791]: I1210 23:06:24.932470 4791 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/9ce8dbeb-dce8-47ac-98f1-fe92f74aa3de-rabbitmq-confd\") pod \"rabbitmq-server-0\" (UID: \"9ce8dbeb-dce8-47ac-98f1-fe92f74aa3de\") " pod="openstack/rabbitmq-server-0" Dec 10 23:06:24 crc kubenswrapper[4791]: I1210 23:06:24.950539 4791 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-85rrv\" (UniqueName: \"kubernetes.io/projected/9ce8dbeb-dce8-47ac-98f1-fe92f74aa3de-kube-api-access-85rrv\") pod \"rabbitmq-server-0\" (UID: \"9ce8dbeb-dce8-47ac-98f1-fe92f74aa3de\") " pod="openstack/rabbitmq-server-0" Dec 10 23:06:24 crc kubenswrapper[4791]: I1210 23:06:24.950721 4791 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-57d769cc4f-tnfbj" event={"ID":"b2c7ad15-ca5a-4042-9d55-31524249a851","Type":"ContainerStarted","Data":"2a8408d791d9146497875b91274e2afd7e291d826983d0300b16368d46cdfe20"} Dec 10 23:06:24 crc kubenswrapper[4791]: I1210 23:06:24.950957 4791 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-kmc8x\" (UniqueName: \"kubernetes.io/projected/47eb9200-b0bc-41a6-abb9-f6167bd69c66-kube-api-access-kmc8x\") pod \"rabbitmq-cell1-server-0\" (UID: \"47eb9200-b0bc-41a6-abb9-f6167bd69c66\") " pod="openstack/rabbitmq-cell1-server-0" Dec 10 23:06:24 crc kubenswrapper[4791]: I1210 23:06:24.951206 4791 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/47eb9200-b0bc-41a6-abb9-f6167bd69c66-rabbitmq-confd\") pod \"rabbitmq-cell1-server-0\" (UID: \"47eb9200-b0bc-41a6-abb9-f6167bd69c66\") " pod="openstack/rabbitmq-cell1-server-0" Dec 10 23:06:24 crc kubenswrapper[4791]: I1210 23:06:24.969059 4791 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage07-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage07-crc\") pod \"rabbitmq-server-0\" (UID: \"9ce8dbeb-dce8-47ac-98f1-fe92f74aa3de\") " pod="openstack/rabbitmq-server-0" Dec 10 23:06:24 crc kubenswrapper[4791]: I1210 23:06:24.969721 4791 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage08-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage08-crc\") pod \"rabbitmq-cell1-server-0\" (UID: \"47eb9200-b0bc-41a6-abb9-f6167bd69c66\") " pod="openstack/rabbitmq-cell1-server-0" Dec 10 23:06:24 crc kubenswrapper[4791]: I1210 23:06:24.990945 4791 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/rabbitmq-cell1-server-0" Dec 10 23:06:25 crc kubenswrapper[4791]: I1210 23:06:25.038003 4791 patch_prober.go:28] interesting pod/machine-config-daemon-5rb5l container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 10 23:06:25 crc kubenswrapper[4791]: I1210 23:06:25.038061 4791 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-5rb5l" podUID="ba35653c-6e06-4cee-a4d6-137764090d18" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 10 23:06:25 crc kubenswrapper[4791]: I1210 23:06:25.038110 4791 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-5rb5l" Dec 10 23:06:25 crc kubenswrapper[4791]: I1210 23:06:25.038799 4791 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"6bffb04499032491e4580b048fb3dc3c71ca5a93b7a98e3a62cb913d0da4b5c3"} pod="openshift-machine-config-operator/machine-config-daemon-5rb5l" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Dec 10 23:06:25 crc kubenswrapper[4791]: I1210 23:06:25.038901 4791 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-5rb5l" podUID="ba35653c-6e06-4cee-a4d6-137764090d18" containerName="machine-config-daemon" containerID="cri-o://6bffb04499032491e4580b048fb3dc3c71ca5a93b7a98e3a62cb913d0da4b5c3" gracePeriod=600 Dec 10 23:06:25 crc kubenswrapper[4791]: I1210 23:06:25.172353 4791 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/openstack-galera-0"] Dec 10 23:06:25 crc kubenswrapper[4791]: I1210 23:06:25.173770 4791 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/openstack-galera-0" Dec 10 23:06:25 crc kubenswrapper[4791]: I1210 23:06:25.177696 4791 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-scripts" Dec 10 23:06:25 crc kubenswrapper[4791]: I1210 23:06:25.178017 4791 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-config-data" Dec 10 23:06:25 crc kubenswrapper[4791]: I1210 23:06:25.178939 4791 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-galera-openstack-svc" Dec 10 23:06:25 crc kubenswrapper[4791]: I1210 23:06:25.179077 4791 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"galera-openstack-dockercfg-nccf4" Dec 10 23:06:25 crc kubenswrapper[4791]: I1210 23:06:25.184530 4791 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"combined-ca-bundle" Dec 10 23:06:25 crc kubenswrapper[4791]: I1210 23:06:25.198763 4791 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/openstack-galera-0"] Dec 10 23:06:25 crc kubenswrapper[4791]: I1210 23:06:25.275849 4791 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/rabbitmq-server-0" Dec 10 23:06:25 crc kubenswrapper[4791]: I1210 23:06:25.318778 4791 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kolla-config\" (UniqueName: \"kubernetes.io/configmap/aed2b986-c040-4191-864b-47f29c5c8537-kolla-config\") pod \"openstack-galera-0\" (UID: \"aed2b986-c040-4191-864b-47f29c5c8537\") " pod="openstack/openstack-galera-0" Dec 10 23:06:25 crc kubenswrapper[4791]: I1210 23:06:25.318892 4791 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-generated\" (UniqueName: \"kubernetes.io/empty-dir/aed2b986-c040-4191-864b-47f29c5c8537-config-data-generated\") pod \"openstack-galera-0\" (UID: \"aed2b986-c040-4191-864b-47f29c5c8537\") " pod="openstack/openstack-galera-0" Dec 10 23:06:25 crc kubenswrapper[4791]: I1210 23:06:25.318925 4791 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/aed2b986-c040-4191-864b-47f29c5c8537-operator-scripts\") pod \"openstack-galera-0\" (UID: \"aed2b986-c040-4191-864b-47f29c5c8537\") " pod="openstack/openstack-galera-0" Dec 10 23:06:25 crc kubenswrapper[4791]: I1210 23:06:25.318977 4791 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage02-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage02-crc\") pod \"openstack-galera-0\" (UID: \"aed2b986-c040-4191-864b-47f29c5c8537\") " pod="openstack/openstack-galera-0" Dec 10 23:06:25 crc kubenswrapper[4791]: I1210 23:06:25.319006 4791 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-phmhb\" (UniqueName: \"kubernetes.io/projected/aed2b986-c040-4191-864b-47f29c5c8537-kube-api-access-phmhb\") pod \"openstack-galera-0\" (UID: \"aed2b986-c040-4191-864b-47f29c5c8537\") " pod="openstack/openstack-galera-0" Dec 10 23:06:25 crc kubenswrapper[4791]: I1210 23:06:25.319031 4791 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"galera-tls-certs\" (UniqueName: \"kubernetes.io/secret/aed2b986-c040-4191-864b-47f29c5c8537-galera-tls-certs\") pod \"openstack-galera-0\" (UID: \"aed2b986-c040-4191-864b-47f29c5c8537\") " pod="openstack/openstack-galera-0" Dec 10 23:06:25 crc kubenswrapper[4791]: I1210 23:06:25.319083 4791 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-default\" (UniqueName: \"kubernetes.io/configmap/aed2b986-c040-4191-864b-47f29c5c8537-config-data-default\") pod \"openstack-galera-0\" (UID: \"aed2b986-c040-4191-864b-47f29c5c8537\") " pod="openstack/openstack-galera-0" Dec 10 23:06:25 crc kubenswrapper[4791]: I1210 23:06:25.319128 4791 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/aed2b986-c040-4191-864b-47f29c5c8537-combined-ca-bundle\") pod \"openstack-galera-0\" (UID: \"aed2b986-c040-4191-864b-47f29c5c8537\") " pod="openstack/openstack-galera-0" Dec 10 23:06:25 crc kubenswrapper[4791]: I1210 23:06:25.339612 4791 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/rabbitmq-cell1-server-0"] Dec 10 23:06:25 crc kubenswrapper[4791]: W1210 23:06:25.389521 4791 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod47eb9200_b0bc_41a6_abb9_f6167bd69c66.slice/crio-01e09f7fc253f735d96b6208f87608bbd0ff2e0759cea509621b849b58aad1f4 WatchSource:0}: Error finding container 01e09f7fc253f735d96b6208f87608bbd0ff2e0759cea509621b849b58aad1f4: Status 404 returned error can't find the container with id 01e09f7fc253f735d96b6208f87608bbd0ff2e0759cea509621b849b58aad1f4 Dec 10 23:06:25 crc kubenswrapper[4791]: I1210 23:06:25.421107 4791 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kolla-config\" (UniqueName: \"kubernetes.io/configmap/aed2b986-c040-4191-864b-47f29c5c8537-kolla-config\") pod \"openstack-galera-0\" (UID: \"aed2b986-c040-4191-864b-47f29c5c8537\") " pod="openstack/openstack-galera-0" Dec 10 23:06:25 crc kubenswrapper[4791]: I1210 23:06:25.421150 4791 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-generated\" (UniqueName: \"kubernetes.io/empty-dir/aed2b986-c040-4191-864b-47f29c5c8537-config-data-generated\") pod \"openstack-galera-0\" (UID: \"aed2b986-c040-4191-864b-47f29c5c8537\") " pod="openstack/openstack-galera-0" Dec 10 23:06:25 crc kubenswrapper[4791]: I1210 23:06:25.421171 4791 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/aed2b986-c040-4191-864b-47f29c5c8537-operator-scripts\") pod \"openstack-galera-0\" (UID: \"aed2b986-c040-4191-864b-47f29c5c8537\") " pod="openstack/openstack-galera-0" Dec 10 23:06:25 crc kubenswrapper[4791]: I1210 23:06:25.421210 4791 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage02-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage02-crc\") pod \"openstack-galera-0\" (UID: \"aed2b986-c040-4191-864b-47f29c5c8537\") " pod="openstack/openstack-galera-0" Dec 10 23:06:25 crc kubenswrapper[4791]: I1210 23:06:25.421232 4791 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-phmhb\" (UniqueName: \"kubernetes.io/projected/aed2b986-c040-4191-864b-47f29c5c8537-kube-api-access-phmhb\") pod \"openstack-galera-0\" (UID: \"aed2b986-c040-4191-864b-47f29c5c8537\") " pod="openstack/openstack-galera-0" Dec 10 23:06:25 crc kubenswrapper[4791]: I1210 23:06:25.421251 4791 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"galera-tls-certs\" (UniqueName: \"kubernetes.io/secret/aed2b986-c040-4191-864b-47f29c5c8537-galera-tls-certs\") pod \"openstack-galera-0\" (UID: \"aed2b986-c040-4191-864b-47f29c5c8537\") " pod="openstack/openstack-galera-0" Dec 10 23:06:25 crc kubenswrapper[4791]: I1210 23:06:25.421292 4791 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-default\" (UniqueName: \"kubernetes.io/configmap/aed2b986-c040-4191-864b-47f29c5c8537-config-data-default\") pod \"openstack-galera-0\" (UID: \"aed2b986-c040-4191-864b-47f29c5c8537\") " pod="openstack/openstack-galera-0" Dec 10 23:06:25 crc kubenswrapper[4791]: I1210 23:06:25.421348 4791 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/aed2b986-c040-4191-864b-47f29c5c8537-combined-ca-bundle\") pod \"openstack-galera-0\" (UID: \"aed2b986-c040-4191-864b-47f29c5c8537\") " pod="openstack/openstack-galera-0" Dec 10 23:06:25 crc kubenswrapper[4791]: I1210 23:06:25.421917 4791 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage02-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage02-crc\") pod \"openstack-galera-0\" (UID: \"aed2b986-c040-4191-864b-47f29c5c8537\") device mount path \"/mnt/openstack/pv02\"" pod="openstack/openstack-galera-0" Dec 10 23:06:25 crc kubenswrapper[4791]: I1210 23:06:25.426629 4791 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"galera-tls-certs\" (UniqueName: \"kubernetes.io/secret/aed2b986-c040-4191-864b-47f29c5c8537-galera-tls-certs\") pod \"openstack-galera-0\" (UID: \"aed2b986-c040-4191-864b-47f29c5c8537\") " pod="openstack/openstack-galera-0" Dec 10 23:06:25 crc kubenswrapper[4791]: I1210 23:06:25.426904 4791 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/aed2b986-c040-4191-864b-47f29c5c8537-combined-ca-bundle\") pod \"openstack-galera-0\" (UID: \"aed2b986-c040-4191-864b-47f29c5c8537\") " pod="openstack/openstack-galera-0" Dec 10 23:06:25 crc kubenswrapper[4791]: I1210 23:06:25.427167 4791 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-generated\" (UniqueName: \"kubernetes.io/empty-dir/aed2b986-c040-4191-864b-47f29c5c8537-config-data-generated\") pod \"openstack-galera-0\" (UID: \"aed2b986-c040-4191-864b-47f29c5c8537\") " pod="openstack/openstack-galera-0" Dec 10 23:06:25 crc kubenswrapper[4791]: I1210 23:06:25.427941 4791 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kolla-config\" (UniqueName: \"kubernetes.io/configmap/aed2b986-c040-4191-864b-47f29c5c8537-kolla-config\") pod \"openstack-galera-0\" (UID: \"aed2b986-c040-4191-864b-47f29c5c8537\") " pod="openstack/openstack-galera-0" Dec 10 23:06:25 crc kubenswrapper[4791]: I1210 23:06:25.428312 4791 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-default\" (UniqueName: \"kubernetes.io/configmap/aed2b986-c040-4191-864b-47f29c5c8537-config-data-default\") pod \"openstack-galera-0\" (UID: \"aed2b986-c040-4191-864b-47f29c5c8537\") " pod="openstack/openstack-galera-0" Dec 10 23:06:25 crc kubenswrapper[4791]: I1210 23:06:25.428977 4791 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/aed2b986-c040-4191-864b-47f29c5c8537-operator-scripts\") pod \"openstack-galera-0\" (UID: \"aed2b986-c040-4191-864b-47f29c5c8537\") " pod="openstack/openstack-galera-0" Dec 10 23:06:25 crc kubenswrapper[4791]: I1210 23:06:25.446079 4791 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage02-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage02-crc\") pod \"openstack-galera-0\" (UID: \"aed2b986-c040-4191-864b-47f29c5c8537\") " pod="openstack/openstack-galera-0" Dec 10 23:06:25 crc kubenswrapper[4791]: I1210 23:06:25.449502 4791 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-phmhb\" (UniqueName: \"kubernetes.io/projected/aed2b986-c040-4191-864b-47f29c5c8537-kube-api-access-phmhb\") pod \"openstack-galera-0\" (UID: \"aed2b986-c040-4191-864b-47f29c5c8537\") " pod="openstack/openstack-galera-0" Dec 10 23:06:25 crc kubenswrapper[4791]: I1210 23:06:25.508833 4791 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/openstack-galera-0" Dec 10 23:06:25 crc kubenswrapper[4791]: I1210 23:06:25.952873 4791 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/rabbitmq-server-0"] Dec 10 23:06:26 crc kubenswrapper[4791]: I1210 23:06:26.052828 4791 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-cell1-server-0" event={"ID":"47eb9200-b0bc-41a6-abb9-f6167bd69c66","Type":"ContainerStarted","Data":"01e09f7fc253f735d96b6208f87608bbd0ff2e0759cea509621b849b58aad1f4"} Dec 10 23:06:26 crc kubenswrapper[4791]: I1210 23:06:26.096430 4791 generic.go:334] "Generic (PLEG): container finished" podID="ba35653c-6e06-4cee-a4d6-137764090d18" containerID="6bffb04499032491e4580b048fb3dc3c71ca5a93b7a98e3a62cb913d0da4b5c3" exitCode=0 Dec 10 23:06:26 crc kubenswrapper[4791]: I1210 23:06:26.096496 4791 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-5rb5l" event={"ID":"ba35653c-6e06-4cee-a4d6-137764090d18","Type":"ContainerDied","Data":"6bffb04499032491e4580b048fb3dc3c71ca5a93b7a98e3a62cb913d0da4b5c3"} Dec 10 23:06:26 crc kubenswrapper[4791]: I1210 23:06:26.096525 4791 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-5rb5l" event={"ID":"ba35653c-6e06-4cee-a4d6-137764090d18","Type":"ContainerStarted","Data":"2b6bbc8a626f5244441ef712b0db42022f2a15236c1846adb0b1f8c097905a73"} Dec 10 23:06:26 crc kubenswrapper[4791]: I1210 23:06:26.096541 4791 scope.go:117] "RemoveContainer" containerID="cb32d752921f31f5faa2ab64f1cbd0e43f7e3f3e95a29f0f2fbc2ed42452d743" Dec 10 23:06:26 crc kubenswrapper[4791]: I1210 23:06:26.433055 4791 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/openstack-galera-0"] Dec 10 23:06:26 crc kubenswrapper[4791]: W1210 23:06:26.439221 4791 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podaed2b986_c040_4191_864b_47f29c5c8537.slice/crio-0e09b079c12ae520b3498fa633b2983990b5bdc9a47746f43d65d9281b20acae WatchSource:0}: Error finding container 0e09b079c12ae520b3498fa633b2983990b5bdc9a47746f43d65d9281b20acae: Status 404 returned error can't find the container with id 0e09b079c12ae520b3498fa633b2983990b5bdc9a47746f43d65d9281b20acae Dec 10 23:06:26 crc kubenswrapper[4791]: I1210 23:06:26.632634 4791 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/openstack-cell1-galera-0"] Dec 10 23:06:26 crc kubenswrapper[4791]: I1210 23:06:26.633736 4791 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/openstack-cell1-galera-0" Dec 10 23:06:26 crc kubenswrapper[4791]: I1210 23:06:26.638003 4791 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"galera-openstack-cell1-dockercfg-k5fxh" Dec 10 23:06:26 crc kubenswrapper[4791]: I1210 23:06:26.638274 4791 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-cell1-scripts" Dec 10 23:06:26 crc kubenswrapper[4791]: I1210 23:06:26.638457 4791 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-galera-openstack-cell1-svc" Dec 10 23:06:26 crc kubenswrapper[4791]: I1210 23:06:26.638970 4791 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-cell1-config-data" Dec 10 23:06:26 crc kubenswrapper[4791]: I1210 23:06:26.663934 4791 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/openstack-cell1-galera-0"] Dec 10 23:06:26 crc kubenswrapper[4791]: I1210 23:06:26.719265 4791 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/memcached-0"] Dec 10 23:06:26 crc kubenswrapper[4791]: I1210 23:06:26.720276 4791 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/memcached-0" Dec 10 23:06:26 crc kubenswrapper[4791]: I1210 23:06:26.722982 4791 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"memcached-memcached-dockercfg-qz928" Dec 10 23:06:26 crc kubenswrapper[4791]: I1210 23:06:26.723302 4791 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"memcached-config-data" Dec 10 23:06:26 crc kubenswrapper[4791]: I1210 23:06:26.725286 4791 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-memcached-svc" Dec 10 23:06:26 crc kubenswrapper[4791]: I1210 23:06:26.761451 4791 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/memcached-0"] Dec 10 23:06:26 crc kubenswrapper[4791]: I1210 23:06:26.762990 4791 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f04fa829-3ebe-445d-a4e5-f7592ac682f3-combined-ca-bundle\") pod \"openstack-cell1-galera-0\" (UID: \"f04fa829-3ebe-445d-a4e5-f7592ac682f3\") " pod="openstack/openstack-cell1-galera-0" Dec 10 23:06:26 crc kubenswrapper[4791]: I1210 23:06:26.763115 4791 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-default\" (UniqueName: \"kubernetes.io/configmap/f04fa829-3ebe-445d-a4e5-f7592ac682f3-config-data-default\") pod \"openstack-cell1-galera-0\" (UID: \"f04fa829-3ebe-445d-a4e5-f7592ac682f3\") " pod="openstack/openstack-cell1-galera-0" Dec 10 23:06:26 crc kubenswrapper[4791]: I1210 23:06:26.763193 4791 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage12-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage12-crc\") pod \"openstack-cell1-galera-0\" (UID: \"f04fa829-3ebe-445d-a4e5-f7592ac682f3\") " pod="openstack/openstack-cell1-galera-0" Dec 10 23:06:26 crc kubenswrapper[4791]: I1210 23:06:26.763277 4791 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kolla-config\" (UniqueName: \"kubernetes.io/configmap/f04fa829-3ebe-445d-a4e5-f7592ac682f3-kolla-config\") pod \"openstack-cell1-galera-0\" (UID: \"f04fa829-3ebe-445d-a4e5-f7592ac682f3\") " pod="openstack/openstack-cell1-galera-0" Dec 10 23:06:26 crc kubenswrapper[4791]: I1210 23:06:26.763349 4791 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-generated\" (UniqueName: \"kubernetes.io/empty-dir/f04fa829-3ebe-445d-a4e5-f7592ac682f3-config-data-generated\") pod \"openstack-cell1-galera-0\" (UID: \"f04fa829-3ebe-445d-a4e5-f7592ac682f3\") " pod="openstack/openstack-cell1-galera-0" Dec 10 23:06:26 crc kubenswrapper[4791]: I1210 23:06:26.763417 4791 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-7sphf\" (UniqueName: \"kubernetes.io/projected/f04fa829-3ebe-445d-a4e5-f7592ac682f3-kube-api-access-7sphf\") pod \"openstack-cell1-galera-0\" (UID: \"f04fa829-3ebe-445d-a4e5-f7592ac682f3\") " pod="openstack/openstack-cell1-galera-0" Dec 10 23:06:26 crc kubenswrapper[4791]: I1210 23:06:26.763458 4791 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/f04fa829-3ebe-445d-a4e5-f7592ac682f3-operator-scripts\") pod \"openstack-cell1-galera-0\" (UID: \"f04fa829-3ebe-445d-a4e5-f7592ac682f3\") " pod="openstack/openstack-cell1-galera-0" Dec 10 23:06:26 crc kubenswrapper[4791]: I1210 23:06:26.763519 4791 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"galera-tls-certs\" (UniqueName: \"kubernetes.io/secret/f04fa829-3ebe-445d-a4e5-f7592ac682f3-galera-tls-certs\") pod \"openstack-cell1-galera-0\" (UID: \"f04fa829-3ebe-445d-a4e5-f7592ac682f3\") " pod="openstack/openstack-cell1-galera-0" Dec 10 23:06:26 crc kubenswrapper[4791]: I1210 23:06:26.865233 4791 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"galera-tls-certs\" (UniqueName: \"kubernetes.io/secret/f04fa829-3ebe-445d-a4e5-f7592ac682f3-galera-tls-certs\") pod \"openstack-cell1-galera-0\" (UID: \"f04fa829-3ebe-445d-a4e5-f7592ac682f3\") " pod="openstack/openstack-cell1-galera-0" Dec 10 23:06:26 crc kubenswrapper[4791]: I1210 23:06:26.865297 4791 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/49839437-3696-4ac8-98cf-2cc16048f21a-combined-ca-bundle\") pod \"memcached-0\" (UID: \"49839437-3696-4ac8-98cf-2cc16048f21a\") " pod="openstack/memcached-0" Dec 10 23:06:26 crc kubenswrapper[4791]: I1210 23:06:26.865360 4791 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f04fa829-3ebe-445d-a4e5-f7592ac682f3-combined-ca-bundle\") pod \"openstack-cell1-galera-0\" (UID: \"f04fa829-3ebe-445d-a4e5-f7592ac682f3\") " pod="openstack/openstack-cell1-galera-0" Dec 10 23:06:26 crc kubenswrapper[4791]: I1210 23:06:26.865391 4791 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/49839437-3696-4ac8-98cf-2cc16048f21a-config-data\") pod \"memcached-0\" (UID: \"49839437-3696-4ac8-98cf-2cc16048f21a\") " pod="openstack/memcached-0" Dec 10 23:06:26 crc kubenswrapper[4791]: I1210 23:06:26.865415 4791 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-rwlqb\" (UniqueName: \"kubernetes.io/projected/49839437-3696-4ac8-98cf-2cc16048f21a-kube-api-access-rwlqb\") pod \"memcached-0\" (UID: \"49839437-3696-4ac8-98cf-2cc16048f21a\") " pod="openstack/memcached-0" Dec 10 23:06:26 crc kubenswrapper[4791]: I1210 23:06:26.865438 4791 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"memcached-tls-certs\" (UniqueName: \"kubernetes.io/secret/49839437-3696-4ac8-98cf-2cc16048f21a-memcached-tls-certs\") pod \"memcached-0\" (UID: \"49839437-3696-4ac8-98cf-2cc16048f21a\") " pod="openstack/memcached-0" Dec 10 23:06:26 crc kubenswrapper[4791]: I1210 23:06:26.865465 4791 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-default\" (UniqueName: \"kubernetes.io/configmap/f04fa829-3ebe-445d-a4e5-f7592ac682f3-config-data-default\") pod \"openstack-cell1-galera-0\" (UID: \"f04fa829-3ebe-445d-a4e5-f7592ac682f3\") " pod="openstack/openstack-cell1-galera-0" Dec 10 23:06:26 crc kubenswrapper[4791]: I1210 23:06:26.865492 4791 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage12-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage12-crc\") pod \"openstack-cell1-galera-0\" (UID: \"f04fa829-3ebe-445d-a4e5-f7592ac682f3\") " pod="openstack/openstack-cell1-galera-0" Dec 10 23:06:26 crc kubenswrapper[4791]: I1210 23:06:26.865525 4791 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kolla-config\" (UniqueName: \"kubernetes.io/configmap/f04fa829-3ebe-445d-a4e5-f7592ac682f3-kolla-config\") pod \"openstack-cell1-galera-0\" (UID: \"f04fa829-3ebe-445d-a4e5-f7592ac682f3\") " pod="openstack/openstack-cell1-galera-0" Dec 10 23:06:26 crc kubenswrapper[4791]: I1210 23:06:26.865543 4791 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-generated\" (UniqueName: \"kubernetes.io/empty-dir/f04fa829-3ebe-445d-a4e5-f7592ac682f3-config-data-generated\") pod \"openstack-cell1-galera-0\" (UID: \"f04fa829-3ebe-445d-a4e5-f7592ac682f3\") " pod="openstack/openstack-cell1-galera-0" Dec 10 23:06:26 crc kubenswrapper[4791]: I1210 23:06:26.865565 4791 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kolla-config\" (UniqueName: \"kubernetes.io/configmap/49839437-3696-4ac8-98cf-2cc16048f21a-kolla-config\") pod \"memcached-0\" (UID: \"49839437-3696-4ac8-98cf-2cc16048f21a\") " pod="openstack/memcached-0" Dec 10 23:06:26 crc kubenswrapper[4791]: I1210 23:06:26.865583 4791 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-7sphf\" (UniqueName: \"kubernetes.io/projected/f04fa829-3ebe-445d-a4e5-f7592ac682f3-kube-api-access-7sphf\") pod \"openstack-cell1-galera-0\" (UID: \"f04fa829-3ebe-445d-a4e5-f7592ac682f3\") " pod="openstack/openstack-cell1-galera-0" Dec 10 23:06:26 crc kubenswrapper[4791]: I1210 23:06:26.865603 4791 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/f04fa829-3ebe-445d-a4e5-f7592ac682f3-operator-scripts\") pod \"openstack-cell1-galera-0\" (UID: \"f04fa829-3ebe-445d-a4e5-f7592ac682f3\") " pod="openstack/openstack-cell1-galera-0" Dec 10 23:06:26 crc kubenswrapper[4791]: I1210 23:06:26.866293 4791 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage12-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage12-crc\") pod \"openstack-cell1-galera-0\" (UID: \"f04fa829-3ebe-445d-a4e5-f7592ac682f3\") device mount path \"/mnt/openstack/pv12\"" pod="openstack/openstack-cell1-galera-0" Dec 10 23:06:26 crc kubenswrapper[4791]: I1210 23:06:26.866903 4791 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-generated\" (UniqueName: \"kubernetes.io/empty-dir/f04fa829-3ebe-445d-a4e5-f7592ac682f3-config-data-generated\") pod \"openstack-cell1-galera-0\" (UID: \"f04fa829-3ebe-445d-a4e5-f7592ac682f3\") " pod="openstack/openstack-cell1-galera-0" Dec 10 23:06:26 crc kubenswrapper[4791]: I1210 23:06:26.867248 4791 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-default\" (UniqueName: \"kubernetes.io/configmap/f04fa829-3ebe-445d-a4e5-f7592ac682f3-config-data-default\") pod \"openstack-cell1-galera-0\" (UID: \"f04fa829-3ebe-445d-a4e5-f7592ac682f3\") " pod="openstack/openstack-cell1-galera-0" Dec 10 23:06:26 crc kubenswrapper[4791]: I1210 23:06:26.867404 4791 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/f04fa829-3ebe-445d-a4e5-f7592ac682f3-operator-scripts\") pod \"openstack-cell1-galera-0\" (UID: \"f04fa829-3ebe-445d-a4e5-f7592ac682f3\") " pod="openstack/openstack-cell1-galera-0" Dec 10 23:06:26 crc kubenswrapper[4791]: I1210 23:06:26.874288 4791 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"galera-tls-certs\" (UniqueName: \"kubernetes.io/secret/f04fa829-3ebe-445d-a4e5-f7592ac682f3-galera-tls-certs\") pod \"openstack-cell1-galera-0\" (UID: \"f04fa829-3ebe-445d-a4e5-f7592ac682f3\") " pod="openstack/openstack-cell1-galera-0" Dec 10 23:06:26 crc kubenswrapper[4791]: I1210 23:06:26.878380 4791 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kolla-config\" (UniqueName: \"kubernetes.io/configmap/f04fa829-3ebe-445d-a4e5-f7592ac682f3-kolla-config\") pod \"openstack-cell1-galera-0\" (UID: \"f04fa829-3ebe-445d-a4e5-f7592ac682f3\") " pod="openstack/openstack-cell1-galera-0" Dec 10 23:06:26 crc kubenswrapper[4791]: I1210 23:06:26.888644 4791 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f04fa829-3ebe-445d-a4e5-f7592ac682f3-combined-ca-bundle\") pod \"openstack-cell1-galera-0\" (UID: \"f04fa829-3ebe-445d-a4e5-f7592ac682f3\") " pod="openstack/openstack-cell1-galera-0" Dec 10 23:06:26 crc kubenswrapper[4791]: I1210 23:06:26.898539 4791 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage12-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage12-crc\") pod \"openstack-cell1-galera-0\" (UID: \"f04fa829-3ebe-445d-a4e5-f7592ac682f3\") " pod="openstack/openstack-cell1-galera-0" Dec 10 23:06:26 crc kubenswrapper[4791]: I1210 23:06:26.908561 4791 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-7sphf\" (UniqueName: \"kubernetes.io/projected/f04fa829-3ebe-445d-a4e5-f7592ac682f3-kube-api-access-7sphf\") pod \"openstack-cell1-galera-0\" (UID: \"f04fa829-3ebe-445d-a4e5-f7592ac682f3\") " pod="openstack/openstack-cell1-galera-0" Dec 10 23:06:26 crc kubenswrapper[4791]: I1210 23:06:26.960463 4791 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/openstack-cell1-galera-0" Dec 10 23:06:26 crc kubenswrapper[4791]: I1210 23:06:26.969026 4791 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-rwlqb\" (UniqueName: \"kubernetes.io/projected/49839437-3696-4ac8-98cf-2cc16048f21a-kube-api-access-rwlqb\") pod \"memcached-0\" (UID: \"49839437-3696-4ac8-98cf-2cc16048f21a\") " pod="openstack/memcached-0" Dec 10 23:06:26 crc kubenswrapper[4791]: I1210 23:06:26.969084 4791 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"memcached-tls-certs\" (UniqueName: \"kubernetes.io/secret/49839437-3696-4ac8-98cf-2cc16048f21a-memcached-tls-certs\") pod \"memcached-0\" (UID: \"49839437-3696-4ac8-98cf-2cc16048f21a\") " pod="openstack/memcached-0" Dec 10 23:06:26 crc kubenswrapper[4791]: I1210 23:06:26.969146 4791 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kolla-config\" (UniqueName: \"kubernetes.io/configmap/49839437-3696-4ac8-98cf-2cc16048f21a-kolla-config\") pod \"memcached-0\" (UID: \"49839437-3696-4ac8-98cf-2cc16048f21a\") " pod="openstack/memcached-0" Dec 10 23:06:26 crc kubenswrapper[4791]: I1210 23:06:26.969174 4791 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/49839437-3696-4ac8-98cf-2cc16048f21a-combined-ca-bundle\") pod \"memcached-0\" (UID: \"49839437-3696-4ac8-98cf-2cc16048f21a\") " pod="openstack/memcached-0" Dec 10 23:06:26 crc kubenswrapper[4791]: I1210 23:06:26.969216 4791 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/49839437-3696-4ac8-98cf-2cc16048f21a-config-data\") pod \"memcached-0\" (UID: \"49839437-3696-4ac8-98cf-2cc16048f21a\") " pod="openstack/memcached-0" Dec 10 23:06:26 crc kubenswrapper[4791]: I1210 23:06:26.969952 4791 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/49839437-3696-4ac8-98cf-2cc16048f21a-config-data\") pod \"memcached-0\" (UID: \"49839437-3696-4ac8-98cf-2cc16048f21a\") " pod="openstack/memcached-0" Dec 10 23:06:26 crc kubenswrapper[4791]: I1210 23:06:26.970903 4791 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kolla-config\" (UniqueName: \"kubernetes.io/configmap/49839437-3696-4ac8-98cf-2cc16048f21a-kolla-config\") pod \"memcached-0\" (UID: \"49839437-3696-4ac8-98cf-2cc16048f21a\") " pod="openstack/memcached-0" Dec 10 23:06:26 crc kubenswrapper[4791]: I1210 23:06:26.983428 4791 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"memcached-tls-certs\" (UniqueName: \"kubernetes.io/secret/49839437-3696-4ac8-98cf-2cc16048f21a-memcached-tls-certs\") pod \"memcached-0\" (UID: \"49839437-3696-4ac8-98cf-2cc16048f21a\") " pod="openstack/memcached-0" Dec 10 23:06:27 crc kubenswrapper[4791]: I1210 23:06:27.024493 4791 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-rwlqb\" (UniqueName: \"kubernetes.io/projected/49839437-3696-4ac8-98cf-2cc16048f21a-kube-api-access-rwlqb\") pod \"memcached-0\" (UID: \"49839437-3696-4ac8-98cf-2cc16048f21a\") " pod="openstack/memcached-0" Dec 10 23:06:27 crc kubenswrapper[4791]: I1210 23:06:27.024864 4791 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/49839437-3696-4ac8-98cf-2cc16048f21a-combined-ca-bundle\") pod \"memcached-0\" (UID: \"49839437-3696-4ac8-98cf-2cc16048f21a\") " pod="openstack/memcached-0" Dec 10 23:06:27 crc kubenswrapper[4791]: I1210 23:06:27.056927 4791 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/memcached-0" Dec 10 23:06:27 crc kubenswrapper[4791]: I1210 23:06:27.115553 4791 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-server-0" event={"ID":"9ce8dbeb-dce8-47ac-98f1-fe92f74aa3de","Type":"ContainerStarted","Data":"8b25683d606ab47b6eb65d1f8cd0a149b5a68f9bfbcb6f1b892270e6316094af"} Dec 10 23:06:27 crc kubenswrapper[4791]: I1210 23:06:27.117919 4791 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/openstack-galera-0" event={"ID":"aed2b986-c040-4191-864b-47f29c5c8537","Type":"ContainerStarted","Data":"0e09b079c12ae520b3498fa633b2983990b5bdc9a47746f43d65d9281b20acae"} Dec 10 23:06:27 crc kubenswrapper[4791]: I1210 23:06:27.738809 4791 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/openstack-cell1-galera-0"] Dec 10 23:06:27 crc kubenswrapper[4791]: W1210 23:06:27.761890 4791 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podf04fa829_3ebe_445d_a4e5_f7592ac682f3.slice/crio-e0becfab4499f45f87501c9a7fe5182f3c3880bd2d7d39cb12f3c4c9c8c29f98 WatchSource:0}: Error finding container e0becfab4499f45f87501c9a7fe5182f3c3880bd2d7d39cb12f3c4c9c8c29f98: Status 404 returned error can't find the container with id e0becfab4499f45f87501c9a7fe5182f3c3880bd2d7d39cb12f3c4c9c8c29f98 Dec 10 23:06:27 crc kubenswrapper[4791]: I1210 23:06:27.837915 4791 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/memcached-0"] Dec 10 23:06:27 crc kubenswrapper[4791]: W1210 23:06:27.870747 4791 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod49839437_3696_4ac8_98cf_2cc16048f21a.slice/crio-cef6667c8cb777d7e59c6d982121a312b34583995d1addbdc5d62eeda19cf2bc WatchSource:0}: Error finding container cef6667c8cb777d7e59c6d982121a312b34583995d1addbdc5d62eeda19cf2bc: Status 404 returned error can't find the container with id cef6667c8cb777d7e59c6d982121a312b34583995d1addbdc5d62eeda19cf2bc Dec 10 23:06:28 crc kubenswrapper[4791]: I1210 23:06:28.130898 4791 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/openstack-cell1-galera-0" event={"ID":"f04fa829-3ebe-445d-a4e5-f7592ac682f3","Type":"ContainerStarted","Data":"e0becfab4499f45f87501c9a7fe5182f3c3880bd2d7d39cb12f3c4c9c8c29f98"} Dec 10 23:06:28 crc kubenswrapper[4791]: I1210 23:06:28.132744 4791 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/memcached-0" event={"ID":"49839437-3696-4ac8-98cf-2cc16048f21a","Type":"ContainerStarted","Data":"cef6667c8cb777d7e59c6d982121a312b34583995d1addbdc5d62eeda19cf2bc"} Dec 10 23:06:28 crc kubenswrapper[4791]: I1210 23:06:28.472023 4791 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/kube-state-metrics-0"] Dec 10 23:06:28 crc kubenswrapper[4791]: I1210 23:06:28.473266 4791 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/kube-state-metrics-0" Dec 10 23:06:28 crc kubenswrapper[4791]: I1210 23:06:28.475525 4791 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"telemetry-ceilometer-dockercfg-bhjlm" Dec 10 23:06:28 crc kubenswrapper[4791]: I1210 23:06:28.498534 4791 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/kube-state-metrics-0"] Dec 10 23:06:28 crc kubenswrapper[4791]: I1210 23:06:28.613465 4791 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-tlj7g\" (UniqueName: \"kubernetes.io/projected/247efe24-6d59-4c93-ab04-6e249e1005a8-kube-api-access-tlj7g\") pod \"kube-state-metrics-0\" (UID: \"247efe24-6d59-4c93-ab04-6e249e1005a8\") " pod="openstack/kube-state-metrics-0" Dec 10 23:06:28 crc kubenswrapper[4791]: I1210 23:06:28.717116 4791 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-tlj7g\" (UniqueName: \"kubernetes.io/projected/247efe24-6d59-4c93-ab04-6e249e1005a8-kube-api-access-tlj7g\") pod \"kube-state-metrics-0\" (UID: \"247efe24-6d59-4c93-ab04-6e249e1005a8\") " pod="openstack/kube-state-metrics-0" Dec 10 23:06:28 crc kubenswrapper[4791]: I1210 23:06:28.773168 4791 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-tlj7g\" (UniqueName: \"kubernetes.io/projected/247efe24-6d59-4c93-ab04-6e249e1005a8-kube-api-access-tlj7g\") pod \"kube-state-metrics-0\" (UID: \"247efe24-6d59-4c93-ab04-6e249e1005a8\") " pod="openstack/kube-state-metrics-0" Dec 10 23:06:28 crc kubenswrapper[4791]: I1210 23:06:28.800714 4791 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/kube-state-metrics-0" Dec 10 23:06:29 crc kubenswrapper[4791]: I1210 23:06:29.631162 4791 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/kube-state-metrics-0"] Dec 10 23:06:30 crc kubenswrapper[4791]: I1210 23:06:30.152602 4791 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/kube-state-metrics-0" event={"ID":"247efe24-6d59-4c93-ab04-6e249e1005a8","Type":"ContainerStarted","Data":"675f475dd72ea2082913c5c6f79644c0372948ff1f0345a38a944d833d3ece2c"} Dec 10 23:06:32 crc kubenswrapper[4791]: I1210 23:06:32.692954 4791 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ovn-controller-xb29h"] Dec 10 23:06:32 crc kubenswrapper[4791]: I1210 23:06:32.695019 4791 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-controller-xb29h" Dec 10 23:06:32 crc kubenswrapper[4791]: I1210 23:06:32.699232 4791 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ovncontroller-ovncontroller-dockercfg-785n6" Dec 10 23:06:32 crc kubenswrapper[4791]: I1210 23:06:32.699241 4791 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-ovncontroller-ovndbs" Dec 10 23:06:32 crc kubenswrapper[4791]: I1210 23:06:32.700379 4791 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"ovncontroller-scripts" Dec 10 23:06:32 crc kubenswrapper[4791]: I1210 23:06:32.715654 4791 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovn-controller-xb29h"] Dec 10 23:06:32 crc kubenswrapper[4791]: I1210 23:06:32.715699 4791 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ovn-controller-ovs-phsfg"] Dec 10 23:06:32 crc kubenswrapper[4791]: I1210 23:06:32.717093 4791 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-controller-ovs-phsfg" Dec 10 23:06:32 crc kubenswrapper[4791]: I1210 23:06:32.746039 4791 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovn-controller-ovs-phsfg"] Dec 10 23:06:32 crc kubenswrapper[4791]: I1210 23:06:32.786168 4791 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/780fd8a7-cfb0-4958-afb3-1d060398d3ea-combined-ca-bundle\") pod \"ovn-controller-xb29h\" (UID: \"780fd8a7-cfb0-4958-afb3-1d060398d3ea\") " pod="openstack/ovn-controller-xb29h" Dec 10 23:06:32 crc kubenswrapper[4791]: I1210 23:06:32.786287 4791 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-run\" (UniqueName: \"kubernetes.io/host-path/ac28ec72-8d7b-4576-b2ba-c93731d50267-var-run\") pod \"ovn-controller-ovs-phsfg\" (UID: \"ac28ec72-8d7b-4576-b2ba-c93731d50267\") " pod="openstack/ovn-controller-ovs-phsfg" Dec 10 23:06:32 crc kubenswrapper[4791]: I1210 23:06:32.786330 4791 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-log\" (UniqueName: \"kubernetes.io/host-path/ac28ec72-8d7b-4576-b2ba-c93731d50267-var-log\") pod \"ovn-controller-ovs-phsfg\" (UID: \"ac28ec72-8d7b-4576-b2ba-c93731d50267\") " pod="openstack/ovn-controller-ovs-phsfg" Dec 10 23:06:32 crc kubenswrapper[4791]: I1210 23:06:32.786871 4791 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-lib\" (UniqueName: \"kubernetes.io/host-path/ac28ec72-8d7b-4576-b2ba-c93731d50267-var-lib\") pod \"ovn-controller-ovs-phsfg\" (UID: \"ac28ec72-8d7b-4576-b2ba-c93731d50267\") " pod="openstack/ovn-controller-ovs-phsfg" Dec 10 23:06:32 crc kubenswrapper[4791]: I1210 23:06:32.786899 4791 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-run\" (UniqueName: \"kubernetes.io/host-path/780fd8a7-cfb0-4958-afb3-1d060398d3ea-var-run\") pod \"ovn-controller-xb29h\" (UID: \"780fd8a7-cfb0-4958-afb3-1d060398d3ea\") " pod="openstack/ovn-controller-xb29h" Dec 10 23:06:32 crc kubenswrapper[4791]: I1210 23:06:32.786918 4791 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovn-controller-tls-certs\" (UniqueName: \"kubernetes.io/secret/780fd8a7-cfb0-4958-afb3-1d060398d3ea-ovn-controller-tls-certs\") pod \"ovn-controller-xb29h\" (UID: \"780fd8a7-cfb0-4958-afb3-1d060398d3ea\") " pod="openstack/ovn-controller-xb29h" Dec 10 23:06:32 crc kubenswrapper[4791]: I1210 23:06:32.786940 4791 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-run-ovn\" (UniqueName: \"kubernetes.io/host-path/780fd8a7-cfb0-4958-afb3-1d060398d3ea-var-run-ovn\") pod \"ovn-controller-xb29h\" (UID: \"780fd8a7-cfb0-4958-afb3-1d060398d3ea\") " pod="openstack/ovn-controller-xb29h" Dec 10 23:06:32 crc kubenswrapper[4791]: I1210 23:06:32.786981 4791 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-ovs\" (UniqueName: \"kubernetes.io/host-path/ac28ec72-8d7b-4576-b2ba-c93731d50267-etc-ovs\") pod \"ovn-controller-ovs-phsfg\" (UID: \"ac28ec72-8d7b-4576-b2ba-c93731d50267\") " pod="openstack/ovn-controller-ovs-phsfg" Dec 10 23:06:32 crc kubenswrapper[4791]: I1210 23:06:32.787017 4791 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/ac28ec72-8d7b-4576-b2ba-c93731d50267-scripts\") pod \"ovn-controller-ovs-phsfg\" (UID: \"ac28ec72-8d7b-4576-b2ba-c93731d50267\") " pod="openstack/ovn-controller-ovs-phsfg" Dec 10 23:06:32 crc kubenswrapper[4791]: I1210 23:06:32.787039 4791 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-swdq4\" (UniqueName: \"kubernetes.io/projected/780fd8a7-cfb0-4958-afb3-1d060398d3ea-kube-api-access-swdq4\") pod \"ovn-controller-xb29h\" (UID: \"780fd8a7-cfb0-4958-afb3-1d060398d3ea\") " pod="openstack/ovn-controller-xb29h" Dec 10 23:06:32 crc kubenswrapper[4791]: I1210 23:06:32.787064 4791 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-log-ovn\" (UniqueName: \"kubernetes.io/host-path/780fd8a7-cfb0-4958-afb3-1d060398d3ea-var-log-ovn\") pod \"ovn-controller-xb29h\" (UID: \"780fd8a7-cfb0-4958-afb3-1d060398d3ea\") " pod="openstack/ovn-controller-xb29h" Dec 10 23:06:32 crc kubenswrapper[4791]: I1210 23:06:32.787079 4791 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-dd2fh\" (UniqueName: \"kubernetes.io/projected/ac28ec72-8d7b-4576-b2ba-c93731d50267-kube-api-access-dd2fh\") pod \"ovn-controller-ovs-phsfg\" (UID: \"ac28ec72-8d7b-4576-b2ba-c93731d50267\") " pod="openstack/ovn-controller-ovs-phsfg" Dec 10 23:06:32 crc kubenswrapper[4791]: I1210 23:06:32.787094 4791 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/780fd8a7-cfb0-4958-afb3-1d060398d3ea-scripts\") pod \"ovn-controller-xb29h\" (UID: \"780fd8a7-cfb0-4958-afb3-1d060398d3ea\") " pod="openstack/ovn-controller-xb29h" Dec 10 23:06:32 crc kubenswrapper[4791]: I1210 23:06:32.888224 4791 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-log\" (UniqueName: \"kubernetes.io/host-path/ac28ec72-8d7b-4576-b2ba-c93731d50267-var-log\") pod \"ovn-controller-ovs-phsfg\" (UID: \"ac28ec72-8d7b-4576-b2ba-c93731d50267\") " pod="openstack/ovn-controller-ovs-phsfg" Dec 10 23:06:32 crc kubenswrapper[4791]: I1210 23:06:32.888260 4791 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-lib\" (UniqueName: \"kubernetes.io/host-path/ac28ec72-8d7b-4576-b2ba-c93731d50267-var-lib\") pod \"ovn-controller-ovs-phsfg\" (UID: \"ac28ec72-8d7b-4576-b2ba-c93731d50267\") " pod="openstack/ovn-controller-ovs-phsfg" Dec 10 23:06:32 crc kubenswrapper[4791]: I1210 23:06:32.888278 4791 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-run\" (UniqueName: \"kubernetes.io/host-path/780fd8a7-cfb0-4958-afb3-1d060398d3ea-var-run\") pod \"ovn-controller-xb29h\" (UID: \"780fd8a7-cfb0-4958-afb3-1d060398d3ea\") " pod="openstack/ovn-controller-xb29h" Dec 10 23:06:32 crc kubenswrapper[4791]: I1210 23:06:32.888300 4791 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovn-controller-tls-certs\" (UniqueName: \"kubernetes.io/secret/780fd8a7-cfb0-4958-afb3-1d060398d3ea-ovn-controller-tls-certs\") pod \"ovn-controller-xb29h\" (UID: \"780fd8a7-cfb0-4958-afb3-1d060398d3ea\") " pod="openstack/ovn-controller-xb29h" Dec 10 23:06:32 crc kubenswrapper[4791]: I1210 23:06:32.888320 4791 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-run-ovn\" (UniqueName: \"kubernetes.io/host-path/780fd8a7-cfb0-4958-afb3-1d060398d3ea-var-run-ovn\") pod \"ovn-controller-xb29h\" (UID: \"780fd8a7-cfb0-4958-afb3-1d060398d3ea\") " pod="openstack/ovn-controller-xb29h" Dec 10 23:06:32 crc kubenswrapper[4791]: I1210 23:06:32.888358 4791 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-ovs\" (UniqueName: \"kubernetes.io/host-path/ac28ec72-8d7b-4576-b2ba-c93731d50267-etc-ovs\") pod \"ovn-controller-ovs-phsfg\" (UID: \"ac28ec72-8d7b-4576-b2ba-c93731d50267\") " pod="openstack/ovn-controller-ovs-phsfg" Dec 10 23:06:32 crc kubenswrapper[4791]: I1210 23:06:32.888389 4791 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/ac28ec72-8d7b-4576-b2ba-c93731d50267-scripts\") pod \"ovn-controller-ovs-phsfg\" (UID: \"ac28ec72-8d7b-4576-b2ba-c93731d50267\") " pod="openstack/ovn-controller-ovs-phsfg" Dec 10 23:06:32 crc kubenswrapper[4791]: I1210 23:06:32.888412 4791 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-swdq4\" (UniqueName: \"kubernetes.io/projected/780fd8a7-cfb0-4958-afb3-1d060398d3ea-kube-api-access-swdq4\") pod \"ovn-controller-xb29h\" (UID: \"780fd8a7-cfb0-4958-afb3-1d060398d3ea\") " pod="openstack/ovn-controller-xb29h" Dec 10 23:06:32 crc kubenswrapper[4791]: I1210 23:06:32.888437 4791 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-log-ovn\" (UniqueName: \"kubernetes.io/host-path/780fd8a7-cfb0-4958-afb3-1d060398d3ea-var-log-ovn\") pod \"ovn-controller-xb29h\" (UID: \"780fd8a7-cfb0-4958-afb3-1d060398d3ea\") " pod="openstack/ovn-controller-xb29h" Dec 10 23:06:32 crc kubenswrapper[4791]: I1210 23:06:32.888453 4791 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/780fd8a7-cfb0-4958-afb3-1d060398d3ea-scripts\") pod \"ovn-controller-xb29h\" (UID: \"780fd8a7-cfb0-4958-afb3-1d060398d3ea\") " pod="openstack/ovn-controller-xb29h" Dec 10 23:06:32 crc kubenswrapper[4791]: I1210 23:06:32.888468 4791 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-dd2fh\" (UniqueName: \"kubernetes.io/projected/ac28ec72-8d7b-4576-b2ba-c93731d50267-kube-api-access-dd2fh\") pod \"ovn-controller-ovs-phsfg\" (UID: \"ac28ec72-8d7b-4576-b2ba-c93731d50267\") " pod="openstack/ovn-controller-ovs-phsfg" Dec 10 23:06:32 crc kubenswrapper[4791]: I1210 23:06:32.888513 4791 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/780fd8a7-cfb0-4958-afb3-1d060398d3ea-combined-ca-bundle\") pod \"ovn-controller-xb29h\" (UID: \"780fd8a7-cfb0-4958-afb3-1d060398d3ea\") " pod="openstack/ovn-controller-xb29h" Dec 10 23:06:32 crc kubenswrapper[4791]: I1210 23:06:32.888539 4791 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-run\" (UniqueName: \"kubernetes.io/host-path/ac28ec72-8d7b-4576-b2ba-c93731d50267-var-run\") pod \"ovn-controller-ovs-phsfg\" (UID: \"ac28ec72-8d7b-4576-b2ba-c93731d50267\") " pod="openstack/ovn-controller-ovs-phsfg" Dec 10 23:06:32 crc kubenswrapper[4791]: I1210 23:06:32.888761 4791 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-log\" (UniqueName: \"kubernetes.io/host-path/ac28ec72-8d7b-4576-b2ba-c93731d50267-var-log\") pod \"ovn-controller-ovs-phsfg\" (UID: \"ac28ec72-8d7b-4576-b2ba-c93731d50267\") " pod="openstack/ovn-controller-ovs-phsfg" Dec 10 23:06:32 crc kubenswrapper[4791]: I1210 23:06:32.888849 4791 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-run\" (UniqueName: \"kubernetes.io/host-path/ac28ec72-8d7b-4576-b2ba-c93731d50267-var-run\") pod \"ovn-controller-ovs-phsfg\" (UID: \"ac28ec72-8d7b-4576-b2ba-c93731d50267\") " pod="openstack/ovn-controller-ovs-phsfg" Dec 10 23:06:32 crc kubenswrapper[4791]: I1210 23:06:32.888857 4791 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-run\" (UniqueName: \"kubernetes.io/host-path/780fd8a7-cfb0-4958-afb3-1d060398d3ea-var-run\") pod \"ovn-controller-xb29h\" (UID: \"780fd8a7-cfb0-4958-afb3-1d060398d3ea\") " pod="openstack/ovn-controller-xb29h" Dec 10 23:06:32 crc kubenswrapper[4791]: I1210 23:06:32.888894 4791 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-lib\" (UniqueName: \"kubernetes.io/host-path/ac28ec72-8d7b-4576-b2ba-c93731d50267-var-lib\") pod \"ovn-controller-ovs-phsfg\" (UID: \"ac28ec72-8d7b-4576-b2ba-c93731d50267\") " pod="openstack/ovn-controller-ovs-phsfg" Dec 10 23:06:32 crc kubenswrapper[4791]: I1210 23:06:32.888963 4791 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-log-ovn\" (UniqueName: \"kubernetes.io/host-path/780fd8a7-cfb0-4958-afb3-1d060398d3ea-var-log-ovn\") pod \"ovn-controller-xb29h\" (UID: \"780fd8a7-cfb0-4958-afb3-1d060398d3ea\") " pod="openstack/ovn-controller-xb29h" Dec 10 23:06:32 crc kubenswrapper[4791]: I1210 23:06:32.889059 4791 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-run-ovn\" (UniqueName: \"kubernetes.io/host-path/780fd8a7-cfb0-4958-afb3-1d060398d3ea-var-run-ovn\") pod \"ovn-controller-xb29h\" (UID: \"780fd8a7-cfb0-4958-afb3-1d060398d3ea\") " pod="openstack/ovn-controller-xb29h" Dec 10 23:06:32 crc kubenswrapper[4791]: I1210 23:06:32.889172 4791 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-ovs\" (UniqueName: \"kubernetes.io/host-path/ac28ec72-8d7b-4576-b2ba-c93731d50267-etc-ovs\") pod \"ovn-controller-ovs-phsfg\" (UID: \"ac28ec72-8d7b-4576-b2ba-c93731d50267\") " pod="openstack/ovn-controller-ovs-phsfg" Dec 10 23:06:32 crc kubenswrapper[4791]: I1210 23:06:32.892717 4791 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/780fd8a7-cfb0-4958-afb3-1d060398d3ea-scripts\") pod \"ovn-controller-xb29h\" (UID: \"780fd8a7-cfb0-4958-afb3-1d060398d3ea\") " pod="openstack/ovn-controller-xb29h" Dec 10 23:06:32 crc kubenswrapper[4791]: I1210 23:06:32.892910 4791 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/ac28ec72-8d7b-4576-b2ba-c93731d50267-scripts\") pod \"ovn-controller-ovs-phsfg\" (UID: \"ac28ec72-8d7b-4576-b2ba-c93731d50267\") " pod="openstack/ovn-controller-ovs-phsfg" Dec 10 23:06:32 crc kubenswrapper[4791]: I1210 23:06:32.895477 4791 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/780fd8a7-cfb0-4958-afb3-1d060398d3ea-combined-ca-bundle\") pod \"ovn-controller-xb29h\" (UID: \"780fd8a7-cfb0-4958-afb3-1d060398d3ea\") " pod="openstack/ovn-controller-xb29h" Dec 10 23:06:32 crc kubenswrapper[4791]: I1210 23:06:32.900930 4791 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovn-controller-tls-certs\" (UniqueName: \"kubernetes.io/secret/780fd8a7-cfb0-4958-afb3-1d060398d3ea-ovn-controller-tls-certs\") pod \"ovn-controller-xb29h\" (UID: \"780fd8a7-cfb0-4958-afb3-1d060398d3ea\") " pod="openstack/ovn-controller-xb29h" Dec 10 23:06:32 crc kubenswrapper[4791]: I1210 23:06:32.906851 4791 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-swdq4\" (UniqueName: \"kubernetes.io/projected/780fd8a7-cfb0-4958-afb3-1d060398d3ea-kube-api-access-swdq4\") pod \"ovn-controller-xb29h\" (UID: \"780fd8a7-cfb0-4958-afb3-1d060398d3ea\") " pod="openstack/ovn-controller-xb29h" Dec 10 23:06:32 crc kubenswrapper[4791]: I1210 23:06:32.907133 4791 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-dd2fh\" (UniqueName: \"kubernetes.io/projected/ac28ec72-8d7b-4576-b2ba-c93731d50267-kube-api-access-dd2fh\") pod \"ovn-controller-ovs-phsfg\" (UID: \"ac28ec72-8d7b-4576-b2ba-c93731d50267\") " pod="openstack/ovn-controller-ovs-phsfg" Dec 10 23:06:33 crc kubenswrapper[4791]: I1210 23:06:33.035326 4791 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-controller-xb29h" Dec 10 23:06:33 crc kubenswrapper[4791]: I1210 23:06:33.047739 4791 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-controller-ovs-phsfg" Dec 10 23:06:33 crc kubenswrapper[4791]: I1210 23:06:33.060763 4791 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ovsdbserver-nb-0"] Dec 10 23:06:33 crc kubenswrapper[4791]: I1210 23:06:33.074193 4791 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovsdbserver-nb-0" Dec 10 23:06:33 crc kubenswrapper[4791]: I1210 23:06:33.080417 4791 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"ovndbcluster-nb-config" Dec 10 23:06:33 crc kubenswrapper[4791]: I1210 23:06:33.080482 4791 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ovncluster-ovndbcluster-nb-dockercfg-k5kt4" Dec 10 23:06:33 crc kubenswrapper[4791]: I1210 23:06:33.082273 4791 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-ovndbcluster-nb-ovndbs" Dec 10 23:06:33 crc kubenswrapper[4791]: I1210 23:06:33.082474 4791 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"ovndbcluster-nb-scripts" Dec 10 23:06:33 crc kubenswrapper[4791]: I1210 23:06:33.082609 4791 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-ovn-metrics" Dec 10 23:06:33 crc kubenswrapper[4791]: I1210 23:06:33.094849 4791 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovsdbserver-nb-0"] Dec 10 23:06:33 crc kubenswrapper[4791]: I1210 23:06:33.201202 4791 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage11-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage11-crc\") pod \"ovsdbserver-nb-0\" (UID: \"a4bfff72-3a66-40f0-a06a-432d3cb4fff4\") " pod="openstack/ovsdbserver-nb-0" Dec 10 23:06:33 crc kubenswrapper[4791]: I1210 23:06:33.201280 4791 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/a4bfff72-3a66-40f0-a06a-432d3cb4fff4-scripts\") pod \"ovsdbserver-nb-0\" (UID: \"a4bfff72-3a66-40f0-a06a-432d3cb4fff4\") " pod="openstack/ovsdbserver-nb-0" Dec 10 23:06:33 crc kubenswrapper[4791]: I1210 23:06:33.201351 4791 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a4bfff72-3a66-40f0-a06a-432d3cb4fff4-combined-ca-bundle\") pod \"ovsdbserver-nb-0\" (UID: \"a4bfff72-3a66-40f0-a06a-432d3cb4fff4\") " pod="openstack/ovsdbserver-nb-0" Dec 10 23:06:33 crc kubenswrapper[4791]: I1210 23:06:33.201389 4791 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics-certs-tls-certs\" (UniqueName: \"kubernetes.io/secret/a4bfff72-3a66-40f0-a06a-432d3cb4fff4-metrics-certs-tls-certs\") pod \"ovsdbserver-nb-0\" (UID: \"a4bfff72-3a66-40f0-a06a-432d3cb4fff4\") " pod="openstack/ovsdbserver-nb-0" Dec 10 23:06:33 crc kubenswrapper[4791]: I1210 23:06:33.201424 4791 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-vqqg6\" (UniqueName: \"kubernetes.io/projected/a4bfff72-3a66-40f0-a06a-432d3cb4fff4-kube-api-access-vqqg6\") pod \"ovsdbserver-nb-0\" (UID: \"a4bfff72-3a66-40f0-a06a-432d3cb4fff4\") " pod="openstack/ovsdbserver-nb-0" Dec 10 23:06:33 crc kubenswrapper[4791]: I1210 23:06:33.201451 4791 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/a4bfff72-3a66-40f0-a06a-432d3cb4fff4-config\") pod \"ovsdbserver-nb-0\" (UID: \"a4bfff72-3a66-40f0-a06a-432d3cb4fff4\") " pod="openstack/ovsdbserver-nb-0" Dec 10 23:06:33 crc kubenswrapper[4791]: I1210 23:06:33.201481 4791 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdb-rundir\" (UniqueName: \"kubernetes.io/empty-dir/a4bfff72-3a66-40f0-a06a-432d3cb4fff4-ovsdb-rundir\") pod \"ovsdbserver-nb-0\" (UID: \"a4bfff72-3a66-40f0-a06a-432d3cb4fff4\") " pod="openstack/ovsdbserver-nb-0" Dec 10 23:06:33 crc kubenswrapper[4791]: I1210 23:06:33.201606 4791 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-nb-tls-certs\" (UniqueName: \"kubernetes.io/secret/a4bfff72-3a66-40f0-a06a-432d3cb4fff4-ovsdbserver-nb-tls-certs\") pod \"ovsdbserver-nb-0\" (UID: \"a4bfff72-3a66-40f0-a06a-432d3cb4fff4\") " pod="openstack/ovsdbserver-nb-0" Dec 10 23:06:33 crc kubenswrapper[4791]: I1210 23:06:33.302916 4791 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-nb-tls-certs\" (UniqueName: \"kubernetes.io/secret/a4bfff72-3a66-40f0-a06a-432d3cb4fff4-ovsdbserver-nb-tls-certs\") pod \"ovsdbserver-nb-0\" (UID: \"a4bfff72-3a66-40f0-a06a-432d3cb4fff4\") " pod="openstack/ovsdbserver-nb-0" Dec 10 23:06:33 crc kubenswrapper[4791]: I1210 23:06:33.302985 4791 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage11-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage11-crc\") pod \"ovsdbserver-nb-0\" (UID: \"a4bfff72-3a66-40f0-a06a-432d3cb4fff4\") " pod="openstack/ovsdbserver-nb-0" Dec 10 23:06:33 crc kubenswrapper[4791]: I1210 23:06:33.303034 4791 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/a4bfff72-3a66-40f0-a06a-432d3cb4fff4-scripts\") pod \"ovsdbserver-nb-0\" (UID: \"a4bfff72-3a66-40f0-a06a-432d3cb4fff4\") " pod="openstack/ovsdbserver-nb-0" Dec 10 23:06:33 crc kubenswrapper[4791]: I1210 23:06:33.303078 4791 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a4bfff72-3a66-40f0-a06a-432d3cb4fff4-combined-ca-bundle\") pod \"ovsdbserver-nb-0\" (UID: \"a4bfff72-3a66-40f0-a06a-432d3cb4fff4\") " pod="openstack/ovsdbserver-nb-0" Dec 10 23:06:33 crc kubenswrapper[4791]: I1210 23:06:33.303120 4791 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs-tls-certs\" (UniqueName: \"kubernetes.io/secret/a4bfff72-3a66-40f0-a06a-432d3cb4fff4-metrics-certs-tls-certs\") pod \"ovsdbserver-nb-0\" (UID: \"a4bfff72-3a66-40f0-a06a-432d3cb4fff4\") " pod="openstack/ovsdbserver-nb-0" Dec 10 23:06:33 crc kubenswrapper[4791]: I1210 23:06:33.303147 4791 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-vqqg6\" (UniqueName: \"kubernetes.io/projected/a4bfff72-3a66-40f0-a06a-432d3cb4fff4-kube-api-access-vqqg6\") pod \"ovsdbserver-nb-0\" (UID: \"a4bfff72-3a66-40f0-a06a-432d3cb4fff4\") " pod="openstack/ovsdbserver-nb-0" Dec 10 23:06:33 crc kubenswrapper[4791]: I1210 23:06:33.303175 4791 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/a4bfff72-3a66-40f0-a06a-432d3cb4fff4-config\") pod \"ovsdbserver-nb-0\" (UID: \"a4bfff72-3a66-40f0-a06a-432d3cb4fff4\") " pod="openstack/ovsdbserver-nb-0" Dec 10 23:06:33 crc kubenswrapper[4791]: I1210 23:06:33.303218 4791 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdb-rundir\" (UniqueName: \"kubernetes.io/empty-dir/a4bfff72-3a66-40f0-a06a-432d3cb4fff4-ovsdb-rundir\") pod \"ovsdbserver-nb-0\" (UID: \"a4bfff72-3a66-40f0-a06a-432d3cb4fff4\") " pod="openstack/ovsdbserver-nb-0" Dec 10 23:06:33 crc kubenswrapper[4791]: I1210 23:06:33.303694 4791 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdb-rundir\" (UniqueName: \"kubernetes.io/empty-dir/a4bfff72-3a66-40f0-a06a-432d3cb4fff4-ovsdb-rundir\") pod \"ovsdbserver-nb-0\" (UID: \"a4bfff72-3a66-40f0-a06a-432d3cb4fff4\") " pod="openstack/ovsdbserver-nb-0" Dec 10 23:06:33 crc kubenswrapper[4791]: I1210 23:06:33.305846 4791 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage11-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage11-crc\") pod \"ovsdbserver-nb-0\" (UID: \"a4bfff72-3a66-40f0-a06a-432d3cb4fff4\") device mount path \"/mnt/openstack/pv11\"" pod="openstack/ovsdbserver-nb-0" Dec 10 23:06:33 crc kubenswrapper[4791]: I1210 23:06:33.305891 4791 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/a4bfff72-3a66-40f0-a06a-432d3cb4fff4-scripts\") pod \"ovsdbserver-nb-0\" (UID: \"a4bfff72-3a66-40f0-a06a-432d3cb4fff4\") " pod="openstack/ovsdbserver-nb-0" Dec 10 23:06:33 crc kubenswrapper[4791]: I1210 23:06:33.306775 4791 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/a4bfff72-3a66-40f0-a06a-432d3cb4fff4-config\") pod \"ovsdbserver-nb-0\" (UID: \"a4bfff72-3a66-40f0-a06a-432d3cb4fff4\") " pod="openstack/ovsdbserver-nb-0" Dec 10 23:06:33 crc kubenswrapper[4791]: I1210 23:06:33.316404 4791 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-certs-tls-certs\" (UniqueName: \"kubernetes.io/secret/a4bfff72-3a66-40f0-a06a-432d3cb4fff4-metrics-certs-tls-certs\") pod \"ovsdbserver-nb-0\" (UID: \"a4bfff72-3a66-40f0-a06a-432d3cb4fff4\") " pod="openstack/ovsdbserver-nb-0" Dec 10 23:06:33 crc kubenswrapper[4791]: I1210 23:06:33.316775 4791 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-nb-tls-certs\" (UniqueName: \"kubernetes.io/secret/a4bfff72-3a66-40f0-a06a-432d3cb4fff4-ovsdbserver-nb-tls-certs\") pod \"ovsdbserver-nb-0\" (UID: \"a4bfff72-3a66-40f0-a06a-432d3cb4fff4\") " pod="openstack/ovsdbserver-nb-0" Dec 10 23:06:33 crc kubenswrapper[4791]: I1210 23:06:33.316921 4791 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a4bfff72-3a66-40f0-a06a-432d3cb4fff4-combined-ca-bundle\") pod \"ovsdbserver-nb-0\" (UID: \"a4bfff72-3a66-40f0-a06a-432d3cb4fff4\") " pod="openstack/ovsdbserver-nb-0" Dec 10 23:06:33 crc kubenswrapper[4791]: I1210 23:06:33.329279 4791 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-vqqg6\" (UniqueName: \"kubernetes.io/projected/a4bfff72-3a66-40f0-a06a-432d3cb4fff4-kube-api-access-vqqg6\") pod \"ovsdbserver-nb-0\" (UID: \"a4bfff72-3a66-40f0-a06a-432d3cb4fff4\") " pod="openstack/ovsdbserver-nb-0" Dec 10 23:06:33 crc kubenswrapper[4791]: I1210 23:06:33.332113 4791 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage11-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage11-crc\") pod \"ovsdbserver-nb-0\" (UID: \"a4bfff72-3a66-40f0-a06a-432d3cb4fff4\") " pod="openstack/ovsdbserver-nb-0" Dec 10 23:06:33 crc kubenswrapper[4791]: I1210 23:06:33.508598 4791 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovsdbserver-nb-0" Dec 10 23:06:36 crc kubenswrapper[4791]: I1210 23:06:36.660163 4791 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ovsdbserver-sb-0"] Dec 10 23:06:36 crc kubenswrapper[4791]: I1210 23:06:36.664267 4791 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovsdbserver-sb-0" Dec 10 23:06:36 crc kubenswrapper[4791]: I1210 23:06:36.670126 4791 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ovncluster-ovndbcluster-sb-dockercfg-8cczh" Dec 10 23:06:36 crc kubenswrapper[4791]: I1210 23:06:36.671179 4791 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-ovndbcluster-sb-ovndbs" Dec 10 23:06:36 crc kubenswrapper[4791]: I1210 23:06:36.671540 4791 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"ovndbcluster-sb-scripts" Dec 10 23:06:36 crc kubenswrapper[4791]: I1210 23:06:36.677632 4791 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"ovndbcluster-sb-config" Dec 10 23:06:36 crc kubenswrapper[4791]: I1210 23:06:36.687984 4791 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovsdbserver-sb-0"] Dec 10 23:06:36 crc kubenswrapper[4791]: I1210 23:06:36.770354 4791 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage04-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage04-crc\") pod \"ovsdbserver-sb-0\" (UID: \"61769416-0728-4475-b965-dedb50d4455d\") " pod="openstack/ovsdbserver-sb-0" Dec 10 23:06:36 crc kubenswrapper[4791]: I1210 23:06:36.770410 4791 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics-certs-tls-certs\" (UniqueName: \"kubernetes.io/secret/61769416-0728-4475-b965-dedb50d4455d-metrics-certs-tls-certs\") pod \"ovsdbserver-sb-0\" (UID: \"61769416-0728-4475-b965-dedb50d4455d\") " pod="openstack/ovsdbserver-sb-0" Dec 10 23:06:36 crc kubenswrapper[4791]: I1210 23:06:36.770449 4791 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-tgv9h\" (UniqueName: \"kubernetes.io/projected/61769416-0728-4475-b965-dedb50d4455d-kube-api-access-tgv9h\") pod \"ovsdbserver-sb-0\" (UID: \"61769416-0728-4475-b965-dedb50d4455d\") " pod="openstack/ovsdbserver-sb-0" Dec 10 23:06:36 crc kubenswrapper[4791]: I1210 23:06:36.770481 4791 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/61769416-0728-4475-b965-dedb50d4455d-scripts\") pod \"ovsdbserver-sb-0\" (UID: \"61769416-0728-4475-b965-dedb50d4455d\") " pod="openstack/ovsdbserver-sb-0" Dec 10 23:06:36 crc kubenswrapper[4791]: I1210 23:06:36.770514 4791 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-sb-tls-certs\" (UniqueName: \"kubernetes.io/secret/61769416-0728-4475-b965-dedb50d4455d-ovsdbserver-sb-tls-certs\") pod \"ovsdbserver-sb-0\" (UID: \"61769416-0728-4475-b965-dedb50d4455d\") " pod="openstack/ovsdbserver-sb-0" Dec 10 23:06:36 crc kubenswrapper[4791]: I1210 23:06:36.770537 4791 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/61769416-0728-4475-b965-dedb50d4455d-combined-ca-bundle\") pod \"ovsdbserver-sb-0\" (UID: \"61769416-0728-4475-b965-dedb50d4455d\") " pod="openstack/ovsdbserver-sb-0" Dec 10 23:06:36 crc kubenswrapper[4791]: I1210 23:06:36.770554 4791 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdb-rundir\" (UniqueName: \"kubernetes.io/empty-dir/61769416-0728-4475-b965-dedb50d4455d-ovsdb-rundir\") pod \"ovsdbserver-sb-0\" (UID: \"61769416-0728-4475-b965-dedb50d4455d\") " pod="openstack/ovsdbserver-sb-0" Dec 10 23:06:36 crc kubenswrapper[4791]: I1210 23:06:36.770571 4791 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/61769416-0728-4475-b965-dedb50d4455d-config\") pod \"ovsdbserver-sb-0\" (UID: \"61769416-0728-4475-b965-dedb50d4455d\") " pod="openstack/ovsdbserver-sb-0" Dec 10 23:06:36 crc kubenswrapper[4791]: I1210 23:06:36.872135 4791 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdb-rundir\" (UniqueName: \"kubernetes.io/empty-dir/61769416-0728-4475-b965-dedb50d4455d-ovsdb-rundir\") pod \"ovsdbserver-sb-0\" (UID: \"61769416-0728-4475-b965-dedb50d4455d\") " pod="openstack/ovsdbserver-sb-0" Dec 10 23:06:36 crc kubenswrapper[4791]: I1210 23:06:36.872205 4791 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/61769416-0728-4475-b965-dedb50d4455d-config\") pod \"ovsdbserver-sb-0\" (UID: \"61769416-0728-4475-b965-dedb50d4455d\") " pod="openstack/ovsdbserver-sb-0" Dec 10 23:06:36 crc kubenswrapper[4791]: I1210 23:06:36.872274 4791 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage04-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage04-crc\") pod \"ovsdbserver-sb-0\" (UID: \"61769416-0728-4475-b965-dedb50d4455d\") " pod="openstack/ovsdbserver-sb-0" Dec 10 23:06:36 crc kubenswrapper[4791]: I1210 23:06:36.872301 4791 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs-tls-certs\" (UniqueName: \"kubernetes.io/secret/61769416-0728-4475-b965-dedb50d4455d-metrics-certs-tls-certs\") pod \"ovsdbserver-sb-0\" (UID: \"61769416-0728-4475-b965-dedb50d4455d\") " pod="openstack/ovsdbserver-sb-0" Dec 10 23:06:36 crc kubenswrapper[4791]: I1210 23:06:36.872833 4791 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdb-rundir\" (UniqueName: \"kubernetes.io/empty-dir/61769416-0728-4475-b965-dedb50d4455d-ovsdb-rundir\") pod \"ovsdbserver-sb-0\" (UID: \"61769416-0728-4475-b965-dedb50d4455d\") " pod="openstack/ovsdbserver-sb-0" Dec 10 23:06:36 crc kubenswrapper[4791]: I1210 23:06:36.872692 4791 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage04-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage04-crc\") pod \"ovsdbserver-sb-0\" (UID: \"61769416-0728-4475-b965-dedb50d4455d\") device mount path \"/mnt/openstack/pv04\"" pod="openstack/ovsdbserver-sb-0" Dec 10 23:06:36 crc kubenswrapper[4791]: I1210 23:06:36.873239 4791 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/61769416-0728-4475-b965-dedb50d4455d-config\") pod \"ovsdbserver-sb-0\" (UID: \"61769416-0728-4475-b965-dedb50d4455d\") " pod="openstack/ovsdbserver-sb-0" Dec 10 23:06:36 crc kubenswrapper[4791]: I1210 23:06:36.873906 4791 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-tgv9h\" (UniqueName: \"kubernetes.io/projected/61769416-0728-4475-b965-dedb50d4455d-kube-api-access-tgv9h\") pod \"ovsdbserver-sb-0\" (UID: \"61769416-0728-4475-b965-dedb50d4455d\") " pod="openstack/ovsdbserver-sb-0" Dec 10 23:06:36 crc kubenswrapper[4791]: I1210 23:06:36.874209 4791 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/61769416-0728-4475-b965-dedb50d4455d-scripts\") pod \"ovsdbserver-sb-0\" (UID: \"61769416-0728-4475-b965-dedb50d4455d\") " pod="openstack/ovsdbserver-sb-0" Dec 10 23:06:36 crc kubenswrapper[4791]: I1210 23:06:36.874473 4791 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-sb-tls-certs\" (UniqueName: \"kubernetes.io/secret/61769416-0728-4475-b965-dedb50d4455d-ovsdbserver-sb-tls-certs\") pod \"ovsdbserver-sb-0\" (UID: \"61769416-0728-4475-b965-dedb50d4455d\") " pod="openstack/ovsdbserver-sb-0" Dec 10 23:06:36 crc kubenswrapper[4791]: I1210 23:06:36.874674 4791 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/61769416-0728-4475-b965-dedb50d4455d-combined-ca-bundle\") pod \"ovsdbserver-sb-0\" (UID: \"61769416-0728-4475-b965-dedb50d4455d\") " pod="openstack/ovsdbserver-sb-0" Dec 10 23:06:36 crc kubenswrapper[4791]: I1210 23:06:36.875870 4791 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/61769416-0728-4475-b965-dedb50d4455d-scripts\") pod \"ovsdbserver-sb-0\" (UID: \"61769416-0728-4475-b965-dedb50d4455d\") " pod="openstack/ovsdbserver-sb-0" Dec 10 23:06:36 crc kubenswrapper[4791]: I1210 23:06:36.881560 4791 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-certs-tls-certs\" (UniqueName: \"kubernetes.io/secret/61769416-0728-4475-b965-dedb50d4455d-metrics-certs-tls-certs\") pod \"ovsdbserver-sb-0\" (UID: \"61769416-0728-4475-b965-dedb50d4455d\") " pod="openstack/ovsdbserver-sb-0" Dec 10 23:06:36 crc kubenswrapper[4791]: I1210 23:06:36.884701 4791 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/61769416-0728-4475-b965-dedb50d4455d-combined-ca-bundle\") pod \"ovsdbserver-sb-0\" (UID: \"61769416-0728-4475-b965-dedb50d4455d\") " pod="openstack/ovsdbserver-sb-0" Dec 10 23:06:36 crc kubenswrapper[4791]: I1210 23:06:36.888586 4791 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-sb-tls-certs\" (UniqueName: \"kubernetes.io/secret/61769416-0728-4475-b965-dedb50d4455d-ovsdbserver-sb-tls-certs\") pod \"ovsdbserver-sb-0\" (UID: \"61769416-0728-4475-b965-dedb50d4455d\") " pod="openstack/ovsdbserver-sb-0" Dec 10 23:06:36 crc kubenswrapper[4791]: I1210 23:06:36.896239 4791 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-tgv9h\" (UniqueName: \"kubernetes.io/projected/61769416-0728-4475-b965-dedb50d4455d-kube-api-access-tgv9h\") pod \"ovsdbserver-sb-0\" (UID: \"61769416-0728-4475-b965-dedb50d4455d\") " pod="openstack/ovsdbserver-sb-0" Dec 10 23:06:36 crc kubenswrapper[4791]: I1210 23:06:36.896287 4791 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage04-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage04-crc\") pod \"ovsdbserver-sb-0\" (UID: \"61769416-0728-4475-b965-dedb50d4455d\") " pod="openstack/ovsdbserver-sb-0" Dec 10 23:06:36 crc kubenswrapper[4791]: I1210 23:06:36.995240 4791 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovsdbserver-sb-0" Dec 10 23:06:46 crc kubenswrapper[4791]: E1210 23:06:46.209175 4791 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="quay.io/podified-antelope-centos9/openstack-memcached:current-podified" Dec 10 23:06:46 crc kubenswrapper[4791]: E1210 23:06:46.209938 4791 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:memcached,Image:quay.io/podified-antelope-centos9/openstack-memcached:current-podified,Command:[/usr/bin/dumb-init -- /usr/local/bin/kolla_start],Args:[],WorkingDir:,Ports:[]ContainerPort{ContainerPort{Name:memcached,HostPort:0,ContainerPort:11211,Protocol:TCP,HostIP:,},ContainerPort{Name:memcached-tls,HostPort:0,ContainerPort:11212,Protocol:TCP,HostIP:,},},Env:[]EnvVar{EnvVar{Name:KOLLA_CONFIG_STRATEGY,Value:COPY_ALWAYS,ValueFrom:nil,},EnvVar{Name:POD_IPS,Value:,ValueFrom:&EnvVarSource{FieldRef:&ObjectFieldSelector{APIVersion:v1,FieldPath:status.podIPs,},ResourceFieldRef:nil,ConfigMapKeyRef:nil,SecretKeyRef:nil,},},EnvVar{Name:CONFIG_HASH,Value:n646h668h5c6h69h654h559h687h56bh5f4h5bh588hc7h5ffh55h68h676hb4h696h664h549h5f4hbfhcdh65bh76h588h55h6dh5b8h647h579h6cq,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:config-data,ReadOnly:true,MountPath:/var/lib/kolla/config_files/src,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kolla-config,ReadOnly:true,MountPath:/var/lib/kolla/config_files,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:memcached-tls-certs,ReadOnly:true,MountPath:/var/lib/config-data/tls/certs/memcached.crt,SubPath:tls.crt,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:memcached-tls-certs,ReadOnly:true,MountPath:/var/lib/config-data/tls/private/memcached.key,SubPath:tls.key,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:combined-ca-bundle,ReadOnly:true,MountPath:/etc/pki/ca-trust/extracted/pem/tls-ca-bundle.pem,SubPath:tls-ca-bundle.pem,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-rwlqb,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:nil,TCPSocket:&TCPSocketAction{Port:{0 11211 },Host:,},GRPC:nil,},InitialDelaySeconds:3,TimeoutSeconds:5,PeriodSeconds:3,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},ReadinessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:nil,TCPSocket:&TCPSocketAction{Port:{0 11211 },Host:,},GRPC:nil,},InitialDelaySeconds:5,TimeoutSeconds:5,PeriodSeconds:5,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*42457,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:nil,RunAsGroup:*42457,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod memcached-0_openstack(49839437-3696-4ac8-98cf-2cc16048f21a): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError" Dec 10 23:06:46 crc kubenswrapper[4791]: E1210 23:06:46.212736 4791 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"memcached\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\"" pod="openstack/memcached-0" podUID="49839437-3696-4ac8-98cf-2cc16048f21a" Dec 10 23:06:46 crc kubenswrapper[4791]: E1210 23:06:46.309908 4791 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"memcached\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/podified-antelope-centos9/openstack-memcached:current-podified\\\"\"" pod="openstack/memcached-0" podUID="49839437-3696-4ac8-98cf-2cc16048f21a" Dec 10 23:06:48 crc kubenswrapper[4791]: E1210 23:06:48.241317 4791 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="quay.io/podified-antelope-centos9/openstack-mariadb:current-podified" Dec 10 23:06:48 crc kubenswrapper[4791]: E1210 23:06:48.241808 4791 kuberuntime_manager.go:1274] "Unhandled Error" err="init container &Container{Name:mysql-bootstrap,Image:quay.io/podified-antelope-centos9/openstack-mariadb:current-podified,Command:[bash /var/lib/operator-scripts/mysql_bootstrap.sh],Args:[],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:KOLLA_BOOTSTRAP,Value:True,ValueFrom:nil,},EnvVar{Name:KOLLA_CONFIG_STRATEGY,Value:COPY_ALWAYS,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:mysql-db,ReadOnly:false,MountPath:/var/lib/mysql,SubPath:mysql,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:config-data-default,ReadOnly:true,MountPath:/var/lib/config-data/default,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:config-data-generated,ReadOnly:false,MountPath:/var/lib/config-data/generated,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:operator-scripts,ReadOnly:true,MountPath:/var/lib/operator-scripts,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kolla-config,ReadOnly:true,MountPath:/var/lib/kolla/config_files,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-7sphf,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:nil,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod openstack-cell1-galera-0_openstack(f04fa829-3ebe-445d-a4e5-f7592ac682f3): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError" Dec 10 23:06:48 crc kubenswrapper[4791]: E1210 23:06:48.243000 4791 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"mysql-bootstrap\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\"" pod="openstack/openstack-cell1-galera-0" podUID="f04fa829-3ebe-445d-a4e5-f7592ac682f3" Dec 10 23:06:48 crc kubenswrapper[4791]: E1210 23:06:48.331097 4791 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"mysql-bootstrap\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/podified-antelope-centos9/openstack-mariadb:current-podified\\\"\"" pod="openstack/openstack-cell1-galera-0" podUID="f04fa829-3ebe-445d-a4e5-f7592ac682f3" Dec 10 23:06:49 crc kubenswrapper[4791]: E1210 23:06:49.182700 4791 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="quay.io/podified-antelope-centos9/openstack-rabbitmq:current-podified" Dec 10 23:06:49 crc kubenswrapper[4791]: E1210 23:06:49.182906 4791 kuberuntime_manager.go:1274] "Unhandled Error" err="init container &Container{Name:setup-container,Image:quay.io/podified-antelope-centos9/openstack-rabbitmq:current-podified,Command:[sh -c cp /tmp/erlang-cookie-secret/.erlang.cookie /var/lib/rabbitmq/.erlang.cookie && chmod 600 /var/lib/rabbitmq/.erlang.cookie ; cp /tmp/rabbitmq-plugins/enabled_plugins /operator/enabled_plugins ; echo '[default]' > /var/lib/rabbitmq/.rabbitmqadmin.conf && sed -e 's/default_user/username/' -e 's/default_pass/password/' /tmp/default_user.conf >> /var/lib/rabbitmq/.rabbitmqadmin.conf && chmod 600 /var/lib/rabbitmq/.rabbitmqadmin.conf ; sleep 30],Args:[],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{cpu: {{20 -3} {} 20m DecimalSI},memory: {{67108864 0} {} BinarySI},},Requests:ResourceList{cpu: {{20 -3} {} 20m DecimalSI},memory: {{67108864 0} {} BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:plugins-conf,ReadOnly:false,MountPath:/tmp/rabbitmq-plugins/,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:rabbitmq-erlang-cookie,ReadOnly:false,MountPath:/var/lib/rabbitmq/,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:erlang-cookie-secret,ReadOnly:false,MountPath:/tmp/erlang-cookie-secret/,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:rabbitmq-plugins,ReadOnly:false,MountPath:/operator,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:persistence,ReadOnly:false,MountPath:/var/lib/rabbitmq/mnesia/,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:rabbitmq-confd,ReadOnly:false,MountPath:/tmp/default_user.conf,SubPath:default_user.conf,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-kmc8x,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000650000,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod rabbitmq-cell1-server-0_openstack(47eb9200-b0bc-41a6-abb9-f6167bd69c66): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError" Dec 10 23:06:49 crc kubenswrapper[4791]: E1210 23:06:49.184084 4791 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"setup-container\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\"" pod="openstack/rabbitmq-cell1-server-0" podUID="47eb9200-b0bc-41a6-abb9-f6167bd69c66" Dec 10 23:06:49 crc kubenswrapper[4791]: E1210 23:06:49.333521 4791 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"setup-container\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/podified-antelope-centos9/openstack-rabbitmq:current-podified\\\"\"" pod="openstack/rabbitmq-cell1-server-0" podUID="47eb9200-b0bc-41a6-abb9-f6167bd69c66" Dec 10 23:06:53 crc kubenswrapper[4791]: E1210 23:06:53.551124 4791 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="quay.io/podified-antelope-centos9/openstack-mariadb:current-podified" Dec 10 23:06:53 crc kubenswrapper[4791]: E1210 23:06:53.551529 4791 kuberuntime_manager.go:1274] "Unhandled Error" err="init container &Container{Name:mysql-bootstrap,Image:quay.io/podified-antelope-centos9/openstack-mariadb:current-podified,Command:[bash /var/lib/operator-scripts/mysql_bootstrap.sh],Args:[],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:KOLLA_BOOTSTRAP,Value:True,ValueFrom:nil,},EnvVar{Name:KOLLA_CONFIG_STRATEGY,Value:COPY_ALWAYS,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:mysql-db,ReadOnly:false,MountPath:/var/lib/mysql,SubPath:mysql,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:config-data-default,ReadOnly:true,MountPath:/var/lib/config-data/default,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:config-data-generated,ReadOnly:false,MountPath:/var/lib/config-data/generated,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:operator-scripts,ReadOnly:true,MountPath:/var/lib/operator-scripts,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kolla-config,ReadOnly:true,MountPath:/var/lib/kolla/config_files,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-phmhb,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:nil,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod openstack-galera-0_openstack(aed2b986-c040-4191-864b-47f29c5c8537): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError" Dec 10 23:06:53 crc kubenswrapper[4791]: E1210 23:06:53.552753 4791 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"mysql-bootstrap\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\"" pod="openstack/openstack-galera-0" podUID="aed2b986-c040-4191-864b-47f29c5c8537" Dec 10 23:06:53 crc kubenswrapper[4791]: E1210 23:06:53.596036 4791 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"mysql-bootstrap\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/podified-antelope-centos9/openstack-mariadb:current-podified\\\"\"" pod="openstack/openstack-galera-0" podUID="aed2b986-c040-4191-864b-47f29c5c8537" Dec 10 23:06:53 crc kubenswrapper[4791]: E1210 23:06:53.622540 4791 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="quay.io/podified-antelope-centos9/openstack-rabbitmq:current-podified" Dec 10 23:06:53 crc kubenswrapper[4791]: E1210 23:06:53.622731 4791 kuberuntime_manager.go:1274] "Unhandled Error" err="init container &Container{Name:setup-container,Image:quay.io/podified-antelope-centos9/openstack-rabbitmq:current-podified,Command:[sh -c cp /tmp/erlang-cookie-secret/.erlang.cookie /var/lib/rabbitmq/.erlang.cookie && chmod 600 /var/lib/rabbitmq/.erlang.cookie ; cp /tmp/rabbitmq-plugins/enabled_plugins /operator/enabled_plugins ; echo '[default]' > /var/lib/rabbitmq/.rabbitmqadmin.conf && sed -e 's/default_user/username/' -e 's/default_pass/password/' /tmp/default_user.conf >> /var/lib/rabbitmq/.rabbitmqadmin.conf && chmod 600 /var/lib/rabbitmq/.rabbitmqadmin.conf ; sleep 30],Args:[],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{cpu: {{20 -3} {} 20m DecimalSI},memory: {{67108864 0} {} BinarySI},},Requests:ResourceList{cpu: {{20 -3} {} 20m DecimalSI},memory: {{67108864 0} {} BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:plugins-conf,ReadOnly:false,MountPath:/tmp/rabbitmq-plugins/,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:rabbitmq-erlang-cookie,ReadOnly:false,MountPath:/var/lib/rabbitmq/,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:erlang-cookie-secret,ReadOnly:false,MountPath:/tmp/erlang-cookie-secret/,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:rabbitmq-plugins,ReadOnly:false,MountPath:/operator,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:persistence,ReadOnly:false,MountPath:/var/lib/rabbitmq/mnesia/,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:rabbitmq-confd,ReadOnly:false,MountPath:/tmp/default_user.conf,SubPath:default_user.conf,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-85rrv,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000650000,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod rabbitmq-server-0_openstack(9ce8dbeb-dce8-47ac-98f1-fe92f74aa3de): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError" Dec 10 23:06:53 crc kubenswrapper[4791]: E1210 23:06:53.623904 4791 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"setup-container\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\"" pod="openstack/rabbitmq-server-0" podUID="9ce8dbeb-dce8-47ac-98f1-fe92f74aa3de" Dec 10 23:06:54 crc kubenswrapper[4791]: I1210 23:06:54.076299 4791 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovn-controller-xb29h"] Dec 10 23:06:54 crc kubenswrapper[4791]: E1210 23:06:54.445578 4791 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="quay.io/podified-antelope-centos9/openstack-neutron-server:current-podified" Dec 10 23:06:54 crc kubenswrapper[4791]: E1210 23:06:54.445730 4791 kuberuntime_manager.go:1274] "Unhandled Error" err="init container &Container{Name:init,Image:quay.io/podified-antelope-centos9/openstack-neutron-server:current-podified,Command:[/bin/bash],Args:[-c dnsmasq --interface=* --conf-dir=/etc/dnsmasq.d --hostsdir=/etc/dnsmasq.d/hosts --keep-in-foreground --log-debug --bind-interfaces --listen-address=$(POD_IP) --port 5353 --log-facility=- --no-hosts --domain-needed --no-resolv --bogus-priv --log-queries --test],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:CONFIG_HASH,Value:ndfhb5h667h568h584h5f9h58dh565h664h587h597h577h64bh5c4h66fh647hbdh68ch5c5h68dh686h5f7h64hd7hc6h55fh57bh98h57fh87h5fh57fq,ValueFrom:nil,},EnvVar{Name:POD_IP,Value:,ValueFrom:&EnvVarSource{FieldRef:&ObjectFieldSelector{APIVersion:v1,FieldPath:status.podIP,},ResourceFieldRef:nil,ConfigMapKeyRef:nil,SecretKeyRef:nil,},},},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:config,ReadOnly:true,MountPath:/etc/dnsmasq.d/config.cfg,SubPath:dns,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:dns-svc,ReadOnly:true,MountPath:/etc/dnsmasq.d/hosts/dns-svc,SubPath:dns-svc,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-fbfng,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000650000,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:&SeccompProfile{Type:RuntimeDefault,LocalhostProfile:nil,},AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod dnsmasq-dns-78dd6ddcc-gg848_openstack(bb91eb46-389f-46d4-8fc3-a84266bb5aff): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError" Dec 10 23:06:54 crc kubenswrapper[4791]: E1210 23:06:54.446925 4791 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"init\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\"" pod="openstack/dnsmasq-dns-78dd6ddcc-gg848" podUID="bb91eb46-389f-46d4-8fc3-a84266bb5aff" Dec 10 23:06:54 crc kubenswrapper[4791]: E1210 23:06:54.453333 4791 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="quay.io/podified-antelope-centos9/openstack-neutron-server:current-podified" Dec 10 23:06:54 crc kubenswrapper[4791]: E1210 23:06:54.453478 4791 kuberuntime_manager.go:1274] "Unhandled Error" err="init container &Container{Name:init,Image:quay.io/podified-antelope-centos9/openstack-neutron-server:current-podified,Command:[/bin/bash],Args:[-c dnsmasq --interface=* --conf-dir=/etc/dnsmasq.d --hostsdir=/etc/dnsmasq.d/hosts --keep-in-foreground --log-debug --bind-interfaces --listen-address=$(POD_IP) --port 5353 --log-facility=- --no-hosts --domain-needed --no-resolv --bogus-priv --log-queries --test],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:CONFIG_HASH,Value:n659h4h664hbh658h587h67ch89h587h8fh679hc6hf9h55fh644h5d5h698h68dh5cdh5ffh669h54ch9h689hb8hd4h5bfhd8h5d7h5fh665h574q,ValueFrom:nil,},EnvVar{Name:POD_IP,Value:,ValueFrom:&EnvVarSource{FieldRef:&ObjectFieldSelector{APIVersion:v1,FieldPath:status.podIP,},ResourceFieldRef:nil,ConfigMapKeyRef:nil,SecretKeyRef:nil,},},},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:config,ReadOnly:true,MountPath:/etc/dnsmasq.d/config.cfg,SubPath:dns,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:dns-svc,ReadOnly:true,MountPath:/etc/dnsmasq.d/hosts/dns-svc,SubPath:dns-svc,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-kwvlk,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000650000,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:&SeccompProfile{Type:RuntimeDefault,LocalhostProfile:nil,},AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod dnsmasq-dns-57d769cc4f-tnfbj_openstack(b2c7ad15-ca5a-4042-9d55-31524249a851): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError" Dec 10 23:06:54 crc kubenswrapper[4791]: E1210 23:06:54.454801 4791 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"init\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\"" pod="openstack/dnsmasq-dns-57d769cc4f-tnfbj" podUID="b2c7ad15-ca5a-4042-9d55-31524249a851" Dec 10 23:06:54 crc kubenswrapper[4791]: E1210 23:06:54.527540 4791 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="quay.io/podified-antelope-centos9/openstack-neutron-server:current-podified" Dec 10 23:06:54 crc kubenswrapper[4791]: E1210 23:06:54.527683 4791 kuberuntime_manager.go:1274] "Unhandled Error" err="init container &Container{Name:init,Image:quay.io/podified-antelope-centos9/openstack-neutron-server:current-podified,Command:[/bin/bash],Args:[-c dnsmasq --interface=* --conf-dir=/etc/dnsmasq.d --hostsdir=/etc/dnsmasq.d/hosts --keep-in-foreground --log-debug --bind-interfaces --listen-address=$(POD_IP) --port 5353 --log-facility=- --no-hosts --domain-needed --no-resolv --bogus-priv --log-queries --test],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:CONFIG_HASH,Value:nffh5bdhf4h5f8h79h55h77h58fh56dh7bh6fh578hbch55dh68h56bhd9h65dh57ch658hc9h566h666h688h58h65dh684h5d7h6ch575h5d6h88q,ValueFrom:nil,},EnvVar{Name:POD_IP,Value:,ValueFrom:&EnvVarSource{FieldRef:&ObjectFieldSelector{APIVersion:v1,FieldPath:status.podIP,},ResourceFieldRef:nil,ConfigMapKeyRef:nil,SecretKeyRef:nil,},},},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:config,ReadOnly:true,MountPath:/etc/dnsmasq.d/config.cfg,SubPath:dns,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-578pp,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000650000,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:&SeccompProfile{Type:RuntimeDefault,LocalhostProfile:nil,},AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod dnsmasq-dns-675f4bcbfc-68bpm_openstack(6dd522dc-1554-4d27-a289-78b0650df8d1): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError" Dec 10 23:06:54 crc kubenswrapper[4791]: E1210 23:06:54.530769 4791 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"init\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\"" pod="openstack/dnsmasq-dns-675f4bcbfc-68bpm" podUID="6dd522dc-1554-4d27-a289-78b0650df8d1" Dec 10 23:06:54 crc kubenswrapper[4791]: E1210 23:06:54.557531 4791 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="quay.io/podified-antelope-centos9/openstack-neutron-server:current-podified" Dec 10 23:06:54 crc kubenswrapper[4791]: E1210 23:06:54.557714 4791 kuberuntime_manager.go:1274] "Unhandled Error" err="init container &Container{Name:init,Image:quay.io/podified-antelope-centos9/openstack-neutron-server:current-podified,Command:[/bin/bash],Args:[-c dnsmasq --interface=* --conf-dir=/etc/dnsmasq.d --hostsdir=/etc/dnsmasq.d/hosts --keep-in-foreground --log-debug --bind-interfaces --listen-address=$(POD_IP) --port 5353 --log-facility=- --no-hosts --domain-needed --no-resolv --bogus-priv --log-queries --test],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:CONFIG_HASH,Value:nfdh5dfhb6h64h676hc4h78h97h669h54chfbh696hb5h54bh5d4h6bh64h644h677h584h5cbh698h9dh5bbh5f8h5b8hcdh644h5c7h694hbfh589q,ValueFrom:nil,},EnvVar{Name:POD_IP,Value:,ValueFrom:&EnvVarSource{FieldRef:&ObjectFieldSelector{APIVersion:v1,FieldPath:status.podIP,},ResourceFieldRef:nil,ConfigMapKeyRef:nil,SecretKeyRef:nil,},},},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:config,ReadOnly:true,MountPath:/etc/dnsmasq.d/config.cfg,SubPath:dns,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:dns-svc,ReadOnly:true,MountPath:/etc/dnsmasq.d/hosts/dns-svc,SubPath:dns-svc,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-zn4rz,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000650000,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:&SeccompProfile{Type:RuntimeDefault,LocalhostProfile:nil,},AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod dnsmasq-dns-5ccc8479f9-8w9hb_openstack(73f16e68-0124-4784-977f-72ea5d6d04bf): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError" Dec 10 23:06:54 crc kubenswrapper[4791]: E1210 23:06:54.560019 4791 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"init\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\"" pod="openstack/dnsmasq-dns-5ccc8479f9-8w9hb" podUID="73f16e68-0124-4784-977f-72ea5d6d04bf" Dec 10 23:06:54 crc kubenswrapper[4791]: I1210 23:06:54.599025 4791 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-controller-xb29h" event={"ID":"780fd8a7-cfb0-4958-afb3-1d060398d3ea","Type":"ContainerStarted","Data":"9d57084873a08827475d5ab9bc93d58f0a049a42c10e3e5d69940a9f77368aba"} Dec 10 23:06:54 crc kubenswrapper[4791]: E1210 23:06:54.608819 4791 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"init\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/podified-antelope-centos9/openstack-neutron-server:current-podified\\\"\"" pod="openstack/dnsmasq-dns-5ccc8479f9-8w9hb" podUID="73f16e68-0124-4784-977f-72ea5d6d04bf" Dec 10 23:06:54 crc kubenswrapper[4791]: E1210 23:06:54.609122 4791 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"setup-container\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/podified-antelope-centos9/openstack-rabbitmq:current-podified\\\"\"" pod="openstack/rabbitmq-server-0" podUID="9ce8dbeb-dce8-47ac-98f1-fe92f74aa3de" Dec 10 23:06:54 crc kubenswrapper[4791]: E1210 23:06:54.609226 4791 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"init\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/podified-antelope-centos9/openstack-neutron-server:current-podified\\\"\"" pod="openstack/dnsmasq-dns-57d769cc4f-tnfbj" podUID="b2c7ad15-ca5a-4042-9d55-31524249a851" Dec 10 23:06:54 crc kubenswrapper[4791]: I1210 23:06:54.801734 4791 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovsdbserver-nb-0"] Dec 10 23:06:55 crc kubenswrapper[4791]: I1210 23:06:55.065406 4791 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-78dd6ddcc-gg848" Dec 10 23:06:55 crc kubenswrapper[4791]: I1210 23:06:55.092394 4791 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovn-controller-ovs-phsfg"] Dec 10 23:06:55 crc kubenswrapper[4791]: I1210 23:06:55.218167 4791 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-fbfng\" (UniqueName: \"kubernetes.io/projected/bb91eb46-389f-46d4-8fc3-a84266bb5aff-kube-api-access-fbfng\") pod \"bb91eb46-389f-46d4-8fc3-a84266bb5aff\" (UID: \"bb91eb46-389f-46d4-8fc3-a84266bb5aff\") " Dec 10 23:06:55 crc kubenswrapper[4791]: I1210 23:06:55.218352 4791 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/bb91eb46-389f-46d4-8fc3-a84266bb5aff-dns-svc\") pod \"bb91eb46-389f-46d4-8fc3-a84266bb5aff\" (UID: \"bb91eb46-389f-46d4-8fc3-a84266bb5aff\") " Dec 10 23:06:55 crc kubenswrapper[4791]: I1210 23:06:55.218449 4791 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/bb91eb46-389f-46d4-8fc3-a84266bb5aff-config\") pod \"bb91eb46-389f-46d4-8fc3-a84266bb5aff\" (UID: \"bb91eb46-389f-46d4-8fc3-a84266bb5aff\") " Dec 10 23:06:55 crc kubenswrapper[4791]: I1210 23:06:55.219440 4791 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/bb91eb46-389f-46d4-8fc3-a84266bb5aff-config" (OuterVolumeSpecName: "config") pod "bb91eb46-389f-46d4-8fc3-a84266bb5aff" (UID: "bb91eb46-389f-46d4-8fc3-a84266bb5aff"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 10 23:06:55 crc kubenswrapper[4791]: I1210 23:06:55.219842 4791 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/bb91eb46-389f-46d4-8fc3-a84266bb5aff-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "bb91eb46-389f-46d4-8fc3-a84266bb5aff" (UID: "bb91eb46-389f-46d4-8fc3-a84266bb5aff"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 10 23:06:55 crc kubenswrapper[4791]: I1210 23:06:55.225751 4791 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/bb91eb46-389f-46d4-8fc3-a84266bb5aff-kube-api-access-fbfng" (OuterVolumeSpecName: "kube-api-access-fbfng") pod "bb91eb46-389f-46d4-8fc3-a84266bb5aff" (UID: "bb91eb46-389f-46d4-8fc3-a84266bb5aff"). InnerVolumeSpecName "kube-api-access-fbfng". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 10 23:06:55 crc kubenswrapper[4791]: I1210 23:06:55.272286 4791 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-675f4bcbfc-68bpm" Dec 10 23:06:55 crc kubenswrapper[4791]: E1210 23:06:55.296588 4791 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying system image from manifest list: copying layer: context canceled" image="registry.k8s.io/kube-state-metrics/kube-state-metrics:v2.15.0" Dec 10 23:06:55 crc kubenswrapper[4791]: E1210 23:06:55.296644 4791 kuberuntime_image.go:55] "Failed to pull image" err="rpc error: code = Canceled desc = copying system image from manifest list: copying layer: context canceled" image="registry.k8s.io/kube-state-metrics/kube-state-metrics:v2.15.0" Dec 10 23:06:55 crc kubenswrapper[4791]: E1210 23:06:55.296775 4791 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:kube-state-metrics,Image:registry.k8s.io/kube-state-metrics/kube-state-metrics:v2.15.0,Command:[],Args:[--resources=pods --namespaces=openstack],WorkingDir:,Ports:[]ContainerPort{ContainerPort{Name:http-metrics,HostPort:0,ContainerPort:8080,Protocol:TCP,HostIP:,},ContainerPort{Name:telemetry,HostPort:0,ContainerPort:8081,Protocol:TCP,HostIP:,},},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kube-api-access-tlj7g,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/livez,Port:{0 8080 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:5,TimeoutSeconds:5,PeriodSeconds:10,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},ReadinessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/readyz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:5,TimeoutSeconds:5,PeriodSeconds:10,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:Always,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000650000,RunAsNonRoot:*true,ReadOnlyRootFilesystem:*true,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:&SeccompProfile{Type:RuntimeDefault,LocalhostProfile:nil,},AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod kube-state-metrics-0_openstack(247efe24-6d59-4c93-ab04-6e249e1005a8): ErrImagePull: rpc error: code = Canceled desc = copying system image from manifest list: copying layer: context canceled" logger="UnhandledError" Dec 10 23:06:55 crc kubenswrapper[4791]: E1210 23:06:55.298113 4791 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"kube-state-metrics\" with ErrImagePull: \"rpc error: code = Canceled desc = copying system image from manifest list: copying layer: context canceled\"" pod="openstack/kube-state-metrics-0" podUID="247efe24-6d59-4c93-ab04-6e249e1005a8" Dec 10 23:06:55 crc kubenswrapper[4791]: I1210 23:06:55.319761 4791 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/bb91eb46-389f-46d4-8fc3-a84266bb5aff-config\") on node \"crc\" DevicePath \"\"" Dec 10 23:06:55 crc kubenswrapper[4791]: I1210 23:06:55.319805 4791 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-fbfng\" (UniqueName: \"kubernetes.io/projected/bb91eb46-389f-46d4-8fc3-a84266bb5aff-kube-api-access-fbfng\") on node \"crc\" DevicePath \"\"" Dec 10 23:06:55 crc kubenswrapper[4791]: I1210 23:06:55.319824 4791 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/bb91eb46-389f-46d4-8fc3-a84266bb5aff-dns-svc\") on node \"crc\" DevicePath \"\"" Dec 10 23:06:55 crc kubenswrapper[4791]: I1210 23:06:55.421247 4791 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/6dd522dc-1554-4d27-a289-78b0650df8d1-config\") pod \"6dd522dc-1554-4d27-a289-78b0650df8d1\" (UID: \"6dd522dc-1554-4d27-a289-78b0650df8d1\") " Dec 10 23:06:55 crc kubenswrapper[4791]: I1210 23:06:55.421404 4791 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-578pp\" (UniqueName: \"kubernetes.io/projected/6dd522dc-1554-4d27-a289-78b0650df8d1-kube-api-access-578pp\") pod \"6dd522dc-1554-4d27-a289-78b0650df8d1\" (UID: \"6dd522dc-1554-4d27-a289-78b0650df8d1\") " Dec 10 23:06:55 crc kubenswrapper[4791]: I1210 23:06:55.421789 4791 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/6dd522dc-1554-4d27-a289-78b0650df8d1-config" (OuterVolumeSpecName: "config") pod "6dd522dc-1554-4d27-a289-78b0650df8d1" (UID: "6dd522dc-1554-4d27-a289-78b0650df8d1"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 10 23:06:55 crc kubenswrapper[4791]: I1210 23:06:55.425262 4791 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/6dd522dc-1554-4d27-a289-78b0650df8d1-kube-api-access-578pp" (OuterVolumeSpecName: "kube-api-access-578pp") pod "6dd522dc-1554-4d27-a289-78b0650df8d1" (UID: "6dd522dc-1554-4d27-a289-78b0650df8d1"). InnerVolumeSpecName "kube-api-access-578pp". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 10 23:06:55 crc kubenswrapper[4791]: I1210 23:06:55.499988 4791 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovsdbserver-sb-0"] Dec 10 23:06:55 crc kubenswrapper[4791]: W1210 23:06:55.502798 4791 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod61769416_0728_4475_b965_dedb50d4455d.slice/crio-1a1b43ef1cca7f232ad987afc18f7a7ac4fbdf8b2fa97729304c0e63718eb886 WatchSource:0}: Error finding container 1a1b43ef1cca7f232ad987afc18f7a7ac4fbdf8b2fa97729304c0e63718eb886: Status 404 returned error can't find the container with id 1a1b43ef1cca7f232ad987afc18f7a7ac4fbdf8b2fa97729304c0e63718eb886 Dec 10 23:06:55 crc kubenswrapper[4791]: I1210 23:06:55.523129 4791 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/6dd522dc-1554-4d27-a289-78b0650df8d1-config\") on node \"crc\" DevicePath \"\"" Dec 10 23:06:55 crc kubenswrapper[4791]: I1210 23:06:55.523170 4791 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-578pp\" (UniqueName: \"kubernetes.io/projected/6dd522dc-1554-4d27-a289-78b0650df8d1-kube-api-access-578pp\") on node \"crc\" DevicePath \"\"" Dec 10 23:06:55 crc kubenswrapper[4791]: I1210 23:06:55.608784 4791 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-675f4bcbfc-68bpm" event={"ID":"6dd522dc-1554-4d27-a289-78b0650df8d1","Type":"ContainerDied","Data":"7338da6d30033fbd48e70f4fc83255e34b5b5143034535de719e8f6318120a52"} Dec 10 23:06:55 crc kubenswrapper[4791]: I1210 23:06:55.608866 4791 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-675f4bcbfc-68bpm" Dec 10 23:06:55 crc kubenswrapper[4791]: I1210 23:06:55.612083 4791 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-78dd6ddcc-gg848" event={"ID":"bb91eb46-389f-46d4-8fc3-a84266bb5aff","Type":"ContainerDied","Data":"c73c67e9a3cab9f4f79fce0e6c2a2b49d4fb828ea36b11a377c77404dc287ce7"} Dec 10 23:06:55 crc kubenswrapper[4791]: I1210 23:06:55.612178 4791 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-78dd6ddcc-gg848" Dec 10 23:06:55 crc kubenswrapper[4791]: I1210 23:06:55.615087 4791 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovsdbserver-nb-0" event={"ID":"a4bfff72-3a66-40f0-a06a-432d3cb4fff4","Type":"ContainerStarted","Data":"ecc1a1ba5b01cab1a8890abaf35fd1c38daab817f9dccbf36382692b2db1a5e8"} Dec 10 23:06:55 crc kubenswrapper[4791]: I1210 23:06:55.616707 4791 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovsdbserver-sb-0" event={"ID":"61769416-0728-4475-b965-dedb50d4455d","Type":"ContainerStarted","Data":"1a1b43ef1cca7f232ad987afc18f7a7ac4fbdf8b2fa97729304c0e63718eb886"} Dec 10 23:06:55 crc kubenswrapper[4791]: I1210 23:06:55.618511 4791 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-controller-ovs-phsfg" event={"ID":"ac28ec72-8d7b-4576-b2ba-c93731d50267","Type":"ContainerStarted","Data":"56b1701cd4307a59cb4ca07144a992ca7270682f9f4a96da176b0deb01c64896"} Dec 10 23:06:55 crc kubenswrapper[4791]: E1210 23:06:55.620475 4791 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"kube-state-metrics\" with ImagePullBackOff: \"Back-off pulling image \\\"registry.k8s.io/kube-state-metrics/kube-state-metrics:v2.15.0\\\"\"" pod="openstack/kube-state-metrics-0" podUID="247efe24-6d59-4c93-ab04-6e249e1005a8" Dec 10 23:06:55 crc kubenswrapper[4791]: I1210 23:06:55.679476 4791 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-675f4bcbfc-68bpm"] Dec 10 23:06:55 crc kubenswrapper[4791]: I1210 23:06:55.685986 4791 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-675f4bcbfc-68bpm"] Dec 10 23:06:55 crc kubenswrapper[4791]: I1210 23:06:55.705469 4791 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-78dd6ddcc-gg848"] Dec 10 23:06:55 crc kubenswrapper[4791]: I1210 23:06:55.712390 4791 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-78dd6ddcc-gg848"] Dec 10 23:06:55 crc kubenswrapper[4791]: I1210 23:06:55.897029 4791 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="6dd522dc-1554-4d27-a289-78b0650df8d1" path="/var/lib/kubelet/pods/6dd522dc-1554-4d27-a289-78b0650df8d1/volumes" Dec 10 23:06:55 crc kubenswrapper[4791]: I1210 23:06:55.897433 4791 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="bb91eb46-389f-46d4-8fc3-a84266bb5aff" path="/var/lib/kubelet/pods/bb91eb46-389f-46d4-8fc3-a84266bb5aff/volumes" Dec 10 23:06:58 crc kubenswrapper[4791]: I1210 23:06:58.640944 4791 generic.go:334] "Generic (PLEG): container finished" podID="ac28ec72-8d7b-4576-b2ba-c93731d50267" containerID="d777687f74d5a6163317900ae59451321bf5e302956432921a7d6f4a45c86ea8" exitCode=0 Dec 10 23:06:58 crc kubenswrapper[4791]: I1210 23:06:58.641009 4791 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-controller-ovs-phsfg" event={"ID":"ac28ec72-8d7b-4576-b2ba-c93731d50267","Type":"ContainerDied","Data":"d777687f74d5a6163317900ae59451321bf5e302956432921a7d6f4a45c86ea8"} Dec 10 23:06:58 crc kubenswrapper[4791]: I1210 23:06:58.644323 4791 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-controller-xb29h" event={"ID":"780fd8a7-cfb0-4958-afb3-1d060398d3ea","Type":"ContainerStarted","Data":"97decd949e429513015c82e7f5ae69b0e4cf07df6af9d8e2f4247102fba027b0"} Dec 10 23:06:58 crc kubenswrapper[4791]: I1210 23:06:58.644474 4791 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/ovn-controller-xb29h" Dec 10 23:06:58 crc kubenswrapper[4791]: I1210 23:06:58.646887 4791 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovsdbserver-nb-0" event={"ID":"a4bfff72-3a66-40f0-a06a-432d3cb4fff4","Type":"ContainerStarted","Data":"ace8d9f9b723b04b9f57de4c9213c53e1e096f57af2fd652ab35512c22f23ddb"} Dec 10 23:06:58 crc kubenswrapper[4791]: I1210 23:06:58.648146 4791 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovsdbserver-sb-0" event={"ID":"61769416-0728-4475-b965-dedb50d4455d","Type":"ContainerStarted","Data":"79e33c1f3bd1d9e27d6d719c31e922c4b54fc417f01b19d821fb676ee3d89205"} Dec 10 23:06:58 crc kubenswrapper[4791]: I1210 23:06:58.677978 4791 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ovn-controller-xb29h" podStartSLOduration=23.073278057 podStartE2EDuration="26.677963517s" podCreationTimestamp="2025-12-10 23:06:32 +0000 UTC" firstStartedPulling="2025-12-10 23:06:54.481143429 +0000 UTC m=+1048.910761042" lastFinishedPulling="2025-12-10 23:06:58.085828889 +0000 UTC m=+1052.515446502" observedRunningTime="2025-12-10 23:06:58.672603772 +0000 UTC m=+1053.102221395" watchObservedRunningTime="2025-12-10 23:06:58.677963517 +0000 UTC m=+1053.107581130" Dec 10 23:06:59 crc kubenswrapper[4791]: I1210 23:06:59.658148 4791 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-controller-ovs-phsfg" event={"ID":"ac28ec72-8d7b-4576-b2ba-c93731d50267","Type":"ContainerStarted","Data":"23ab2a837a71462f0cfff693869b59a735d1662cce56f2d2decf579c51ea716b"} Dec 10 23:06:59 crc kubenswrapper[4791]: I1210 23:06:59.660743 4791 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/ovn-controller-ovs-phsfg" Dec 10 23:06:59 crc kubenswrapper[4791]: I1210 23:06:59.660854 4791 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-controller-ovs-phsfg" event={"ID":"ac28ec72-8d7b-4576-b2ba-c93731d50267","Type":"ContainerStarted","Data":"6c343fcc98a921142781ffcbe8874d3e6e39823f6512a0d7d7ea9fe725b5ac89"} Dec 10 23:06:59 crc kubenswrapper[4791]: I1210 23:06:59.675849 4791 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ovn-controller-ovs-phsfg" podStartSLOduration=24.815926855 podStartE2EDuration="27.675826918s" podCreationTimestamp="2025-12-10 23:06:32 +0000 UTC" firstStartedPulling="2025-12-10 23:06:55.204239388 +0000 UTC m=+1049.633857001" lastFinishedPulling="2025-12-10 23:06:58.064139441 +0000 UTC m=+1052.493757064" observedRunningTime="2025-12-10 23:06:59.674295383 +0000 UTC m=+1054.103912996" watchObservedRunningTime="2025-12-10 23:06:59.675826918 +0000 UTC m=+1054.105444541" Dec 10 23:07:00 crc kubenswrapper[4791]: I1210 23:07:00.665967 4791 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/ovn-controller-ovs-phsfg" Dec 10 23:07:01 crc kubenswrapper[4791]: I1210 23:07:01.681272 4791 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovsdbserver-sb-0" event={"ID":"61769416-0728-4475-b965-dedb50d4455d","Type":"ContainerStarted","Data":"ee84233ebbb7a374586fea93c2ef5c78d1243962275b87c3e40dc410cb70974b"} Dec 10 23:07:01 crc kubenswrapper[4791]: I1210 23:07:01.687632 4791 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovsdbserver-nb-0" event={"ID":"a4bfff72-3a66-40f0-a06a-432d3cb4fff4","Type":"ContainerStarted","Data":"64878f1c15ce806cdbd476451e4dc1411471266a0e966dd1508b641ed787cbd5"} Dec 10 23:07:01 crc kubenswrapper[4791]: I1210 23:07:01.718448 4791 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ovsdbserver-sb-0" podStartSLOduration=20.978282751 podStartE2EDuration="26.718428766s" podCreationTimestamp="2025-12-10 23:06:35 +0000 UTC" firstStartedPulling="2025-12-10 23:06:55.504940251 +0000 UTC m=+1049.934557864" lastFinishedPulling="2025-12-10 23:07:01.245086266 +0000 UTC m=+1055.674703879" observedRunningTime="2025-12-10 23:07:01.711544697 +0000 UTC m=+1056.141162340" watchObservedRunningTime="2025-12-10 23:07:01.718428766 +0000 UTC m=+1056.148046389" Dec 10 23:07:01 crc kubenswrapper[4791]: I1210 23:07:01.753811 4791 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ovsdbserver-nb-0" podStartSLOduration=23.514185889 podStartE2EDuration="29.753785319s" podCreationTimestamp="2025-12-10 23:06:32 +0000 UTC" firstStartedPulling="2025-12-10 23:06:55.00528761 +0000 UTC m=+1049.434905223" lastFinishedPulling="2025-12-10 23:07:01.24488704 +0000 UTC m=+1055.674504653" observedRunningTime="2025-12-10 23:07:01.742151012 +0000 UTC m=+1056.171768675" watchObservedRunningTime="2025-12-10 23:07:01.753785319 +0000 UTC m=+1056.183402962" Dec 10 23:07:01 crc kubenswrapper[4791]: I1210 23:07:01.996403 4791 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/ovsdbserver-sb-0" Dec 10 23:07:03 crc kubenswrapper[4791]: I1210 23:07:03.509027 4791 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/ovsdbserver-nb-0" Dec 10 23:07:03 crc kubenswrapper[4791]: I1210 23:07:03.509607 4791 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/ovsdbserver-nb-0" Dec 10 23:07:03 crc kubenswrapper[4791]: I1210 23:07:03.560913 4791 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/ovsdbserver-nb-0" Dec 10 23:07:03 crc kubenswrapper[4791]: I1210 23:07:03.703721 4791 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-cell1-server-0" event={"ID":"47eb9200-b0bc-41a6-abb9-f6167bd69c66","Type":"ContainerStarted","Data":"f0d087440ba482240de855bdc96648601e34f74962b3bf2123c4205414310fef"} Dec 10 23:07:03 crc kubenswrapper[4791]: I1210 23:07:03.706459 4791 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/openstack-cell1-galera-0" event={"ID":"f04fa829-3ebe-445d-a4e5-f7592ac682f3","Type":"ContainerStarted","Data":"4ac34c10ef97f846f1cbd4e7bd46b995220d082b77e94ede2ad536e90f578da4"} Dec 10 23:07:03 crc kubenswrapper[4791]: I1210 23:07:03.708688 4791 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/memcached-0" event={"ID":"49839437-3696-4ac8-98cf-2cc16048f21a","Type":"ContainerStarted","Data":"609de6066c505aa72c57245e5d9a4e76dd0a4aae59e59b7d9440152520e50b97"} Dec 10 23:07:03 crc kubenswrapper[4791]: I1210 23:07:03.709203 4791 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/memcached-0" Dec 10 23:07:03 crc kubenswrapper[4791]: I1210 23:07:03.751632 4791 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/ovsdbserver-nb-0" Dec 10 23:07:03 crc kubenswrapper[4791]: I1210 23:07:03.832945 4791 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/memcached-0" podStartSLOduration=3.200089687 podStartE2EDuration="37.832928115s" podCreationTimestamp="2025-12-10 23:06:26 +0000 UTC" firstStartedPulling="2025-12-10 23:06:27.874527812 +0000 UTC m=+1022.304145425" lastFinishedPulling="2025-12-10 23:07:02.50736624 +0000 UTC m=+1056.936983853" observedRunningTime="2025-12-10 23:07:03.809062395 +0000 UTC m=+1058.238679998" watchObservedRunningTime="2025-12-10 23:07:03.832928115 +0000 UTC m=+1058.262545728" Dec 10 23:07:03 crc kubenswrapper[4791]: I1210 23:07:03.960159 4791 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-57d769cc4f-tnfbj"] Dec 10 23:07:04 crc kubenswrapper[4791]: I1210 23:07:03.996270 4791 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/ovsdbserver-sb-0" Dec 10 23:07:04 crc kubenswrapper[4791]: I1210 23:07:04.031392 4791 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-7fd796d7df-7z8ld"] Dec 10 23:07:04 crc kubenswrapper[4791]: I1210 23:07:04.032785 4791 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-7fd796d7df-7z8ld" Dec 10 23:07:04 crc kubenswrapper[4791]: I1210 23:07:04.038635 4791 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"ovsdbserver-nb" Dec 10 23:07:04 crc kubenswrapper[4791]: I1210 23:07:04.054156 4791 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-7fd796d7df-7z8ld"] Dec 10 23:07:04 crc kubenswrapper[4791]: I1210 23:07:04.055388 4791 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-r7979\" (UniqueName: \"kubernetes.io/projected/60c7a68b-a75f-4cdb-ab6c-703ceccd2036-kube-api-access-r7979\") pod \"dnsmasq-dns-7fd796d7df-7z8ld\" (UID: \"60c7a68b-a75f-4cdb-ab6c-703ceccd2036\") " pod="openstack/dnsmasq-dns-7fd796d7df-7z8ld" Dec 10 23:07:04 crc kubenswrapper[4791]: I1210 23:07:04.055580 4791 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/60c7a68b-a75f-4cdb-ab6c-703ceccd2036-config\") pod \"dnsmasq-dns-7fd796d7df-7z8ld\" (UID: \"60c7a68b-a75f-4cdb-ab6c-703ceccd2036\") " pod="openstack/dnsmasq-dns-7fd796d7df-7z8ld" Dec 10 23:07:04 crc kubenswrapper[4791]: I1210 23:07:04.055608 4791 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/60c7a68b-a75f-4cdb-ab6c-703ceccd2036-dns-svc\") pod \"dnsmasq-dns-7fd796d7df-7z8ld\" (UID: \"60c7a68b-a75f-4cdb-ab6c-703ceccd2036\") " pod="openstack/dnsmasq-dns-7fd796d7df-7z8ld" Dec 10 23:07:04 crc kubenswrapper[4791]: I1210 23:07:04.055679 4791 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/60c7a68b-a75f-4cdb-ab6c-703ceccd2036-ovsdbserver-nb\") pod \"dnsmasq-dns-7fd796d7df-7z8ld\" (UID: \"60c7a68b-a75f-4cdb-ab6c-703ceccd2036\") " pod="openstack/dnsmasq-dns-7fd796d7df-7z8ld" Dec 10 23:07:04 crc kubenswrapper[4791]: I1210 23:07:04.114131 4791 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ovn-controller-metrics-djrh2"] Dec 10 23:07:04 crc kubenswrapper[4791]: I1210 23:07:04.115116 4791 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-controller-metrics-djrh2" Dec 10 23:07:04 crc kubenswrapper[4791]: I1210 23:07:04.127408 4791 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"ovncontroller-metrics-config" Dec 10 23:07:04 crc kubenswrapper[4791]: I1210 23:07:04.148044 4791 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/ovsdbserver-sb-0" Dec 10 23:07:04 crc kubenswrapper[4791]: I1210 23:07:04.148093 4791 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovn-controller-metrics-djrh2"] Dec 10 23:07:04 crc kubenswrapper[4791]: I1210 23:07:04.158235 4791 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/53ceaa05-0398-4cf9-a787-874adab015aa-config\") pod \"ovn-controller-metrics-djrh2\" (UID: \"53ceaa05-0398-4cf9-a787-874adab015aa\") " pod="openstack/ovn-controller-metrics-djrh2" Dec 10 23:07:04 crc kubenswrapper[4791]: I1210 23:07:04.158305 4791 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics-certs-tls-certs\" (UniqueName: \"kubernetes.io/secret/53ceaa05-0398-4cf9-a787-874adab015aa-metrics-certs-tls-certs\") pod \"ovn-controller-metrics-djrh2\" (UID: \"53ceaa05-0398-4cf9-a787-874adab015aa\") " pod="openstack/ovn-controller-metrics-djrh2" Dec 10 23:07:04 crc kubenswrapper[4791]: I1210 23:07:04.158362 4791 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/60c7a68b-a75f-4cdb-ab6c-703ceccd2036-ovsdbserver-nb\") pod \"dnsmasq-dns-7fd796d7df-7z8ld\" (UID: \"60c7a68b-a75f-4cdb-ab6c-703ceccd2036\") " pod="openstack/dnsmasq-dns-7fd796d7df-7z8ld" Dec 10 23:07:04 crc kubenswrapper[4791]: I1210 23:07:04.158402 4791 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovn-rundir\" (UniqueName: \"kubernetes.io/host-path/53ceaa05-0398-4cf9-a787-874adab015aa-ovn-rundir\") pod \"ovn-controller-metrics-djrh2\" (UID: \"53ceaa05-0398-4cf9-a787-874adab015aa\") " pod="openstack/ovn-controller-metrics-djrh2" Dec 10 23:07:04 crc kubenswrapper[4791]: I1210 23:07:04.158463 4791 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-r7979\" (UniqueName: \"kubernetes.io/projected/60c7a68b-a75f-4cdb-ab6c-703ceccd2036-kube-api-access-r7979\") pod \"dnsmasq-dns-7fd796d7df-7z8ld\" (UID: \"60c7a68b-a75f-4cdb-ab6c-703ceccd2036\") " pod="openstack/dnsmasq-dns-7fd796d7df-7z8ld" Dec 10 23:07:04 crc kubenswrapper[4791]: I1210 23:07:04.158486 4791 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovs-rundir\" (UniqueName: \"kubernetes.io/host-path/53ceaa05-0398-4cf9-a787-874adab015aa-ovs-rundir\") pod \"ovn-controller-metrics-djrh2\" (UID: \"53ceaa05-0398-4cf9-a787-874adab015aa\") " pod="openstack/ovn-controller-metrics-djrh2" Dec 10 23:07:04 crc kubenswrapper[4791]: I1210 23:07:04.158545 4791 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-pnslj\" (UniqueName: \"kubernetes.io/projected/53ceaa05-0398-4cf9-a787-874adab015aa-kube-api-access-pnslj\") pod \"ovn-controller-metrics-djrh2\" (UID: \"53ceaa05-0398-4cf9-a787-874adab015aa\") " pod="openstack/ovn-controller-metrics-djrh2" Dec 10 23:07:04 crc kubenswrapper[4791]: I1210 23:07:04.158601 4791 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/60c7a68b-a75f-4cdb-ab6c-703ceccd2036-config\") pod \"dnsmasq-dns-7fd796d7df-7z8ld\" (UID: \"60c7a68b-a75f-4cdb-ab6c-703ceccd2036\") " pod="openstack/dnsmasq-dns-7fd796d7df-7z8ld" Dec 10 23:07:04 crc kubenswrapper[4791]: I1210 23:07:04.158623 4791 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/60c7a68b-a75f-4cdb-ab6c-703ceccd2036-dns-svc\") pod \"dnsmasq-dns-7fd796d7df-7z8ld\" (UID: \"60c7a68b-a75f-4cdb-ab6c-703ceccd2036\") " pod="openstack/dnsmasq-dns-7fd796d7df-7z8ld" Dec 10 23:07:04 crc kubenswrapper[4791]: I1210 23:07:04.158697 4791 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/53ceaa05-0398-4cf9-a787-874adab015aa-combined-ca-bundle\") pod \"ovn-controller-metrics-djrh2\" (UID: \"53ceaa05-0398-4cf9-a787-874adab015aa\") " pod="openstack/ovn-controller-metrics-djrh2" Dec 10 23:07:04 crc kubenswrapper[4791]: I1210 23:07:04.160470 4791 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/60c7a68b-a75f-4cdb-ab6c-703ceccd2036-config\") pod \"dnsmasq-dns-7fd796d7df-7z8ld\" (UID: \"60c7a68b-a75f-4cdb-ab6c-703ceccd2036\") " pod="openstack/dnsmasq-dns-7fd796d7df-7z8ld" Dec 10 23:07:04 crc kubenswrapper[4791]: I1210 23:07:04.160822 4791 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/60c7a68b-a75f-4cdb-ab6c-703ceccd2036-dns-svc\") pod \"dnsmasq-dns-7fd796d7df-7z8ld\" (UID: \"60c7a68b-a75f-4cdb-ab6c-703ceccd2036\") " pod="openstack/dnsmasq-dns-7fd796d7df-7z8ld" Dec 10 23:07:04 crc kubenswrapper[4791]: I1210 23:07:04.161329 4791 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/60c7a68b-a75f-4cdb-ab6c-703ceccd2036-ovsdbserver-nb\") pod \"dnsmasq-dns-7fd796d7df-7z8ld\" (UID: \"60c7a68b-a75f-4cdb-ab6c-703ceccd2036\") " pod="openstack/dnsmasq-dns-7fd796d7df-7z8ld" Dec 10 23:07:04 crc kubenswrapper[4791]: I1210 23:07:04.190800 4791 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-r7979\" (UniqueName: \"kubernetes.io/projected/60c7a68b-a75f-4cdb-ab6c-703ceccd2036-kube-api-access-r7979\") pod \"dnsmasq-dns-7fd796d7df-7z8ld\" (UID: \"60c7a68b-a75f-4cdb-ab6c-703ceccd2036\") " pod="openstack/dnsmasq-dns-7fd796d7df-7z8ld" Dec 10 23:07:04 crc kubenswrapper[4791]: I1210 23:07:04.261435 4791 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-pnslj\" (UniqueName: \"kubernetes.io/projected/53ceaa05-0398-4cf9-a787-874adab015aa-kube-api-access-pnslj\") pod \"ovn-controller-metrics-djrh2\" (UID: \"53ceaa05-0398-4cf9-a787-874adab015aa\") " pod="openstack/ovn-controller-metrics-djrh2" Dec 10 23:07:04 crc kubenswrapper[4791]: I1210 23:07:04.261524 4791 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/53ceaa05-0398-4cf9-a787-874adab015aa-combined-ca-bundle\") pod \"ovn-controller-metrics-djrh2\" (UID: \"53ceaa05-0398-4cf9-a787-874adab015aa\") " pod="openstack/ovn-controller-metrics-djrh2" Dec 10 23:07:04 crc kubenswrapper[4791]: I1210 23:07:04.261548 4791 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/53ceaa05-0398-4cf9-a787-874adab015aa-config\") pod \"ovn-controller-metrics-djrh2\" (UID: \"53ceaa05-0398-4cf9-a787-874adab015aa\") " pod="openstack/ovn-controller-metrics-djrh2" Dec 10 23:07:04 crc kubenswrapper[4791]: I1210 23:07:04.261571 4791 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs-tls-certs\" (UniqueName: \"kubernetes.io/secret/53ceaa05-0398-4cf9-a787-874adab015aa-metrics-certs-tls-certs\") pod \"ovn-controller-metrics-djrh2\" (UID: \"53ceaa05-0398-4cf9-a787-874adab015aa\") " pod="openstack/ovn-controller-metrics-djrh2" Dec 10 23:07:04 crc kubenswrapper[4791]: I1210 23:07:04.261607 4791 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovn-rundir\" (UniqueName: \"kubernetes.io/host-path/53ceaa05-0398-4cf9-a787-874adab015aa-ovn-rundir\") pod \"ovn-controller-metrics-djrh2\" (UID: \"53ceaa05-0398-4cf9-a787-874adab015aa\") " pod="openstack/ovn-controller-metrics-djrh2" Dec 10 23:07:04 crc kubenswrapper[4791]: I1210 23:07:04.261784 4791 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovs-rundir\" (UniqueName: \"kubernetes.io/host-path/53ceaa05-0398-4cf9-a787-874adab015aa-ovs-rundir\") pod \"ovn-controller-metrics-djrh2\" (UID: \"53ceaa05-0398-4cf9-a787-874adab015aa\") " pod="openstack/ovn-controller-metrics-djrh2" Dec 10 23:07:04 crc kubenswrapper[4791]: I1210 23:07:04.262224 4791 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/53ceaa05-0398-4cf9-a787-874adab015aa-config\") pod \"ovn-controller-metrics-djrh2\" (UID: \"53ceaa05-0398-4cf9-a787-874adab015aa\") " pod="openstack/ovn-controller-metrics-djrh2" Dec 10 23:07:04 crc kubenswrapper[4791]: I1210 23:07:04.262483 4791 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovs-rundir\" (UniqueName: \"kubernetes.io/host-path/53ceaa05-0398-4cf9-a787-874adab015aa-ovs-rundir\") pod \"ovn-controller-metrics-djrh2\" (UID: \"53ceaa05-0398-4cf9-a787-874adab015aa\") " pod="openstack/ovn-controller-metrics-djrh2" Dec 10 23:07:04 crc kubenswrapper[4791]: I1210 23:07:04.262520 4791 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovn-rundir\" (UniqueName: \"kubernetes.io/host-path/53ceaa05-0398-4cf9-a787-874adab015aa-ovn-rundir\") pod \"ovn-controller-metrics-djrh2\" (UID: \"53ceaa05-0398-4cf9-a787-874adab015aa\") " pod="openstack/ovn-controller-metrics-djrh2" Dec 10 23:07:04 crc kubenswrapper[4791]: I1210 23:07:04.268322 4791 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/53ceaa05-0398-4cf9-a787-874adab015aa-combined-ca-bundle\") pod \"ovn-controller-metrics-djrh2\" (UID: \"53ceaa05-0398-4cf9-a787-874adab015aa\") " pod="openstack/ovn-controller-metrics-djrh2" Dec 10 23:07:04 crc kubenswrapper[4791]: I1210 23:07:04.277276 4791 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-certs-tls-certs\" (UniqueName: \"kubernetes.io/secret/53ceaa05-0398-4cf9-a787-874adab015aa-metrics-certs-tls-certs\") pod \"ovn-controller-metrics-djrh2\" (UID: \"53ceaa05-0398-4cf9-a787-874adab015aa\") " pod="openstack/ovn-controller-metrics-djrh2" Dec 10 23:07:04 crc kubenswrapper[4791]: I1210 23:07:04.281906 4791 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-pnslj\" (UniqueName: \"kubernetes.io/projected/53ceaa05-0398-4cf9-a787-874adab015aa-kube-api-access-pnslj\") pod \"ovn-controller-metrics-djrh2\" (UID: \"53ceaa05-0398-4cf9-a787-874adab015aa\") " pod="openstack/ovn-controller-metrics-djrh2" Dec 10 23:07:04 crc kubenswrapper[4791]: I1210 23:07:04.372628 4791 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-7fd796d7df-7z8ld" Dec 10 23:07:04 crc kubenswrapper[4791]: I1210 23:07:04.446520 4791 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-controller-metrics-djrh2" Dec 10 23:07:04 crc kubenswrapper[4791]: I1210 23:07:04.450824 4791 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-57d769cc4f-tnfbj" Dec 10 23:07:04 crc kubenswrapper[4791]: I1210 23:07:04.452863 4791 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-5ccc8479f9-8w9hb"] Dec 10 23:07:04 crc kubenswrapper[4791]: I1210 23:07:04.469752 4791 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/b2c7ad15-ca5a-4042-9d55-31524249a851-dns-svc\") pod \"b2c7ad15-ca5a-4042-9d55-31524249a851\" (UID: \"b2c7ad15-ca5a-4042-9d55-31524249a851\") " Dec 10 23:07:04 crc kubenswrapper[4791]: I1210 23:07:04.469923 4791 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-kwvlk\" (UniqueName: \"kubernetes.io/projected/b2c7ad15-ca5a-4042-9d55-31524249a851-kube-api-access-kwvlk\") pod \"b2c7ad15-ca5a-4042-9d55-31524249a851\" (UID: \"b2c7ad15-ca5a-4042-9d55-31524249a851\") " Dec 10 23:07:04 crc kubenswrapper[4791]: I1210 23:07:04.469992 4791 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/b2c7ad15-ca5a-4042-9d55-31524249a851-config\") pod \"b2c7ad15-ca5a-4042-9d55-31524249a851\" (UID: \"b2c7ad15-ca5a-4042-9d55-31524249a851\") " Dec 10 23:07:04 crc kubenswrapper[4791]: I1210 23:07:04.470830 4791 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/b2c7ad15-ca5a-4042-9d55-31524249a851-config" (OuterVolumeSpecName: "config") pod "b2c7ad15-ca5a-4042-9d55-31524249a851" (UID: "b2c7ad15-ca5a-4042-9d55-31524249a851"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 10 23:07:04 crc kubenswrapper[4791]: I1210 23:07:04.471284 4791 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/b2c7ad15-ca5a-4042-9d55-31524249a851-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "b2c7ad15-ca5a-4042-9d55-31524249a851" (UID: "b2c7ad15-ca5a-4042-9d55-31524249a851"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 10 23:07:04 crc kubenswrapper[4791]: I1210 23:07:04.474736 4791 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/b2c7ad15-ca5a-4042-9d55-31524249a851-kube-api-access-kwvlk" (OuterVolumeSpecName: "kube-api-access-kwvlk") pod "b2c7ad15-ca5a-4042-9d55-31524249a851" (UID: "b2c7ad15-ca5a-4042-9d55-31524249a851"). InnerVolumeSpecName "kube-api-access-kwvlk". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 10 23:07:04 crc kubenswrapper[4791]: I1210 23:07:04.484165 4791 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-86db49b7ff-2f6xm"] Dec 10 23:07:04 crc kubenswrapper[4791]: I1210 23:07:04.502059 4791 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-86db49b7ff-2f6xm" Dec 10 23:07:04 crc kubenswrapper[4791]: I1210 23:07:04.525509 4791 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-86db49b7ff-2f6xm"] Dec 10 23:07:04 crc kubenswrapper[4791]: I1210 23:07:04.528730 4791 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"ovsdbserver-sb" Dec 10 23:07:04 crc kubenswrapper[4791]: I1210 23:07:04.571768 4791 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/1f9954ba-068e-4a98-bcc1-1a6cd477b858-ovsdbserver-sb\") pod \"dnsmasq-dns-86db49b7ff-2f6xm\" (UID: \"1f9954ba-068e-4a98-bcc1-1a6cd477b858\") " pod="openstack/dnsmasq-dns-86db49b7ff-2f6xm" Dec 10 23:07:04 crc kubenswrapper[4791]: I1210 23:07:04.571818 4791 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/1f9954ba-068e-4a98-bcc1-1a6cd477b858-ovsdbserver-nb\") pod \"dnsmasq-dns-86db49b7ff-2f6xm\" (UID: \"1f9954ba-068e-4a98-bcc1-1a6cd477b858\") " pod="openstack/dnsmasq-dns-86db49b7ff-2f6xm" Dec 10 23:07:04 crc kubenswrapper[4791]: I1210 23:07:04.571850 4791 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/1f9954ba-068e-4a98-bcc1-1a6cd477b858-config\") pod \"dnsmasq-dns-86db49b7ff-2f6xm\" (UID: \"1f9954ba-068e-4a98-bcc1-1a6cd477b858\") " pod="openstack/dnsmasq-dns-86db49b7ff-2f6xm" Dec 10 23:07:04 crc kubenswrapper[4791]: I1210 23:07:04.571865 4791 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-xnhdv\" (UniqueName: \"kubernetes.io/projected/1f9954ba-068e-4a98-bcc1-1a6cd477b858-kube-api-access-xnhdv\") pod \"dnsmasq-dns-86db49b7ff-2f6xm\" (UID: \"1f9954ba-068e-4a98-bcc1-1a6cd477b858\") " pod="openstack/dnsmasq-dns-86db49b7ff-2f6xm" Dec 10 23:07:04 crc kubenswrapper[4791]: I1210 23:07:04.571893 4791 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/1f9954ba-068e-4a98-bcc1-1a6cd477b858-dns-svc\") pod \"dnsmasq-dns-86db49b7ff-2f6xm\" (UID: \"1f9954ba-068e-4a98-bcc1-1a6cd477b858\") " pod="openstack/dnsmasq-dns-86db49b7ff-2f6xm" Dec 10 23:07:04 crc kubenswrapper[4791]: I1210 23:07:04.572206 4791 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-kwvlk\" (UniqueName: \"kubernetes.io/projected/b2c7ad15-ca5a-4042-9d55-31524249a851-kube-api-access-kwvlk\") on node \"crc\" DevicePath \"\"" Dec 10 23:07:04 crc kubenswrapper[4791]: I1210 23:07:04.572254 4791 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/b2c7ad15-ca5a-4042-9d55-31524249a851-config\") on node \"crc\" DevicePath \"\"" Dec 10 23:07:04 crc kubenswrapper[4791]: I1210 23:07:04.572283 4791 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/b2c7ad15-ca5a-4042-9d55-31524249a851-dns-svc\") on node \"crc\" DevicePath \"\"" Dec 10 23:07:04 crc kubenswrapper[4791]: I1210 23:07:04.673868 4791 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/1f9954ba-068e-4a98-bcc1-1a6cd477b858-ovsdbserver-sb\") pod \"dnsmasq-dns-86db49b7ff-2f6xm\" (UID: \"1f9954ba-068e-4a98-bcc1-1a6cd477b858\") " pod="openstack/dnsmasq-dns-86db49b7ff-2f6xm" Dec 10 23:07:04 crc kubenswrapper[4791]: I1210 23:07:04.674539 4791 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/1f9954ba-068e-4a98-bcc1-1a6cd477b858-ovsdbserver-nb\") pod \"dnsmasq-dns-86db49b7ff-2f6xm\" (UID: \"1f9954ba-068e-4a98-bcc1-1a6cd477b858\") " pod="openstack/dnsmasq-dns-86db49b7ff-2f6xm" Dec 10 23:07:04 crc kubenswrapper[4791]: I1210 23:07:04.674598 4791 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/1f9954ba-068e-4a98-bcc1-1a6cd477b858-config\") pod \"dnsmasq-dns-86db49b7ff-2f6xm\" (UID: \"1f9954ba-068e-4a98-bcc1-1a6cd477b858\") " pod="openstack/dnsmasq-dns-86db49b7ff-2f6xm" Dec 10 23:07:04 crc kubenswrapper[4791]: I1210 23:07:04.674617 4791 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-xnhdv\" (UniqueName: \"kubernetes.io/projected/1f9954ba-068e-4a98-bcc1-1a6cd477b858-kube-api-access-xnhdv\") pod \"dnsmasq-dns-86db49b7ff-2f6xm\" (UID: \"1f9954ba-068e-4a98-bcc1-1a6cd477b858\") " pod="openstack/dnsmasq-dns-86db49b7ff-2f6xm" Dec 10 23:07:04 crc kubenswrapper[4791]: I1210 23:07:04.674655 4791 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/1f9954ba-068e-4a98-bcc1-1a6cd477b858-dns-svc\") pod \"dnsmasq-dns-86db49b7ff-2f6xm\" (UID: \"1f9954ba-068e-4a98-bcc1-1a6cd477b858\") " pod="openstack/dnsmasq-dns-86db49b7ff-2f6xm" Dec 10 23:07:04 crc kubenswrapper[4791]: I1210 23:07:04.675308 4791 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/1f9954ba-068e-4a98-bcc1-1a6cd477b858-ovsdbserver-nb\") pod \"dnsmasq-dns-86db49b7ff-2f6xm\" (UID: \"1f9954ba-068e-4a98-bcc1-1a6cd477b858\") " pod="openstack/dnsmasq-dns-86db49b7ff-2f6xm" Dec 10 23:07:04 crc kubenswrapper[4791]: I1210 23:07:04.675477 4791 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/1f9954ba-068e-4a98-bcc1-1a6cd477b858-dns-svc\") pod \"dnsmasq-dns-86db49b7ff-2f6xm\" (UID: \"1f9954ba-068e-4a98-bcc1-1a6cd477b858\") " pod="openstack/dnsmasq-dns-86db49b7ff-2f6xm" Dec 10 23:07:04 crc kubenswrapper[4791]: I1210 23:07:04.675926 4791 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/1f9954ba-068e-4a98-bcc1-1a6cd477b858-config\") pod \"dnsmasq-dns-86db49b7ff-2f6xm\" (UID: \"1f9954ba-068e-4a98-bcc1-1a6cd477b858\") " pod="openstack/dnsmasq-dns-86db49b7ff-2f6xm" Dec 10 23:07:04 crc kubenswrapper[4791]: I1210 23:07:04.676524 4791 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/1f9954ba-068e-4a98-bcc1-1a6cd477b858-ovsdbserver-sb\") pod \"dnsmasq-dns-86db49b7ff-2f6xm\" (UID: \"1f9954ba-068e-4a98-bcc1-1a6cd477b858\") " pod="openstack/dnsmasq-dns-86db49b7ff-2f6xm" Dec 10 23:07:04 crc kubenswrapper[4791]: I1210 23:07:04.712614 4791 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-xnhdv\" (UniqueName: \"kubernetes.io/projected/1f9954ba-068e-4a98-bcc1-1a6cd477b858-kube-api-access-xnhdv\") pod \"dnsmasq-dns-86db49b7ff-2f6xm\" (UID: \"1f9954ba-068e-4a98-bcc1-1a6cd477b858\") " pod="openstack/dnsmasq-dns-86db49b7ff-2f6xm" Dec 10 23:07:04 crc kubenswrapper[4791]: I1210 23:07:04.722935 4791 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-57d769cc4f-tnfbj" Dec 10 23:07:04 crc kubenswrapper[4791]: I1210 23:07:04.727264 4791 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-57d769cc4f-tnfbj" event={"ID":"b2c7ad15-ca5a-4042-9d55-31524249a851","Type":"ContainerDied","Data":"2a8408d791d9146497875b91274e2afd7e291d826983d0300b16368d46cdfe20"} Dec 10 23:07:04 crc kubenswrapper[4791]: I1210 23:07:04.800010 4791 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-57d769cc4f-tnfbj"] Dec 10 23:07:04 crc kubenswrapper[4791]: I1210 23:07:04.803256 4791 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/ovsdbserver-sb-0" Dec 10 23:07:04 crc kubenswrapper[4791]: I1210 23:07:04.809031 4791 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-57d769cc4f-tnfbj"] Dec 10 23:07:04 crc kubenswrapper[4791]: I1210 23:07:04.864034 4791 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-86db49b7ff-2f6xm" Dec 10 23:07:04 crc kubenswrapper[4791]: I1210 23:07:04.975824 4791 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ovn-northd-0"] Dec 10 23:07:04 crc kubenswrapper[4791]: I1210 23:07:04.981900 4791 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-northd-0" Dec 10 23:07:04 crc kubenswrapper[4791]: I1210 23:07:04.986846 4791 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"ovnnorthd-config" Dec 10 23:07:04 crc kubenswrapper[4791]: I1210 23:07:04.987159 4791 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ovnnorthd-ovnnorthd-dockercfg-wczx9" Dec 10 23:07:04 crc kubenswrapper[4791]: I1210 23:07:04.989400 4791 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"ovnnorthd-scripts" Dec 10 23:07:04 crc kubenswrapper[4791]: I1210 23:07:04.990576 4791 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-ovnnorthd-ovndbs" Dec 10 23:07:04 crc kubenswrapper[4791]: I1210 23:07:04.991058 4791 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-5ccc8479f9-8w9hb" Dec 10 23:07:05 crc kubenswrapper[4791]: I1210 23:07:05.008166 4791 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovn-northd-0"] Dec 10 23:07:05 crc kubenswrapper[4791]: W1210 23:07:05.086119 4791 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod60c7a68b_a75f_4cdb_ab6c_703ceccd2036.slice/crio-083b13f26dca27aaf95c41beb2de42ee6244cf0c7a879eb3ada4000cc7e12746 WatchSource:0}: Error finding container 083b13f26dca27aaf95c41beb2de42ee6244cf0c7a879eb3ada4000cc7e12746: Status 404 returned error can't find the container with id 083b13f26dca27aaf95c41beb2de42ee6244cf0c7a879eb3ada4000cc7e12746 Dec 10 23:07:05 crc kubenswrapper[4791]: I1210 23:07:05.086463 4791 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-7fd796d7df-7z8ld"] Dec 10 23:07:05 crc kubenswrapper[4791]: I1210 23:07:05.090726 4791 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/73f16e68-0124-4784-977f-72ea5d6d04bf-config\") pod \"73f16e68-0124-4784-977f-72ea5d6d04bf\" (UID: \"73f16e68-0124-4784-977f-72ea5d6d04bf\") " Dec 10 23:07:05 crc kubenswrapper[4791]: I1210 23:07:05.090794 4791 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/73f16e68-0124-4784-977f-72ea5d6d04bf-dns-svc\") pod \"73f16e68-0124-4784-977f-72ea5d6d04bf\" (UID: \"73f16e68-0124-4784-977f-72ea5d6d04bf\") " Dec 10 23:07:05 crc kubenswrapper[4791]: I1210 23:07:05.090858 4791 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-zn4rz\" (UniqueName: \"kubernetes.io/projected/73f16e68-0124-4784-977f-72ea5d6d04bf-kube-api-access-zn4rz\") pod \"73f16e68-0124-4784-977f-72ea5d6d04bf\" (UID: \"73f16e68-0124-4784-977f-72ea5d6d04bf\") " Dec 10 23:07:05 crc kubenswrapper[4791]: I1210 23:07:05.091192 4791 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/b73f3e7f-f699-460a-8d40-6abc63f13a1c-scripts\") pod \"ovn-northd-0\" (UID: \"b73f3e7f-f699-460a-8d40-6abc63f13a1c\") " pod="openstack/ovn-northd-0" Dec 10 23:07:05 crc kubenswrapper[4791]: I1210 23:07:05.091229 4791 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/b73f3e7f-f699-460a-8d40-6abc63f13a1c-config\") pod \"ovn-northd-0\" (UID: \"b73f3e7f-f699-460a-8d40-6abc63f13a1c\") " pod="openstack/ovn-northd-0" Dec 10 23:07:05 crc kubenswrapper[4791]: I1210 23:07:05.091308 4791 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovn-northd-tls-certs\" (UniqueName: \"kubernetes.io/secret/b73f3e7f-f699-460a-8d40-6abc63f13a1c-ovn-northd-tls-certs\") pod \"ovn-northd-0\" (UID: \"b73f3e7f-f699-460a-8d40-6abc63f13a1c\") " pod="openstack/ovn-northd-0" Dec 10 23:07:05 crc kubenswrapper[4791]: I1210 23:07:05.091359 4791 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/b73f3e7f-f699-460a-8d40-6abc63f13a1c-combined-ca-bundle\") pod \"ovn-northd-0\" (UID: \"b73f3e7f-f699-460a-8d40-6abc63f13a1c\") " pod="openstack/ovn-northd-0" Dec 10 23:07:05 crc kubenswrapper[4791]: I1210 23:07:05.091371 4791 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/73f16e68-0124-4784-977f-72ea5d6d04bf-config" (OuterVolumeSpecName: "config") pod "73f16e68-0124-4784-977f-72ea5d6d04bf" (UID: "73f16e68-0124-4784-977f-72ea5d6d04bf"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 10 23:07:05 crc kubenswrapper[4791]: I1210 23:07:05.091383 4791 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/73f16e68-0124-4784-977f-72ea5d6d04bf-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "73f16e68-0124-4784-977f-72ea5d6d04bf" (UID: "73f16e68-0124-4784-977f-72ea5d6d04bf"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 10 23:07:05 crc kubenswrapper[4791]: I1210 23:07:05.091480 4791 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics-certs-tls-certs\" (UniqueName: \"kubernetes.io/secret/b73f3e7f-f699-460a-8d40-6abc63f13a1c-metrics-certs-tls-certs\") pod \"ovn-northd-0\" (UID: \"b73f3e7f-f699-460a-8d40-6abc63f13a1c\") " pod="openstack/ovn-northd-0" Dec 10 23:07:05 crc kubenswrapper[4791]: I1210 23:07:05.091620 4791 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovn-rundir\" (UniqueName: \"kubernetes.io/empty-dir/b73f3e7f-f699-460a-8d40-6abc63f13a1c-ovn-rundir\") pod \"ovn-northd-0\" (UID: \"b73f3e7f-f699-460a-8d40-6abc63f13a1c\") " pod="openstack/ovn-northd-0" Dec 10 23:07:05 crc kubenswrapper[4791]: I1210 23:07:05.091640 4791 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-wvr2c\" (UniqueName: \"kubernetes.io/projected/b73f3e7f-f699-460a-8d40-6abc63f13a1c-kube-api-access-wvr2c\") pod \"ovn-northd-0\" (UID: \"b73f3e7f-f699-460a-8d40-6abc63f13a1c\") " pod="openstack/ovn-northd-0" Dec 10 23:07:05 crc kubenswrapper[4791]: I1210 23:07:05.091737 4791 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/73f16e68-0124-4784-977f-72ea5d6d04bf-config\") on node \"crc\" DevicePath \"\"" Dec 10 23:07:05 crc kubenswrapper[4791]: I1210 23:07:05.091749 4791 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/73f16e68-0124-4784-977f-72ea5d6d04bf-dns-svc\") on node \"crc\" DevicePath \"\"" Dec 10 23:07:05 crc kubenswrapper[4791]: I1210 23:07:05.096195 4791 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/73f16e68-0124-4784-977f-72ea5d6d04bf-kube-api-access-zn4rz" (OuterVolumeSpecName: "kube-api-access-zn4rz") pod "73f16e68-0124-4784-977f-72ea5d6d04bf" (UID: "73f16e68-0124-4784-977f-72ea5d6d04bf"). InnerVolumeSpecName "kube-api-access-zn4rz". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 10 23:07:05 crc kubenswrapper[4791]: I1210 23:07:05.115710 4791 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovn-controller-metrics-djrh2"] Dec 10 23:07:05 crc kubenswrapper[4791]: I1210 23:07:05.193611 4791 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/b73f3e7f-f699-460a-8d40-6abc63f13a1c-scripts\") pod \"ovn-northd-0\" (UID: \"b73f3e7f-f699-460a-8d40-6abc63f13a1c\") " pod="openstack/ovn-northd-0" Dec 10 23:07:05 crc kubenswrapper[4791]: I1210 23:07:05.193927 4791 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/b73f3e7f-f699-460a-8d40-6abc63f13a1c-config\") pod \"ovn-northd-0\" (UID: \"b73f3e7f-f699-460a-8d40-6abc63f13a1c\") " pod="openstack/ovn-northd-0" Dec 10 23:07:05 crc kubenswrapper[4791]: I1210 23:07:05.193978 4791 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovn-northd-tls-certs\" (UniqueName: \"kubernetes.io/secret/b73f3e7f-f699-460a-8d40-6abc63f13a1c-ovn-northd-tls-certs\") pod \"ovn-northd-0\" (UID: \"b73f3e7f-f699-460a-8d40-6abc63f13a1c\") " pod="openstack/ovn-northd-0" Dec 10 23:07:05 crc kubenswrapper[4791]: I1210 23:07:05.194000 4791 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/b73f3e7f-f699-460a-8d40-6abc63f13a1c-combined-ca-bundle\") pod \"ovn-northd-0\" (UID: \"b73f3e7f-f699-460a-8d40-6abc63f13a1c\") " pod="openstack/ovn-northd-0" Dec 10 23:07:05 crc kubenswrapper[4791]: I1210 23:07:05.194038 4791 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs-tls-certs\" (UniqueName: \"kubernetes.io/secret/b73f3e7f-f699-460a-8d40-6abc63f13a1c-metrics-certs-tls-certs\") pod \"ovn-northd-0\" (UID: \"b73f3e7f-f699-460a-8d40-6abc63f13a1c\") " pod="openstack/ovn-northd-0" Dec 10 23:07:05 crc kubenswrapper[4791]: I1210 23:07:05.194084 4791 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovn-rundir\" (UniqueName: \"kubernetes.io/empty-dir/b73f3e7f-f699-460a-8d40-6abc63f13a1c-ovn-rundir\") pod \"ovn-northd-0\" (UID: \"b73f3e7f-f699-460a-8d40-6abc63f13a1c\") " pod="openstack/ovn-northd-0" Dec 10 23:07:05 crc kubenswrapper[4791]: I1210 23:07:05.194106 4791 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-wvr2c\" (UniqueName: \"kubernetes.io/projected/b73f3e7f-f699-460a-8d40-6abc63f13a1c-kube-api-access-wvr2c\") pod \"ovn-northd-0\" (UID: \"b73f3e7f-f699-460a-8d40-6abc63f13a1c\") " pod="openstack/ovn-northd-0" Dec 10 23:07:05 crc kubenswrapper[4791]: I1210 23:07:05.194177 4791 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-zn4rz\" (UniqueName: \"kubernetes.io/projected/73f16e68-0124-4784-977f-72ea5d6d04bf-kube-api-access-zn4rz\") on node \"crc\" DevicePath \"\"" Dec 10 23:07:05 crc kubenswrapper[4791]: I1210 23:07:05.195005 4791 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovn-rundir\" (UniqueName: \"kubernetes.io/empty-dir/b73f3e7f-f699-460a-8d40-6abc63f13a1c-ovn-rundir\") pod \"ovn-northd-0\" (UID: \"b73f3e7f-f699-460a-8d40-6abc63f13a1c\") " pod="openstack/ovn-northd-0" Dec 10 23:07:05 crc kubenswrapper[4791]: I1210 23:07:05.195118 4791 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/b73f3e7f-f699-460a-8d40-6abc63f13a1c-config\") pod \"ovn-northd-0\" (UID: \"b73f3e7f-f699-460a-8d40-6abc63f13a1c\") " pod="openstack/ovn-northd-0" Dec 10 23:07:05 crc kubenswrapper[4791]: I1210 23:07:05.195575 4791 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/b73f3e7f-f699-460a-8d40-6abc63f13a1c-scripts\") pod \"ovn-northd-0\" (UID: \"b73f3e7f-f699-460a-8d40-6abc63f13a1c\") " pod="openstack/ovn-northd-0" Dec 10 23:07:05 crc kubenswrapper[4791]: I1210 23:07:05.197760 4791 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/b73f3e7f-f699-460a-8d40-6abc63f13a1c-combined-ca-bundle\") pod \"ovn-northd-0\" (UID: \"b73f3e7f-f699-460a-8d40-6abc63f13a1c\") " pod="openstack/ovn-northd-0" Dec 10 23:07:05 crc kubenswrapper[4791]: I1210 23:07:05.197768 4791 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovn-northd-tls-certs\" (UniqueName: \"kubernetes.io/secret/b73f3e7f-f699-460a-8d40-6abc63f13a1c-ovn-northd-tls-certs\") pod \"ovn-northd-0\" (UID: \"b73f3e7f-f699-460a-8d40-6abc63f13a1c\") " pod="openstack/ovn-northd-0" Dec 10 23:07:05 crc kubenswrapper[4791]: I1210 23:07:05.202191 4791 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-certs-tls-certs\" (UniqueName: \"kubernetes.io/secret/b73f3e7f-f699-460a-8d40-6abc63f13a1c-metrics-certs-tls-certs\") pod \"ovn-northd-0\" (UID: \"b73f3e7f-f699-460a-8d40-6abc63f13a1c\") " pod="openstack/ovn-northd-0" Dec 10 23:07:05 crc kubenswrapper[4791]: I1210 23:07:05.210470 4791 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-wvr2c\" (UniqueName: \"kubernetes.io/projected/b73f3e7f-f699-460a-8d40-6abc63f13a1c-kube-api-access-wvr2c\") pod \"ovn-northd-0\" (UID: \"b73f3e7f-f699-460a-8d40-6abc63f13a1c\") " pod="openstack/ovn-northd-0" Dec 10 23:07:05 crc kubenswrapper[4791]: I1210 23:07:05.301751 4791 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-northd-0" Dec 10 23:07:05 crc kubenswrapper[4791]: I1210 23:07:05.405222 4791 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-86db49b7ff-2f6xm"] Dec 10 23:07:05 crc kubenswrapper[4791]: I1210 23:07:05.712491 4791 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovn-northd-0"] Dec 10 23:07:05 crc kubenswrapper[4791]: W1210 23:07:05.712723 4791 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podb73f3e7f_f699_460a_8d40_6abc63f13a1c.slice/crio-ecb127106396f7a236ca1fb6d85895471dc8d07fa38a5b4c3f1bafca036bbfd1 WatchSource:0}: Error finding container ecb127106396f7a236ca1fb6d85895471dc8d07fa38a5b4c3f1bafca036bbfd1: Status 404 returned error can't find the container with id ecb127106396f7a236ca1fb6d85895471dc8d07fa38a5b4c3f1bafca036bbfd1 Dec 10 23:07:05 crc kubenswrapper[4791]: I1210 23:07:05.747484 4791 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-northd-0" event={"ID":"b73f3e7f-f699-460a-8d40-6abc63f13a1c","Type":"ContainerStarted","Data":"ecb127106396f7a236ca1fb6d85895471dc8d07fa38a5b4c3f1bafca036bbfd1"} Dec 10 23:07:05 crc kubenswrapper[4791]: I1210 23:07:05.748910 4791 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-5ccc8479f9-8w9hb" Dec 10 23:07:05 crc kubenswrapper[4791]: I1210 23:07:05.748912 4791 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-5ccc8479f9-8w9hb" event={"ID":"73f16e68-0124-4784-977f-72ea5d6d04bf","Type":"ContainerDied","Data":"1cf836f70d0946a02cb65df3a7c27da984eb3de1f80c9e4d62a08bc0b144969b"} Dec 10 23:07:05 crc kubenswrapper[4791]: I1210 23:07:05.750785 4791 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-7fd796d7df-7z8ld" event={"ID":"60c7a68b-a75f-4cdb-ab6c-703ceccd2036","Type":"ContainerStarted","Data":"083b13f26dca27aaf95c41beb2de42ee6244cf0c7a879eb3ada4000cc7e12746"} Dec 10 23:07:05 crc kubenswrapper[4791]: I1210 23:07:05.754374 4791 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-controller-metrics-djrh2" event={"ID":"53ceaa05-0398-4cf9-a787-874adab015aa","Type":"ContainerStarted","Data":"bee3c1d4bb0e36e0f61898d86c8df05d9e05695ba2dae3a77308408782d55e9c"} Dec 10 23:07:05 crc kubenswrapper[4791]: I1210 23:07:05.754480 4791 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-controller-metrics-djrh2" event={"ID":"53ceaa05-0398-4cf9-a787-874adab015aa","Type":"ContainerStarted","Data":"f9d9225d09061ba2ee3e64407a64deddf2a9afe6529aef46bb3879ba13e520b1"} Dec 10 23:07:05 crc kubenswrapper[4791]: I1210 23:07:05.756625 4791 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-86db49b7ff-2f6xm" event={"ID":"1f9954ba-068e-4a98-bcc1-1a6cd477b858","Type":"ContainerStarted","Data":"d6a6298c825337fa4c702d1d74fcd4b7e2db5e4a65bfdc78a828d7f758c425da"} Dec 10 23:07:05 crc kubenswrapper[4791]: I1210 23:07:05.775823 4791 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ovn-controller-metrics-djrh2" podStartSLOduration=1.775803357 podStartE2EDuration="1.775803357s" podCreationTimestamp="2025-12-10 23:07:04 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-10 23:07:05.768133625 +0000 UTC m=+1060.197751238" watchObservedRunningTime="2025-12-10 23:07:05.775803357 +0000 UTC m=+1060.205420970" Dec 10 23:07:05 crc kubenswrapper[4791]: I1210 23:07:05.836401 4791 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-5ccc8479f9-8w9hb"] Dec 10 23:07:05 crc kubenswrapper[4791]: I1210 23:07:05.843171 4791 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-5ccc8479f9-8w9hb"] Dec 10 23:07:05 crc kubenswrapper[4791]: I1210 23:07:05.896419 4791 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="73f16e68-0124-4784-977f-72ea5d6d04bf" path="/var/lib/kubelet/pods/73f16e68-0124-4784-977f-72ea5d6d04bf/volumes" Dec 10 23:07:05 crc kubenswrapper[4791]: I1210 23:07:05.897371 4791 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="b2c7ad15-ca5a-4042-9d55-31524249a851" path="/var/lib/kubelet/pods/b2c7ad15-ca5a-4042-9d55-31524249a851/volumes" Dec 10 23:07:06 crc kubenswrapper[4791]: I1210 23:07:06.767440 4791 generic.go:334] "Generic (PLEG): container finished" podID="60c7a68b-a75f-4cdb-ab6c-703ceccd2036" containerID="6348f9788adcd912dd9bf9a96922c01f22a40a8b931c9cec96ac56d758935f85" exitCode=0 Dec 10 23:07:06 crc kubenswrapper[4791]: I1210 23:07:06.767524 4791 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-7fd796d7df-7z8ld" event={"ID":"60c7a68b-a75f-4cdb-ab6c-703ceccd2036","Type":"ContainerDied","Data":"6348f9788adcd912dd9bf9a96922c01f22a40a8b931c9cec96ac56d758935f85"} Dec 10 23:07:06 crc kubenswrapper[4791]: I1210 23:07:06.769320 4791 generic.go:334] "Generic (PLEG): container finished" podID="1f9954ba-068e-4a98-bcc1-1a6cd477b858" containerID="d27b53cb65a001e1228aaff2c2f4c1105609b6ad3dc01e7b73887396f8779f14" exitCode=0 Dec 10 23:07:06 crc kubenswrapper[4791]: I1210 23:07:06.769400 4791 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-86db49b7ff-2f6xm" event={"ID":"1f9954ba-068e-4a98-bcc1-1a6cd477b858","Type":"ContainerDied","Data":"d27b53cb65a001e1228aaff2c2f4c1105609b6ad3dc01e7b73887396f8779f14"} Dec 10 23:07:06 crc kubenswrapper[4791]: I1210 23:07:06.771704 4791 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/openstack-galera-0" event={"ID":"aed2b986-c040-4191-864b-47f29c5c8537","Type":"ContainerStarted","Data":"541308ce28a0a022d8881af9f292d84d11f4505e1c54cde5b0cd371c4282559f"} Dec 10 23:07:07 crc kubenswrapper[4791]: I1210 23:07:07.059097 4791 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/memcached-0" Dec 10 23:07:08 crc kubenswrapper[4791]: I1210 23:07:08.810275 4791 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-server-0" event={"ID":"9ce8dbeb-dce8-47ac-98f1-fe92f74aa3de","Type":"ContainerStarted","Data":"d8a93d7e8bfe0416ef17df92a9875113cc702aca941f1f73df50121135822c94"} Dec 10 23:07:08 crc kubenswrapper[4791]: I1210 23:07:08.882896 4791 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-7fd796d7df-7z8ld"] Dec 10 23:07:08 crc kubenswrapper[4791]: I1210 23:07:08.953516 4791 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-698758b865-72nqc"] Dec 10 23:07:08 crc kubenswrapper[4791]: I1210 23:07:08.959847 4791 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-698758b865-72nqc" Dec 10 23:07:08 crc kubenswrapper[4791]: I1210 23:07:08.964050 4791 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-698758b865-72nqc"] Dec 10 23:07:09 crc kubenswrapper[4791]: I1210 23:07:09.057465 4791 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/7e284c1c-b650-4488-8c24-f9f5900b2326-dns-svc\") pod \"dnsmasq-dns-698758b865-72nqc\" (UID: \"7e284c1c-b650-4488-8c24-f9f5900b2326\") " pod="openstack/dnsmasq-dns-698758b865-72nqc" Dec 10 23:07:09 crc kubenswrapper[4791]: I1210 23:07:09.057536 4791 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-8jj9k\" (UniqueName: \"kubernetes.io/projected/7e284c1c-b650-4488-8c24-f9f5900b2326-kube-api-access-8jj9k\") pod \"dnsmasq-dns-698758b865-72nqc\" (UID: \"7e284c1c-b650-4488-8c24-f9f5900b2326\") " pod="openstack/dnsmasq-dns-698758b865-72nqc" Dec 10 23:07:09 crc kubenswrapper[4791]: I1210 23:07:09.057569 4791 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/7e284c1c-b650-4488-8c24-f9f5900b2326-config\") pod \"dnsmasq-dns-698758b865-72nqc\" (UID: \"7e284c1c-b650-4488-8c24-f9f5900b2326\") " pod="openstack/dnsmasq-dns-698758b865-72nqc" Dec 10 23:07:09 crc kubenswrapper[4791]: I1210 23:07:09.057598 4791 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/7e284c1c-b650-4488-8c24-f9f5900b2326-ovsdbserver-sb\") pod \"dnsmasq-dns-698758b865-72nqc\" (UID: \"7e284c1c-b650-4488-8c24-f9f5900b2326\") " pod="openstack/dnsmasq-dns-698758b865-72nqc" Dec 10 23:07:09 crc kubenswrapper[4791]: I1210 23:07:09.057623 4791 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/7e284c1c-b650-4488-8c24-f9f5900b2326-ovsdbserver-nb\") pod \"dnsmasq-dns-698758b865-72nqc\" (UID: \"7e284c1c-b650-4488-8c24-f9f5900b2326\") " pod="openstack/dnsmasq-dns-698758b865-72nqc" Dec 10 23:07:09 crc kubenswrapper[4791]: I1210 23:07:09.158589 4791 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/7e284c1c-b650-4488-8c24-f9f5900b2326-config\") pod \"dnsmasq-dns-698758b865-72nqc\" (UID: \"7e284c1c-b650-4488-8c24-f9f5900b2326\") " pod="openstack/dnsmasq-dns-698758b865-72nqc" Dec 10 23:07:09 crc kubenswrapper[4791]: I1210 23:07:09.158655 4791 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/7e284c1c-b650-4488-8c24-f9f5900b2326-ovsdbserver-sb\") pod \"dnsmasq-dns-698758b865-72nqc\" (UID: \"7e284c1c-b650-4488-8c24-f9f5900b2326\") " pod="openstack/dnsmasq-dns-698758b865-72nqc" Dec 10 23:07:09 crc kubenswrapper[4791]: I1210 23:07:09.158684 4791 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/7e284c1c-b650-4488-8c24-f9f5900b2326-ovsdbserver-nb\") pod \"dnsmasq-dns-698758b865-72nqc\" (UID: \"7e284c1c-b650-4488-8c24-f9f5900b2326\") " pod="openstack/dnsmasq-dns-698758b865-72nqc" Dec 10 23:07:09 crc kubenswrapper[4791]: I1210 23:07:09.158746 4791 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/7e284c1c-b650-4488-8c24-f9f5900b2326-dns-svc\") pod \"dnsmasq-dns-698758b865-72nqc\" (UID: \"7e284c1c-b650-4488-8c24-f9f5900b2326\") " pod="openstack/dnsmasq-dns-698758b865-72nqc" Dec 10 23:07:09 crc kubenswrapper[4791]: I1210 23:07:09.158794 4791 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-8jj9k\" (UniqueName: \"kubernetes.io/projected/7e284c1c-b650-4488-8c24-f9f5900b2326-kube-api-access-8jj9k\") pod \"dnsmasq-dns-698758b865-72nqc\" (UID: \"7e284c1c-b650-4488-8c24-f9f5900b2326\") " pod="openstack/dnsmasq-dns-698758b865-72nqc" Dec 10 23:07:09 crc kubenswrapper[4791]: I1210 23:07:09.160107 4791 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/7e284c1c-b650-4488-8c24-f9f5900b2326-ovsdbserver-sb\") pod \"dnsmasq-dns-698758b865-72nqc\" (UID: \"7e284c1c-b650-4488-8c24-f9f5900b2326\") " pod="openstack/dnsmasq-dns-698758b865-72nqc" Dec 10 23:07:09 crc kubenswrapper[4791]: I1210 23:07:09.160151 4791 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/7e284c1c-b650-4488-8c24-f9f5900b2326-ovsdbserver-nb\") pod \"dnsmasq-dns-698758b865-72nqc\" (UID: \"7e284c1c-b650-4488-8c24-f9f5900b2326\") " pod="openstack/dnsmasq-dns-698758b865-72nqc" Dec 10 23:07:09 crc kubenswrapper[4791]: I1210 23:07:09.160174 4791 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/7e284c1c-b650-4488-8c24-f9f5900b2326-dns-svc\") pod \"dnsmasq-dns-698758b865-72nqc\" (UID: \"7e284c1c-b650-4488-8c24-f9f5900b2326\") " pod="openstack/dnsmasq-dns-698758b865-72nqc" Dec 10 23:07:09 crc kubenswrapper[4791]: I1210 23:07:09.160195 4791 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/7e284c1c-b650-4488-8c24-f9f5900b2326-config\") pod \"dnsmasq-dns-698758b865-72nqc\" (UID: \"7e284c1c-b650-4488-8c24-f9f5900b2326\") " pod="openstack/dnsmasq-dns-698758b865-72nqc" Dec 10 23:07:09 crc kubenswrapper[4791]: I1210 23:07:09.180990 4791 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-8jj9k\" (UniqueName: \"kubernetes.io/projected/7e284c1c-b650-4488-8c24-f9f5900b2326-kube-api-access-8jj9k\") pod \"dnsmasq-dns-698758b865-72nqc\" (UID: \"7e284c1c-b650-4488-8c24-f9f5900b2326\") " pod="openstack/dnsmasq-dns-698758b865-72nqc" Dec 10 23:07:09 crc kubenswrapper[4791]: I1210 23:07:09.389094 4791 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-698758b865-72nqc" Dec 10 23:07:09 crc kubenswrapper[4791]: I1210 23:07:09.823068 4791 generic.go:334] "Generic (PLEG): container finished" podID="f04fa829-3ebe-445d-a4e5-f7592ac682f3" containerID="4ac34c10ef97f846f1cbd4e7bd46b995220d082b77e94ede2ad536e90f578da4" exitCode=0 Dec 10 23:07:09 crc kubenswrapper[4791]: I1210 23:07:09.823305 4791 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/openstack-cell1-galera-0" event={"ID":"f04fa829-3ebe-445d-a4e5-f7592ac682f3","Type":"ContainerDied","Data":"4ac34c10ef97f846f1cbd4e7bd46b995220d082b77e94ede2ad536e90f578da4"} Dec 10 23:07:09 crc kubenswrapper[4791]: I1210 23:07:09.828105 4791 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-northd-0" event={"ID":"b73f3e7f-f699-460a-8d40-6abc63f13a1c","Type":"ContainerStarted","Data":"0d7d01c190013a1d37f861a948b0e5f2a5d840720a399f2e5504446bb385e856"} Dec 10 23:07:09 crc kubenswrapper[4791]: I1210 23:07:09.828166 4791 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-northd-0" event={"ID":"b73f3e7f-f699-460a-8d40-6abc63f13a1c","Type":"ContainerStarted","Data":"f7bd57658db32667c6e82ee34411aa8bcfcd20478d7825815d6d5254d0a05f8d"} Dec 10 23:07:09 crc kubenswrapper[4791]: I1210 23:07:09.828256 4791 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/ovn-northd-0" Dec 10 23:07:09 crc kubenswrapper[4791]: I1210 23:07:09.832020 4791 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-7fd796d7df-7z8ld" event={"ID":"60c7a68b-a75f-4cdb-ab6c-703ceccd2036","Type":"ContainerStarted","Data":"64dcea04ea039f947ecfb6c9f3899bdc0f8f0205e31bd717c63fe893428fc2c1"} Dec 10 23:07:09 crc kubenswrapper[4791]: I1210 23:07:09.832153 4791 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/dnsmasq-dns-7fd796d7df-7z8ld" podUID="60c7a68b-a75f-4cdb-ab6c-703ceccd2036" containerName="dnsmasq-dns" containerID="cri-o://64dcea04ea039f947ecfb6c9f3899bdc0f8f0205e31bd717c63fe893428fc2c1" gracePeriod=10 Dec 10 23:07:09 crc kubenswrapper[4791]: I1210 23:07:09.832227 4791 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/dnsmasq-dns-7fd796d7df-7z8ld" Dec 10 23:07:09 crc kubenswrapper[4791]: I1210 23:07:09.836864 4791 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-86db49b7ff-2f6xm" event={"ID":"1f9954ba-068e-4a98-bcc1-1a6cd477b858","Type":"ContainerStarted","Data":"22433271d016baa99aaef1cf8a84c3ff2124b51b98f1cbbb4bf1e537938483c8"} Dec 10 23:07:09 crc kubenswrapper[4791]: I1210 23:07:09.837630 4791 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/dnsmasq-dns-86db49b7ff-2f6xm" Dec 10 23:07:09 crc kubenswrapper[4791]: I1210 23:07:09.852563 4791 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-698758b865-72nqc"] Dec 10 23:07:09 crc kubenswrapper[4791]: I1210 23:07:09.882242 4791 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/dnsmasq-dns-7fd796d7df-7z8ld" podStartSLOduration=6.401495064 podStartE2EDuration="6.882222098s" podCreationTimestamp="2025-12-10 23:07:03 +0000 UTC" firstStartedPulling="2025-12-10 23:07:05.088889556 +0000 UTC m=+1059.518507169" lastFinishedPulling="2025-12-10 23:07:05.56961659 +0000 UTC m=+1059.999234203" observedRunningTime="2025-12-10 23:07:09.881543418 +0000 UTC m=+1064.311161051" watchObservedRunningTime="2025-12-10 23:07:09.882222098 +0000 UTC m=+1064.311839731" Dec 10 23:07:09 crc kubenswrapper[4791]: I1210 23:07:09.907062 4791 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ovn-northd-0" podStartSLOduration=4.705919572 podStartE2EDuration="5.907048256s" podCreationTimestamp="2025-12-10 23:07:04 +0000 UTC" firstStartedPulling="2025-12-10 23:07:05.715113971 +0000 UTC m=+1060.144731584" lastFinishedPulling="2025-12-10 23:07:06.916242655 +0000 UTC m=+1061.345860268" observedRunningTime="2025-12-10 23:07:09.902075622 +0000 UTC m=+1064.331693235" watchObservedRunningTime="2025-12-10 23:07:09.907048256 +0000 UTC m=+1064.336665869" Dec 10 23:07:09 crc kubenswrapper[4791]: I1210 23:07:09.962323 4791 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/dnsmasq-dns-86db49b7ff-2f6xm" podStartSLOduration=5.51570879 podStartE2EDuration="5.962307596s" podCreationTimestamp="2025-12-10 23:07:04 +0000 UTC" firstStartedPulling="2025-12-10 23:07:05.416728124 +0000 UTC m=+1059.846345737" lastFinishedPulling="2025-12-10 23:07:05.86332693 +0000 UTC m=+1060.292944543" observedRunningTime="2025-12-10 23:07:09.942513013 +0000 UTC m=+1064.372130626" watchObservedRunningTime="2025-12-10 23:07:09.962307596 +0000 UTC m=+1064.391925209" Dec 10 23:07:10 crc kubenswrapper[4791]: I1210 23:07:10.079881 4791 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/swift-storage-0"] Dec 10 23:07:10 crc kubenswrapper[4791]: I1210 23:07:10.087892 4791 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/swift-storage-0"] Dec 10 23:07:10 crc kubenswrapper[4791]: I1210 23:07:10.088020 4791 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/swift-storage-0" Dec 10 23:07:10 crc kubenswrapper[4791]: I1210 23:07:10.090358 4791 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"swift-swift-dockercfg-dmjmp" Dec 10 23:07:10 crc kubenswrapper[4791]: I1210 23:07:10.090393 4791 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"swift-conf" Dec 10 23:07:10 crc kubenswrapper[4791]: I1210 23:07:10.090434 4791 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"swift-ring-files" Dec 10 23:07:10 crc kubenswrapper[4791]: I1210 23:07:10.091214 4791 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"swift-storage-config-data" Dec 10 23:07:10 crc kubenswrapper[4791]: I1210 23:07:10.277598 4791 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"lock\" (UniqueName: \"kubernetes.io/empty-dir/69a6c950-8a3b-4a7c-b284-ebd20157eb20-lock\") pod \"swift-storage-0\" (UID: \"69a6c950-8a3b-4a7c-b284-ebd20157eb20\") " pod="openstack/swift-storage-0" Dec 10 23:07:10 crc kubenswrapper[4791]: I1210 23:07:10.277674 4791 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cache\" (UniqueName: \"kubernetes.io/empty-dir/69a6c950-8a3b-4a7c-b284-ebd20157eb20-cache\") pod \"swift-storage-0\" (UID: \"69a6c950-8a3b-4a7c-b284-ebd20157eb20\") " pod="openstack/swift-storage-0" Dec 10 23:07:10 crc kubenswrapper[4791]: I1210 23:07:10.277700 4791 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/projected/69a6c950-8a3b-4a7c-b284-ebd20157eb20-etc-swift\") pod \"swift-storage-0\" (UID: \"69a6c950-8a3b-4a7c-b284-ebd20157eb20\") " pod="openstack/swift-storage-0" Dec 10 23:07:10 crc kubenswrapper[4791]: I1210 23:07:10.277723 4791 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-8fgq8\" (UniqueName: \"kubernetes.io/projected/69a6c950-8a3b-4a7c-b284-ebd20157eb20-kube-api-access-8fgq8\") pod \"swift-storage-0\" (UID: \"69a6c950-8a3b-4a7c-b284-ebd20157eb20\") " pod="openstack/swift-storage-0" Dec 10 23:07:10 crc kubenswrapper[4791]: I1210 23:07:10.277745 4791 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage01-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage01-crc\") pod \"swift-storage-0\" (UID: \"69a6c950-8a3b-4a7c-b284-ebd20157eb20\") " pod="openstack/swift-storage-0" Dec 10 23:07:10 crc kubenswrapper[4791]: I1210 23:07:10.350253 4791 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-7fd796d7df-7z8ld" Dec 10 23:07:10 crc kubenswrapper[4791]: I1210 23:07:10.379591 4791 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cache\" (UniqueName: \"kubernetes.io/empty-dir/69a6c950-8a3b-4a7c-b284-ebd20157eb20-cache\") pod \"swift-storage-0\" (UID: \"69a6c950-8a3b-4a7c-b284-ebd20157eb20\") " pod="openstack/swift-storage-0" Dec 10 23:07:10 crc kubenswrapper[4791]: I1210 23:07:10.379645 4791 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/projected/69a6c950-8a3b-4a7c-b284-ebd20157eb20-etc-swift\") pod \"swift-storage-0\" (UID: \"69a6c950-8a3b-4a7c-b284-ebd20157eb20\") " pod="openstack/swift-storage-0" Dec 10 23:07:10 crc kubenswrapper[4791]: I1210 23:07:10.379723 4791 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-8fgq8\" (UniqueName: \"kubernetes.io/projected/69a6c950-8a3b-4a7c-b284-ebd20157eb20-kube-api-access-8fgq8\") pod \"swift-storage-0\" (UID: \"69a6c950-8a3b-4a7c-b284-ebd20157eb20\") " pod="openstack/swift-storage-0" Dec 10 23:07:10 crc kubenswrapper[4791]: I1210 23:07:10.379758 4791 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage01-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage01-crc\") pod \"swift-storage-0\" (UID: \"69a6c950-8a3b-4a7c-b284-ebd20157eb20\") " pod="openstack/swift-storage-0" Dec 10 23:07:10 crc kubenswrapper[4791]: I1210 23:07:10.379848 4791 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"lock\" (UniqueName: \"kubernetes.io/empty-dir/69a6c950-8a3b-4a7c-b284-ebd20157eb20-lock\") pod \"swift-storage-0\" (UID: \"69a6c950-8a3b-4a7c-b284-ebd20157eb20\") " pod="openstack/swift-storage-0" Dec 10 23:07:10 crc kubenswrapper[4791]: E1210 23:07:10.380356 4791 projected.go:288] Couldn't get configMap openstack/swift-ring-files: configmap "swift-ring-files" not found Dec 10 23:07:10 crc kubenswrapper[4791]: E1210 23:07:10.380395 4791 projected.go:194] Error preparing data for projected volume etc-swift for pod openstack/swift-storage-0: configmap "swift-ring-files" not found Dec 10 23:07:10 crc kubenswrapper[4791]: E1210 23:07:10.380460 4791 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/69a6c950-8a3b-4a7c-b284-ebd20157eb20-etc-swift podName:69a6c950-8a3b-4a7c-b284-ebd20157eb20 nodeName:}" failed. No retries permitted until 2025-12-10 23:07:10.880437087 +0000 UTC m=+1065.310054710 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "etc-swift" (UniqueName: "kubernetes.io/projected/69a6c950-8a3b-4a7c-b284-ebd20157eb20-etc-swift") pod "swift-storage-0" (UID: "69a6c950-8a3b-4a7c-b284-ebd20157eb20") : configmap "swift-ring-files" not found Dec 10 23:07:10 crc kubenswrapper[4791]: I1210 23:07:10.380717 4791 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cache\" (UniqueName: \"kubernetes.io/empty-dir/69a6c950-8a3b-4a7c-b284-ebd20157eb20-cache\") pod \"swift-storage-0\" (UID: \"69a6c950-8a3b-4a7c-b284-ebd20157eb20\") " pod="openstack/swift-storage-0" Dec 10 23:07:10 crc kubenswrapper[4791]: I1210 23:07:10.380723 4791 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"lock\" (UniqueName: \"kubernetes.io/empty-dir/69a6c950-8a3b-4a7c-b284-ebd20157eb20-lock\") pod \"swift-storage-0\" (UID: \"69a6c950-8a3b-4a7c-b284-ebd20157eb20\") " pod="openstack/swift-storage-0" Dec 10 23:07:10 crc kubenswrapper[4791]: I1210 23:07:10.380990 4791 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage01-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage01-crc\") pod \"swift-storage-0\" (UID: \"69a6c950-8a3b-4a7c-b284-ebd20157eb20\") device mount path \"/mnt/openstack/pv01\"" pod="openstack/swift-storage-0" Dec 10 23:07:10 crc kubenswrapper[4791]: I1210 23:07:10.383966 4791 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/swift-ring-rebalance-pzkk7"] Dec 10 23:07:10 crc kubenswrapper[4791]: E1210 23:07:10.384508 4791 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="60c7a68b-a75f-4cdb-ab6c-703ceccd2036" containerName="init" Dec 10 23:07:10 crc kubenswrapper[4791]: I1210 23:07:10.384533 4791 state_mem.go:107] "Deleted CPUSet assignment" podUID="60c7a68b-a75f-4cdb-ab6c-703ceccd2036" containerName="init" Dec 10 23:07:10 crc kubenswrapper[4791]: E1210 23:07:10.384575 4791 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="60c7a68b-a75f-4cdb-ab6c-703ceccd2036" containerName="dnsmasq-dns" Dec 10 23:07:10 crc kubenswrapper[4791]: I1210 23:07:10.384582 4791 state_mem.go:107] "Deleted CPUSet assignment" podUID="60c7a68b-a75f-4cdb-ab6c-703ceccd2036" containerName="dnsmasq-dns" Dec 10 23:07:10 crc kubenswrapper[4791]: I1210 23:07:10.384758 4791 memory_manager.go:354] "RemoveStaleState removing state" podUID="60c7a68b-a75f-4cdb-ab6c-703ceccd2036" containerName="dnsmasq-dns" Dec 10 23:07:10 crc kubenswrapper[4791]: I1210 23:07:10.385559 4791 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/swift-ring-rebalance-pzkk7" Dec 10 23:07:10 crc kubenswrapper[4791]: I1210 23:07:10.391654 4791 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"swift-ring-config-data" Dec 10 23:07:10 crc kubenswrapper[4791]: I1210 23:07:10.391759 4791 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"swift-ring-scripts" Dec 10 23:07:10 crc kubenswrapper[4791]: I1210 23:07:10.392037 4791 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"swift-proxy-config-data" Dec 10 23:07:10 crc kubenswrapper[4791]: I1210 23:07:10.415224 4791 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage01-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage01-crc\") pod \"swift-storage-0\" (UID: \"69a6c950-8a3b-4a7c-b284-ebd20157eb20\") " pod="openstack/swift-storage-0" Dec 10 23:07:10 crc kubenswrapper[4791]: I1210 23:07:10.440442 4791 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/swift-ring-rebalance-pzkk7"] Dec 10 23:07:10 crc kubenswrapper[4791]: E1210 23:07:10.441166 4791 pod_workers.go:1301] "Error syncing pod, skipping" err="unmounted volumes=[combined-ca-bundle dispersionconf etc-swift kube-api-access-s5w5w ring-data-devices scripts swiftconf], unattached volumes=[], failed to process volumes=[]: context canceled" pod="openstack/swift-ring-rebalance-pzkk7" podUID="7f35b2d5-de89-4b42-85fe-8a7eb461c2ce" Dec 10 23:07:10 crc kubenswrapper[4791]: I1210 23:07:10.455321 4791 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-8fgq8\" (UniqueName: \"kubernetes.io/projected/69a6c950-8a3b-4a7c-b284-ebd20157eb20-kube-api-access-8fgq8\") pod \"swift-storage-0\" (UID: \"69a6c950-8a3b-4a7c-b284-ebd20157eb20\") " pod="openstack/swift-storage-0" Dec 10 23:07:10 crc kubenswrapper[4791]: I1210 23:07:10.469730 4791 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/swift-ring-rebalance-bqplg"] Dec 10 23:07:10 crc kubenswrapper[4791]: I1210 23:07:10.471023 4791 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/swift-ring-rebalance-bqplg" Dec 10 23:07:10 crc kubenswrapper[4791]: I1210 23:07:10.481690 4791 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/60c7a68b-a75f-4cdb-ab6c-703ceccd2036-dns-svc\") pod \"60c7a68b-a75f-4cdb-ab6c-703ceccd2036\" (UID: \"60c7a68b-a75f-4cdb-ab6c-703ceccd2036\") " Dec 10 23:07:10 crc kubenswrapper[4791]: I1210 23:07:10.483520 4791 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/60c7a68b-a75f-4cdb-ab6c-703ceccd2036-config\") pod \"60c7a68b-a75f-4cdb-ab6c-703ceccd2036\" (UID: \"60c7a68b-a75f-4cdb-ab6c-703ceccd2036\") " Dec 10 23:07:10 crc kubenswrapper[4791]: I1210 23:07:10.483616 4791 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-r7979\" (UniqueName: \"kubernetes.io/projected/60c7a68b-a75f-4cdb-ab6c-703ceccd2036-kube-api-access-r7979\") pod \"60c7a68b-a75f-4cdb-ab6c-703ceccd2036\" (UID: \"60c7a68b-a75f-4cdb-ab6c-703ceccd2036\") " Dec 10 23:07:10 crc kubenswrapper[4791]: I1210 23:07:10.483643 4791 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/60c7a68b-a75f-4cdb-ab6c-703ceccd2036-ovsdbserver-nb\") pod \"60c7a68b-a75f-4cdb-ab6c-703ceccd2036\" (UID: \"60c7a68b-a75f-4cdb-ab6c-703ceccd2036\") " Dec 10 23:07:10 crc kubenswrapper[4791]: I1210 23:07:10.483892 4791 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-s5w5w\" (UniqueName: \"kubernetes.io/projected/7f35b2d5-de89-4b42-85fe-8a7eb461c2ce-kube-api-access-s5w5w\") pod \"swift-ring-rebalance-pzkk7\" (UID: \"7f35b2d5-de89-4b42-85fe-8a7eb461c2ce\") " pod="openstack/swift-ring-rebalance-pzkk7" Dec 10 23:07:10 crc kubenswrapper[4791]: I1210 23:07:10.483964 4791 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dispersionconf\" (UniqueName: \"kubernetes.io/secret/7f35b2d5-de89-4b42-85fe-8a7eb461c2ce-dispersionconf\") pod \"swift-ring-rebalance-pzkk7\" (UID: \"7f35b2d5-de89-4b42-85fe-8a7eb461c2ce\") " pod="openstack/swift-ring-rebalance-pzkk7" Dec 10 23:07:10 crc kubenswrapper[4791]: I1210 23:07:10.483991 4791 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/empty-dir/7f35b2d5-de89-4b42-85fe-8a7eb461c2ce-etc-swift\") pod \"swift-ring-rebalance-pzkk7\" (UID: \"7f35b2d5-de89-4b42-85fe-8a7eb461c2ce\") " pod="openstack/swift-ring-rebalance-pzkk7" Dec 10 23:07:10 crc kubenswrapper[4791]: I1210 23:07:10.484012 4791 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ring-data-devices\" (UniqueName: \"kubernetes.io/configmap/7f35b2d5-de89-4b42-85fe-8a7eb461c2ce-ring-data-devices\") pod \"swift-ring-rebalance-pzkk7\" (UID: \"7f35b2d5-de89-4b42-85fe-8a7eb461c2ce\") " pod="openstack/swift-ring-rebalance-pzkk7" Dec 10 23:07:10 crc kubenswrapper[4791]: I1210 23:07:10.484097 4791 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"swiftconf\" (UniqueName: \"kubernetes.io/secret/7f35b2d5-de89-4b42-85fe-8a7eb461c2ce-swiftconf\") pod \"swift-ring-rebalance-pzkk7\" (UID: \"7f35b2d5-de89-4b42-85fe-8a7eb461c2ce\") " pod="openstack/swift-ring-rebalance-pzkk7" Dec 10 23:07:10 crc kubenswrapper[4791]: I1210 23:07:10.484116 4791 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/7f35b2d5-de89-4b42-85fe-8a7eb461c2ce-scripts\") pod \"swift-ring-rebalance-pzkk7\" (UID: \"7f35b2d5-de89-4b42-85fe-8a7eb461c2ce\") " pod="openstack/swift-ring-rebalance-pzkk7" Dec 10 23:07:10 crc kubenswrapper[4791]: I1210 23:07:10.484185 4791 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/7f35b2d5-de89-4b42-85fe-8a7eb461c2ce-combined-ca-bundle\") pod \"swift-ring-rebalance-pzkk7\" (UID: \"7f35b2d5-de89-4b42-85fe-8a7eb461c2ce\") " pod="openstack/swift-ring-rebalance-pzkk7" Dec 10 23:07:10 crc kubenswrapper[4791]: I1210 23:07:10.512991 4791 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/60c7a68b-a75f-4cdb-ab6c-703ceccd2036-kube-api-access-r7979" (OuterVolumeSpecName: "kube-api-access-r7979") pod "60c7a68b-a75f-4cdb-ab6c-703ceccd2036" (UID: "60c7a68b-a75f-4cdb-ab6c-703ceccd2036"). InnerVolumeSpecName "kube-api-access-r7979". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 10 23:07:10 crc kubenswrapper[4791]: I1210 23:07:10.536115 4791 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/swift-ring-rebalance-bqplg"] Dec 10 23:07:10 crc kubenswrapper[4791]: I1210 23:07:10.551853 4791 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/swift-ring-rebalance-pzkk7"] Dec 10 23:07:10 crc kubenswrapper[4791]: I1210 23:07:10.552765 4791 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/60c7a68b-a75f-4cdb-ab6c-703ceccd2036-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "60c7a68b-a75f-4cdb-ab6c-703ceccd2036" (UID: "60c7a68b-a75f-4cdb-ab6c-703ceccd2036"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 10 23:07:10 crc kubenswrapper[4791]: I1210 23:07:10.568882 4791 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/60c7a68b-a75f-4cdb-ab6c-703ceccd2036-ovsdbserver-nb" (OuterVolumeSpecName: "ovsdbserver-nb") pod "60c7a68b-a75f-4cdb-ab6c-703ceccd2036" (UID: "60c7a68b-a75f-4cdb-ab6c-703ceccd2036"). InnerVolumeSpecName "ovsdbserver-nb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 10 23:07:10 crc kubenswrapper[4791]: I1210 23:07:10.574297 4791 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/60c7a68b-a75f-4cdb-ab6c-703ceccd2036-config" (OuterVolumeSpecName: "config") pod "60c7a68b-a75f-4cdb-ab6c-703ceccd2036" (UID: "60c7a68b-a75f-4cdb-ab6c-703ceccd2036"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 10 23:07:10 crc kubenswrapper[4791]: I1210 23:07:10.585753 4791 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dispersionconf\" (UniqueName: \"kubernetes.io/secret/16bd9b65-844e-4537-8d5b-c0a73666c2b2-dispersionconf\") pod \"swift-ring-rebalance-bqplg\" (UID: \"16bd9b65-844e-4537-8d5b-c0a73666c2b2\") " pod="openstack/swift-ring-rebalance-bqplg" Dec 10 23:07:10 crc kubenswrapper[4791]: I1210 23:07:10.585826 4791 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/16bd9b65-844e-4537-8d5b-c0a73666c2b2-combined-ca-bundle\") pod \"swift-ring-rebalance-bqplg\" (UID: \"16bd9b65-844e-4537-8d5b-c0a73666c2b2\") " pod="openstack/swift-ring-rebalance-bqplg" Dec 10 23:07:10 crc kubenswrapper[4791]: I1210 23:07:10.585872 4791 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/7f35b2d5-de89-4b42-85fe-8a7eb461c2ce-combined-ca-bundle\") pod \"swift-ring-rebalance-pzkk7\" (UID: \"7f35b2d5-de89-4b42-85fe-8a7eb461c2ce\") " pod="openstack/swift-ring-rebalance-pzkk7" Dec 10 23:07:10 crc kubenswrapper[4791]: I1210 23:07:10.586105 4791 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"swiftconf\" (UniqueName: \"kubernetes.io/secret/16bd9b65-844e-4537-8d5b-c0a73666c2b2-swiftconf\") pod \"swift-ring-rebalance-bqplg\" (UID: \"16bd9b65-844e-4537-8d5b-c0a73666c2b2\") " pod="openstack/swift-ring-rebalance-bqplg" Dec 10 23:07:10 crc kubenswrapper[4791]: I1210 23:07:10.586222 4791 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-vnfxn\" (UniqueName: \"kubernetes.io/projected/16bd9b65-844e-4537-8d5b-c0a73666c2b2-kube-api-access-vnfxn\") pod \"swift-ring-rebalance-bqplg\" (UID: \"16bd9b65-844e-4537-8d5b-c0a73666c2b2\") " pod="openstack/swift-ring-rebalance-bqplg" Dec 10 23:07:10 crc kubenswrapper[4791]: I1210 23:07:10.586264 4791 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/empty-dir/16bd9b65-844e-4537-8d5b-c0a73666c2b2-etc-swift\") pod \"swift-ring-rebalance-bqplg\" (UID: \"16bd9b65-844e-4537-8d5b-c0a73666c2b2\") " pod="openstack/swift-ring-rebalance-bqplg" Dec 10 23:07:10 crc kubenswrapper[4791]: I1210 23:07:10.586519 4791 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/16bd9b65-844e-4537-8d5b-c0a73666c2b2-scripts\") pod \"swift-ring-rebalance-bqplg\" (UID: \"16bd9b65-844e-4537-8d5b-c0a73666c2b2\") " pod="openstack/swift-ring-rebalance-bqplg" Dec 10 23:07:10 crc kubenswrapper[4791]: I1210 23:07:10.586570 4791 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-s5w5w\" (UniqueName: \"kubernetes.io/projected/7f35b2d5-de89-4b42-85fe-8a7eb461c2ce-kube-api-access-s5w5w\") pod \"swift-ring-rebalance-pzkk7\" (UID: \"7f35b2d5-de89-4b42-85fe-8a7eb461c2ce\") " pod="openstack/swift-ring-rebalance-pzkk7" Dec 10 23:07:10 crc kubenswrapper[4791]: I1210 23:07:10.586661 4791 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dispersionconf\" (UniqueName: \"kubernetes.io/secret/7f35b2d5-de89-4b42-85fe-8a7eb461c2ce-dispersionconf\") pod \"swift-ring-rebalance-pzkk7\" (UID: \"7f35b2d5-de89-4b42-85fe-8a7eb461c2ce\") " pod="openstack/swift-ring-rebalance-pzkk7" Dec 10 23:07:10 crc kubenswrapper[4791]: I1210 23:07:10.586694 4791 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/empty-dir/7f35b2d5-de89-4b42-85fe-8a7eb461c2ce-etc-swift\") pod \"swift-ring-rebalance-pzkk7\" (UID: \"7f35b2d5-de89-4b42-85fe-8a7eb461c2ce\") " pod="openstack/swift-ring-rebalance-pzkk7" Dec 10 23:07:10 crc kubenswrapper[4791]: I1210 23:07:10.586732 4791 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ring-data-devices\" (UniqueName: \"kubernetes.io/configmap/7f35b2d5-de89-4b42-85fe-8a7eb461c2ce-ring-data-devices\") pod \"swift-ring-rebalance-pzkk7\" (UID: \"7f35b2d5-de89-4b42-85fe-8a7eb461c2ce\") " pod="openstack/swift-ring-rebalance-pzkk7" Dec 10 23:07:10 crc kubenswrapper[4791]: I1210 23:07:10.586837 4791 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ring-data-devices\" (UniqueName: \"kubernetes.io/configmap/16bd9b65-844e-4537-8d5b-c0a73666c2b2-ring-data-devices\") pod \"swift-ring-rebalance-bqplg\" (UID: \"16bd9b65-844e-4537-8d5b-c0a73666c2b2\") " pod="openstack/swift-ring-rebalance-bqplg" Dec 10 23:07:10 crc kubenswrapper[4791]: I1210 23:07:10.586927 4791 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"swiftconf\" (UniqueName: \"kubernetes.io/secret/7f35b2d5-de89-4b42-85fe-8a7eb461c2ce-swiftconf\") pod \"swift-ring-rebalance-pzkk7\" (UID: \"7f35b2d5-de89-4b42-85fe-8a7eb461c2ce\") " pod="openstack/swift-ring-rebalance-pzkk7" Dec 10 23:07:10 crc kubenswrapper[4791]: I1210 23:07:10.586956 4791 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/7f35b2d5-de89-4b42-85fe-8a7eb461c2ce-scripts\") pod \"swift-ring-rebalance-pzkk7\" (UID: \"7f35b2d5-de89-4b42-85fe-8a7eb461c2ce\") " pod="openstack/swift-ring-rebalance-pzkk7" Dec 10 23:07:10 crc kubenswrapper[4791]: I1210 23:07:10.587061 4791 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-r7979\" (UniqueName: \"kubernetes.io/projected/60c7a68b-a75f-4cdb-ab6c-703ceccd2036-kube-api-access-r7979\") on node \"crc\" DevicePath \"\"" Dec 10 23:07:10 crc kubenswrapper[4791]: I1210 23:07:10.587083 4791 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/60c7a68b-a75f-4cdb-ab6c-703ceccd2036-ovsdbserver-nb\") on node \"crc\" DevicePath \"\"" Dec 10 23:07:10 crc kubenswrapper[4791]: I1210 23:07:10.587096 4791 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/60c7a68b-a75f-4cdb-ab6c-703ceccd2036-dns-svc\") on node \"crc\" DevicePath \"\"" Dec 10 23:07:10 crc kubenswrapper[4791]: I1210 23:07:10.587110 4791 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/60c7a68b-a75f-4cdb-ab6c-703ceccd2036-config\") on node \"crc\" DevicePath \"\"" Dec 10 23:07:10 crc kubenswrapper[4791]: I1210 23:07:10.588002 4791 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/7f35b2d5-de89-4b42-85fe-8a7eb461c2ce-scripts\") pod \"swift-ring-rebalance-pzkk7\" (UID: \"7f35b2d5-de89-4b42-85fe-8a7eb461c2ce\") " pod="openstack/swift-ring-rebalance-pzkk7" Dec 10 23:07:10 crc kubenswrapper[4791]: I1210 23:07:10.588933 4791 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-swift\" (UniqueName: \"kubernetes.io/empty-dir/7f35b2d5-de89-4b42-85fe-8a7eb461c2ce-etc-swift\") pod \"swift-ring-rebalance-pzkk7\" (UID: \"7f35b2d5-de89-4b42-85fe-8a7eb461c2ce\") " pod="openstack/swift-ring-rebalance-pzkk7" Dec 10 23:07:10 crc kubenswrapper[4791]: I1210 23:07:10.589547 4791 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ring-data-devices\" (UniqueName: \"kubernetes.io/configmap/7f35b2d5-de89-4b42-85fe-8a7eb461c2ce-ring-data-devices\") pod \"swift-ring-rebalance-pzkk7\" (UID: \"7f35b2d5-de89-4b42-85fe-8a7eb461c2ce\") " pod="openstack/swift-ring-rebalance-pzkk7" Dec 10 23:07:10 crc kubenswrapper[4791]: I1210 23:07:10.590458 4791 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"swiftconf\" (UniqueName: \"kubernetes.io/secret/7f35b2d5-de89-4b42-85fe-8a7eb461c2ce-swiftconf\") pod \"swift-ring-rebalance-pzkk7\" (UID: \"7f35b2d5-de89-4b42-85fe-8a7eb461c2ce\") " pod="openstack/swift-ring-rebalance-pzkk7" Dec 10 23:07:10 crc kubenswrapper[4791]: I1210 23:07:10.591978 4791 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dispersionconf\" (UniqueName: \"kubernetes.io/secret/7f35b2d5-de89-4b42-85fe-8a7eb461c2ce-dispersionconf\") pod \"swift-ring-rebalance-pzkk7\" (UID: \"7f35b2d5-de89-4b42-85fe-8a7eb461c2ce\") " pod="openstack/swift-ring-rebalance-pzkk7" Dec 10 23:07:10 crc kubenswrapper[4791]: I1210 23:07:10.592505 4791 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/7f35b2d5-de89-4b42-85fe-8a7eb461c2ce-combined-ca-bundle\") pod \"swift-ring-rebalance-pzkk7\" (UID: \"7f35b2d5-de89-4b42-85fe-8a7eb461c2ce\") " pod="openstack/swift-ring-rebalance-pzkk7" Dec 10 23:07:10 crc kubenswrapper[4791]: I1210 23:07:10.605816 4791 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-s5w5w\" (UniqueName: \"kubernetes.io/projected/7f35b2d5-de89-4b42-85fe-8a7eb461c2ce-kube-api-access-s5w5w\") pod \"swift-ring-rebalance-pzkk7\" (UID: \"7f35b2d5-de89-4b42-85fe-8a7eb461c2ce\") " pod="openstack/swift-ring-rebalance-pzkk7" Dec 10 23:07:10 crc kubenswrapper[4791]: I1210 23:07:10.688590 4791 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dispersionconf\" (UniqueName: \"kubernetes.io/secret/16bd9b65-844e-4537-8d5b-c0a73666c2b2-dispersionconf\") pod \"swift-ring-rebalance-bqplg\" (UID: \"16bd9b65-844e-4537-8d5b-c0a73666c2b2\") " pod="openstack/swift-ring-rebalance-bqplg" Dec 10 23:07:10 crc kubenswrapper[4791]: I1210 23:07:10.688672 4791 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/16bd9b65-844e-4537-8d5b-c0a73666c2b2-combined-ca-bundle\") pod \"swift-ring-rebalance-bqplg\" (UID: \"16bd9b65-844e-4537-8d5b-c0a73666c2b2\") " pod="openstack/swift-ring-rebalance-bqplg" Dec 10 23:07:10 crc kubenswrapper[4791]: I1210 23:07:10.688715 4791 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"swiftconf\" (UniqueName: \"kubernetes.io/secret/16bd9b65-844e-4537-8d5b-c0a73666c2b2-swiftconf\") pod \"swift-ring-rebalance-bqplg\" (UID: \"16bd9b65-844e-4537-8d5b-c0a73666c2b2\") " pod="openstack/swift-ring-rebalance-bqplg" Dec 10 23:07:10 crc kubenswrapper[4791]: I1210 23:07:10.688745 4791 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-vnfxn\" (UniqueName: \"kubernetes.io/projected/16bd9b65-844e-4537-8d5b-c0a73666c2b2-kube-api-access-vnfxn\") pod \"swift-ring-rebalance-bqplg\" (UID: \"16bd9b65-844e-4537-8d5b-c0a73666c2b2\") " pod="openstack/swift-ring-rebalance-bqplg" Dec 10 23:07:10 crc kubenswrapper[4791]: I1210 23:07:10.688766 4791 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/empty-dir/16bd9b65-844e-4537-8d5b-c0a73666c2b2-etc-swift\") pod \"swift-ring-rebalance-bqplg\" (UID: \"16bd9b65-844e-4537-8d5b-c0a73666c2b2\") " pod="openstack/swift-ring-rebalance-bqplg" Dec 10 23:07:10 crc kubenswrapper[4791]: I1210 23:07:10.689406 4791 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/16bd9b65-844e-4537-8d5b-c0a73666c2b2-scripts\") pod \"swift-ring-rebalance-bqplg\" (UID: \"16bd9b65-844e-4537-8d5b-c0a73666c2b2\") " pod="openstack/swift-ring-rebalance-bqplg" Dec 10 23:07:10 crc kubenswrapper[4791]: I1210 23:07:10.689799 4791 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-swift\" (UniqueName: \"kubernetes.io/empty-dir/16bd9b65-844e-4537-8d5b-c0a73666c2b2-etc-swift\") pod \"swift-ring-rebalance-bqplg\" (UID: \"16bd9b65-844e-4537-8d5b-c0a73666c2b2\") " pod="openstack/swift-ring-rebalance-bqplg" Dec 10 23:07:10 crc kubenswrapper[4791]: I1210 23:07:10.691403 4791 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ring-data-devices\" (UniqueName: \"kubernetes.io/configmap/16bd9b65-844e-4537-8d5b-c0a73666c2b2-ring-data-devices\") pod \"swift-ring-rebalance-bqplg\" (UID: \"16bd9b65-844e-4537-8d5b-c0a73666c2b2\") " pod="openstack/swift-ring-rebalance-bqplg" Dec 10 23:07:10 crc kubenswrapper[4791]: I1210 23:07:10.692537 4791 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dispersionconf\" (UniqueName: \"kubernetes.io/secret/16bd9b65-844e-4537-8d5b-c0a73666c2b2-dispersionconf\") pod \"swift-ring-rebalance-bqplg\" (UID: \"16bd9b65-844e-4537-8d5b-c0a73666c2b2\") " pod="openstack/swift-ring-rebalance-bqplg" Dec 10 23:07:10 crc kubenswrapper[4791]: I1210 23:07:10.692659 4791 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"swiftconf\" (UniqueName: \"kubernetes.io/secret/16bd9b65-844e-4537-8d5b-c0a73666c2b2-swiftconf\") pod \"swift-ring-rebalance-bqplg\" (UID: \"16bd9b65-844e-4537-8d5b-c0a73666c2b2\") " pod="openstack/swift-ring-rebalance-bqplg" Dec 10 23:07:10 crc kubenswrapper[4791]: I1210 23:07:10.693809 4791 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/16bd9b65-844e-4537-8d5b-c0a73666c2b2-combined-ca-bundle\") pod \"swift-ring-rebalance-bqplg\" (UID: \"16bd9b65-844e-4537-8d5b-c0a73666c2b2\") " pod="openstack/swift-ring-rebalance-bqplg" Dec 10 23:07:10 crc kubenswrapper[4791]: I1210 23:07:10.695178 4791 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ring-data-devices\" (UniqueName: \"kubernetes.io/configmap/16bd9b65-844e-4537-8d5b-c0a73666c2b2-ring-data-devices\") pod \"swift-ring-rebalance-bqplg\" (UID: \"16bd9b65-844e-4537-8d5b-c0a73666c2b2\") " pod="openstack/swift-ring-rebalance-bqplg" Dec 10 23:07:10 crc kubenswrapper[4791]: I1210 23:07:10.695754 4791 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/16bd9b65-844e-4537-8d5b-c0a73666c2b2-scripts\") pod \"swift-ring-rebalance-bqplg\" (UID: \"16bd9b65-844e-4537-8d5b-c0a73666c2b2\") " pod="openstack/swift-ring-rebalance-bqplg" Dec 10 23:07:10 crc kubenswrapper[4791]: I1210 23:07:10.710811 4791 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-vnfxn\" (UniqueName: \"kubernetes.io/projected/16bd9b65-844e-4537-8d5b-c0a73666c2b2-kube-api-access-vnfxn\") pod \"swift-ring-rebalance-bqplg\" (UID: \"16bd9b65-844e-4537-8d5b-c0a73666c2b2\") " pod="openstack/swift-ring-rebalance-bqplg" Dec 10 23:07:10 crc kubenswrapper[4791]: I1210 23:07:10.831078 4791 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/swift-ring-rebalance-bqplg" Dec 10 23:07:10 crc kubenswrapper[4791]: I1210 23:07:10.846744 4791 generic.go:334] "Generic (PLEG): container finished" podID="60c7a68b-a75f-4cdb-ab6c-703ceccd2036" containerID="64dcea04ea039f947ecfb6c9f3899bdc0f8f0205e31bd717c63fe893428fc2c1" exitCode=0 Dec 10 23:07:10 crc kubenswrapper[4791]: I1210 23:07:10.846781 4791 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-7fd796d7df-7z8ld" Dec 10 23:07:10 crc kubenswrapper[4791]: I1210 23:07:10.846811 4791 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-7fd796d7df-7z8ld" event={"ID":"60c7a68b-a75f-4cdb-ab6c-703ceccd2036","Type":"ContainerDied","Data":"64dcea04ea039f947ecfb6c9f3899bdc0f8f0205e31bd717c63fe893428fc2c1"} Dec 10 23:07:10 crc kubenswrapper[4791]: I1210 23:07:10.846836 4791 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-7fd796d7df-7z8ld" event={"ID":"60c7a68b-a75f-4cdb-ab6c-703ceccd2036","Type":"ContainerDied","Data":"083b13f26dca27aaf95c41beb2de42ee6244cf0c7a879eb3ada4000cc7e12746"} Dec 10 23:07:10 crc kubenswrapper[4791]: I1210 23:07:10.846852 4791 scope.go:117] "RemoveContainer" containerID="64dcea04ea039f947ecfb6c9f3899bdc0f8f0205e31bd717c63fe893428fc2c1" Dec 10 23:07:10 crc kubenswrapper[4791]: I1210 23:07:10.851207 4791 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/openstack-cell1-galera-0" event={"ID":"f04fa829-3ebe-445d-a4e5-f7592ac682f3","Type":"ContainerStarted","Data":"beec18e9158625aadb166cb9a0b8a93e29ea941b9c14e13cf90ac6213c02e695"} Dec 10 23:07:10 crc kubenswrapper[4791]: I1210 23:07:10.857797 4791 generic.go:334] "Generic (PLEG): container finished" podID="aed2b986-c040-4191-864b-47f29c5c8537" containerID="541308ce28a0a022d8881af9f292d84d11f4505e1c54cde5b0cd371c4282559f" exitCode=0 Dec 10 23:07:10 crc kubenswrapper[4791]: I1210 23:07:10.857899 4791 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/openstack-galera-0" event={"ID":"aed2b986-c040-4191-864b-47f29c5c8537","Type":"ContainerDied","Data":"541308ce28a0a022d8881af9f292d84d11f4505e1c54cde5b0cd371c4282559f"} Dec 10 23:07:10 crc kubenswrapper[4791]: I1210 23:07:10.862696 4791 generic.go:334] "Generic (PLEG): container finished" podID="7e284c1c-b650-4488-8c24-f9f5900b2326" containerID="ae9f763a9fb2de89ba33dc56f9ef91103a82f777e1d9bcd04f3adef9f108e092" exitCode=0 Dec 10 23:07:10 crc kubenswrapper[4791]: I1210 23:07:10.862797 4791 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-698758b865-72nqc" event={"ID":"7e284c1c-b650-4488-8c24-f9f5900b2326","Type":"ContainerDied","Data":"ae9f763a9fb2de89ba33dc56f9ef91103a82f777e1d9bcd04f3adef9f108e092"} Dec 10 23:07:10 crc kubenswrapper[4791]: I1210 23:07:10.862847 4791 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-698758b865-72nqc" event={"ID":"7e284c1c-b650-4488-8c24-f9f5900b2326","Type":"ContainerStarted","Data":"8b882073e14aa2aa82987269ad3d80ced15404f389b70490a115782671386def"} Dec 10 23:07:10 crc kubenswrapper[4791]: I1210 23:07:10.863331 4791 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/swift-ring-rebalance-pzkk7" Dec 10 23:07:10 crc kubenswrapper[4791]: I1210 23:07:10.875075 4791 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/openstack-cell1-galera-0" podStartSLOduration=10.305018418 podStartE2EDuration="45.875053873s" podCreationTimestamp="2025-12-10 23:06:25 +0000 UTC" firstStartedPulling="2025-12-10 23:06:27.765993529 +0000 UTC m=+1022.195611142" lastFinishedPulling="2025-12-10 23:07:03.336028974 +0000 UTC m=+1057.765646597" observedRunningTime="2025-12-10 23:07:10.872933092 +0000 UTC m=+1065.302550715" watchObservedRunningTime="2025-12-10 23:07:10.875053873 +0000 UTC m=+1065.304671486" Dec 10 23:07:10 crc kubenswrapper[4791]: I1210 23:07:10.888476 4791 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/swift-ring-rebalance-pzkk7" Dec 10 23:07:10 crc kubenswrapper[4791]: I1210 23:07:10.896393 4791 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/projected/69a6c950-8a3b-4a7c-b284-ebd20157eb20-etc-swift\") pod \"swift-storage-0\" (UID: \"69a6c950-8a3b-4a7c-b284-ebd20157eb20\") " pod="openstack/swift-storage-0" Dec 10 23:07:10 crc kubenswrapper[4791]: E1210 23:07:10.896599 4791 projected.go:288] Couldn't get configMap openstack/swift-ring-files: configmap "swift-ring-files" not found Dec 10 23:07:10 crc kubenswrapper[4791]: E1210 23:07:10.896626 4791 projected.go:194] Error preparing data for projected volume etc-swift for pod openstack/swift-storage-0: configmap "swift-ring-files" not found Dec 10 23:07:10 crc kubenswrapper[4791]: E1210 23:07:10.896698 4791 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/69a6c950-8a3b-4a7c-b284-ebd20157eb20-etc-swift podName:69a6c950-8a3b-4a7c-b284-ebd20157eb20 nodeName:}" failed. No retries permitted until 2025-12-10 23:07:11.896680899 +0000 UTC m=+1066.326298512 (durationBeforeRetry 1s). Error: MountVolume.SetUp failed for volume "etc-swift" (UniqueName: "kubernetes.io/projected/69a6c950-8a3b-4a7c-b284-ebd20157eb20-etc-swift") pod "swift-storage-0" (UID: "69a6c950-8a3b-4a7c-b284-ebd20157eb20") : configmap "swift-ring-files" not found Dec 10 23:07:10 crc kubenswrapper[4791]: I1210 23:07:10.945839 4791 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-7fd796d7df-7z8ld"] Dec 10 23:07:10 crc kubenswrapper[4791]: I1210 23:07:10.950063 4791 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-7fd796d7df-7z8ld"] Dec 10 23:07:10 crc kubenswrapper[4791]: I1210 23:07:10.997219 4791 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"swiftconf\" (UniqueName: \"kubernetes.io/secret/7f35b2d5-de89-4b42-85fe-8a7eb461c2ce-swiftconf\") pod \"7f35b2d5-de89-4b42-85fe-8a7eb461c2ce\" (UID: \"7f35b2d5-de89-4b42-85fe-8a7eb461c2ce\") " Dec 10 23:07:10 crc kubenswrapper[4791]: I1210 23:07:10.997276 4791 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-s5w5w\" (UniqueName: \"kubernetes.io/projected/7f35b2d5-de89-4b42-85fe-8a7eb461c2ce-kube-api-access-s5w5w\") pod \"7f35b2d5-de89-4b42-85fe-8a7eb461c2ce\" (UID: \"7f35b2d5-de89-4b42-85fe-8a7eb461c2ce\") " Dec 10 23:07:10 crc kubenswrapper[4791]: I1210 23:07:10.997364 4791 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ring-data-devices\" (UniqueName: \"kubernetes.io/configmap/7f35b2d5-de89-4b42-85fe-8a7eb461c2ce-ring-data-devices\") pod \"7f35b2d5-de89-4b42-85fe-8a7eb461c2ce\" (UID: \"7f35b2d5-de89-4b42-85fe-8a7eb461c2ce\") " Dec 10 23:07:10 crc kubenswrapper[4791]: I1210 23:07:10.997420 4791 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/7f35b2d5-de89-4b42-85fe-8a7eb461c2ce-combined-ca-bundle\") pod \"7f35b2d5-de89-4b42-85fe-8a7eb461c2ce\" (UID: \"7f35b2d5-de89-4b42-85fe-8a7eb461c2ce\") " Dec 10 23:07:10 crc kubenswrapper[4791]: I1210 23:07:10.997476 4791 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dispersionconf\" (UniqueName: \"kubernetes.io/secret/7f35b2d5-de89-4b42-85fe-8a7eb461c2ce-dispersionconf\") pod \"7f35b2d5-de89-4b42-85fe-8a7eb461c2ce\" (UID: \"7f35b2d5-de89-4b42-85fe-8a7eb461c2ce\") " Dec 10 23:07:10 crc kubenswrapper[4791]: I1210 23:07:10.997543 4791 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/empty-dir/7f35b2d5-de89-4b42-85fe-8a7eb461c2ce-etc-swift\") pod \"7f35b2d5-de89-4b42-85fe-8a7eb461c2ce\" (UID: \"7f35b2d5-de89-4b42-85fe-8a7eb461c2ce\") " Dec 10 23:07:10 crc kubenswrapper[4791]: I1210 23:07:10.997600 4791 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/7f35b2d5-de89-4b42-85fe-8a7eb461c2ce-scripts\") pod \"7f35b2d5-de89-4b42-85fe-8a7eb461c2ce\" (UID: \"7f35b2d5-de89-4b42-85fe-8a7eb461c2ce\") " Dec 10 23:07:10 crc kubenswrapper[4791]: I1210 23:07:10.998656 4791 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/7f35b2d5-de89-4b42-85fe-8a7eb461c2ce-ring-data-devices" (OuterVolumeSpecName: "ring-data-devices") pod "7f35b2d5-de89-4b42-85fe-8a7eb461c2ce" (UID: "7f35b2d5-de89-4b42-85fe-8a7eb461c2ce"). InnerVolumeSpecName "ring-data-devices". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 10 23:07:10 crc kubenswrapper[4791]: I1210 23:07:10.998783 4791 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/7f35b2d5-de89-4b42-85fe-8a7eb461c2ce-scripts" (OuterVolumeSpecName: "scripts") pod "7f35b2d5-de89-4b42-85fe-8a7eb461c2ce" (UID: "7f35b2d5-de89-4b42-85fe-8a7eb461c2ce"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 10 23:07:10 crc kubenswrapper[4791]: I1210 23:07:10.999056 4791 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/7f35b2d5-de89-4b42-85fe-8a7eb461c2ce-etc-swift" (OuterVolumeSpecName: "etc-swift") pod "7f35b2d5-de89-4b42-85fe-8a7eb461c2ce" (UID: "7f35b2d5-de89-4b42-85fe-8a7eb461c2ce"). InnerVolumeSpecName "etc-swift". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 10 23:07:11 crc kubenswrapper[4791]: I1210 23:07:11.001196 4791 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/7f35b2d5-de89-4b42-85fe-8a7eb461c2ce-dispersionconf" (OuterVolumeSpecName: "dispersionconf") pod "7f35b2d5-de89-4b42-85fe-8a7eb461c2ce" (UID: "7f35b2d5-de89-4b42-85fe-8a7eb461c2ce"). InnerVolumeSpecName "dispersionconf". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 10 23:07:11 crc kubenswrapper[4791]: I1210 23:07:11.001231 4791 reconciler_common.go:293] "Volume detached for volume \"etc-swift\" (UniqueName: \"kubernetes.io/empty-dir/7f35b2d5-de89-4b42-85fe-8a7eb461c2ce-etc-swift\") on node \"crc\" DevicePath \"\"" Dec 10 23:07:11 crc kubenswrapper[4791]: I1210 23:07:11.001249 4791 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/7f35b2d5-de89-4b42-85fe-8a7eb461c2ce-scripts\") on node \"crc\" DevicePath \"\"" Dec 10 23:07:11 crc kubenswrapper[4791]: I1210 23:07:11.001260 4791 reconciler_common.go:293] "Volume detached for volume \"ring-data-devices\" (UniqueName: \"kubernetes.io/configmap/7f35b2d5-de89-4b42-85fe-8a7eb461c2ce-ring-data-devices\") on node \"crc\" DevicePath \"\"" Dec 10 23:07:11 crc kubenswrapper[4791]: I1210 23:07:11.002146 4791 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/7f35b2d5-de89-4b42-85fe-8a7eb461c2ce-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "7f35b2d5-de89-4b42-85fe-8a7eb461c2ce" (UID: "7f35b2d5-de89-4b42-85fe-8a7eb461c2ce"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 10 23:07:11 crc kubenswrapper[4791]: I1210 23:07:11.002926 4791 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/7f35b2d5-de89-4b42-85fe-8a7eb461c2ce-swiftconf" (OuterVolumeSpecName: "swiftconf") pod "7f35b2d5-de89-4b42-85fe-8a7eb461c2ce" (UID: "7f35b2d5-de89-4b42-85fe-8a7eb461c2ce"). InnerVolumeSpecName "swiftconf". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 10 23:07:11 crc kubenswrapper[4791]: I1210 23:07:11.011185 4791 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/7f35b2d5-de89-4b42-85fe-8a7eb461c2ce-kube-api-access-s5w5w" (OuterVolumeSpecName: "kube-api-access-s5w5w") pod "7f35b2d5-de89-4b42-85fe-8a7eb461c2ce" (UID: "7f35b2d5-de89-4b42-85fe-8a7eb461c2ce"). InnerVolumeSpecName "kube-api-access-s5w5w". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 10 23:07:11 crc kubenswrapper[4791]: I1210 23:07:11.102373 4791 reconciler_common.go:293] "Volume detached for volume \"swiftconf\" (UniqueName: \"kubernetes.io/secret/7f35b2d5-de89-4b42-85fe-8a7eb461c2ce-swiftconf\") on node \"crc\" DevicePath \"\"" Dec 10 23:07:11 crc kubenswrapper[4791]: I1210 23:07:11.102742 4791 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-s5w5w\" (UniqueName: \"kubernetes.io/projected/7f35b2d5-de89-4b42-85fe-8a7eb461c2ce-kube-api-access-s5w5w\") on node \"crc\" DevicePath \"\"" Dec 10 23:07:11 crc kubenswrapper[4791]: I1210 23:07:11.102761 4791 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/7f35b2d5-de89-4b42-85fe-8a7eb461c2ce-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 10 23:07:11 crc kubenswrapper[4791]: I1210 23:07:11.102773 4791 reconciler_common.go:293] "Volume detached for volume \"dispersionconf\" (UniqueName: \"kubernetes.io/secret/7f35b2d5-de89-4b42-85fe-8a7eb461c2ce-dispersionconf\") on node \"crc\" DevicePath \"\"" Dec 10 23:07:11 crc kubenswrapper[4791]: I1210 23:07:11.489081 4791 scope.go:117] "RemoveContainer" containerID="6348f9788adcd912dd9bf9a96922c01f22a40a8b931c9cec96ac56d758935f85" Dec 10 23:07:11 crc kubenswrapper[4791]: I1210 23:07:11.528792 4791 scope.go:117] "RemoveContainer" containerID="64dcea04ea039f947ecfb6c9f3899bdc0f8f0205e31bd717c63fe893428fc2c1" Dec 10 23:07:11 crc kubenswrapper[4791]: E1210 23:07:11.529266 4791 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"64dcea04ea039f947ecfb6c9f3899bdc0f8f0205e31bd717c63fe893428fc2c1\": container with ID starting with 64dcea04ea039f947ecfb6c9f3899bdc0f8f0205e31bd717c63fe893428fc2c1 not found: ID does not exist" containerID="64dcea04ea039f947ecfb6c9f3899bdc0f8f0205e31bd717c63fe893428fc2c1" Dec 10 23:07:11 crc kubenswrapper[4791]: I1210 23:07:11.529306 4791 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"64dcea04ea039f947ecfb6c9f3899bdc0f8f0205e31bd717c63fe893428fc2c1"} err="failed to get container status \"64dcea04ea039f947ecfb6c9f3899bdc0f8f0205e31bd717c63fe893428fc2c1\": rpc error: code = NotFound desc = could not find container \"64dcea04ea039f947ecfb6c9f3899bdc0f8f0205e31bd717c63fe893428fc2c1\": container with ID starting with 64dcea04ea039f947ecfb6c9f3899bdc0f8f0205e31bd717c63fe893428fc2c1 not found: ID does not exist" Dec 10 23:07:11 crc kubenswrapper[4791]: I1210 23:07:11.529452 4791 scope.go:117] "RemoveContainer" containerID="6348f9788adcd912dd9bf9a96922c01f22a40a8b931c9cec96ac56d758935f85" Dec 10 23:07:11 crc kubenswrapper[4791]: E1210 23:07:11.529884 4791 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"6348f9788adcd912dd9bf9a96922c01f22a40a8b931c9cec96ac56d758935f85\": container with ID starting with 6348f9788adcd912dd9bf9a96922c01f22a40a8b931c9cec96ac56d758935f85 not found: ID does not exist" containerID="6348f9788adcd912dd9bf9a96922c01f22a40a8b931c9cec96ac56d758935f85" Dec 10 23:07:11 crc kubenswrapper[4791]: I1210 23:07:11.529914 4791 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"6348f9788adcd912dd9bf9a96922c01f22a40a8b931c9cec96ac56d758935f85"} err="failed to get container status \"6348f9788adcd912dd9bf9a96922c01f22a40a8b931c9cec96ac56d758935f85\": rpc error: code = NotFound desc = could not find container \"6348f9788adcd912dd9bf9a96922c01f22a40a8b931c9cec96ac56d758935f85\": container with ID starting with 6348f9788adcd912dd9bf9a96922c01f22a40a8b931c9cec96ac56d758935f85 not found: ID does not exist" Dec 10 23:07:11 crc kubenswrapper[4791]: I1210 23:07:11.871907 4791 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/openstack-galera-0" event={"ID":"aed2b986-c040-4191-864b-47f29c5c8537","Type":"ContainerStarted","Data":"514c80d5f583dfc376cbc0d14897609af074d4a55ec22d60e4484c3d2f6f6808"} Dec 10 23:07:11 crc kubenswrapper[4791]: I1210 23:07:11.876495 4791 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-698758b865-72nqc" event={"ID":"7e284c1c-b650-4488-8c24-f9f5900b2326","Type":"ContainerStarted","Data":"064f69cbe089bbf9a8cab03b9f340ca2bb059356287fac23429c1823da83654d"} Dec 10 23:07:11 crc kubenswrapper[4791]: I1210 23:07:11.876842 4791 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/dnsmasq-dns-698758b865-72nqc" Dec 10 23:07:11 crc kubenswrapper[4791]: I1210 23:07:11.879826 4791 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/kube-state-metrics-0" event={"ID":"247efe24-6d59-4c93-ab04-6e249e1005a8","Type":"ContainerStarted","Data":"98e8f3bcbc5efc1d138b8c561352d547944c2c4e43d105c7bd3827406ac310dd"} Dec 10 23:07:11 crc kubenswrapper[4791]: I1210 23:07:11.880214 4791 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/kube-state-metrics-0" Dec 10 23:07:11 crc kubenswrapper[4791]: I1210 23:07:11.880263 4791 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/swift-ring-rebalance-pzkk7" Dec 10 23:07:11 crc kubenswrapper[4791]: I1210 23:07:11.904371 4791 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="60c7a68b-a75f-4cdb-ab6c-703ceccd2036" path="/var/lib/kubelet/pods/60c7a68b-a75f-4cdb-ab6c-703ceccd2036/volumes" Dec 10 23:07:11 crc kubenswrapper[4791]: I1210 23:07:11.915250 4791 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/projected/69a6c950-8a3b-4a7c-b284-ebd20157eb20-etc-swift\") pod \"swift-storage-0\" (UID: \"69a6c950-8a3b-4a7c-b284-ebd20157eb20\") " pod="openstack/swift-storage-0" Dec 10 23:07:11 crc kubenswrapper[4791]: E1210 23:07:11.915433 4791 projected.go:288] Couldn't get configMap openstack/swift-ring-files: configmap "swift-ring-files" not found Dec 10 23:07:11 crc kubenswrapper[4791]: E1210 23:07:11.915462 4791 projected.go:194] Error preparing data for projected volume etc-swift for pod openstack/swift-storage-0: configmap "swift-ring-files" not found Dec 10 23:07:11 crc kubenswrapper[4791]: E1210 23:07:11.915519 4791 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/69a6c950-8a3b-4a7c-b284-ebd20157eb20-etc-swift podName:69a6c950-8a3b-4a7c-b284-ebd20157eb20 nodeName:}" failed. No retries permitted until 2025-12-10 23:07:13.915500555 +0000 UTC m=+1068.345118168 (durationBeforeRetry 2s). Error: MountVolume.SetUp failed for volume "etc-swift" (UniqueName: "kubernetes.io/projected/69a6c950-8a3b-4a7c-b284-ebd20157eb20-etc-swift") pod "swift-storage-0" (UID: "69a6c950-8a3b-4a7c-b284-ebd20157eb20") : configmap "swift-ring-files" not found Dec 10 23:07:11 crc kubenswrapper[4791]: I1210 23:07:11.915847 4791 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/openstack-galera-0" podStartSLOduration=-9223371988.938944 podStartE2EDuration="47.915831904s" podCreationTimestamp="2025-12-10 23:06:24 +0000 UTC" firstStartedPulling="2025-12-10 23:06:26.441093853 +0000 UTC m=+1020.870711466" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-10 23:07:11.910803309 +0000 UTC m=+1066.340420932" watchObservedRunningTime="2025-12-10 23:07:11.915831904 +0000 UTC m=+1066.345449517" Dec 10 23:07:11 crc kubenswrapper[4791]: I1210 23:07:11.942557 4791 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/swift-ring-rebalance-pzkk7"] Dec 10 23:07:11 crc kubenswrapper[4791]: I1210 23:07:11.949911 4791 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/swift-ring-rebalance-pzkk7"] Dec 10 23:07:11 crc kubenswrapper[4791]: I1210 23:07:11.969729 4791 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/swift-ring-rebalance-bqplg"] Dec 10 23:07:11 crc kubenswrapper[4791]: I1210 23:07:11.988762 4791 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/dnsmasq-dns-698758b865-72nqc" podStartSLOduration=3.9887379149999997 podStartE2EDuration="3.988737915s" podCreationTimestamp="2025-12-10 23:07:08 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-10 23:07:11.983642067 +0000 UTC m=+1066.413259680" watchObservedRunningTime="2025-12-10 23:07:11.988737915 +0000 UTC m=+1066.418355528" Dec 10 23:07:12 crc kubenswrapper[4791]: I1210 23:07:12.008524 4791 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/kube-state-metrics-0" podStartSLOduration=2.227105074 podStartE2EDuration="44.008507227s" podCreationTimestamp="2025-12-10 23:06:28 +0000 UTC" firstStartedPulling="2025-12-10 23:06:29.720478081 +0000 UTC m=+1024.150095694" lastFinishedPulling="2025-12-10 23:07:11.501880234 +0000 UTC m=+1065.931497847" observedRunningTime="2025-12-10 23:07:12.002929115 +0000 UTC m=+1066.432546728" watchObservedRunningTime="2025-12-10 23:07:12.008507227 +0000 UTC m=+1066.438124840" Dec 10 23:07:12 crc kubenswrapper[4791]: I1210 23:07:12.894605 4791 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-ring-rebalance-bqplg" event={"ID":"16bd9b65-844e-4537-8d5b-c0a73666c2b2","Type":"ContainerStarted","Data":"34ea794495eb03aa6267efe080a7cf22b6133e29a959f42c12df14189b7a8600"} Dec 10 23:07:13 crc kubenswrapper[4791]: I1210 23:07:13.896383 4791 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="7f35b2d5-de89-4b42-85fe-8a7eb461c2ce" path="/var/lib/kubelet/pods/7f35b2d5-de89-4b42-85fe-8a7eb461c2ce/volumes" Dec 10 23:07:13 crc kubenswrapper[4791]: I1210 23:07:13.953616 4791 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/projected/69a6c950-8a3b-4a7c-b284-ebd20157eb20-etc-swift\") pod \"swift-storage-0\" (UID: \"69a6c950-8a3b-4a7c-b284-ebd20157eb20\") " pod="openstack/swift-storage-0" Dec 10 23:07:13 crc kubenswrapper[4791]: E1210 23:07:13.953904 4791 projected.go:288] Couldn't get configMap openstack/swift-ring-files: configmap "swift-ring-files" not found Dec 10 23:07:13 crc kubenswrapper[4791]: E1210 23:07:13.953949 4791 projected.go:194] Error preparing data for projected volume etc-swift for pod openstack/swift-storage-0: configmap "swift-ring-files" not found Dec 10 23:07:13 crc kubenswrapper[4791]: E1210 23:07:13.954021 4791 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/69a6c950-8a3b-4a7c-b284-ebd20157eb20-etc-swift podName:69a6c950-8a3b-4a7c-b284-ebd20157eb20 nodeName:}" failed. No retries permitted until 2025-12-10 23:07:17.953998284 +0000 UTC m=+1072.383615897 (durationBeforeRetry 4s). Error: MountVolume.SetUp failed for volume "etc-swift" (UniqueName: "kubernetes.io/projected/69a6c950-8a3b-4a7c-b284-ebd20157eb20-etc-swift") pod "swift-storage-0" (UID: "69a6c950-8a3b-4a7c-b284-ebd20157eb20") : configmap "swift-ring-files" not found Dec 10 23:07:14 crc kubenswrapper[4791]: I1210 23:07:14.865468 4791 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/dnsmasq-dns-86db49b7ff-2f6xm" Dec 10 23:07:15 crc kubenswrapper[4791]: I1210 23:07:15.509930 4791 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/openstack-galera-0" Dec 10 23:07:15 crc kubenswrapper[4791]: I1210 23:07:15.510440 4791 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/openstack-galera-0" Dec 10 23:07:15 crc kubenswrapper[4791]: I1210 23:07:15.918732 4791 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-ring-rebalance-bqplg" event={"ID":"16bd9b65-844e-4537-8d5b-c0a73666c2b2","Type":"ContainerStarted","Data":"8bc1c7c4515e4cf9ea2a83f392ce1bea0c24e41deab5877f637edc11094fc6ca"} Dec 10 23:07:15 crc kubenswrapper[4791]: I1210 23:07:15.943325 4791 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/swift-ring-rebalance-bqplg" podStartSLOduration=2.435860126 podStartE2EDuration="5.943291788s" podCreationTimestamp="2025-12-10 23:07:10 +0000 UTC" firstStartedPulling="2025-12-10 23:07:11.976370077 +0000 UTC m=+1066.405987680" lastFinishedPulling="2025-12-10 23:07:15.483801729 +0000 UTC m=+1069.913419342" observedRunningTime="2025-12-10 23:07:15.933183875 +0000 UTC m=+1070.362801488" watchObservedRunningTime="2025-12-10 23:07:15.943291788 +0000 UTC m=+1070.372909411" Dec 10 23:07:16 crc kubenswrapper[4791]: I1210 23:07:16.961389 4791 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/openstack-cell1-galera-0" Dec 10 23:07:16 crc kubenswrapper[4791]: I1210 23:07:16.961448 4791 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/openstack-cell1-galera-0" Dec 10 23:07:17 crc kubenswrapper[4791]: I1210 23:07:17.059979 4791 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/openstack-cell1-galera-0" Dec 10 23:07:17 crc kubenswrapper[4791]: I1210 23:07:17.720793 4791 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/openstack-galera-0" Dec 10 23:07:17 crc kubenswrapper[4791]: I1210 23:07:17.805655 4791 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/openstack-galera-0" Dec 10 23:07:18 crc kubenswrapper[4791]: I1210 23:07:18.017676 4791 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/openstack-cell1-galera-0" Dec 10 23:07:18 crc kubenswrapper[4791]: I1210 23:07:18.022274 4791 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/projected/69a6c950-8a3b-4a7c-b284-ebd20157eb20-etc-swift\") pod \"swift-storage-0\" (UID: \"69a6c950-8a3b-4a7c-b284-ebd20157eb20\") " pod="openstack/swift-storage-0" Dec 10 23:07:18 crc kubenswrapper[4791]: E1210 23:07:18.024601 4791 projected.go:288] Couldn't get configMap openstack/swift-ring-files: configmap "swift-ring-files" not found Dec 10 23:07:18 crc kubenswrapper[4791]: E1210 23:07:18.024647 4791 projected.go:194] Error preparing data for projected volume etc-swift for pod openstack/swift-storage-0: configmap "swift-ring-files" not found Dec 10 23:07:18 crc kubenswrapper[4791]: E1210 23:07:18.024721 4791 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/69a6c950-8a3b-4a7c-b284-ebd20157eb20-etc-swift podName:69a6c950-8a3b-4a7c-b284-ebd20157eb20 nodeName:}" failed. No retries permitted until 2025-12-10 23:07:26.02469159 +0000 UTC m=+1080.454309243 (durationBeforeRetry 8s). Error: MountVolume.SetUp failed for volume "etc-swift" (UniqueName: "kubernetes.io/projected/69a6c950-8a3b-4a7c-b284-ebd20157eb20-etc-swift") pod "swift-storage-0" (UID: "69a6c950-8a3b-4a7c-b284-ebd20157eb20") : configmap "swift-ring-files" not found Dec 10 23:07:18 crc kubenswrapper[4791]: I1210 23:07:18.812353 4791 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/kube-state-metrics-0" Dec 10 23:07:19 crc kubenswrapper[4791]: I1210 23:07:19.391599 4791 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/dnsmasq-dns-698758b865-72nqc" Dec 10 23:07:19 crc kubenswrapper[4791]: I1210 23:07:19.441629 4791 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-86db49b7ff-2f6xm"] Dec 10 23:07:19 crc kubenswrapper[4791]: I1210 23:07:19.441912 4791 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/dnsmasq-dns-86db49b7ff-2f6xm" podUID="1f9954ba-068e-4a98-bcc1-1a6cd477b858" containerName="dnsmasq-dns" containerID="cri-o://22433271d016baa99aaef1cf8a84c3ff2124b51b98f1cbbb4bf1e537938483c8" gracePeriod=10 Dec 10 23:07:19 crc kubenswrapper[4791]: I1210 23:07:19.865519 4791 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/dnsmasq-dns-86db49b7ff-2f6xm" podUID="1f9954ba-068e-4a98-bcc1-1a6cd477b858" containerName="dnsmasq-dns" probeResult="failure" output="dial tcp 10.217.0.110:5353: connect: connection refused" Dec 10 23:07:19 crc kubenswrapper[4791]: I1210 23:07:19.947101 4791 generic.go:334] "Generic (PLEG): container finished" podID="1f9954ba-068e-4a98-bcc1-1a6cd477b858" containerID="22433271d016baa99aaef1cf8a84c3ff2124b51b98f1cbbb4bf1e537938483c8" exitCode=0 Dec 10 23:07:19 crc kubenswrapper[4791]: I1210 23:07:19.947149 4791 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-86db49b7ff-2f6xm" event={"ID":"1f9954ba-068e-4a98-bcc1-1a6cd477b858","Type":"ContainerDied","Data":"22433271d016baa99aaef1cf8a84c3ff2124b51b98f1cbbb4bf1e537938483c8"} Dec 10 23:07:20 crc kubenswrapper[4791]: I1210 23:07:20.362148 4791 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/ovn-northd-0" Dec 10 23:07:20 crc kubenswrapper[4791]: I1210 23:07:20.911225 4791 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-86db49b7ff-2f6xm" Dec 10 23:07:20 crc kubenswrapper[4791]: I1210 23:07:20.959980 4791 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-86db49b7ff-2f6xm" event={"ID":"1f9954ba-068e-4a98-bcc1-1a6cd477b858","Type":"ContainerDied","Data":"d6a6298c825337fa4c702d1d74fcd4b7e2db5e4a65bfdc78a828d7f758c425da"} Dec 10 23:07:20 crc kubenswrapper[4791]: I1210 23:07:20.960044 4791 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-86db49b7ff-2f6xm" Dec 10 23:07:20 crc kubenswrapper[4791]: I1210 23:07:20.960060 4791 scope.go:117] "RemoveContainer" containerID="22433271d016baa99aaef1cf8a84c3ff2124b51b98f1cbbb4bf1e537938483c8" Dec 10 23:07:20 crc kubenswrapper[4791]: I1210 23:07:20.975136 4791 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-xnhdv\" (UniqueName: \"kubernetes.io/projected/1f9954ba-068e-4a98-bcc1-1a6cd477b858-kube-api-access-xnhdv\") pod \"1f9954ba-068e-4a98-bcc1-1a6cd477b858\" (UID: \"1f9954ba-068e-4a98-bcc1-1a6cd477b858\") " Dec 10 23:07:20 crc kubenswrapper[4791]: I1210 23:07:20.975200 4791 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/1f9954ba-068e-4a98-bcc1-1a6cd477b858-config\") pod \"1f9954ba-068e-4a98-bcc1-1a6cd477b858\" (UID: \"1f9954ba-068e-4a98-bcc1-1a6cd477b858\") " Dec 10 23:07:20 crc kubenswrapper[4791]: I1210 23:07:20.975258 4791 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/1f9954ba-068e-4a98-bcc1-1a6cd477b858-ovsdbserver-sb\") pod \"1f9954ba-068e-4a98-bcc1-1a6cd477b858\" (UID: \"1f9954ba-068e-4a98-bcc1-1a6cd477b858\") " Dec 10 23:07:20 crc kubenswrapper[4791]: I1210 23:07:20.975276 4791 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/1f9954ba-068e-4a98-bcc1-1a6cd477b858-ovsdbserver-nb\") pod \"1f9954ba-068e-4a98-bcc1-1a6cd477b858\" (UID: \"1f9954ba-068e-4a98-bcc1-1a6cd477b858\") " Dec 10 23:07:20 crc kubenswrapper[4791]: I1210 23:07:20.975353 4791 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/1f9954ba-068e-4a98-bcc1-1a6cd477b858-dns-svc\") pod \"1f9954ba-068e-4a98-bcc1-1a6cd477b858\" (UID: \"1f9954ba-068e-4a98-bcc1-1a6cd477b858\") " Dec 10 23:07:20 crc kubenswrapper[4791]: I1210 23:07:20.983450 4791 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/1f9954ba-068e-4a98-bcc1-1a6cd477b858-kube-api-access-xnhdv" (OuterVolumeSpecName: "kube-api-access-xnhdv") pod "1f9954ba-068e-4a98-bcc1-1a6cd477b858" (UID: "1f9954ba-068e-4a98-bcc1-1a6cd477b858"). InnerVolumeSpecName "kube-api-access-xnhdv". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 10 23:07:20 crc kubenswrapper[4791]: I1210 23:07:20.983959 4791 scope.go:117] "RemoveContainer" containerID="d27b53cb65a001e1228aaff2c2f4c1105609b6ad3dc01e7b73887396f8779f14" Dec 10 23:07:21 crc kubenswrapper[4791]: I1210 23:07:21.019497 4791 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/1f9954ba-068e-4a98-bcc1-1a6cd477b858-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "1f9954ba-068e-4a98-bcc1-1a6cd477b858" (UID: "1f9954ba-068e-4a98-bcc1-1a6cd477b858"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 10 23:07:21 crc kubenswrapper[4791]: I1210 23:07:21.030792 4791 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/1f9954ba-068e-4a98-bcc1-1a6cd477b858-config" (OuterVolumeSpecName: "config") pod "1f9954ba-068e-4a98-bcc1-1a6cd477b858" (UID: "1f9954ba-068e-4a98-bcc1-1a6cd477b858"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 10 23:07:21 crc kubenswrapper[4791]: I1210 23:07:21.033304 4791 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/1f9954ba-068e-4a98-bcc1-1a6cd477b858-ovsdbserver-nb" (OuterVolumeSpecName: "ovsdbserver-nb") pod "1f9954ba-068e-4a98-bcc1-1a6cd477b858" (UID: "1f9954ba-068e-4a98-bcc1-1a6cd477b858"). InnerVolumeSpecName "ovsdbserver-nb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 10 23:07:21 crc kubenswrapper[4791]: I1210 23:07:21.038371 4791 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/1f9954ba-068e-4a98-bcc1-1a6cd477b858-ovsdbserver-sb" (OuterVolumeSpecName: "ovsdbserver-sb") pod "1f9954ba-068e-4a98-bcc1-1a6cd477b858" (UID: "1f9954ba-068e-4a98-bcc1-1a6cd477b858"). InnerVolumeSpecName "ovsdbserver-sb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 10 23:07:21 crc kubenswrapper[4791]: I1210 23:07:21.077711 4791 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/1f9954ba-068e-4a98-bcc1-1a6cd477b858-dns-svc\") on node \"crc\" DevicePath \"\"" Dec 10 23:07:21 crc kubenswrapper[4791]: I1210 23:07:21.077981 4791 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-xnhdv\" (UniqueName: \"kubernetes.io/projected/1f9954ba-068e-4a98-bcc1-1a6cd477b858-kube-api-access-xnhdv\") on node \"crc\" DevicePath \"\"" Dec 10 23:07:21 crc kubenswrapper[4791]: I1210 23:07:21.078100 4791 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/1f9954ba-068e-4a98-bcc1-1a6cd477b858-config\") on node \"crc\" DevicePath \"\"" Dec 10 23:07:21 crc kubenswrapper[4791]: I1210 23:07:21.078176 4791 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/1f9954ba-068e-4a98-bcc1-1a6cd477b858-ovsdbserver-sb\") on node \"crc\" DevicePath \"\"" Dec 10 23:07:21 crc kubenswrapper[4791]: I1210 23:07:21.078248 4791 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/1f9954ba-068e-4a98-bcc1-1a6cd477b858-ovsdbserver-nb\") on node \"crc\" DevicePath \"\"" Dec 10 23:07:21 crc kubenswrapper[4791]: I1210 23:07:21.290878 4791 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-86db49b7ff-2f6xm"] Dec 10 23:07:21 crc kubenswrapper[4791]: I1210 23:07:21.299996 4791 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-86db49b7ff-2f6xm"] Dec 10 23:07:21 crc kubenswrapper[4791]: I1210 23:07:21.909600 4791 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="1f9954ba-068e-4a98-bcc1-1a6cd477b858" path="/var/lib/kubelet/pods/1f9954ba-068e-4a98-bcc1-1a6cd477b858/volumes" Dec 10 23:07:23 crc kubenswrapper[4791]: I1210 23:07:23.997914 4791 generic.go:334] "Generic (PLEG): container finished" podID="16bd9b65-844e-4537-8d5b-c0a73666c2b2" containerID="8bc1c7c4515e4cf9ea2a83f392ce1bea0c24e41deab5877f637edc11094fc6ca" exitCode=0 Dec 10 23:07:23 crc kubenswrapper[4791]: I1210 23:07:23.997974 4791 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-ring-rebalance-bqplg" event={"ID":"16bd9b65-844e-4537-8d5b-c0a73666c2b2","Type":"ContainerDied","Data":"8bc1c7c4515e4cf9ea2a83f392ce1bea0c24e41deab5877f637edc11094fc6ca"} Dec 10 23:07:25 crc kubenswrapper[4791]: I1210 23:07:25.330322 4791 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/swift-ring-rebalance-bqplg" Dec 10 23:07:25 crc kubenswrapper[4791]: I1210 23:07:25.503811 4791 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/16bd9b65-844e-4537-8d5b-c0a73666c2b2-scripts\") pod \"16bd9b65-844e-4537-8d5b-c0a73666c2b2\" (UID: \"16bd9b65-844e-4537-8d5b-c0a73666c2b2\") " Dec 10 23:07:25 crc kubenswrapper[4791]: I1210 23:07:25.504292 4791 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/16bd9b65-844e-4537-8d5b-c0a73666c2b2-combined-ca-bundle\") pod \"16bd9b65-844e-4537-8d5b-c0a73666c2b2\" (UID: \"16bd9b65-844e-4537-8d5b-c0a73666c2b2\") " Dec 10 23:07:25 crc kubenswrapper[4791]: I1210 23:07:25.504412 4791 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ring-data-devices\" (UniqueName: \"kubernetes.io/configmap/16bd9b65-844e-4537-8d5b-c0a73666c2b2-ring-data-devices\") pod \"16bd9b65-844e-4537-8d5b-c0a73666c2b2\" (UID: \"16bd9b65-844e-4537-8d5b-c0a73666c2b2\") " Dec 10 23:07:25 crc kubenswrapper[4791]: I1210 23:07:25.504476 4791 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/empty-dir/16bd9b65-844e-4537-8d5b-c0a73666c2b2-etc-swift\") pod \"16bd9b65-844e-4537-8d5b-c0a73666c2b2\" (UID: \"16bd9b65-844e-4537-8d5b-c0a73666c2b2\") " Dec 10 23:07:25 crc kubenswrapper[4791]: I1210 23:07:25.505457 4791 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/16bd9b65-844e-4537-8d5b-c0a73666c2b2-ring-data-devices" (OuterVolumeSpecName: "ring-data-devices") pod "16bd9b65-844e-4537-8d5b-c0a73666c2b2" (UID: "16bd9b65-844e-4537-8d5b-c0a73666c2b2"). InnerVolumeSpecName "ring-data-devices". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 10 23:07:25 crc kubenswrapper[4791]: I1210 23:07:25.506187 4791 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"swiftconf\" (UniqueName: \"kubernetes.io/secret/16bd9b65-844e-4537-8d5b-c0a73666c2b2-swiftconf\") pod \"16bd9b65-844e-4537-8d5b-c0a73666c2b2\" (UID: \"16bd9b65-844e-4537-8d5b-c0a73666c2b2\") " Dec 10 23:07:25 crc kubenswrapper[4791]: I1210 23:07:25.506191 4791 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/16bd9b65-844e-4537-8d5b-c0a73666c2b2-etc-swift" (OuterVolumeSpecName: "etc-swift") pod "16bd9b65-844e-4537-8d5b-c0a73666c2b2" (UID: "16bd9b65-844e-4537-8d5b-c0a73666c2b2"). InnerVolumeSpecName "etc-swift". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 10 23:07:25 crc kubenswrapper[4791]: I1210 23:07:25.506230 4791 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-vnfxn\" (UniqueName: \"kubernetes.io/projected/16bd9b65-844e-4537-8d5b-c0a73666c2b2-kube-api-access-vnfxn\") pod \"16bd9b65-844e-4537-8d5b-c0a73666c2b2\" (UID: \"16bd9b65-844e-4537-8d5b-c0a73666c2b2\") " Dec 10 23:07:25 crc kubenswrapper[4791]: I1210 23:07:25.506600 4791 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dispersionconf\" (UniqueName: \"kubernetes.io/secret/16bd9b65-844e-4537-8d5b-c0a73666c2b2-dispersionconf\") pod \"16bd9b65-844e-4537-8d5b-c0a73666c2b2\" (UID: \"16bd9b65-844e-4537-8d5b-c0a73666c2b2\") " Dec 10 23:07:25 crc kubenswrapper[4791]: I1210 23:07:25.510522 4791 reconciler_common.go:293] "Volume detached for volume \"ring-data-devices\" (UniqueName: \"kubernetes.io/configmap/16bd9b65-844e-4537-8d5b-c0a73666c2b2-ring-data-devices\") on node \"crc\" DevicePath \"\"" Dec 10 23:07:25 crc kubenswrapper[4791]: I1210 23:07:25.510581 4791 reconciler_common.go:293] "Volume detached for volume \"etc-swift\" (UniqueName: \"kubernetes.io/empty-dir/16bd9b65-844e-4537-8d5b-c0a73666c2b2-etc-swift\") on node \"crc\" DevicePath \"\"" Dec 10 23:07:25 crc kubenswrapper[4791]: I1210 23:07:25.515323 4791 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/16bd9b65-844e-4537-8d5b-c0a73666c2b2-kube-api-access-vnfxn" (OuterVolumeSpecName: "kube-api-access-vnfxn") pod "16bd9b65-844e-4537-8d5b-c0a73666c2b2" (UID: "16bd9b65-844e-4537-8d5b-c0a73666c2b2"). InnerVolumeSpecName "kube-api-access-vnfxn". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 10 23:07:25 crc kubenswrapper[4791]: I1210 23:07:25.520488 4791 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/16bd9b65-844e-4537-8d5b-c0a73666c2b2-dispersionconf" (OuterVolumeSpecName: "dispersionconf") pod "16bd9b65-844e-4537-8d5b-c0a73666c2b2" (UID: "16bd9b65-844e-4537-8d5b-c0a73666c2b2"). InnerVolumeSpecName "dispersionconf". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 10 23:07:25 crc kubenswrapper[4791]: I1210 23:07:25.529865 4791 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/16bd9b65-844e-4537-8d5b-c0a73666c2b2-scripts" (OuterVolumeSpecName: "scripts") pod "16bd9b65-844e-4537-8d5b-c0a73666c2b2" (UID: "16bd9b65-844e-4537-8d5b-c0a73666c2b2"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 10 23:07:25 crc kubenswrapper[4791]: I1210 23:07:25.535843 4791 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/16bd9b65-844e-4537-8d5b-c0a73666c2b2-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "16bd9b65-844e-4537-8d5b-c0a73666c2b2" (UID: "16bd9b65-844e-4537-8d5b-c0a73666c2b2"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 10 23:07:25 crc kubenswrapper[4791]: I1210 23:07:25.542749 4791 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/16bd9b65-844e-4537-8d5b-c0a73666c2b2-swiftconf" (OuterVolumeSpecName: "swiftconf") pod "16bd9b65-844e-4537-8d5b-c0a73666c2b2" (UID: "16bd9b65-844e-4537-8d5b-c0a73666c2b2"). InnerVolumeSpecName "swiftconf". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 10 23:07:25 crc kubenswrapper[4791]: I1210 23:07:25.611807 4791 reconciler_common.go:293] "Volume detached for volume \"swiftconf\" (UniqueName: \"kubernetes.io/secret/16bd9b65-844e-4537-8d5b-c0a73666c2b2-swiftconf\") on node \"crc\" DevicePath \"\"" Dec 10 23:07:25 crc kubenswrapper[4791]: I1210 23:07:25.611840 4791 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-vnfxn\" (UniqueName: \"kubernetes.io/projected/16bd9b65-844e-4537-8d5b-c0a73666c2b2-kube-api-access-vnfxn\") on node \"crc\" DevicePath \"\"" Dec 10 23:07:25 crc kubenswrapper[4791]: I1210 23:07:25.611852 4791 reconciler_common.go:293] "Volume detached for volume \"dispersionconf\" (UniqueName: \"kubernetes.io/secret/16bd9b65-844e-4537-8d5b-c0a73666c2b2-dispersionconf\") on node \"crc\" DevicePath \"\"" Dec 10 23:07:25 crc kubenswrapper[4791]: I1210 23:07:25.611861 4791 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/16bd9b65-844e-4537-8d5b-c0a73666c2b2-scripts\") on node \"crc\" DevicePath \"\"" Dec 10 23:07:25 crc kubenswrapper[4791]: I1210 23:07:25.611870 4791 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/16bd9b65-844e-4537-8d5b-c0a73666c2b2-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 10 23:07:26 crc kubenswrapper[4791]: I1210 23:07:26.013872 4791 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-ring-rebalance-bqplg" event={"ID":"16bd9b65-844e-4537-8d5b-c0a73666c2b2","Type":"ContainerDied","Data":"34ea794495eb03aa6267efe080a7cf22b6133e29a959f42c12df14189b7a8600"} Dec 10 23:07:26 crc kubenswrapper[4791]: I1210 23:07:26.013910 4791 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="34ea794495eb03aa6267efe080a7cf22b6133e29a959f42c12df14189b7a8600" Dec 10 23:07:26 crc kubenswrapper[4791]: I1210 23:07:26.013949 4791 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/swift-ring-rebalance-bqplg" Dec 10 23:07:26 crc kubenswrapper[4791]: I1210 23:07:26.118632 4791 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/projected/69a6c950-8a3b-4a7c-b284-ebd20157eb20-etc-swift\") pod \"swift-storage-0\" (UID: \"69a6c950-8a3b-4a7c-b284-ebd20157eb20\") " pod="openstack/swift-storage-0" Dec 10 23:07:26 crc kubenswrapper[4791]: I1210 23:07:26.122269 4791 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-swift\" (UniqueName: \"kubernetes.io/projected/69a6c950-8a3b-4a7c-b284-ebd20157eb20-etc-swift\") pod \"swift-storage-0\" (UID: \"69a6c950-8a3b-4a7c-b284-ebd20157eb20\") " pod="openstack/swift-storage-0" Dec 10 23:07:26 crc kubenswrapper[4791]: I1210 23:07:26.337468 4791 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/swift-storage-0" Dec 10 23:07:26 crc kubenswrapper[4791]: I1210 23:07:26.695348 4791 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/keystone-85e7-account-create-update-25zns"] Dec 10 23:07:26 crc kubenswrapper[4791]: E1210 23:07:26.695944 4791 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="1f9954ba-068e-4a98-bcc1-1a6cd477b858" containerName="dnsmasq-dns" Dec 10 23:07:26 crc kubenswrapper[4791]: I1210 23:07:26.695965 4791 state_mem.go:107] "Deleted CPUSet assignment" podUID="1f9954ba-068e-4a98-bcc1-1a6cd477b858" containerName="dnsmasq-dns" Dec 10 23:07:26 crc kubenswrapper[4791]: E1210 23:07:26.696004 4791 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="1f9954ba-068e-4a98-bcc1-1a6cd477b858" containerName="init" Dec 10 23:07:26 crc kubenswrapper[4791]: I1210 23:07:26.696013 4791 state_mem.go:107] "Deleted CPUSet assignment" podUID="1f9954ba-068e-4a98-bcc1-1a6cd477b858" containerName="init" Dec 10 23:07:26 crc kubenswrapper[4791]: E1210 23:07:26.696028 4791 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="16bd9b65-844e-4537-8d5b-c0a73666c2b2" containerName="swift-ring-rebalance" Dec 10 23:07:26 crc kubenswrapper[4791]: I1210 23:07:26.696037 4791 state_mem.go:107] "Deleted CPUSet assignment" podUID="16bd9b65-844e-4537-8d5b-c0a73666c2b2" containerName="swift-ring-rebalance" Dec 10 23:07:26 crc kubenswrapper[4791]: I1210 23:07:26.696237 4791 memory_manager.go:354] "RemoveStaleState removing state" podUID="16bd9b65-844e-4537-8d5b-c0a73666c2b2" containerName="swift-ring-rebalance" Dec 10 23:07:26 crc kubenswrapper[4791]: I1210 23:07:26.696267 4791 memory_manager.go:354] "RemoveStaleState removing state" podUID="1f9954ba-068e-4a98-bcc1-1a6cd477b858" containerName="dnsmasq-dns" Dec 10 23:07:26 crc kubenswrapper[4791]: I1210 23:07:26.696911 4791 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-85e7-account-create-update-25zns" Dec 10 23:07:26 crc kubenswrapper[4791]: I1210 23:07:26.699457 4791 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone-db-secret" Dec 10 23:07:26 crc kubenswrapper[4791]: I1210 23:07:26.709544 4791 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/keystone-85e7-account-create-update-25zns"] Dec 10 23:07:26 crc kubenswrapper[4791]: I1210 23:07:26.727612 4791 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/4d16f575-dc6d-4a78-8303-10a1c8202c0e-operator-scripts\") pod \"keystone-85e7-account-create-update-25zns\" (UID: \"4d16f575-dc6d-4a78-8303-10a1c8202c0e\") " pod="openstack/keystone-85e7-account-create-update-25zns" Dec 10 23:07:26 crc kubenswrapper[4791]: I1210 23:07:26.727662 4791 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-c256w\" (UniqueName: \"kubernetes.io/projected/4d16f575-dc6d-4a78-8303-10a1c8202c0e-kube-api-access-c256w\") pod \"keystone-85e7-account-create-update-25zns\" (UID: \"4d16f575-dc6d-4a78-8303-10a1c8202c0e\") " pod="openstack/keystone-85e7-account-create-update-25zns" Dec 10 23:07:26 crc kubenswrapper[4791]: I1210 23:07:26.729628 4791 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/keystone-db-create-qfvp4"] Dec 10 23:07:26 crc kubenswrapper[4791]: I1210 23:07:26.731302 4791 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-db-create-qfvp4" Dec 10 23:07:26 crc kubenswrapper[4791]: I1210 23:07:26.745687 4791 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/keystone-db-create-qfvp4"] Dec 10 23:07:26 crc kubenswrapper[4791]: I1210 23:07:26.828809 4791 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/afcdc302-63c5-45c5-bf5b-46ef1caf953d-operator-scripts\") pod \"keystone-db-create-qfvp4\" (UID: \"afcdc302-63c5-45c5-bf5b-46ef1caf953d\") " pod="openstack/keystone-db-create-qfvp4" Dec 10 23:07:26 crc kubenswrapper[4791]: I1210 23:07:26.828893 4791 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-fdr7j\" (UniqueName: \"kubernetes.io/projected/afcdc302-63c5-45c5-bf5b-46ef1caf953d-kube-api-access-fdr7j\") pod \"keystone-db-create-qfvp4\" (UID: \"afcdc302-63c5-45c5-bf5b-46ef1caf953d\") " pod="openstack/keystone-db-create-qfvp4" Dec 10 23:07:26 crc kubenswrapper[4791]: I1210 23:07:26.828933 4791 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/4d16f575-dc6d-4a78-8303-10a1c8202c0e-operator-scripts\") pod \"keystone-85e7-account-create-update-25zns\" (UID: \"4d16f575-dc6d-4a78-8303-10a1c8202c0e\") " pod="openstack/keystone-85e7-account-create-update-25zns" Dec 10 23:07:26 crc kubenswrapper[4791]: I1210 23:07:26.828967 4791 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-c256w\" (UniqueName: \"kubernetes.io/projected/4d16f575-dc6d-4a78-8303-10a1c8202c0e-kube-api-access-c256w\") pod \"keystone-85e7-account-create-update-25zns\" (UID: \"4d16f575-dc6d-4a78-8303-10a1c8202c0e\") " pod="openstack/keystone-85e7-account-create-update-25zns" Dec 10 23:07:26 crc kubenswrapper[4791]: I1210 23:07:26.829695 4791 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/4d16f575-dc6d-4a78-8303-10a1c8202c0e-operator-scripts\") pod \"keystone-85e7-account-create-update-25zns\" (UID: \"4d16f575-dc6d-4a78-8303-10a1c8202c0e\") " pod="openstack/keystone-85e7-account-create-update-25zns" Dec 10 23:07:26 crc kubenswrapper[4791]: I1210 23:07:26.851006 4791 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-c256w\" (UniqueName: \"kubernetes.io/projected/4d16f575-dc6d-4a78-8303-10a1c8202c0e-kube-api-access-c256w\") pod \"keystone-85e7-account-create-update-25zns\" (UID: \"4d16f575-dc6d-4a78-8303-10a1c8202c0e\") " pod="openstack/keystone-85e7-account-create-update-25zns" Dec 10 23:07:26 crc kubenswrapper[4791]: W1210 23:07:26.857606 4791 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod69a6c950_8a3b_4a7c_b284_ebd20157eb20.slice/crio-45b655726b37dcd905184e8c9a35d8d88ad41ac2f6fc016dcc41b83d46ae210a WatchSource:0}: Error finding container 45b655726b37dcd905184e8c9a35d8d88ad41ac2f6fc016dcc41b83d46ae210a: Status 404 returned error can't find the container with id 45b655726b37dcd905184e8c9a35d8d88ad41ac2f6fc016dcc41b83d46ae210a Dec 10 23:07:26 crc kubenswrapper[4791]: I1210 23:07:26.859833 4791 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/swift-storage-0"] Dec 10 23:07:26 crc kubenswrapper[4791]: I1210 23:07:26.930757 4791 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-fdr7j\" (UniqueName: \"kubernetes.io/projected/afcdc302-63c5-45c5-bf5b-46ef1caf953d-kube-api-access-fdr7j\") pod \"keystone-db-create-qfvp4\" (UID: \"afcdc302-63c5-45c5-bf5b-46ef1caf953d\") " pod="openstack/keystone-db-create-qfvp4" Dec 10 23:07:26 crc kubenswrapper[4791]: I1210 23:07:26.930920 4791 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/afcdc302-63c5-45c5-bf5b-46ef1caf953d-operator-scripts\") pod \"keystone-db-create-qfvp4\" (UID: \"afcdc302-63c5-45c5-bf5b-46ef1caf953d\") " pod="openstack/keystone-db-create-qfvp4" Dec 10 23:07:26 crc kubenswrapper[4791]: I1210 23:07:26.931575 4791 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/afcdc302-63c5-45c5-bf5b-46ef1caf953d-operator-scripts\") pod \"keystone-db-create-qfvp4\" (UID: \"afcdc302-63c5-45c5-bf5b-46ef1caf953d\") " pod="openstack/keystone-db-create-qfvp4" Dec 10 23:07:26 crc kubenswrapper[4791]: I1210 23:07:26.940501 4791 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/placement-db-create-g4swg"] Dec 10 23:07:26 crc kubenswrapper[4791]: I1210 23:07:26.941724 4791 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/placement-db-create-g4swg" Dec 10 23:07:26 crc kubenswrapper[4791]: I1210 23:07:26.949534 4791 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/placement-db-create-g4swg"] Dec 10 23:07:26 crc kubenswrapper[4791]: I1210 23:07:26.959238 4791 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-fdr7j\" (UniqueName: \"kubernetes.io/projected/afcdc302-63c5-45c5-bf5b-46ef1caf953d-kube-api-access-fdr7j\") pod \"keystone-db-create-qfvp4\" (UID: \"afcdc302-63c5-45c5-bf5b-46ef1caf953d\") " pod="openstack/keystone-db-create-qfvp4" Dec 10 23:07:27 crc kubenswrapper[4791]: I1210 23:07:27.009407 4791 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/placement-492e-account-create-update-xfsh4"] Dec 10 23:07:27 crc kubenswrapper[4791]: I1210 23:07:27.010422 4791 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/placement-492e-account-create-update-xfsh4" Dec 10 23:07:27 crc kubenswrapper[4791]: I1210 23:07:27.012942 4791 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"placement-db-secret" Dec 10 23:07:27 crc kubenswrapper[4791]: I1210 23:07:27.018273 4791 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-85e7-account-create-update-25zns" Dec 10 23:07:27 crc kubenswrapper[4791]: I1210 23:07:27.020731 4791 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/placement-492e-account-create-update-xfsh4"] Dec 10 23:07:27 crc kubenswrapper[4791]: I1210 23:07:27.038245 4791 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"69a6c950-8a3b-4a7c-b284-ebd20157eb20","Type":"ContainerStarted","Data":"45b655726b37dcd905184e8c9a35d8d88ad41ac2f6fc016dcc41b83d46ae210a"} Dec 10 23:07:27 crc kubenswrapper[4791]: I1210 23:07:27.055067 4791 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-db-create-qfvp4" Dec 10 23:07:27 crc kubenswrapper[4791]: I1210 23:07:27.134552 4791 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/831d2f93-31c5-469f-b4bf-1dd85fa21cbf-operator-scripts\") pod \"placement-db-create-g4swg\" (UID: \"831d2f93-31c5-469f-b4bf-1dd85fa21cbf\") " pod="openstack/placement-db-create-g4swg" Dec 10 23:07:27 crc kubenswrapper[4791]: I1210 23:07:27.134587 4791 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/2c828334-1c90-4871-b203-71f82ac6fb8b-operator-scripts\") pod \"placement-492e-account-create-update-xfsh4\" (UID: \"2c828334-1c90-4871-b203-71f82ac6fb8b\") " pod="openstack/placement-492e-account-create-update-xfsh4" Dec 10 23:07:27 crc kubenswrapper[4791]: I1210 23:07:27.134917 4791 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-2x6pm\" (UniqueName: \"kubernetes.io/projected/2c828334-1c90-4871-b203-71f82ac6fb8b-kube-api-access-2x6pm\") pod \"placement-492e-account-create-update-xfsh4\" (UID: \"2c828334-1c90-4871-b203-71f82ac6fb8b\") " pod="openstack/placement-492e-account-create-update-xfsh4" Dec 10 23:07:27 crc kubenswrapper[4791]: I1210 23:07:27.135237 4791 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-mrrss\" (UniqueName: \"kubernetes.io/projected/831d2f93-31c5-469f-b4bf-1dd85fa21cbf-kube-api-access-mrrss\") pod \"placement-db-create-g4swg\" (UID: \"831d2f93-31c5-469f-b4bf-1dd85fa21cbf\") " pod="openstack/placement-db-create-g4swg" Dec 10 23:07:27 crc kubenswrapper[4791]: I1210 23:07:27.223965 4791 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/glance-db-create-czcrm"] Dec 10 23:07:27 crc kubenswrapper[4791]: I1210 23:07:27.224933 4791 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-db-create-czcrm" Dec 10 23:07:27 crc kubenswrapper[4791]: I1210 23:07:27.236732 4791 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/2a8736f4-b93d-4bde-a270-0acb0c730ab0-operator-scripts\") pod \"glance-db-create-czcrm\" (UID: \"2a8736f4-b93d-4bde-a270-0acb0c730ab0\") " pod="openstack/glance-db-create-czcrm" Dec 10 23:07:27 crc kubenswrapper[4791]: I1210 23:07:27.236794 4791 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-gcjkw\" (UniqueName: \"kubernetes.io/projected/2a8736f4-b93d-4bde-a270-0acb0c730ab0-kube-api-access-gcjkw\") pod \"glance-db-create-czcrm\" (UID: \"2a8736f4-b93d-4bde-a270-0acb0c730ab0\") " pod="openstack/glance-db-create-czcrm" Dec 10 23:07:27 crc kubenswrapper[4791]: I1210 23:07:27.236830 4791 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-mrrss\" (UniqueName: \"kubernetes.io/projected/831d2f93-31c5-469f-b4bf-1dd85fa21cbf-kube-api-access-mrrss\") pod \"placement-db-create-g4swg\" (UID: \"831d2f93-31c5-469f-b4bf-1dd85fa21cbf\") " pod="openstack/placement-db-create-g4swg" Dec 10 23:07:27 crc kubenswrapper[4791]: I1210 23:07:27.236855 4791 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/831d2f93-31c5-469f-b4bf-1dd85fa21cbf-operator-scripts\") pod \"placement-db-create-g4swg\" (UID: \"831d2f93-31c5-469f-b4bf-1dd85fa21cbf\") " pod="openstack/placement-db-create-g4swg" Dec 10 23:07:27 crc kubenswrapper[4791]: I1210 23:07:27.236877 4791 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/2c828334-1c90-4871-b203-71f82ac6fb8b-operator-scripts\") pod \"placement-492e-account-create-update-xfsh4\" (UID: \"2c828334-1c90-4871-b203-71f82ac6fb8b\") " pod="openstack/placement-492e-account-create-update-xfsh4" Dec 10 23:07:27 crc kubenswrapper[4791]: I1210 23:07:27.236927 4791 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-2x6pm\" (UniqueName: \"kubernetes.io/projected/2c828334-1c90-4871-b203-71f82ac6fb8b-kube-api-access-2x6pm\") pod \"placement-492e-account-create-update-xfsh4\" (UID: \"2c828334-1c90-4871-b203-71f82ac6fb8b\") " pod="openstack/placement-492e-account-create-update-xfsh4" Dec 10 23:07:27 crc kubenswrapper[4791]: I1210 23:07:27.238008 4791 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/831d2f93-31c5-469f-b4bf-1dd85fa21cbf-operator-scripts\") pod \"placement-db-create-g4swg\" (UID: \"831d2f93-31c5-469f-b4bf-1dd85fa21cbf\") " pod="openstack/placement-db-create-g4swg" Dec 10 23:07:27 crc kubenswrapper[4791]: I1210 23:07:27.238088 4791 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-db-create-czcrm"] Dec 10 23:07:27 crc kubenswrapper[4791]: I1210 23:07:27.238361 4791 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/2c828334-1c90-4871-b203-71f82ac6fb8b-operator-scripts\") pod \"placement-492e-account-create-update-xfsh4\" (UID: \"2c828334-1c90-4871-b203-71f82ac6fb8b\") " pod="openstack/placement-492e-account-create-update-xfsh4" Dec 10 23:07:27 crc kubenswrapper[4791]: I1210 23:07:27.264713 4791 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-mrrss\" (UniqueName: \"kubernetes.io/projected/831d2f93-31c5-469f-b4bf-1dd85fa21cbf-kube-api-access-mrrss\") pod \"placement-db-create-g4swg\" (UID: \"831d2f93-31c5-469f-b4bf-1dd85fa21cbf\") " pod="openstack/placement-db-create-g4swg" Dec 10 23:07:27 crc kubenswrapper[4791]: I1210 23:07:27.264728 4791 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-2x6pm\" (UniqueName: \"kubernetes.io/projected/2c828334-1c90-4871-b203-71f82ac6fb8b-kube-api-access-2x6pm\") pod \"placement-492e-account-create-update-xfsh4\" (UID: \"2c828334-1c90-4871-b203-71f82ac6fb8b\") " pod="openstack/placement-492e-account-create-update-xfsh4" Dec 10 23:07:27 crc kubenswrapper[4791]: I1210 23:07:27.294574 4791 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/placement-db-create-g4swg" Dec 10 23:07:27 crc kubenswrapper[4791]: I1210 23:07:27.334478 4791 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/glance-073e-account-create-update-hwgzh"] Dec 10 23:07:27 crc kubenswrapper[4791]: I1210 23:07:27.336122 4791 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-073e-account-create-update-hwgzh" Dec 10 23:07:27 crc kubenswrapper[4791]: I1210 23:07:27.338238 4791 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/2a8736f4-b93d-4bde-a270-0acb0c730ab0-operator-scripts\") pod \"glance-db-create-czcrm\" (UID: \"2a8736f4-b93d-4bde-a270-0acb0c730ab0\") " pod="openstack/glance-db-create-czcrm" Dec 10 23:07:27 crc kubenswrapper[4791]: I1210 23:07:27.338313 4791 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-gcjkw\" (UniqueName: \"kubernetes.io/projected/2a8736f4-b93d-4bde-a270-0acb0c730ab0-kube-api-access-gcjkw\") pod \"glance-db-create-czcrm\" (UID: \"2a8736f4-b93d-4bde-a270-0acb0c730ab0\") " pod="openstack/glance-db-create-czcrm" Dec 10 23:07:27 crc kubenswrapper[4791]: I1210 23:07:27.338368 4791 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/placement-492e-account-create-update-xfsh4" Dec 10 23:07:27 crc kubenswrapper[4791]: I1210 23:07:27.339658 4791 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/2a8736f4-b93d-4bde-a270-0acb0c730ab0-operator-scripts\") pod \"glance-db-create-czcrm\" (UID: \"2a8736f4-b93d-4bde-a270-0acb0c730ab0\") " pod="openstack/glance-db-create-czcrm" Dec 10 23:07:27 crc kubenswrapper[4791]: I1210 23:07:27.340168 4791 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"glance-db-secret" Dec 10 23:07:27 crc kubenswrapper[4791]: I1210 23:07:27.350010 4791 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-073e-account-create-update-hwgzh"] Dec 10 23:07:27 crc kubenswrapper[4791]: I1210 23:07:27.359756 4791 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-gcjkw\" (UniqueName: \"kubernetes.io/projected/2a8736f4-b93d-4bde-a270-0acb0c730ab0-kube-api-access-gcjkw\") pod \"glance-db-create-czcrm\" (UID: \"2a8736f4-b93d-4bde-a270-0acb0c730ab0\") " pod="openstack/glance-db-create-czcrm" Dec 10 23:07:27 crc kubenswrapper[4791]: I1210 23:07:27.441172 4791 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/af2a4d0a-289c-4070-8371-2eca62ff0ca2-operator-scripts\") pod \"glance-073e-account-create-update-hwgzh\" (UID: \"af2a4d0a-289c-4070-8371-2eca62ff0ca2\") " pod="openstack/glance-073e-account-create-update-hwgzh" Dec 10 23:07:27 crc kubenswrapper[4791]: I1210 23:07:27.441243 4791 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/keystone-85e7-account-create-update-25zns"] Dec 10 23:07:27 crc kubenswrapper[4791]: I1210 23:07:27.441275 4791 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-6bh9d\" (UniqueName: \"kubernetes.io/projected/af2a4d0a-289c-4070-8371-2eca62ff0ca2-kube-api-access-6bh9d\") pod \"glance-073e-account-create-update-hwgzh\" (UID: \"af2a4d0a-289c-4070-8371-2eca62ff0ca2\") " pod="openstack/glance-073e-account-create-update-hwgzh" Dec 10 23:07:27 crc kubenswrapper[4791]: W1210 23:07:27.448242 4791 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod4d16f575_dc6d_4a78_8303_10a1c8202c0e.slice/crio-3d62758251e0debcf0a653e55bb0fd69be2e4ea107e55c5e49751a51aee18de6 WatchSource:0}: Error finding container 3d62758251e0debcf0a653e55bb0fd69be2e4ea107e55c5e49751a51aee18de6: Status 404 returned error can't find the container with id 3d62758251e0debcf0a653e55bb0fd69be2e4ea107e55c5e49751a51aee18de6 Dec 10 23:07:27 crc kubenswrapper[4791]: I1210 23:07:27.542480 4791 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/af2a4d0a-289c-4070-8371-2eca62ff0ca2-operator-scripts\") pod \"glance-073e-account-create-update-hwgzh\" (UID: \"af2a4d0a-289c-4070-8371-2eca62ff0ca2\") " pod="openstack/glance-073e-account-create-update-hwgzh" Dec 10 23:07:27 crc kubenswrapper[4791]: I1210 23:07:27.542581 4791 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-6bh9d\" (UniqueName: \"kubernetes.io/projected/af2a4d0a-289c-4070-8371-2eca62ff0ca2-kube-api-access-6bh9d\") pod \"glance-073e-account-create-update-hwgzh\" (UID: \"af2a4d0a-289c-4070-8371-2eca62ff0ca2\") " pod="openstack/glance-073e-account-create-update-hwgzh" Dec 10 23:07:27 crc kubenswrapper[4791]: I1210 23:07:27.543790 4791 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/af2a4d0a-289c-4070-8371-2eca62ff0ca2-operator-scripts\") pod \"glance-073e-account-create-update-hwgzh\" (UID: \"af2a4d0a-289c-4070-8371-2eca62ff0ca2\") " pod="openstack/glance-073e-account-create-update-hwgzh" Dec 10 23:07:27 crc kubenswrapper[4791]: I1210 23:07:27.567219 4791 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-db-create-czcrm" Dec 10 23:07:27 crc kubenswrapper[4791]: I1210 23:07:27.573682 4791 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-6bh9d\" (UniqueName: \"kubernetes.io/projected/af2a4d0a-289c-4070-8371-2eca62ff0ca2-kube-api-access-6bh9d\") pod \"glance-073e-account-create-update-hwgzh\" (UID: \"af2a4d0a-289c-4070-8371-2eca62ff0ca2\") " pod="openstack/glance-073e-account-create-update-hwgzh" Dec 10 23:07:27 crc kubenswrapper[4791]: I1210 23:07:27.600827 4791 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/keystone-db-create-qfvp4"] Dec 10 23:07:27 crc kubenswrapper[4791]: W1210 23:07:27.611644 4791 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podafcdc302_63c5_45c5_bf5b_46ef1caf953d.slice/crio-28d71b866dbe2b9c19fbf7412eedf21fc9e09ba524c0880577f57f16b15b53c0 WatchSource:0}: Error finding container 28d71b866dbe2b9c19fbf7412eedf21fc9e09ba524c0880577f57f16b15b53c0: Status 404 returned error can't find the container with id 28d71b866dbe2b9c19fbf7412eedf21fc9e09ba524c0880577f57f16b15b53c0 Dec 10 23:07:27 crc kubenswrapper[4791]: I1210 23:07:27.735083 4791 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-073e-account-create-update-hwgzh" Dec 10 23:07:27 crc kubenswrapper[4791]: I1210 23:07:27.813176 4791 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/placement-db-create-g4swg"] Dec 10 23:07:27 crc kubenswrapper[4791]: I1210 23:07:27.879394 4791 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/placement-492e-account-create-update-xfsh4"] Dec 10 23:07:27 crc kubenswrapper[4791]: W1210 23:07:27.890784 4791 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod2c828334_1c90_4871_b203_71f82ac6fb8b.slice/crio-694f186b422dfcc334f33065cc5745a051df5c9a91a7801b710f075b3020ca83 WatchSource:0}: Error finding container 694f186b422dfcc334f33065cc5745a051df5c9a91a7801b710f075b3020ca83: Status 404 returned error can't find the container with id 694f186b422dfcc334f33065cc5745a051df5c9a91a7801b710f075b3020ca83 Dec 10 23:07:27 crc kubenswrapper[4791]: I1210 23:07:27.997485 4791 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-db-create-czcrm"] Dec 10 23:07:28 crc kubenswrapper[4791]: I1210 23:07:28.046467 4791 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-492e-account-create-update-xfsh4" event={"ID":"2c828334-1c90-4871-b203-71f82ac6fb8b","Type":"ContainerStarted","Data":"694f186b422dfcc334f33065cc5745a051df5c9a91a7801b710f075b3020ca83"} Dec 10 23:07:28 crc kubenswrapper[4791]: I1210 23:07:28.047453 4791 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-db-create-g4swg" event={"ID":"831d2f93-31c5-469f-b4bf-1dd85fa21cbf","Type":"ContainerStarted","Data":"01367db63ef39fc60b32e43607d7e30c8dd2d4b2b8d75e44b3a6f9136d0a57bf"} Dec 10 23:07:28 crc kubenswrapper[4791]: I1210 23:07:28.048258 4791 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-db-create-czcrm" event={"ID":"2a8736f4-b93d-4bde-a270-0acb0c730ab0","Type":"ContainerStarted","Data":"ca37ac8add1d1e0a3327dd6a31102e8d73ca787d17a7a5bee8afd659b8a5e183"} Dec 10 23:07:28 crc kubenswrapper[4791]: I1210 23:07:28.049736 4791 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-db-create-qfvp4" event={"ID":"afcdc302-63c5-45c5-bf5b-46ef1caf953d","Type":"ContainerStarted","Data":"28d71b866dbe2b9c19fbf7412eedf21fc9e09ba524c0880577f57f16b15b53c0"} Dec 10 23:07:28 crc kubenswrapper[4791]: I1210 23:07:28.051177 4791 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-85e7-account-create-update-25zns" event={"ID":"4d16f575-dc6d-4a78-8303-10a1c8202c0e","Type":"ContainerStarted","Data":"3d62758251e0debcf0a653e55bb0fd69be2e4ea107e55c5e49751a51aee18de6"} Dec 10 23:07:28 crc kubenswrapper[4791]: I1210 23:07:28.071473 4791 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/ovn-controller-xb29h" podUID="780fd8a7-cfb0-4958-afb3-1d060398d3ea" containerName="ovn-controller" probeResult="failure" output=< Dec 10 23:07:28 crc kubenswrapper[4791]: ERROR - ovn-controller connection status is 'not connected', expecting 'connected' status Dec 10 23:07:28 crc kubenswrapper[4791]: > Dec 10 23:07:28 crc kubenswrapper[4791]: I1210 23:07:28.091694 4791 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/ovn-controller-ovs-phsfg" Dec 10 23:07:28 crc kubenswrapper[4791]: I1210 23:07:28.194884 4791 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-073e-account-create-update-hwgzh"] Dec 10 23:07:28 crc kubenswrapper[4791]: W1210 23:07:28.203209 4791 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podaf2a4d0a_289c_4070_8371_2eca62ff0ca2.slice/crio-959b4e151949429dc188fc2c06b2730d278165ea048928e4418e5a44a812c54a WatchSource:0}: Error finding container 959b4e151949429dc188fc2c06b2730d278165ea048928e4418e5a44a812c54a: Status 404 returned error can't find the container with id 959b4e151949429dc188fc2c06b2730d278165ea048928e4418e5a44a812c54a Dec 10 23:07:29 crc kubenswrapper[4791]: I1210 23:07:29.058703 4791 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-073e-account-create-update-hwgzh" event={"ID":"af2a4d0a-289c-4070-8371-2eca62ff0ca2","Type":"ContainerStarted","Data":"959b4e151949429dc188fc2c06b2730d278165ea048928e4418e5a44a812c54a"} Dec 10 23:07:30 crc kubenswrapper[4791]: I1210 23:07:30.078965 4791 generic.go:334] "Generic (PLEG): container finished" podID="2c828334-1c90-4871-b203-71f82ac6fb8b" containerID="8203de309099603138b658f2b92d9f6938a0d58392a05ced22b78b6a5b6d1bcf" exitCode=0 Dec 10 23:07:30 crc kubenswrapper[4791]: I1210 23:07:30.079116 4791 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-492e-account-create-update-xfsh4" event={"ID":"2c828334-1c90-4871-b203-71f82ac6fb8b","Type":"ContainerDied","Data":"8203de309099603138b658f2b92d9f6938a0d58392a05ced22b78b6a5b6d1bcf"} Dec 10 23:07:30 crc kubenswrapper[4791]: I1210 23:07:30.080698 4791 generic.go:334] "Generic (PLEG): container finished" podID="af2a4d0a-289c-4070-8371-2eca62ff0ca2" containerID="b4d9076c09753643536c9ad794e20a5b314f2e20f13af8e7e2cc4329ea297990" exitCode=0 Dec 10 23:07:30 crc kubenswrapper[4791]: I1210 23:07:30.080736 4791 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-073e-account-create-update-hwgzh" event={"ID":"af2a4d0a-289c-4070-8371-2eca62ff0ca2","Type":"ContainerDied","Data":"b4d9076c09753643536c9ad794e20a5b314f2e20f13af8e7e2cc4329ea297990"} Dec 10 23:07:30 crc kubenswrapper[4791]: I1210 23:07:30.082023 4791 generic.go:334] "Generic (PLEG): container finished" podID="831d2f93-31c5-469f-b4bf-1dd85fa21cbf" containerID="eb263c1b91e1c38df8b56867ab31368ec0c95f2353b67c05eb17173caf26ea53" exitCode=0 Dec 10 23:07:30 crc kubenswrapper[4791]: I1210 23:07:30.082065 4791 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-db-create-g4swg" event={"ID":"831d2f93-31c5-469f-b4bf-1dd85fa21cbf","Type":"ContainerDied","Data":"eb263c1b91e1c38df8b56867ab31368ec0c95f2353b67c05eb17173caf26ea53"} Dec 10 23:07:30 crc kubenswrapper[4791]: I1210 23:07:30.085078 4791 generic.go:334] "Generic (PLEG): container finished" podID="2a8736f4-b93d-4bde-a270-0acb0c730ab0" containerID="74c66018017c1765bad98de000d0470b38bafa4c766ec9da4d05f8312538ab4b" exitCode=0 Dec 10 23:07:30 crc kubenswrapper[4791]: I1210 23:07:30.085120 4791 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-db-create-czcrm" event={"ID":"2a8736f4-b93d-4bde-a270-0acb0c730ab0","Type":"ContainerDied","Data":"74c66018017c1765bad98de000d0470b38bafa4c766ec9da4d05f8312538ab4b"} Dec 10 23:07:30 crc kubenswrapper[4791]: I1210 23:07:30.086973 4791 generic.go:334] "Generic (PLEG): container finished" podID="afcdc302-63c5-45c5-bf5b-46ef1caf953d" containerID="8030250db2495a83cbbd9faee36c9964e867a4e460c9bdb7b6cd4661b2928250" exitCode=0 Dec 10 23:07:30 crc kubenswrapper[4791]: I1210 23:07:30.087000 4791 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-db-create-qfvp4" event={"ID":"afcdc302-63c5-45c5-bf5b-46ef1caf953d","Type":"ContainerDied","Data":"8030250db2495a83cbbd9faee36c9964e867a4e460c9bdb7b6cd4661b2928250"} Dec 10 23:07:30 crc kubenswrapper[4791]: I1210 23:07:30.088275 4791 generic.go:334] "Generic (PLEG): container finished" podID="4d16f575-dc6d-4a78-8303-10a1c8202c0e" containerID="54e3ad356567141309d0c6aa74e1bb0becd794cdad3c11cd66345605af663fb3" exitCode=0 Dec 10 23:07:30 crc kubenswrapper[4791]: I1210 23:07:30.088294 4791 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-85e7-account-create-update-25zns" event={"ID":"4d16f575-dc6d-4a78-8303-10a1c8202c0e","Type":"ContainerDied","Data":"54e3ad356567141309d0c6aa74e1bb0becd794cdad3c11cd66345605af663fb3"} Dec 10 23:07:31 crc kubenswrapper[4791]: I1210 23:07:31.100179 4791 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"69a6c950-8a3b-4a7c-b284-ebd20157eb20","Type":"ContainerStarted","Data":"9fce4fd1c27499661b2789d7d27c2caffc25e79720ecfb3444558739b2c29f60"} Dec 10 23:07:31 crc kubenswrapper[4791]: I1210 23:07:31.626659 4791 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/placement-492e-account-create-update-xfsh4" Dec 10 23:07:31 crc kubenswrapper[4791]: I1210 23:07:31.728143 4791 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-db-create-qfvp4" Dec 10 23:07:31 crc kubenswrapper[4791]: I1210 23:07:31.742104 4791 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/placement-db-create-g4swg" Dec 10 23:07:31 crc kubenswrapper[4791]: I1210 23:07:31.792424 4791 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-mrrss\" (UniqueName: \"kubernetes.io/projected/831d2f93-31c5-469f-b4bf-1dd85fa21cbf-kube-api-access-mrrss\") pod \"831d2f93-31c5-469f-b4bf-1dd85fa21cbf\" (UID: \"831d2f93-31c5-469f-b4bf-1dd85fa21cbf\") " Dec 10 23:07:31 crc kubenswrapper[4791]: I1210 23:07:31.792580 4791 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/afcdc302-63c5-45c5-bf5b-46ef1caf953d-operator-scripts\") pod \"afcdc302-63c5-45c5-bf5b-46ef1caf953d\" (UID: \"afcdc302-63c5-45c5-bf5b-46ef1caf953d\") " Dec 10 23:07:31 crc kubenswrapper[4791]: I1210 23:07:31.792606 4791 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/2c828334-1c90-4871-b203-71f82ac6fb8b-operator-scripts\") pod \"2c828334-1c90-4871-b203-71f82ac6fb8b\" (UID: \"2c828334-1c90-4871-b203-71f82ac6fb8b\") " Dec 10 23:07:31 crc kubenswrapper[4791]: I1210 23:07:31.792693 4791 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-2x6pm\" (UniqueName: \"kubernetes.io/projected/2c828334-1c90-4871-b203-71f82ac6fb8b-kube-api-access-2x6pm\") pod \"2c828334-1c90-4871-b203-71f82ac6fb8b\" (UID: \"2c828334-1c90-4871-b203-71f82ac6fb8b\") " Dec 10 23:07:31 crc kubenswrapper[4791]: I1210 23:07:31.795005 4791 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/afcdc302-63c5-45c5-bf5b-46ef1caf953d-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "afcdc302-63c5-45c5-bf5b-46ef1caf953d" (UID: "afcdc302-63c5-45c5-bf5b-46ef1caf953d"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 10 23:07:31 crc kubenswrapper[4791]: I1210 23:07:31.796666 4791 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/2c828334-1c90-4871-b203-71f82ac6fb8b-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "2c828334-1c90-4871-b203-71f82ac6fb8b" (UID: "2c828334-1c90-4871-b203-71f82ac6fb8b"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 10 23:07:31 crc kubenswrapper[4791]: I1210 23:07:31.802206 4791 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/2c828334-1c90-4871-b203-71f82ac6fb8b-kube-api-access-2x6pm" (OuterVolumeSpecName: "kube-api-access-2x6pm") pod "2c828334-1c90-4871-b203-71f82ac6fb8b" (UID: "2c828334-1c90-4871-b203-71f82ac6fb8b"). InnerVolumeSpecName "kube-api-access-2x6pm". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 10 23:07:31 crc kubenswrapper[4791]: I1210 23:07:31.802270 4791 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/831d2f93-31c5-469f-b4bf-1dd85fa21cbf-kube-api-access-mrrss" (OuterVolumeSpecName: "kube-api-access-mrrss") pod "831d2f93-31c5-469f-b4bf-1dd85fa21cbf" (UID: "831d2f93-31c5-469f-b4bf-1dd85fa21cbf"). InnerVolumeSpecName "kube-api-access-mrrss". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 10 23:07:31 crc kubenswrapper[4791]: I1210 23:07:31.897860 4791 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-fdr7j\" (UniqueName: \"kubernetes.io/projected/afcdc302-63c5-45c5-bf5b-46ef1caf953d-kube-api-access-fdr7j\") pod \"afcdc302-63c5-45c5-bf5b-46ef1caf953d\" (UID: \"afcdc302-63c5-45c5-bf5b-46ef1caf953d\") " Dec 10 23:07:31 crc kubenswrapper[4791]: I1210 23:07:31.898165 4791 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/831d2f93-31c5-469f-b4bf-1dd85fa21cbf-operator-scripts\") pod \"831d2f93-31c5-469f-b4bf-1dd85fa21cbf\" (UID: \"831d2f93-31c5-469f-b4bf-1dd85fa21cbf\") " Dec 10 23:07:31 crc kubenswrapper[4791]: I1210 23:07:31.899915 4791 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/831d2f93-31c5-469f-b4bf-1dd85fa21cbf-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "831d2f93-31c5-469f-b4bf-1dd85fa21cbf" (UID: "831d2f93-31c5-469f-b4bf-1dd85fa21cbf"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 10 23:07:31 crc kubenswrapper[4791]: I1210 23:07:31.900045 4791 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-mrrss\" (UniqueName: \"kubernetes.io/projected/831d2f93-31c5-469f-b4bf-1dd85fa21cbf-kube-api-access-mrrss\") on node \"crc\" DevicePath \"\"" Dec 10 23:07:31 crc kubenswrapper[4791]: I1210 23:07:31.900353 4791 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/afcdc302-63c5-45c5-bf5b-46ef1caf953d-operator-scripts\") on node \"crc\" DevicePath \"\"" Dec 10 23:07:31 crc kubenswrapper[4791]: I1210 23:07:31.900656 4791 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/2c828334-1c90-4871-b203-71f82ac6fb8b-operator-scripts\") on node \"crc\" DevicePath \"\"" Dec 10 23:07:31 crc kubenswrapper[4791]: I1210 23:07:31.900677 4791 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-2x6pm\" (UniqueName: \"kubernetes.io/projected/2c828334-1c90-4871-b203-71f82ac6fb8b-kube-api-access-2x6pm\") on node \"crc\" DevicePath \"\"" Dec 10 23:07:31 crc kubenswrapper[4791]: I1210 23:07:31.915909 4791 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/afcdc302-63c5-45c5-bf5b-46ef1caf953d-kube-api-access-fdr7j" (OuterVolumeSpecName: "kube-api-access-fdr7j") pod "afcdc302-63c5-45c5-bf5b-46ef1caf953d" (UID: "afcdc302-63c5-45c5-bf5b-46ef1caf953d"). InnerVolumeSpecName "kube-api-access-fdr7j". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 10 23:07:32 crc kubenswrapper[4791]: I1210 23:07:32.003125 4791 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-fdr7j\" (UniqueName: \"kubernetes.io/projected/afcdc302-63c5-45c5-bf5b-46ef1caf953d-kube-api-access-fdr7j\") on node \"crc\" DevicePath \"\"" Dec 10 23:07:32 crc kubenswrapper[4791]: I1210 23:07:32.003163 4791 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/831d2f93-31c5-469f-b4bf-1dd85fa21cbf-operator-scripts\") on node \"crc\" DevicePath \"\"" Dec 10 23:07:32 crc kubenswrapper[4791]: I1210 23:07:32.029458 4791 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glance-db-create-czcrm" Dec 10 23:07:32 crc kubenswrapper[4791]: I1210 23:07:32.045604 4791 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glance-073e-account-create-update-hwgzh" Dec 10 23:07:32 crc kubenswrapper[4791]: I1210 23:07:32.086158 4791 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-85e7-account-create-update-25zns" Dec 10 23:07:32 crc kubenswrapper[4791]: I1210 23:07:32.149748 4791 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-db-create-qfvp4" event={"ID":"afcdc302-63c5-45c5-bf5b-46ef1caf953d","Type":"ContainerDied","Data":"28d71b866dbe2b9c19fbf7412eedf21fc9e09ba524c0880577f57f16b15b53c0"} Dec 10 23:07:32 crc kubenswrapper[4791]: I1210 23:07:32.150100 4791 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="28d71b866dbe2b9c19fbf7412eedf21fc9e09ba524c0880577f57f16b15b53c0" Dec 10 23:07:32 crc kubenswrapper[4791]: I1210 23:07:32.150191 4791 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-db-create-qfvp4" Dec 10 23:07:32 crc kubenswrapper[4791]: I1210 23:07:32.187474 4791 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-85e7-account-create-update-25zns" event={"ID":"4d16f575-dc6d-4a78-8303-10a1c8202c0e","Type":"ContainerDied","Data":"3d62758251e0debcf0a653e55bb0fd69be2e4ea107e55c5e49751a51aee18de6"} Dec 10 23:07:32 crc kubenswrapper[4791]: I1210 23:07:32.187543 4791 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="3d62758251e0debcf0a653e55bb0fd69be2e4ea107e55c5e49751a51aee18de6" Dec 10 23:07:32 crc kubenswrapper[4791]: I1210 23:07:32.187624 4791 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-85e7-account-create-update-25zns" Dec 10 23:07:32 crc kubenswrapper[4791]: I1210 23:07:32.190761 4791 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"69a6c950-8a3b-4a7c-b284-ebd20157eb20","Type":"ContainerStarted","Data":"d1b4c32111c4121cbcad7ec054feb65120bc7bb41c78df17283c2fab7a832e79"} Dec 10 23:07:32 crc kubenswrapper[4791]: I1210 23:07:32.190812 4791 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"69a6c950-8a3b-4a7c-b284-ebd20157eb20","Type":"ContainerStarted","Data":"ff11cd71b963119e0641bc6d6eb30fe554c80835a2185aa14638da8163f5ab87"} Dec 10 23:07:32 crc kubenswrapper[4791]: I1210 23:07:32.190823 4791 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"69a6c950-8a3b-4a7c-b284-ebd20157eb20","Type":"ContainerStarted","Data":"6f014e191570caf26bf313722d9b55ab595c499ca681e0187c2fe470b286d99e"} Dec 10 23:07:32 crc kubenswrapper[4791]: I1210 23:07:32.194191 4791 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-492e-account-create-update-xfsh4" event={"ID":"2c828334-1c90-4871-b203-71f82ac6fb8b","Type":"ContainerDied","Data":"694f186b422dfcc334f33065cc5745a051df5c9a91a7801b710f075b3020ca83"} Dec 10 23:07:32 crc kubenswrapper[4791]: I1210 23:07:32.194248 4791 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="694f186b422dfcc334f33065cc5745a051df5c9a91a7801b710f075b3020ca83" Dec 10 23:07:32 crc kubenswrapper[4791]: I1210 23:07:32.194329 4791 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/placement-492e-account-create-update-xfsh4" Dec 10 23:07:32 crc kubenswrapper[4791]: I1210 23:07:32.197215 4791 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-073e-account-create-update-hwgzh" event={"ID":"af2a4d0a-289c-4070-8371-2eca62ff0ca2","Type":"ContainerDied","Data":"959b4e151949429dc188fc2c06b2730d278165ea048928e4418e5a44a812c54a"} Dec 10 23:07:32 crc kubenswrapper[4791]: I1210 23:07:32.197248 4791 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="959b4e151949429dc188fc2c06b2730d278165ea048928e4418e5a44a812c54a" Dec 10 23:07:32 crc kubenswrapper[4791]: I1210 23:07:32.197307 4791 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glance-073e-account-create-update-hwgzh" Dec 10 23:07:32 crc kubenswrapper[4791]: I1210 23:07:32.208147 4791 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/2a8736f4-b93d-4bde-a270-0acb0c730ab0-operator-scripts\") pod \"2a8736f4-b93d-4bde-a270-0acb0c730ab0\" (UID: \"2a8736f4-b93d-4bde-a270-0acb0c730ab0\") " Dec 10 23:07:32 crc kubenswrapper[4791]: I1210 23:07:32.208280 4791 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/af2a4d0a-289c-4070-8371-2eca62ff0ca2-operator-scripts\") pod \"af2a4d0a-289c-4070-8371-2eca62ff0ca2\" (UID: \"af2a4d0a-289c-4070-8371-2eca62ff0ca2\") " Dec 10 23:07:32 crc kubenswrapper[4791]: I1210 23:07:32.208369 4791 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-c256w\" (UniqueName: \"kubernetes.io/projected/4d16f575-dc6d-4a78-8303-10a1c8202c0e-kube-api-access-c256w\") pod \"4d16f575-dc6d-4a78-8303-10a1c8202c0e\" (UID: \"4d16f575-dc6d-4a78-8303-10a1c8202c0e\") " Dec 10 23:07:32 crc kubenswrapper[4791]: I1210 23:07:32.208438 4791 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/4d16f575-dc6d-4a78-8303-10a1c8202c0e-operator-scripts\") pod \"4d16f575-dc6d-4a78-8303-10a1c8202c0e\" (UID: \"4d16f575-dc6d-4a78-8303-10a1c8202c0e\") " Dec 10 23:07:32 crc kubenswrapper[4791]: I1210 23:07:32.208479 4791 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-gcjkw\" (UniqueName: \"kubernetes.io/projected/2a8736f4-b93d-4bde-a270-0acb0c730ab0-kube-api-access-gcjkw\") pod \"2a8736f4-b93d-4bde-a270-0acb0c730ab0\" (UID: \"2a8736f4-b93d-4bde-a270-0acb0c730ab0\") " Dec 10 23:07:32 crc kubenswrapper[4791]: I1210 23:07:32.208565 4791 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-6bh9d\" (UniqueName: \"kubernetes.io/projected/af2a4d0a-289c-4070-8371-2eca62ff0ca2-kube-api-access-6bh9d\") pod \"af2a4d0a-289c-4070-8371-2eca62ff0ca2\" (UID: \"af2a4d0a-289c-4070-8371-2eca62ff0ca2\") " Dec 10 23:07:32 crc kubenswrapper[4791]: I1210 23:07:32.216833 4791 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/af2a4d0a-289c-4070-8371-2eca62ff0ca2-kube-api-access-6bh9d" (OuterVolumeSpecName: "kube-api-access-6bh9d") pod "af2a4d0a-289c-4070-8371-2eca62ff0ca2" (UID: "af2a4d0a-289c-4070-8371-2eca62ff0ca2"). InnerVolumeSpecName "kube-api-access-6bh9d". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 10 23:07:32 crc kubenswrapper[4791]: I1210 23:07:32.217724 4791 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/4d16f575-dc6d-4a78-8303-10a1c8202c0e-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "4d16f575-dc6d-4a78-8303-10a1c8202c0e" (UID: "4d16f575-dc6d-4a78-8303-10a1c8202c0e"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 10 23:07:32 crc kubenswrapper[4791]: I1210 23:07:32.218530 4791 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/af2a4d0a-289c-4070-8371-2eca62ff0ca2-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "af2a4d0a-289c-4070-8371-2eca62ff0ca2" (UID: "af2a4d0a-289c-4070-8371-2eca62ff0ca2"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 10 23:07:32 crc kubenswrapper[4791]: I1210 23:07:32.217727 4791 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/2a8736f4-b93d-4bde-a270-0acb0c730ab0-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "2a8736f4-b93d-4bde-a270-0acb0c730ab0" (UID: "2a8736f4-b93d-4bde-a270-0acb0c730ab0"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 10 23:07:32 crc kubenswrapper[4791]: I1210 23:07:32.221652 4791 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/2a8736f4-b93d-4bde-a270-0acb0c730ab0-kube-api-access-gcjkw" (OuterVolumeSpecName: "kube-api-access-gcjkw") pod "2a8736f4-b93d-4bde-a270-0acb0c730ab0" (UID: "2a8736f4-b93d-4bde-a270-0acb0c730ab0"). InnerVolumeSpecName "kube-api-access-gcjkw". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 10 23:07:32 crc kubenswrapper[4791]: I1210 23:07:32.221820 4791 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/4d16f575-dc6d-4a78-8303-10a1c8202c0e-kube-api-access-c256w" (OuterVolumeSpecName: "kube-api-access-c256w") pod "4d16f575-dc6d-4a78-8303-10a1c8202c0e" (UID: "4d16f575-dc6d-4a78-8303-10a1c8202c0e"). InnerVolumeSpecName "kube-api-access-c256w". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 10 23:07:32 crc kubenswrapper[4791]: I1210 23:07:32.249671 4791 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/placement-db-create-g4swg" Dec 10 23:07:32 crc kubenswrapper[4791]: I1210 23:07:32.249676 4791 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-db-create-g4swg" event={"ID":"831d2f93-31c5-469f-b4bf-1dd85fa21cbf","Type":"ContainerDied","Data":"01367db63ef39fc60b32e43607d7e30c8dd2d4b2b8d75e44b3a6f9136d0a57bf"} Dec 10 23:07:32 crc kubenswrapper[4791]: I1210 23:07:32.250140 4791 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="01367db63ef39fc60b32e43607d7e30c8dd2d4b2b8d75e44b3a6f9136d0a57bf" Dec 10 23:07:32 crc kubenswrapper[4791]: I1210 23:07:32.258995 4791 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-db-create-czcrm" event={"ID":"2a8736f4-b93d-4bde-a270-0acb0c730ab0","Type":"ContainerDied","Data":"ca37ac8add1d1e0a3327dd6a31102e8d73ca787d17a7a5bee8afd659b8a5e183"} Dec 10 23:07:32 crc kubenswrapper[4791]: I1210 23:07:32.259034 4791 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="ca37ac8add1d1e0a3327dd6a31102e8d73ca787d17a7a5bee8afd659b8a5e183" Dec 10 23:07:32 crc kubenswrapper[4791]: I1210 23:07:32.259083 4791 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glance-db-create-czcrm" Dec 10 23:07:32 crc kubenswrapper[4791]: I1210 23:07:32.310800 4791 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/2a8736f4-b93d-4bde-a270-0acb0c730ab0-operator-scripts\") on node \"crc\" DevicePath \"\"" Dec 10 23:07:32 crc kubenswrapper[4791]: I1210 23:07:32.310829 4791 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/af2a4d0a-289c-4070-8371-2eca62ff0ca2-operator-scripts\") on node \"crc\" DevicePath \"\"" Dec 10 23:07:32 crc kubenswrapper[4791]: I1210 23:07:32.310839 4791 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-c256w\" (UniqueName: \"kubernetes.io/projected/4d16f575-dc6d-4a78-8303-10a1c8202c0e-kube-api-access-c256w\") on node \"crc\" DevicePath \"\"" Dec 10 23:07:32 crc kubenswrapper[4791]: I1210 23:07:32.310852 4791 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/4d16f575-dc6d-4a78-8303-10a1c8202c0e-operator-scripts\") on node \"crc\" DevicePath \"\"" Dec 10 23:07:32 crc kubenswrapper[4791]: I1210 23:07:32.310861 4791 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-gcjkw\" (UniqueName: \"kubernetes.io/projected/2a8736f4-b93d-4bde-a270-0acb0c730ab0-kube-api-access-gcjkw\") on node \"crc\" DevicePath \"\"" Dec 10 23:07:32 crc kubenswrapper[4791]: I1210 23:07:32.310871 4791 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-6bh9d\" (UniqueName: \"kubernetes.io/projected/af2a4d0a-289c-4070-8371-2eca62ff0ca2-kube-api-access-6bh9d\") on node \"crc\" DevicePath \"\"" Dec 10 23:07:33 crc kubenswrapper[4791]: I1210 23:07:33.255822 4791 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/ovn-controller-xb29h" podUID="780fd8a7-cfb0-4958-afb3-1d060398d3ea" containerName="ovn-controller" probeResult="failure" output=< Dec 10 23:07:33 crc kubenswrapper[4791]: ERROR - ovn-controller connection status is 'not connected', expecting 'connected' status Dec 10 23:07:33 crc kubenswrapper[4791]: > Dec 10 23:07:33 crc kubenswrapper[4791]: I1210 23:07:33.260870 4791 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/ovn-controller-ovs-phsfg" Dec 10 23:07:33 crc kubenswrapper[4791]: I1210 23:07:33.480708 4791 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ovn-controller-xb29h-config-q4hjg"] Dec 10 23:07:33 crc kubenswrapper[4791]: E1210 23:07:33.481057 4791 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="4d16f575-dc6d-4a78-8303-10a1c8202c0e" containerName="mariadb-account-create-update" Dec 10 23:07:33 crc kubenswrapper[4791]: I1210 23:07:33.481073 4791 state_mem.go:107] "Deleted CPUSet assignment" podUID="4d16f575-dc6d-4a78-8303-10a1c8202c0e" containerName="mariadb-account-create-update" Dec 10 23:07:33 crc kubenswrapper[4791]: E1210 23:07:33.481080 4791 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="afcdc302-63c5-45c5-bf5b-46ef1caf953d" containerName="mariadb-database-create" Dec 10 23:07:33 crc kubenswrapper[4791]: I1210 23:07:33.481085 4791 state_mem.go:107] "Deleted CPUSet assignment" podUID="afcdc302-63c5-45c5-bf5b-46ef1caf953d" containerName="mariadb-database-create" Dec 10 23:07:33 crc kubenswrapper[4791]: E1210 23:07:33.481105 4791 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="af2a4d0a-289c-4070-8371-2eca62ff0ca2" containerName="mariadb-account-create-update" Dec 10 23:07:33 crc kubenswrapper[4791]: I1210 23:07:33.481111 4791 state_mem.go:107] "Deleted CPUSet assignment" podUID="af2a4d0a-289c-4070-8371-2eca62ff0ca2" containerName="mariadb-account-create-update" Dec 10 23:07:33 crc kubenswrapper[4791]: E1210 23:07:33.481123 4791 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="831d2f93-31c5-469f-b4bf-1dd85fa21cbf" containerName="mariadb-database-create" Dec 10 23:07:33 crc kubenswrapper[4791]: I1210 23:07:33.481129 4791 state_mem.go:107] "Deleted CPUSet assignment" podUID="831d2f93-31c5-469f-b4bf-1dd85fa21cbf" containerName="mariadb-database-create" Dec 10 23:07:33 crc kubenswrapper[4791]: E1210 23:07:33.481140 4791 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="2a8736f4-b93d-4bde-a270-0acb0c730ab0" containerName="mariadb-database-create" Dec 10 23:07:33 crc kubenswrapper[4791]: I1210 23:07:33.481145 4791 state_mem.go:107] "Deleted CPUSet assignment" podUID="2a8736f4-b93d-4bde-a270-0acb0c730ab0" containerName="mariadb-database-create" Dec 10 23:07:33 crc kubenswrapper[4791]: E1210 23:07:33.481159 4791 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="2c828334-1c90-4871-b203-71f82ac6fb8b" containerName="mariadb-account-create-update" Dec 10 23:07:33 crc kubenswrapper[4791]: I1210 23:07:33.481164 4791 state_mem.go:107] "Deleted CPUSet assignment" podUID="2c828334-1c90-4871-b203-71f82ac6fb8b" containerName="mariadb-account-create-update" Dec 10 23:07:33 crc kubenswrapper[4791]: I1210 23:07:33.481308 4791 memory_manager.go:354] "RemoveStaleState removing state" podUID="2a8736f4-b93d-4bde-a270-0acb0c730ab0" containerName="mariadb-database-create" Dec 10 23:07:33 crc kubenswrapper[4791]: I1210 23:07:33.481351 4791 memory_manager.go:354] "RemoveStaleState removing state" podUID="831d2f93-31c5-469f-b4bf-1dd85fa21cbf" containerName="mariadb-database-create" Dec 10 23:07:33 crc kubenswrapper[4791]: I1210 23:07:33.481371 4791 memory_manager.go:354] "RemoveStaleState removing state" podUID="af2a4d0a-289c-4070-8371-2eca62ff0ca2" containerName="mariadb-account-create-update" Dec 10 23:07:33 crc kubenswrapper[4791]: I1210 23:07:33.481395 4791 memory_manager.go:354] "RemoveStaleState removing state" podUID="2c828334-1c90-4871-b203-71f82ac6fb8b" containerName="mariadb-account-create-update" Dec 10 23:07:33 crc kubenswrapper[4791]: I1210 23:07:33.481410 4791 memory_manager.go:354] "RemoveStaleState removing state" podUID="4d16f575-dc6d-4a78-8303-10a1c8202c0e" containerName="mariadb-account-create-update" Dec 10 23:07:33 crc kubenswrapper[4791]: I1210 23:07:33.481426 4791 memory_manager.go:354] "RemoveStaleState removing state" podUID="afcdc302-63c5-45c5-bf5b-46ef1caf953d" containerName="mariadb-database-create" Dec 10 23:07:33 crc kubenswrapper[4791]: I1210 23:07:33.481943 4791 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-controller-xb29h-config-q4hjg" Dec 10 23:07:33 crc kubenswrapper[4791]: I1210 23:07:33.484772 4791 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"ovncontroller-extra-scripts" Dec 10 23:07:33 crc kubenswrapper[4791]: I1210 23:07:33.488725 4791 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovn-controller-xb29h-config-q4hjg"] Dec 10 23:07:33 crc kubenswrapper[4791]: I1210 23:07:33.541554 4791 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-run-ovn\" (UniqueName: \"kubernetes.io/host-path/d334ac02-5843-4c21-b123-8b3c8870f7b1-var-run-ovn\") pod \"ovn-controller-xb29h-config-q4hjg\" (UID: \"d334ac02-5843-4c21-b123-8b3c8870f7b1\") " pod="openstack/ovn-controller-xb29h-config-q4hjg" Dec 10 23:07:33 crc kubenswrapper[4791]: I1210 23:07:33.541623 4791 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"additional-scripts\" (UniqueName: \"kubernetes.io/configmap/d334ac02-5843-4c21-b123-8b3c8870f7b1-additional-scripts\") pod \"ovn-controller-xb29h-config-q4hjg\" (UID: \"d334ac02-5843-4c21-b123-8b3c8870f7b1\") " pod="openstack/ovn-controller-xb29h-config-q4hjg" Dec 10 23:07:33 crc kubenswrapper[4791]: I1210 23:07:33.541757 4791 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-log-ovn\" (UniqueName: \"kubernetes.io/host-path/d334ac02-5843-4c21-b123-8b3c8870f7b1-var-log-ovn\") pod \"ovn-controller-xb29h-config-q4hjg\" (UID: \"d334ac02-5843-4c21-b123-8b3c8870f7b1\") " pod="openstack/ovn-controller-xb29h-config-q4hjg" Dec 10 23:07:33 crc kubenswrapper[4791]: I1210 23:07:33.541849 4791 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-krxc7\" (UniqueName: \"kubernetes.io/projected/d334ac02-5843-4c21-b123-8b3c8870f7b1-kube-api-access-krxc7\") pod \"ovn-controller-xb29h-config-q4hjg\" (UID: \"d334ac02-5843-4c21-b123-8b3c8870f7b1\") " pod="openstack/ovn-controller-xb29h-config-q4hjg" Dec 10 23:07:33 crc kubenswrapper[4791]: I1210 23:07:33.541902 4791 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-run\" (UniqueName: \"kubernetes.io/host-path/d334ac02-5843-4c21-b123-8b3c8870f7b1-var-run\") pod \"ovn-controller-xb29h-config-q4hjg\" (UID: \"d334ac02-5843-4c21-b123-8b3c8870f7b1\") " pod="openstack/ovn-controller-xb29h-config-q4hjg" Dec 10 23:07:33 crc kubenswrapper[4791]: I1210 23:07:33.541926 4791 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/d334ac02-5843-4c21-b123-8b3c8870f7b1-scripts\") pod \"ovn-controller-xb29h-config-q4hjg\" (UID: \"d334ac02-5843-4c21-b123-8b3c8870f7b1\") " pod="openstack/ovn-controller-xb29h-config-q4hjg" Dec 10 23:07:33 crc kubenswrapper[4791]: I1210 23:07:33.642714 4791 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-run\" (UniqueName: \"kubernetes.io/host-path/d334ac02-5843-4c21-b123-8b3c8870f7b1-var-run\") pod \"ovn-controller-xb29h-config-q4hjg\" (UID: \"d334ac02-5843-4c21-b123-8b3c8870f7b1\") " pod="openstack/ovn-controller-xb29h-config-q4hjg" Dec 10 23:07:33 crc kubenswrapper[4791]: I1210 23:07:33.642765 4791 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/d334ac02-5843-4c21-b123-8b3c8870f7b1-scripts\") pod \"ovn-controller-xb29h-config-q4hjg\" (UID: \"d334ac02-5843-4c21-b123-8b3c8870f7b1\") " pod="openstack/ovn-controller-xb29h-config-q4hjg" Dec 10 23:07:33 crc kubenswrapper[4791]: I1210 23:07:33.642807 4791 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-run-ovn\" (UniqueName: \"kubernetes.io/host-path/d334ac02-5843-4c21-b123-8b3c8870f7b1-var-run-ovn\") pod \"ovn-controller-xb29h-config-q4hjg\" (UID: \"d334ac02-5843-4c21-b123-8b3c8870f7b1\") " pod="openstack/ovn-controller-xb29h-config-q4hjg" Dec 10 23:07:33 crc kubenswrapper[4791]: I1210 23:07:33.642867 4791 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"additional-scripts\" (UniqueName: \"kubernetes.io/configmap/d334ac02-5843-4c21-b123-8b3c8870f7b1-additional-scripts\") pod \"ovn-controller-xb29h-config-q4hjg\" (UID: \"d334ac02-5843-4c21-b123-8b3c8870f7b1\") " pod="openstack/ovn-controller-xb29h-config-q4hjg" Dec 10 23:07:33 crc kubenswrapper[4791]: I1210 23:07:33.642908 4791 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-log-ovn\" (UniqueName: \"kubernetes.io/host-path/d334ac02-5843-4c21-b123-8b3c8870f7b1-var-log-ovn\") pod \"ovn-controller-xb29h-config-q4hjg\" (UID: \"d334ac02-5843-4c21-b123-8b3c8870f7b1\") " pod="openstack/ovn-controller-xb29h-config-q4hjg" Dec 10 23:07:33 crc kubenswrapper[4791]: I1210 23:07:33.642956 4791 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-krxc7\" (UniqueName: \"kubernetes.io/projected/d334ac02-5843-4c21-b123-8b3c8870f7b1-kube-api-access-krxc7\") pod \"ovn-controller-xb29h-config-q4hjg\" (UID: \"d334ac02-5843-4c21-b123-8b3c8870f7b1\") " pod="openstack/ovn-controller-xb29h-config-q4hjg" Dec 10 23:07:33 crc kubenswrapper[4791]: I1210 23:07:33.643007 4791 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-run\" (UniqueName: \"kubernetes.io/host-path/d334ac02-5843-4c21-b123-8b3c8870f7b1-var-run\") pod \"ovn-controller-xb29h-config-q4hjg\" (UID: \"d334ac02-5843-4c21-b123-8b3c8870f7b1\") " pod="openstack/ovn-controller-xb29h-config-q4hjg" Dec 10 23:07:33 crc kubenswrapper[4791]: I1210 23:07:33.643047 4791 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-run-ovn\" (UniqueName: \"kubernetes.io/host-path/d334ac02-5843-4c21-b123-8b3c8870f7b1-var-run-ovn\") pod \"ovn-controller-xb29h-config-q4hjg\" (UID: \"d334ac02-5843-4c21-b123-8b3c8870f7b1\") " pod="openstack/ovn-controller-xb29h-config-q4hjg" Dec 10 23:07:33 crc kubenswrapper[4791]: I1210 23:07:33.643170 4791 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-log-ovn\" (UniqueName: \"kubernetes.io/host-path/d334ac02-5843-4c21-b123-8b3c8870f7b1-var-log-ovn\") pod \"ovn-controller-xb29h-config-q4hjg\" (UID: \"d334ac02-5843-4c21-b123-8b3c8870f7b1\") " pod="openstack/ovn-controller-xb29h-config-q4hjg" Dec 10 23:07:33 crc kubenswrapper[4791]: I1210 23:07:33.643952 4791 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"additional-scripts\" (UniqueName: \"kubernetes.io/configmap/d334ac02-5843-4c21-b123-8b3c8870f7b1-additional-scripts\") pod \"ovn-controller-xb29h-config-q4hjg\" (UID: \"d334ac02-5843-4c21-b123-8b3c8870f7b1\") " pod="openstack/ovn-controller-xb29h-config-q4hjg" Dec 10 23:07:33 crc kubenswrapper[4791]: I1210 23:07:33.645734 4791 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/d334ac02-5843-4c21-b123-8b3c8870f7b1-scripts\") pod \"ovn-controller-xb29h-config-q4hjg\" (UID: \"d334ac02-5843-4c21-b123-8b3c8870f7b1\") " pod="openstack/ovn-controller-xb29h-config-q4hjg" Dec 10 23:07:33 crc kubenswrapper[4791]: I1210 23:07:33.672902 4791 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-krxc7\" (UniqueName: \"kubernetes.io/projected/d334ac02-5843-4c21-b123-8b3c8870f7b1-kube-api-access-krxc7\") pod \"ovn-controller-xb29h-config-q4hjg\" (UID: \"d334ac02-5843-4c21-b123-8b3c8870f7b1\") " pod="openstack/ovn-controller-xb29h-config-q4hjg" Dec 10 23:07:33 crc kubenswrapper[4791]: I1210 23:07:33.802628 4791 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-controller-xb29h-config-q4hjg" Dec 10 23:07:34 crc kubenswrapper[4791]: I1210 23:07:34.302268 4791 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"69a6c950-8a3b-4a7c-b284-ebd20157eb20","Type":"ContainerStarted","Data":"4fdd8665f2ece11da403870646f16502aa04c9b328f1dcb99f0d772287f11ceb"} Dec 10 23:07:34 crc kubenswrapper[4791]: I1210 23:07:34.302557 4791 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"69a6c950-8a3b-4a7c-b284-ebd20157eb20","Type":"ContainerStarted","Data":"9c43cfea1e8e2e1828ab6ae55b9435d8c53c543c76cedf3c47e45937aa8e7375"} Dec 10 23:07:34 crc kubenswrapper[4791]: I1210 23:07:34.302568 4791 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"69a6c950-8a3b-4a7c-b284-ebd20157eb20","Type":"ContainerStarted","Data":"aa2f61d2112ea9a9d580772fba6cc5dea7b1cd81ca4913f8ab86c64d6f2c0b77"} Dec 10 23:07:34 crc kubenswrapper[4791]: I1210 23:07:34.614314 4791 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovn-controller-xb29h-config-q4hjg"] Dec 10 23:07:35 crc kubenswrapper[4791]: I1210 23:07:35.320671 4791 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"69a6c950-8a3b-4a7c-b284-ebd20157eb20","Type":"ContainerStarted","Data":"114a8137ad2bd0132da6f3018d32b1a8cbc8fb698eda8c1fad8014f0b831185d"} Dec 10 23:07:35 crc kubenswrapper[4791]: I1210 23:07:35.322744 4791 generic.go:334] "Generic (PLEG): container finished" podID="d334ac02-5843-4c21-b123-8b3c8870f7b1" containerID="f4aaa5aed3505e556bbf35f76215681ffa6ce51d231f26498e8185599e152873" exitCode=0 Dec 10 23:07:35 crc kubenswrapper[4791]: I1210 23:07:35.322798 4791 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-controller-xb29h-config-q4hjg" event={"ID":"d334ac02-5843-4c21-b123-8b3c8870f7b1","Type":"ContainerDied","Data":"f4aaa5aed3505e556bbf35f76215681ffa6ce51d231f26498e8185599e152873"} Dec 10 23:07:35 crc kubenswrapper[4791]: I1210 23:07:35.322826 4791 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-controller-xb29h-config-q4hjg" event={"ID":"d334ac02-5843-4c21-b123-8b3c8870f7b1","Type":"ContainerStarted","Data":"39933e5bdeaf6f0c0474126d70c8321ae2f5270033d3afd6b71ad0700a7f3432"} Dec 10 23:07:36 crc kubenswrapper[4791]: I1210 23:07:36.331130 4791 generic.go:334] "Generic (PLEG): container finished" podID="47eb9200-b0bc-41a6-abb9-f6167bd69c66" containerID="f0d087440ba482240de855bdc96648601e34f74962b3bf2123c4205414310fef" exitCode=0 Dec 10 23:07:36 crc kubenswrapper[4791]: I1210 23:07:36.331283 4791 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-cell1-server-0" event={"ID":"47eb9200-b0bc-41a6-abb9-f6167bd69c66","Type":"ContainerDied","Data":"f0d087440ba482240de855bdc96648601e34f74962b3bf2123c4205414310fef"} Dec 10 23:07:36 crc kubenswrapper[4791]: I1210 23:07:36.731190 4791 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-controller-xb29h-config-q4hjg" Dec 10 23:07:36 crc kubenswrapper[4791]: I1210 23:07:36.837475 4791 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"var-log-ovn\" (UniqueName: \"kubernetes.io/host-path/d334ac02-5843-4c21-b123-8b3c8870f7b1-var-log-ovn\") pod \"d334ac02-5843-4c21-b123-8b3c8870f7b1\" (UID: \"d334ac02-5843-4c21-b123-8b3c8870f7b1\") " Dec 10 23:07:36 crc kubenswrapper[4791]: I1210 23:07:36.837922 4791 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-krxc7\" (UniqueName: \"kubernetes.io/projected/d334ac02-5843-4c21-b123-8b3c8870f7b1-kube-api-access-krxc7\") pod \"d334ac02-5843-4c21-b123-8b3c8870f7b1\" (UID: \"d334ac02-5843-4c21-b123-8b3c8870f7b1\") " Dec 10 23:07:36 crc kubenswrapper[4791]: I1210 23:07:36.838061 4791 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/d334ac02-5843-4c21-b123-8b3c8870f7b1-scripts\") pod \"d334ac02-5843-4c21-b123-8b3c8870f7b1\" (UID: \"d334ac02-5843-4c21-b123-8b3c8870f7b1\") " Dec 10 23:07:36 crc kubenswrapper[4791]: I1210 23:07:36.838236 4791 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"additional-scripts\" (UniqueName: \"kubernetes.io/configmap/d334ac02-5843-4c21-b123-8b3c8870f7b1-additional-scripts\") pod \"d334ac02-5843-4c21-b123-8b3c8870f7b1\" (UID: \"d334ac02-5843-4c21-b123-8b3c8870f7b1\") " Dec 10 23:07:36 crc kubenswrapper[4791]: I1210 23:07:36.838373 4791 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"var-run\" (UniqueName: \"kubernetes.io/host-path/d334ac02-5843-4c21-b123-8b3c8870f7b1-var-run\") pod \"d334ac02-5843-4c21-b123-8b3c8870f7b1\" (UID: \"d334ac02-5843-4c21-b123-8b3c8870f7b1\") " Dec 10 23:07:36 crc kubenswrapper[4791]: I1210 23:07:36.837616 4791 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/d334ac02-5843-4c21-b123-8b3c8870f7b1-var-log-ovn" (OuterVolumeSpecName: "var-log-ovn") pod "d334ac02-5843-4c21-b123-8b3c8870f7b1" (UID: "d334ac02-5843-4c21-b123-8b3c8870f7b1"). InnerVolumeSpecName "var-log-ovn". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 10 23:07:36 crc kubenswrapper[4791]: I1210 23:07:36.838462 4791 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/d334ac02-5843-4c21-b123-8b3c8870f7b1-var-run" (OuterVolumeSpecName: "var-run") pod "d334ac02-5843-4c21-b123-8b3c8870f7b1" (UID: "d334ac02-5843-4c21-b123-8b3c8870f7b1"). InnerVolumeSpecName "var-run". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 10 23:07:36 crc kubenswrapper[4791]: I1210 23:07:36.838742 4791 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"var-run-ovn\" (UniqueName: \"kubernetes.io/host-path/d334ac02-5843-4c21-b123-8b3c8870f7b1-var-run-ovn\") pod \"d334ac02-5843-4c21-b123-8b3c8870f7b1\" (UID: \"d334ac02-5843-4c21-b123-8b3c8870f7b1\") " Dec 10 23:07:36 crc kubenswrapper[4791]: I1210 23:07:36.838799 4791 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/d334ac02-5843-4c21-b123-8b3c8870f7b1-var-run-ovn" (OuterVolumeSpecName: "var-run-ovn") pod "d334ac02-5843-4c21-b123-8b3c8870f7b1" (UID: "d334ac02-5843-4c21-b123-8b3c8870f7b1"). InnerVolumeSpecName "var-run-ovn". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 10 23:07:36 crc kubenswrapper[4791]: I1210 23:07:36.838821 4791 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/d334ac02-5843-4c21-b123-8b3c8870f7b1-additional-scripts" (OuterVolumeSpecName: "additional-scripts") pod "d334ac02-5843-4c21-b123-8b3c8870f7b1" (UID: "d334ac02-5843-4c21-b123-8b3c8870f7b1"). InnerVolumeSpecName "additional-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 10 23:07:36 crc kubenswrapper[4791]: I1210 23:07:36.838973 4791 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/d334ac02-5843-4c21-b123-8b3c8870f7b1-scripts" (OuterVolumeSpecName: "scripts") pod "d334ac02-5843-4c21-b123-8b3c8870f7b1" (UID: "d334ac02-5843-4c21-b123-8b3c8870f7b1"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 10 23:07:36 crc kubenswrapper[4791]: I1210 23:07:36.839445 4791 reconciler_common.go:293] "Volume detached for volume \"var-log-ovn\" (UniqueName: \"kubernetes.io/host-path/d334ac02-5843-4c21-b123-8b3c8870f7b1-var-log-ovn\") on node \"crc\" DevicePath \"\"" Dec 10 23:07:36 crc kubenswrapper[4791]: I1210 23:07:36.839552 4791 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/d334ac02-5843-4c21-b123-8b3c8870f7b1-scripts\") on node \"crc\" DevicePath \"\"" Dec 10 23:07:36 crc kubenswrapper[4791]: I1210 23:07:36.839640 4791 reconciler_common.go:293] "Volume detached for volume \"additional-scripts\" (UniqueName: \"kubernetes.io/configmap/d334ac02-5843-4c21-b123-8b3c8870f7b1-additional-scripts\") on node \"crc\" DevicePath \"\"" Dec 10 23:07:36 crc kubenswrapper[4791]: I1210 23:07:36.839719 4791 reconciler_common.go:293] "Volume detached for volume \"var-run\" (UniqueName: \"kubernetes.io/host-path/d334ac02-5843-4c21-b123-8b3c8870f7b1-var-run\") on node \"crc\" DevicePath \"\"" Dec 10 23:07:36 crc kubenswrapper[4791]: I1210 23:07:36.839803 4791 reconciler_common.go:293] "Volume detached for volume \"var-run-ovn\" (UniqueName: \"kubernetes.io/host-path/d334ac02-5843-4c21-b123-8b3c8870f7b1-var-run-ovn\") on node \"crc\" DevicePath \"\"" Dec 10 23:07:36 crc kubenswrapper[4791]: I1210 23:07:36.843520 4791 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/d334ac02-5843-4c21-b123-8b3c8870f7b1-kube-api-access-krxc7" (OuterVolumeSpecName: "kube-api-access-krxc7") pod "d334ac02-5843-4c21-b123-8b3c8870f7b1" (UID: "d334ac02-5843-4c21-b123-8b3c8870f7b1"). InnerVolumeSpecName "kube-api-access-krxc7". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 10 23:07:36 crc kubenswrapper[4791]: I1210 23:07:36.941017 4791 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-krxc7\" (UniqueName: \"kubernetes.io/projected/d334ac02-5843-4c21-b123-8b3c8870f7b1-kube-api-access-krxc7\") on node \"crc\" DevicePath \"\"" Dec 10 23:07:37 crc kubenswrapper[4791]: I1210 23:07:37.375429 4791 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-cell1-server-0" event={"ID":"47eb9200-b0bc-41a6-abb9-f6167bd69c66","Type":"ContainerStarted","Data":"8dd3007abca59274d70261742380418ebbe002bfd3c95f64e21f07dcfab98397"} Dec 10 23:07:37 crc kubenswrapper[4791]: I1210 23:07:37.375674 4791 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/rabbitmq-cell1-server-0" Dec 10 23:07:37 crc kubenswrapper[4791]: I1210 23:07:37.377375 4791 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-controller-xb29h-config-q4hjg" event={"ID":"d334ac02-5843-4c21-b123-8b3c8870f7b1","Type":"ContainerDied","Data":"39933e5bdeaf6f0c0474126d70c8321ae2f5270033d3afd6b71ad0700a7f3432"} Dec 10 23:07:37 crc kubenswrapper[4791]: I1210 23:07:37.377404 4791 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-controller-xb29h-config-q4hjg" Dec 10 23:07:37 crc kubenswrapper[4791]: I1210 23:07:37.377413 4791 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="39933e5bdeaf6f0c0474126d70c8321ae2f5270033d3afd6b71ad0700a7f3432" Dec 10 23:07:37 crc kubenswrapper[4791]: I1210 23:07:37.381453 4791 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"69a6c950-8a3b-4a7c-b284-ebd20157eb20","Type":"ContainerStarted","Data":"12a1cca2587c95630efdd69e0a30aa831169d9fddb97957b551f96ae1609a401"} Dec 10 23:07:37 crc kubenswrapper[4791]: I1210 23:07:37.381491 4791 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"69a6c950-8a3b-4a7c-b284-ebd20157eb20","Type":"ContainerStarted","Data":"a6b8fa38ef159bfd2127ebbcbc7698d7b75c11929a4b7ec3f6e864692e11e63a"} Dec 10 23:07:37 crc kubenswrapper[4791]: I1210 23:07:37.381502 4791 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"69a6c950-8a3b-4a7c-b284-ebd20157eb20","Type":"ContainerStarted","Data":"d3059081130b93ba2cfa072f054bfa821cac5e2e51ec6b440c2b107c26200163"} Dec 10 23:07:37 crc kubenswrapper[4791]: I1210 23:07:37.399301 4791 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/rabbitmq-cell1-server-0" podStartSLOduration=38.237441477 podStartE2EDuration="1m15.39928268s" podCreationTimestamp="2025-12-10 23:06:22 +0000 UTC" firstStartedPulling="2025-12-10 23:06:25.394595037 +0000 UTC m=+1019.824212650" lastFinishedPulling="2025-12-10 23:07:02.55643624 +0000 UTC m=+1056.986053853" observedRunningTime="2025-12-10 23:07:37.395301925 +0000 UTC m=+1091.824919558" watchObservedRunningTime="2025-12-10 23:07:37.39928268 +0000 UTC m=+1091.828900293" Dec 10 23:07:37 crc kubenswrapper[4791]: I1210 23:07:37.521621 4791 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/glance-db-sync-ccl4n"] Dec 10 23:07:37 crc kubenswrapper[4791]: E1210 23:07:37.521950 4791 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="d334ac02-5843-4c21-b123-8b3c8870f7b1" containerName="ovn-config" Dec 10 23:07:37 crc kubenswrapper[4791]: I1210 23:07:37.521966 4791 state_mem.go:107] "Deleted CPUSet assignment" podUID="d334ac02-5843-4c21-b123-8b3c8870f7b1" containerName="ovn-config" Dec 10 23:07:37 crc kubenswrapper[4791]: I1210 23:07:37.522121 4791 memory_manager.go:354] "RemoveStaleState removing state" podUID="d334ac02-5843-4c21-b123-8b3c8870f7b1" containerName="ovn-config" Dec 10 23:07:37 crc kubenswrapper[4791]: I1210 23:07:37.522682 4791 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-db-sync-ccl4n" Dec 10 23:07:37 crc kubenswrapper[4791]: I1210 23:07:37.524939 4791 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"glance-config-data" Dec 10 23:07:37 crc kubenswrapper[4791]: I1210 23:07:37.525183 4791 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"glance-glance-dockercfg-rsl4w" Dec 10 23:07:37 crc kubenswrapper[4791]: I1210 23:07:37.532136 4791 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-db-sync-ccl4n"] Dec 10 23:07:37 crc kubenswrapper[4791]: I1210 23:07:37.650650 4791 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-v7lq8\" (UniqueName: \"kubernetes.io/projected/a2c44c96-70c2-414a-9890-cb95b66f10b2-kube-api-access-v7lq8\") pod \"glance-db-sync-ccl4n\" (UID: \"a2c44c96-70c2-414a-9890-cb95b66f10b2\") " pod="openstack/glance-db-sync-ccl4n" Dec 10 23:07:37 crc kubenswrapper[4791]: I1210 23:07:37.650759 4791 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/a2c44c96-70c2-414a-9890-cb95b66f10b2-config-data\") pod \"glance-db-sync-ccl4n\" (UID: \"a2c44c96-70c2-414a-9890-cb95b66f10b2\") " pod="openstack/glance-db-sync-ccl4n" Dec 10 23:07:37 crc kubenswrapper[4791]: I1210 23:07:37.650783 4791 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a2c44c96-70c2-414a-9890-cb95b66f10b2-combined-ca-bundle\") pod \"glance-db-sync-ccl4n\" (UID: \"a2c44c96-70c2-414a-9890-cb95b66f10b2\") " pod="openstack/glance-db-sync-ccl4n" Dec 10 23:07:37 crc kubenswrapper[4791]: I1210 23:07:37.651049 4791 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/a2c44c96-70c2-414a-9890-cb95b66f10b2-db-sync-config-data\") pod \"glance-db-sync-ccl4n\" (UID: \"a2c44c96-70c2-414a-9890-cb95b66f10b2\") " pod="openstack/glance-db-sync-ccl4n" Dec 10 23:07:37 crc kubenswrapper[4791]: I1210 23:07:37.752972 4791 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/a2c44c96-70c2-414a-9890-cb95b66f10b2-config-data\") pod \"glance-db-sync-ccl4n\" (UID: \"a2c44c96-70c2-414a-9890-cb95b66f10b2\") " pod="openstack/glance-db-sync-ccl4n" Dec 10 23:07:37 crc kubenswrapper[4791]: I1210 23:07:37.753022 4791 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a2c44c96-70c2-414a-9890-cb95b66f10b2-combined-ca-bundle\") pod \"glance-db-sync-ccl4n\" (UID: \"a2c44c96-70c2-414a-9890-cb95b66f10b2\") " pod="openstack/glance-db-sync-ccl4n" Dec 10 23:07:37 crc kubenswrapper[4791]: I1210 23:07:37.753120 4791 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/a2c44c96-70c2-414a-9890-cb95b66f10b2-db-sync-config-data\") pod \"glance-db-sync-ccl4n\" (UID: \"a2c44c96-70c2-414a-9890-cb95b66f10b2\") " pod="openstack/glance-db-sync-ccl4n" Dec 10 23:07:37 crc kubenswrapper[4791]: I1210 23:07:37.753182 4791 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-v7lq8\" (UniqueName: \"kubernetes.io/projected/a2c44c96-70c2-414a-9890-cb95b66f10b2-kube-api-access-v7lq8\") pod \"glance-db-sync-ccl4n\" (UID: \"a2c44c96-70c2-414a-9890-cb95b66f10b2\") " pod="openstack/glance-db-sync-ccl4n" Dec 10 23:07:37 crc kubenswrapper[4791]: I1210 23:07:37.765689 4791 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a2c44c96-70c2-414a-9890-cb95b66f10b2-combined-ca-bundle\") pod \"glance-db-sync-ccl4n\" (UID: \"a2c44c96-70c2-414a-9890-cb95b66f10b2\") " pod="openstack/glance-db-sync-ccl4n" Dec 10 23:07:37 crc kubenswrapper[4791]: I1210 23:07:37.766162 4791 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/a2c44c96-70c2-414a-9890-cb95b66f10b2-db-sync-config-data\") pod \"glance-db-sync-ccl4n\" (UID: \"a2c44c96-70c2-414a-9890-cb95b66f10b2\") " pod="openstack/glance-db-sync-ccl4n" Dec 10 23:07:37 crc kubenswrapper[4791]: I1210 23:07:37.766562 4791 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/a2c44c96-70c2-414a-9890-cb95b66f10b2-config-data\") pod \"glance-db-sync-ccl4n\" (UID: \"a2c44c96-70c2-414a-9890-cb95b66f10b2\") " pod="openstack/glance-db-sync-ccl4n" Dec 10 23:07:37 crc kubenswrapper[4791]: I1210 23:07:37.793054 4791 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-v7lq8\" (UniqueName: \"kubernetes.io/projected/a2c44c96-70c2-414a-9890-cb95b66f10b2-kube-api-access-v7lq8\") pod \"glance-db-sync-ccl4n\" (UID: \"a2c44c96-70c2-414a-9890-cb95b66f10b2\") " pod="openstack/glance-db-sync-ccl4n" Dec 10 23:07:37 crc kubenswrapper[4791]: I1210 23:07:37.849628 4791 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-db-sync-ccl4n" Dec 10 23:07:37 crc kubenswrapper[4791]: I1210 23:07:37.880811 4791 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ovn-controller-xb29h-config-q4hjg"] Dec 10 23:07:37 crc kubenswrapper[4791]: I1210 23:07:37.909666 4791 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/ovn-controller-xb29h-config-q4hjg"] Dec 10 23:07:38 crc kubenswrapper[4791]: I1210 23:07:38.227651 4791 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/ovn-controller-xb29h" Dec 10 23:07:38 crc kubenswrapper[4791]: I1210 23:07:38.400756 4791 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"69a6c950-8a3b-4a7c-b284-ebd20157eb20","Type":"ContainerStarted","Data":"baec4e32c03e224189f5de2b652197fc826768ca1891854945efedffa78a2836"} Dec 10 23:07:38 crc kubenswrapper[4791]: I1210 23:07:38.400795 4791 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"69a6c950-8a3b-4a7c-b284-ebd20157eb20","Type":"ContainerStarted","Data":"eb9e771a3788f335d2ef0aceb4976819810258a2312b912f56d19c4f653b5d2c"} Dec 10 23:07:38 crc kubenswrapper[4791]: I1210 23:07:38.400807 4791 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"69a6c950-8a3b-4a7c-b284-ebd20157eb20","Type":"ContainerStarted","Data":"c0166ef050d9120838e08722341ffd4e26ecdc01c5cd23a56a9a6810d804bcf4"} Dec 10 23:07:38 crc kubenswrapper[4791]: I1210 23:07:38.668420 4791 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-db-sync-ccl4n"] Dec 10 23:07:39 crc kubenswrapper[4791]: I1210 23:07:39.412256 4791 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"69a6c950-8a3b-4a7c-b284-ebd20157eb20","Type":"ContainerStarted","Data":"1d33797f95f605b3c6cf17090c44a0d4fb3a9530e39aca2edb3886ab75bc7761"} Dec 10 23:07:39 crc kubenswrapper[4791]: I1210 23:07:39.414701 4791 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-db-sync-ccl4n" event={"ID":"a2c44c96-70c2-414a-9890-cb95b66f10b2","Type":"ContainerStarted","Data":"c32c4e4e6da6ca2408460931b7bc5e7926a5bf56f0c2b87506e93a5fb695ed2c"} Dec 10 23:07:39 crc kubenswrapper[4791]: I1210 23:07:39.463067 4791 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/swift-storage-0" podStartSLOduration=21.794942562 podStartE2EDuration="31.463049512s" podCreationTimestamp="2025-12-10 23:07:08 +0000 UTC" firstStartedPulling="2025-12-10 23:07:26.859405409 +0000 UTC m=+1081.289023022" lastFinishedPulling="2025-12-10 23:07:36.527512339 +0000 UTC m=+1090.957129972" observedRunningTime="2025-12-10 23:07:39.459753806 +0000 UTC m=+1093.889371419" watchObservedRunningTime="2025-12-10 23:07:39.463049512 +0000 UTC m=+1093.892667125" Dec 10 23:07:39 crc kubenswrapper[4791]: I1210 23:07:39.732440 4791 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-77585f5f8c-drnrs"] Dec 10 23:07:39 crc kubenswrapper[4791]: I1210 23:07:39.734054 4791 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-77585f5f8c-drnrs" Dec 10 23:07:39 crc kubenswrapper[4791]: I1210 23:07:39.737763 4791 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"dns-swift-storage-0" Dec 10 23:07:39 crc kubenswrapper[4791]: I1210 23:07:39.747064 4791 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-77585f5f8c-drnrs"] Dec 10 23:07:39 crc kubenswrapper[4791]: I1210 23:07:39.834480 4791 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/defe4ef9-9037-4a53-a860-ce6713ee219c-config\") pod \"dnsmasq-dns-77585f5f8c-drnrs\" (UID: \"defe4ef9-9037-4a53-a860-ce6713ee219c\") " pod="openstack/dnsmasq-dns-77585f5f8c-drnrs" Dec 10 23:07:39 crc kubenswrapper[4791]: I1210 23:07:39.834532 4791 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-9s79j\" (UniqueName: \"kubernetes.io/projected/defe4ef9-9037-4a53-a860-ce6713ee219c-kube-api-access-9s79j\") pod \"dnsmasq-dns-77585f5f8c-drnrs\" (UID: \"defe4ef9-9037-4a53-a860-ce6713ee219c\") " pod="openstack/dnsmasq-dns-77585f5f8c-drnrs" Dec 10 23:07:39 crc kubenswrapper[4791]: I1210 23:07:39.834555 4791 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/defe4ef9-9037-4a53-a860-ce6713ee219c-dns-swift-storage-0\") pod \"dnsmasq-dns-77585f5f8c-drnrs\" (UID: \"defe4ef9-9037-4a53-a860-ce6713ee219c\") " pod="openstack/dnsmasq-dns-77585f5f8c-drnrs" Dec 10 23:07:39 crc kubenswrapper[4791]: I1210 23:07:39.834660 4791 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/defe4ef9-9037-4a53-a860-ce6713ee219c-dns-svc\") pod \"dnsmasq-dns-77585f5f8c-drnrs\" (UID: \"defe4ef9-9037-4a53-a860-ce6713ee219c\") " pod="openstack/dnsmasq-dns-77585f5f8c-drnrs" Dec 10 23:07:39 crc kubenswrapper[4791]: I1210 23:07:39.834682 4791 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/defe4ef9-9037-4a53-a860-ce6713ee219c-ovsdbserver-nb\") pod \"dnsmasq-dns-77585f5f8c-drnrs\" (UID: \"defe4ef9-9037-4a53-a860-ce6713ee219c\") " pod="openstack/dnsmasq-dns-77585f5f8c-drnrs" Dec 10 23:07:39 crc kubenswrapper[4791]: I1210 23:07:39.834702 4791 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/defe4ef9-9037-4a53-a860-ce6713ee219c-ovsdbserver-sb\") pod \"dnsmasq-dns-77585f5f8c-drnrs\" (UID: \"defe4ef9-9037-4a53-a860-ce6713ee219c\") " pod="openstack/dnsmasq-dns-77585f5f8c-drnrs" Dec 10 23:07:39 crc kubenswrapper[4791]: I1210 23:07:39.898298 4791 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="d334ac02-5843-4c21-b123-8b3c8870f7b1" path="/var/lib/kubelet/pods/d334ac02-5843-4c21-b123-8b3c8870f7b1/volumes" Dec 10 23:07:39 crc kubenswrapper[4791]: I1210 23:07:39.936291 4791 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-9s79j\" (UniqueName: \"kubernetes.io/projected/defe4ef9-9037-4a53-a860-ce6713ee219c-kube-api-access-9s79j\") pod \"dnsmasq-dns-77585f5f8c-drnrs\" (UID: \"defe4ef9-9037-4a53-a860-ce6713ee219c\") " pod="openstack/dnsmasq-dns-77585f5f8c-drnrs" Dec 10 23:07:39 crc kubenswrapper[4791]: I1210 23:07:39.936600 4791 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/defe4ef9-9037-4a53-a860-ce6713ee219c-dns-swift-storage-0\") pod \"dnsmasq-dns-77585f5f8c-drnrs\" (UID: \"defe4ef9-9037-4a53-a860-ce6713ee219c\") " pod="openstack/dnsmasq-dns-77585f5f8c-drnrs" Dec 10 23:07:39 crc kubenswrapper[4791]: I1210 23:07:39.936758 4791 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/defe4ef9-9037-4a53-a860-ce6713ee219c-dns-svc\") pod \"dnsmasq-dns-77585f5f8c-drnrs\" (UID: \"defe4ef9-9037-4a53-a860-ce6713ee219c\") " pod="openstack/dnsmasq-dns-77585f5f8c-drnrs" Dec 10 23:07:39 crc kubenswrapper[4791]: I1210 23:07:39.936792 4791 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/defe4ef9-9037-4a53-a860-ce6713ee219c-ovsdbserver-nb\") pod \"dnsmasq-dns-77585f5f8c-drnrs\" (UID: \"defe4ef9-9037-4a53-a860-ce6713ee219c\") " pod="openstack/dnsmasq-dns-77585f5f8c-drnrs" Dec 10 23:07:39 crc kubenswrapper[4791]: I1210 23:07:39.936856 4791 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/defe4ef9-9037-4a53-a860-ce6713ee219c-ovsdbserver-sb\") pod \"dnsmasq-dns-77585f5f8c-drnrs\" (UID: \"defe4ef9-9037-4a53-a860-ce6713ee219c\") " pod="openstack/dnsmasq-dns-77585f5f8c-drnrs" Dec 10 23:07:39 crc kubenswrapper[4791]: I1210 23:07:39.937062 4791 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/defe4ef9-9037-4a53-a860-ce6713ee219c-config\") pod \"dnsmasq-dns-77585f5f8c-drnrs\" (UID: \"defe4ef9-9037-4a53-a860-ce6713ee219c\") " pod="openstack/dnsmasq-dns-77585f5f8c-drnrs" Dec 10 23:07:39 crc kubenswrapper[4791]: I1210 23:07:39.937638 4791 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/defe4ef9-9037-4a53-a860-ce6713ee219c-dns-swift-storage-0\") pod \"dnsmasq-dns-77585f5f8c-drnrs\" (UID: \"defe4ef9-9037-4a53-a860-ce6713ee219c\") " pod="openstack/dnsmasq-dns-77585f5f8c-drnrs" Dec 10 23:07:39 crc kubenswrapper[4791]: I1210 23:07:39.937810 4791 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/defe4ef9-9037-4a53-a860-ce6713ee219c-ovsdbserver-sb\") pod \"dnsmasq-dns-77585f5f8c-drnrs\" (UID: \"defe4ef9-9037-4a53-a860-ce6713ee219c\") " pod="openstack/dnsmasq-dns-77585f5f8c-drnrs" Dec 10 23:07:39 crc kubenswrapper[4791]: I1210 23:07:39.938029 4791 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/defe4ef9-9037-4a53-a860-ce6713ee219c-ovsdbserver-nb\") pod \"dnsmasq-dns-77585f5f8c-drnrs\" (UID: \"defe4ef9-9037-4a53-a860-ce6713ee219c\") " pod="openstack/dnsmasq-dns-77585f5f8c-drnrs" Dec 10 23:07:39 crc kubenswrapper[4791]: I1210 23:07:39.938292 4791 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/defe4ef9-9037-4a53-a860-ce6713ee219c-config\") pod \"dnsmasq-dns-77585f5f8c-drnrs\" (UID: \"defe4ef9-9037-4a53-a860-ce6713ee219c\") " pod="openstack/dnsmasq-dns-77585f5f8c-drnrs" Dec 10 23:07:39 crc kubenswrapper[4791]: I1210 23:07:39.938431 4791 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/defe4ef9-9037-4a53-a860-ce6713ee219c-dns-svc\") pod \"dnsmasq-dns-77585f5f8c-drnrs\" (UID: \"defe4ef9-9037-4a53-a860-ce6713ee219c\") " pod="openstack/dnsmasq-dns-77585f5f8c-drnrs" Dec 10 23:07:39 crc kubenswrapper[4791]: I1210 23:07:39.955042 4791 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-9s79j\" (UniqueName: \"kubernetes.io/projected/defe4ef9-9037-4a53-a860-ce6713ee219c-kube-api-access-9s79j\") pod \"dnsmasq-dns-77585f5f8c-drnrs\" (UID: \"defe4ef9-9037-4a53-a860-ce6713ee219c\") " pod="openstack/dnsmasq-dns-77585f5f8c-drnrs" Dec 10 23:07:40 crc kubenswrapper[4791]: I1210 23:07:40.088438 4791 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-77585f5f8c-drnrs" Dec 10 23:07:40 crc kubenswrapper[4791]: I1210 23:07:40.395163 4791 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-77585f5f8c-drnrs"] Dec 10 23:07:40 crc kubenswrapper[4791]: W1210 23:07:40.408432 4791 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-poddefe4ef9_9037_4a53_a860_ce6713ee219c.slice/crio-7dade463ebac697fb2c29c5b384910908465c1fb5f97888b67133f0c327c79b4 WatchSource:0}: Error finding container 7dade463ebac697fb2c29c5b384910908465c1fb5f97888b67133f0c327c79b4: Status 404 returned error can't find the container with id 7dade463ebac697fb2c29c5b384910908465c1fb5f97888b67133f0c327c79b4 Dec 10 23:07:40 crc kubenswrapper[4791]: I1210 23:07:40.431569 4791 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-77585f5f8c-drnrs" event={"ID":"defe4ef9-9037-4a53-a860-ce6713ee219c","Type":"ContainerStarted","Data":"7dade463ebac697fb2c29c5b384910908465c1fb5f97888b67133f0c327c79b4"} Dec 10 23:07:41 crc kubenswrapper[4791]: I1210 23:07:41.441032 4791 generic.go:334] "Generic (PLEG): container finished" podID="defe4ef9-9037-4a53-a860-ce6713ee219c" containerID="b677977e766e127dbf26f7ec4252e1db8771f1088e3cfb922d4d6f5e770a4a8c" exitCode=0 Dec 10 23:07:41 crc kubenswrapper[4791]: I1210 23:07:41.441406 4791 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-77585f5f8c-drnrs" event={"ID":"defe4ef9-9037-4a53-a860-ce6713ee219c","Type":"ContainerDied","Data":"b677977e766e127dbf26f7ec4252e1db8771f1088e3cfb922d4d6f5e770a4a8c"} Dec 10 23:07:41 crc kubenswrapper[4791]: I1210 23:07:41.446577 4791 generic.go:334] "Generic (PLEG): container finished" podID="9ce8dbeb-dce8-47ac-98f1-fe92f74aa3de" containerID="d8a93d7e8bfe0416ef17df92a9875113cc702aca941f1f73df50121135822c94" exitCode=0 Dec 10 23:07:41 crc kubenswrapper[4791]: I1210 23:07:41.446623 4791 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-server-0" event={"ID":"9ce8dbeb-dce8-47ac-98f1-fe92f74aa3de","Type":"ContainerDied","Data":"d8a93d7e8bfe0416ef17df92a9875113cc702aca941f1f73df50121135822c94"} Dec 10 23:07:42 crc kubenswrapper[4791]: I1210 23:07:42.457762 4791 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-77585f5f8c-drnrs" event={"ID":"defe4ef9-9037-4a53-a860-ce6713ee219c","Type":"ContainerStarted","Data":"799a9c08f13a56a6860ed5b1f58366c2cfea680c89ada5e8e0d0083357e1d1ed"} Dec 10 23:07:42 crc kubenswrapper[4791]: I1210 23:07:42.458078 4791 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/dnsmasq-dns-77585f5f8c-drnrs" Dec 10 23:07:42 crc kubenswrapper[4791]: I1210 23:07:42.462624 4791 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-server-0" event={"ID":"9ce8dbeb-dce8-47ac-98f1-fe92f74aa3de","Type":"ContainerStarted","Data":"f7a414fff9fd9d4a710f40666dd232a7704dad812d497846c5077aa27f08208d"} Dec 10 23:07:42 crc kubenswrapper[4791]: I1210 23:07:42.462875 4791 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/rabbitmq-server-0" Dec 10 23:07:42 crc kubenswrapper[4791]: I1210 23:07:42.477847 4791 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/dnsmasq-dns-77585f5f8c-drnrs" podStartSLOduration=3.477830648 podStartE2EDuration="3.477830648s" podCreationTimestamp="2025-12-10 23:07:39 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-10 23:07:42.476061776 +0000 UTC m=+1096.905679389" watchObservedRunningTime="2025-12-10 23:07:42.477830648 +0000 UTC m=+1096.907448261" Dec 10 23:07:42 crc kubenswrapper[4791]: I1210 23:07:42.508248 4791 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/rabbitmq-server-0" podStartSLOduration=-9223371956.346548 podStartE2EDuration="1m20.508228397s" podCreationTimestamp="2025-12-10 23:06:22 +0000 UTC" firstStartedPulling="2025-12-10 23:06:26.089616067 +0000 UTC m=+1020.519233680" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-10 23:07:42.501586785 +0000 UTC m=+1096.931204398" watchObservedRunningTime="2025-12-10 23:07:42.508228397 +0000 UTC m=+1096.937846010" Dec 10 23:07:50 crc kubenswrapper[4791]: I1210 23:07:50.089500 4791 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/dnsmasq-dns-77585f5f8c-drnrs" Dec 10 23:07:50 crc kubenswrapper[4791]: I1210 23:07:50.155024 4791 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-698758b865-72nqc"] Dec 10 23:07:50 crc kubenswrapper[4791]: I1210 23:07:50.155678 4791 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/dnsmasq-dns-698758b865-72nqc" podUID="7e284c1c-b650-4488-8c24-f9f5900b2326" containerName="dnsmasq-dns" containerID="cri-o://064f69cbe089bbf9a8cab03b9f340ca2bb059356287fac23429c1823da83654d" gracePeriod=10 Dec 10 23:07:50 crc kubenswrapper[4791]: I1210 23:07:50.523370 4791 generic.go:334] "Generic (PLEG): container finished" podID="7e284c1c-b650-4488-8c24-f9f5900b2326" containerID="064f69cbe089bbf9a8cab03b9f340ca2bb059356287fac23429c1823da83654d" exitCode=0 Dec 10 23:07:50 crc kubenswrapper[4791]: I1210 23:07:50.523408 4791 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-698758b865-72nqc" event={"ID":"7e284c1c-b650-4488-8c24-f9f5900b2326","Type":"ContainerDied","Data":"064f69cbe089bbf9a8cab03b9f340ca2bb059356287fac23429c1823da83654d"} Dec 10 23:07:54 crc kubenswrapper[4791]: I1210 23:07:54.390396 4791 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/dnsmasq-dns-698758b865-72nqc" podUID="7e284c1c-b650-4488-8c24-f9f5900b2326" containerName="dnsmasq-dns" probeResult="failure" output="dial tcp 10.217.0.112:5353: connect: connection refused" Dec 10 23:07:54 crc kubenswrapper[4791]: I1210 23:07:54.993588 4791 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/rabbitmq-cell1-server-0" Dec 10 23:07:55 crc kubenswrapper[4791]: I1210 23:07:55.284877 4791 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/rabbitmq-server-0" podUID="9ce8dbeb-dce8-47ac-98f1-fe92f74aa3de" containerName="rabbitmq" probeResult="failure" output="dial tcp 10.217.0.98:5671: connect: connection refused" Dec 10 23:07:56 crc kubenswrapper[4791]: E1210 23:07:56.587482 4791 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="quay.io/podified-antelope-centos9/openstack-glance-api:current-podified" Dec 10 23:07:56 crc kubenswrapper[4791]: E1210 23:07:56.587992 4791 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:glance-db-sync,Image:quay.io/podified-antelope-centos9/openstack-glance-api:current-podified,Command:[/bin/bash],Args:[-c /usr/local/bin/kolla_start],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:KOLLA_BOOTSTRAP,Value:true,ValueFrom:nil,},EnvVar{Name:KOLLA_CONFIG_STRATEGY,Value:COPY_ALWAYS,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:db-sync-config-data,ReadOnly:true,MountPath:/etc/glance/glance.conf.d,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:config-data,ReadOnly:true,MountPath:/etc/my.cnf,SubPath:my.cnf,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:config-data,ReadOnly:true,MountPath:/var/lib/kolla/config_files/config.json,SubPath:db-sync-config.json,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:combined-ca-bundle,ReadOnly:true,MountPath:/etc/pki/ca-trust/extracted/pem/tls-ca-bundle.pem,SubPath:tls-ca-bundle.pem,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-v7lq8,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*42415,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:nil,RunAsGroup:*42415,ProcMount:nil,WindowsOptions:nil,SeccompProfile:&SeccompProfile{Type:RuntimeDefault,LocalhostProfile:nil,},AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod glance-db-sync-ccl4n_openstack(a2c44c96-70c2-414a-9890-cb95b66f10b2): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError" Dec 10 23:07:56 crc kubenswrapper[4791]: E1210 23:07:56.589160 4791 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"glance-db-sync\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\"" pod="openstack/glance-db-sync-ccl4n" podUID="a2c44c96-70c2-414a-9890-cb95b66f10b2" Dec 10 23:07:56 crc kubenswrapper[4791]: I1210 23:07:56.844656 4791 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-698758b865-72nqc" Dec 10 23:07:57 crc kubenswrapper[4791]: I1210 23:07:57.004098 4791 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-8jj9k\" (UniqueName: \"kubernetes.io/projected/7e284c1c-b650-4488-8c24-f9f5900b2326-kube-api-access-8jj9k\") pod \"7e284c1c-b650-4488-8c24-f9f5900b2326\" (UID: \"7e284c1c-b650-4488-8c24-f9f5900b2326\") " Dec 10 23:07:57 crc kubenswrapper[4791]: I1210 23:07:57.004241 4791 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/7e284c1c-b650-4488-8c24-f9f5900b2326-config\") pod \"7e284c1c-b650-4488-8c24-f9f5900b2326\" (UID: \"7e284c1c-b650-4488-8c24-f9f5900b2326\") " Dec 10 23:07:57 crc kubenswrapper[4791]: I1210 23:07:57.004306 4791 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/7e284c1c-b650-4488-8c24-f9f5900b2326-dns-svc\") pod \"7e284c1c-b650-4488-8c24-f9f5900b2326\" (UID: \"7e284c1c-b650-4488-8c24-f9f5900b2326\") " Dec 10 23:07:57 crc kubenswrapper[4791]: I1210 23:07:57.004407 4791 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/7e284c1c-b650-4488-8c24-f9f5900b2326-ovsdbserver-sb\") pod \"7e284c1c-b650-4488-8c24-f9f5900b2326\" (UID: \"7e284c1c-b650-4488-8c24-f9f5900b2326\") " Dec 10 23:07:57 crc kubenswrapper[4791]: I1210 23:07:57.004471 4791 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/7e284c1c-b650-4488-8c24-f9f5900b2326-ovsdbserver-nb\") pod \"7e284c1c-b650-4488-8c24-f9f5900b2326\" (UID: \"7e284c1c-b650-4488-8c24-f9f5900b2326\") " Dec 10 23:07:57 crc kubenswrapper[4791]: I1210 23:07:57.022656 4791 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/7e284c1c-b650-4488-8c24-f9f5900b2326-kube-api-access-8jj9k" (OuterVolumeSpecName: "kube-api-access-8jj9k") pod "7e284c1c-b650-4488-8c24-f9f5900b2326" (UID: "7e284c1c-b650-4488-8c24-f9f5900b2326"). InnerVolumeSpecName "kube-api-access-8jj9k". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 10 23:07:57 crc kubenswrapper[4791]: I1210 23:07:57.050740 4791 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/7e284c1c-b650-4488-8c24-f9f5900b2326-config" (OuterVolumeSpecName: "config") pod "7e284c1c-b650-4488-8c24-f9f5900b2326" (UID: "7e284c1c-b650-4488-8c24-f9f5900b2326"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 10 23:07:57 crc kubenswrapper[4791]: I1210 23:07:57.064002 4791 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/7e284c1c-b650-4488-8c24-f9f5900b2326-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "7e284c1c-b650-4488-8c24-f9f5900b2326" (UID: "7e284c1c-b650-4488-8c24-f9f5900b2326"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 10 23:07:57 crc kubenswrapper[4791]: I1210 23:07:57.071827 4791 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/7e284c1c-b650-4488-8c24-f9f5900b2326-ovsdbserver-nb" (OuterVolumeSpecName: "ovsdbserver-nb") pod "7e284c1c-b650-4488-8c24-f9f5900b2326" (UID: "7e284c1c-b650-4488-8c24-f9f5900b2326"). InnerVolumeSpecName "ovsdbserver-nb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 10 23:07:57 crc kubenswrapper[4791]: I1210 23:07:57.082521 4791 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/7e284c1c-b650-4488-8c24-f9f5900b2326-ovsdbserver-sb" (OuterVolumeSpecName: "ovsdbserver-sb") pod "7e284c1c-b650-4488-8c24-f9f5900b2326" (UID: "7e284c1c-b650-4488-8c24-f9f5900b2326"). InnerVolumeSpecName "ovsdbserver-sb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 10 23:07:57 crc kubenswrapper[4791]: I1210 23:07:57.107398 4791 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-8jj9k\" (UniqueName: \"kubernetes.io/projected/7e284c1c-b650-4488-8c24-f9f5900b2326-kube-api-access-8jj9k\") on node \"crc\" DevicePath \"\"" Dec 10 23:07:57 crc kubenswrapper[4791]: I1210 23:07:57.107429 4791 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/7e284c1c-b650-4488-8c24-f9f5900b2326-config\") on node \"crc\" DevicePath \"\"" Dec 10 23:07:57 crc kubenswrapper[4791]: I1210 23:07:57.107443 4791 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/7e284c1c-b650-4488-8c24-f9f5900b2326-dns-svc\") on node \"crc\" DevicePath \"\"" Dec 10 23:07:57 crc kubenswrapper[4791]: I1210 23:07:57.107454 4791 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/7e284c1c-b650-4488-8c24-f9f5900b2326-ovsdbserver-sb\") on node \"crc\" DevicePath \"\"" Dec 10 23:07:57 crc kubenswrapper[4791]: I1210 23:07:57.107461 4791 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/7e284c1c-b650-4488-8c24-f9f5900b2326-ovsdbserver-nb\") on node \"crc\" DevicePath \"\"" Dec 10 23:07:57 crc kubenswrapper[4791]: I1210 23:07:57.588742 4791 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-698758b865-72nqc" Dec 10 23:07:57 crc kubenswrapper[4791]: I1210 23:07:57.591495 4791 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-698758b865-72nqc" event={"ID":"7e284c1c-b650-4488-8c24-f9f5900b2326","Type":"ContainerDied","Data":"8b882073e14aa2aa82987269ad3d80ced15404f389b70490a115782671386def"} Dec 10 23:07:57 crc kubenswrapper[4791]: I1210 23:07:57.591576 4791 scope.go:117] "RemoveContainer" containerID="064f69cbe089bbf9a8cab03b9f340ca2bb059356287fac23429c1823da83654d" Dec 10 23:07:57 crc kubenswrapper[4791]: E1210 23:07:57.591686 4791 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"glance-db-sync\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/podified-antelope-centos9/openstack-glance-api:current-podified\\\"\"" pod="openstack/glance-db-sync-ccl4n" podUID="a2c44c96-70c2-414a-9890-cb95b66f10b2" Dec 10 23:07:57 crc kubenswrapper[4791]: I1210 23:07:57.632270 4791 scope.go:117] "RemoveContainer" containerID="ae9f763a9fb2de89ba33dc56f9ef91103a82f777e1d9bcd04f3adef9f108e092" Dec 10 23:07:57 crc kubenswrapper[4791]: I1210 23:07:57.638590 4791 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-698758b865-72nqc"] Dec 10 23:07:57 crc kubenswrapper[4791]: I1210 23:07:57.644837 4791 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-698758b865-72nqc"] Dec 10 23:07:57 crc kubenswrapper[4791]: I1210 23:07:57.895106 4791 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="7e284c1c-b650-4488-8c24-f9f5900b2326" path="/var/lib/kubelet/pods/7e284c1c-b650-4488-8c24-f9f5900b2326/volumes" Dec 10 23:08:05 crc kubenswrapper[4791]: I1210 23:08:05.278423 4791 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/rabbitmq-server-0" Dec 10 23:08:05 crc kubenswrapper[4791]: I1210 23:08:05.571438 4791 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/cinder-db-create-58mvg"] Dec 10 23:08:05 crc kubenswrapper[4791]: E1210 23:08:05.571879 4791 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="7e284c1c-b650-4488-8c24-f9f5900b2326" containerName="dnsmasq-dns" Dec 10 23:08:05 crc kubenswrapper[4791]: I1210 23:08:05.571904 4791 state_mem.go:107] "Deleted CPUSet assignment" podUID="7e284c1c-b650-4488-8c24-f9f5900b2326" containerName="dnsmasq-dns" Dec 10 23:08:05 crc kubenswrapper[4791]: E1210 23:08:05.571934 4791 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="7e284c1c-b650-4488-8c24-f9f5900b2326" containerName="init" Dec 10 23:08:05 crc kubenswrapper[4791]: I1210 23:08:05.571943 4791 state_mem.go:107] "Deleted CPUSet assignment" podUID="7e284c1c-b650-4488-8c24-f9f5900b2326" containerName="init" Dec 10 23:08:05 crc kubenswrapper[4791]: I1210 23:08:05.572128 4791 memory_manager.go:354] "RemoveStaleState removing state" podUID="7e284c1c-b650-4488-8c24-f9f5900b2326" containerName="dnsmasq-dns" Dec 10 23:08:05 crc kubenswrapper[4791]: I1210 23:08:05.572809 4791 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-db-create-58mvg" Dec 10 23:08:05 crc kubenswrapper[4791]: I1210 23:08:05.588093 4791 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-db-create-58mvg"] Dec 10 23:08:05 crc kubenswrapper[4791]: I1210 23:08:05.676283 4791 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/barbican-db-create-w8q8q"] Dec 10 23:08:05 crc kubenswrapper[4791]: I1210 23:08:05.677793 4791 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-db-create-w8q8q" Dec 10 23:08:05 crc kubenswrapper[4791]: I1210 23:08:05.690244 4791 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/barbican-db-create-w8q8q"] Dec 10 23:08:05 crc kubenswrapper[4791]: I1210 23:08:05.703322 4791 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/barbican-a84b-account-create-update-7ct6w"] Dec 10 23:08:05 crc kubenswrapper[4791]: I1210 23:08:05.704603 4791 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-a84b-account-create-update-7ct6w" Dec 10 23:08:05 crc kubenswrapper[4791]: I1210 23:08:05.707280 4791 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"barbican-db-secret" Dec 10 23:08:05 crc kubenswrapper[4791]: I1210 23:08:05.728503 4791 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/barbican-a84b-account-create-update-7ct6w"] Dec 10 23:08:05 crc kubenswrapper[4791]: I1210 23:08:05.751316 4791 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/d2eecb39-349b-48d4-8819-aa392095d5d0-operator-scripts\") pod \"cinder-db-create-58mvg\" (UID: \"d2eecb39-349b-48d4-8819-aa392095d5d0\") " pod="openstack/cinder-db-create-58mvg" Dec 10 23:08:05 crc kubenswrapper[4791]: I1210 23:08:05.751443 4791 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-rk79k\" (UniqueName: \"kubernetes.io/projected/d2eecb39-349b-48d4-8819-aa392095d5d0-kube-api-access-rk79k\") pod \"cinder-db-create-58mvg\" (UID: \"d2eecb39-349b-48d4-8819-aa392095d5d0\") " pod="openstack/cinder-db-create-58mvg" Dec 10 23:08:05 crc kubenswrapper[4791]: I1210 23:08:05.795322 4791 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/cinder-8c6c-account-create-update-s9h2s"] Dec 10 23:08:05 crc kubenswrapper[4791]: I1210 23:08:05.796745 4791 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-8c6c-account-create-update-s9h2s" Dec 10 23:08:05 crc kubenswrapper[4791]: I1210 23:08:05.799998 4791 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cinder-db-secret" Dec 10 23:08:05 crc kubenswrapper[4791]: I1210 23:08:05.810302 4791 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-8c6c-account-create-update-s9h2s"] Dec 10 23:08:05 crc kubenswrapper[4791]: I1210 23:08:05.854289 4791 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-rk79k\" (UniqueName: \"kubernetes.io/projected/d2eecb39-349b-48d4-8819-aa392095d5d0-kube-api-access-rk79k\") pod \"cinder-db-create-58mvg\" (UID: \"d2eecb39-349b-48d4-8819-aa392095d5d0\") " pod="openstack/cinder-db-create-58mvg" Dec 10 23:08:05 crc kubenswrapper[4791]: I1210 23:08:05.855056 4791 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-hltjn\" (UniqueName: \"kubernetes.io/projected/ebf742dc-fd16-49ae-a30a-b0f61e4d15fd-kube-api-access-hltjn\") pod \"barbican-db-create-w8q8q\" (UID: \"ebf742dc-fd16-49ae-a30a-b0f61e4d15fd\") " pod="openstack/barbican-db-create-w8q8q" Dec 10 23:08:05 crc kubenswrapper[4791]: I1210 23:08:05.855194 4791 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/25d98ac3-17a9-4885-a64a-2ae593acf177-operator-scripts\") pod \"barbican-a84b-account-create-update-7ct6w\" (UID: \"25d98ac3-17a9-4885-a64a-2ae593acf177\") " pod="openstack/barbican-a84b-account-create-update-7ct6w" Dec 10 23:08:05 crc kubenswrapper[4791]: I1210 23:08:05.855392 4791 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-2w8nc\" (UniqueName: \"kubernetes.io/projected/25d98ac3-17a9-4885-a64a-2ae593acf177-kube-api-access-2w8nc\") pod \"barbican-a84b-account-create-update-7ct6w\" (UID: \"25d98ac3-17a9-4885-a64a-2ae593acf177\") " pod="openstack/barbican-a84b-account-create-update-7ct6w" Dec 10 23:08:05 crc kubenswrapper[4791]: I1210 23:08:05.855560 4791 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/ebf742dc-fd16-49ae-a30a-b0f61e4d15fd-operator-scripts\") pod \"barbican-db-create-w8q8q\" (UID: \"ebf742dc-fd16-49ae-a30a-b0f61e4d15fd\") " pod="openstack/barbican-db-create-w8q8q" Dec 10 23:08:05 crc kubenswrapper[4791]: I1210 23:08:05.855681 4791 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/d2eecb39-349b-48d4-8819-aa392095d5d0-operator-scripts\") pod \"cinder-db-create-58mvg\" (UID: \"d2eecb39-349b-48d4-8819-aa392095d5d0\") " pod="openstack/cinder-db-create-58mvg" Dec 10 23:08:05 crc kubenswrapper[4791]: I1210 23:08:05.856758 4791 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/d2eecb39-349b-48d4-8819-aa392095d5d0-operator-scripts\") pod \"cinder-db-create-58mvg\" (UID: \"d2eecb39-349b-48d4-8819-aa392095d5d0\") " pod="openstack/cinder-db-create-58mvg" Dec 10 23:08:05 crc kubenswrapper[4791]: I1210 23:08:05.879170 4791 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-rk79k\" (UniqueName: \"kubernetes.io/projected/d2eecb39-349b-48d4-8819-aa392095d5d0-kube-api-access-rk79k\") pod \"cinder-db-create-58mvg\" (UID: \"d2eecb39-349b-48d4-8819-aa392095d5d0\") " pod="openstack/cinder-db-create-58mvg" Dec 10 23:08:05 crc kubenswrapper[4791]: I1210 23:08:05.890828 4791 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-db-create-58mvg" Dec 10 23:08:05 crc kubenswrapper[4791]: I1210 23:08:05.899647 4791 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/neutron-db-create-l5xkj"] Dec 10 23:08:05 crc kubenswrapper[4791]: I1210 23:08:05.900763 4791 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-db-create-l5xkj" Dec 10 23:08:05 crc kubenswrapper[4791]: I1210 23:08:05.907540 4791 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/neutron-db-create-l5xkj"] Dec 10 23:08:05 crc kubenswrapper[4791]: I1210 23:08:05.957414 4791 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-hltjn\" (UniqueName: \"kubernetes.io/projected/ebf742dc-fd16-49ae-a30a-b0f61e4d15fd-kube-api-access-hltjn\") pod \"barbican-db-create-w8q8q\" (UID: \"ebf742dc-fd16-49ae-a30a-b0f61e4d15fd\") " pod="openstack/barbican-db-create-w8q8q" Dec 10 23:08:05 crc kubenswrapper[4791]: I1210 23:08:05.957477 4791 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/c289a9e6-9d03-499b-82f6-f0c898d94f04-operator-scripts\") pod \"cinder-8c6c-account-create-update-s9h2s\" (UID: \"c289a9e6-9d03-499b-82f6-f0c898d94f04\") " pod="openstack/cinder-8c6c-account-create-update-s9h2s" Dec 10 23:08:05 crc kubenswrapper[4791]: I1210 23:08:05.957505 4791 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/25d98ac3-17a9-4885-a64a-2ae593acf177-operator-scripts\") pod \"barbican-a84b-account-create-update-7ct6w\" (UID: \"25d98ac3-17a9-4885-a64a-2ae593acf177\") " pod="openstack/barbican-a84b-account-create-update-7ct6w" Dec 10 23:08:05 crc kubenswrapper[4791]: I1210 23:08:05.957530 4791 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-5tqb8\" (UniqueName: \"kubernetes.io/projected/c289a9e6-9d03-499b-82f6-f0c898d94f04-kube-api-access-5tqb8\") pod \"cinder-8c6c-account-create-update-s9h2s\" (UID: \"c289a9e6-9d03-499b-82f6-f0c898d94f04\") " pod="openstack/cinder-8c6c-account-create-update-s9h2s" Dec 10 23:08:05 crc kubenswrapper[4791]: I1210 23:08:05.957564 4791 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-2w8nc\" (UniqueName: \"kubernetes.io/projected/25d98ac3-17a9-4885-a64a-2ae593acf177-kube-api-access-2w8nc\") pod \"barbican-a84b-account-create-update-7ct6w\" (UID: \"25d98ac3-17a9-4885-a64a-2ae593acf177\") " pod="openstack/barbican-a84b-account-create-update-7ct6w" Dec 10 23:08:05 crc kubenswrapper[4791]: I1210 23:08:05.957584 4791 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/ebf742dc-fd16-49ae-a30a-b0f61e4d15fd-operator-scripts\") pod \"barbican-db-create-w8q8q\" (UID: \"ebf742dc-fd16-49ae-a30a-b0f61e4d15fd\") " pod="openstack/barbican-db-create-w8q8q" Dec 10 23:08:05 crc kubenswrapper[4791]: I1210 23:08:05.958332 4791 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/ebf742dc-fd16-49ae-a30a-b0f61e4d15fd-operator-scripts\") pod \"barbican-db-create-w8q8q\" (UID: \"ebf742dc-fd16-49ae-a30a-b0f61e4d15fd\") " pod="openstack/barbican-db-create-w8q8q" Dec 10 23:08:05 crc kubenswrapper[4791]: I1210 23:08:05.958629 4791 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/25d98ac3-17a9-4885-a64a-2ae593acf177-operator-scripts\") pod \"barbican-a84b-account-create-update-7ct6w\" (UID: \"25d98ac3-17a9-4885-a64a-2ae593acf177\") " pod="openstack/barbican-a84b-account-create-update-7ct6w" Dec 10 23:08:05 crc kubenswrapper[4791]: I1210 23:08:05.984534 4791 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/neutron-e982-account-create-update-7fdhc"] Dec 10 23:08:05 crc kubenswrapper[4791]: I1210 23:08:05.985165 4791 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-2w8nc\" (UniqueName: \"kubernetes.io/projected/25d98ac3-17a9-4885-a64a-2ae593acf177-kube-api-access-2w8nc\") pod \"barbican-a84b-account-create-update-7ct6w\" (UID: \"25d98ac3-17a9-4885-a64a-2ae593acf177\") " pod="openstack/barbican-a84b-account-create-update-7ct6w" Dec 10 23:08:05 crc kubenswrapper[4791]: I1210 23:08:05.985933 4791 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-e982-account-create-update-7fdhc" Dec 10 23:08:05 crc kubenswrapper[4791]: I1210 23:08:05.988007 4791 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"neutron-db-secret" Dec 10 23:08:05 crc kubenswrapper[4791]: I1210 23:08:05.988832 4791 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-hltjn\" (UniqueName: \"kubernetes.io/projected/ebf742dc-fd16-49ae-a30a-b0f61e4d15fd-kube-api-access-hltjn\") pod \"barbican-db-create-w8q8q\" (UID: \"ebf742dc-fd16-49ae-a30a-b0f61e4d15fd\") " pod="openstack/barbican-db-create-w8q8q" Dec 10 23:08:05 crc kubenswrapper[4791]: I1210 23:08:05.991880 4791 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-db-create-w8q8q" Dec 10 23:08:06 crc kubenswrapper[4791]: I1210 23:08:06.005886 4791 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/neutron-e982-account-create-update-7fdhc"] Dec 10 23:08:06 crc kubenswrapper[4791]: I1210 23:08:06.028083 4791 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-a84b-account-create-update-7ct6w" Dec 10 23:08:06 crc kubenswrapper[4791]: I1210 23:08:06.059556 4791 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-tjvdl\" (UniqueName: \"kubernetes.io/projected/77886642-7b06-476f-acb0-7e4df70190a8-kube-api-access-tjvdl\") pod \"neutron-db-create-l5xkj\" (UID: \"77886642-7b06-476f-acb0-7e4df70190a8\") " pod="openstack/neutron-db-create-l5xkj" Dec 10 23:08:06 crc kubenswrapper[4791]: I1210 23:08:06.059690 4791 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/c289a9e6-9d03-499b-82f6-f0c898d94f04-operator-scripts\") pod \"cinder-8c6c-account-create-update-s9h2s\" (UID: \"c289a9e6-9d03-499b-82f6-f0c898d94f04\") " pod="openstack/cinder-8c6c-account-create-update-s9h2s" Dec 10 23:08:06 crc kubenswrapper[4791]: I1210 23:08:06.059730 4791 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-5tqb8\" (UniqueName: \"kubernetes.io/projected/c289a9e6-9d03-499b-82f6-f0c898d94f04-kube-api-access-5tqb8\") pod \"cinder-8c6c-account-create-update-s9h2s\" (UID: \"c289a9e6-9d03-499b-82f6-f0c898d94f04\") " pod="openstack/cinder-8c6c-account-create-update-s9h2s" Dec 10 23:08:06 crc kubenswrapper[4791]: I1210 23:08:06.059768 4791 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/77886642-7b06-476f-acb0-7e4df70190a8-operator-scripts\") pod \"neutron-db-create-l5xkj\" (UID: \"77886642-7b06-476f-acb0-7e4df70190a8\") " pod="openstack/neutron-db-create-l5xkj" Dec 10 23:08:06 crc kubenswrapper[4791]: I1210 23:08:06.061000 4791 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/c289a9e6-9d03-499b-82f6-f0c898d94f04-operator-scripts\") pod \"cinder-8c6c-account-create-update-s9h2s\" (UID: \"c289a9e6-9d03-499b-82f6-f0c898d94f04\") " pod="openstack/cinder-8c6c-account-create-update-s9h2s" Dec 10 23:08:06 crc kubenswrapper[4791]: I1210 23:08:06.092000 4791 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-5tqb8\" (UniqueName: \"kubernetes.io/projected/c289a9e6-9d03-499b-82f6-f0c898d94f04-kube-api-access-5tqb8\") pod \"cinder-8c6c-account-create-update-s9h2s\" (UID: \"c289a9e6-9d03-499b-82f6-f0c898d94f04\") " pod="openstack/cinder-8c6c-account-create-update-s9h2s" Dec 10 23:08:06 crc kubenswrapper[4791]: I1210 23:08:06.105184 4791 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/keystone-db-sync-kt8jj"] Dec 10 23:08:06 crc kubenswrapper[4791]: I1210 23:08:06.113552 4791 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-8c6c-account-create-update-s9h2s" Dec 10 23:08:06 crc kubenswrapper[4791]: I1210 23:08:06.115534 4791 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-db-sync-kt8jj" Dec 10 23:08:06 crc kubenswrapper[4791]: I1210 23:08:06.119500 4791 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone" Dec 10 23:08:06 crc kubenswrapper[4791]: I1210 23:08:06.119539 4791 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone-scripts" Dec 10 23:08:06 crc kubenswrapper[4791]: I1210 23:08:06.119541 4791 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone-keystone-dockercfg-t99v8" Dec 10 23:08:06 crc kubenswrapper[4791]: I1210 23:08:06.119670 4791 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone-config-data" Dec 10 23:08:06 crc kubenswrapper[4791]: I1210 23:08:06.121951 4791 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/keystone-db-sync-kt8jj"] Dec 10 23:08:06 crc kubenswrapper[4791]: I1210 23:08:06.164163 4791 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-tjvdl\" (UniqueName: \"kubernetes.io/projected/77886642-7b06-476f-acb0-7e4df70190a8-kube-api-access-tjvdl\") pod \"neutron-db-create-l5xkj\" (UID: \"77886642-7b06-476f-acb0-7e4df70190a8\") " pod="openstack/neutron-db-create-l5xkj" Dec 10 23:08:06 crc kubenswrapper[4791]: I1210 23:08:06.164232 4791 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/e7649319-d8a5-48d8-b9f8-f236602c72a0-operator-scripts\") pod \"neutron-e982-account-create-update-7fdhc\" (UID: \"e7649319-d8a5-48d8-b9f8-f236602c72a0\") " pod="openstack/neutron-e982-account-create-update-7fdhc" Dec 10 23:08:06 crc kubenswrapper[4791]: I1210 23:08:06.164256 4791 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-k9ps6\" (UniqueName: \"kubernetes.io/projected/e7649319-d8a5-48d8-b9f8-f236602c72a0-kube-api-access-k9ps6\") pod \"neutron-e982-account-create-update-7fdhc\" (UID: \"e7649319-d8a5-48d8-b9f8-f236602c72a0\") " pod="openstack/neutron-e982-account-create-update-7fdhc" Dec 10 23:08:06 crc kubenswrapper[4791]: I1210 23:08:06.164329 4791 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/77886642-7b06-476f-acb0-7e4df70190a8-operator-scripts\") pod \"neutron-db-create-l5xkj\" (UID: \"77886642-7b06-476f-acb0-7e4df70190a8\") " pod="openstack/neutron-db-create-l5xkj" Dec 10 23:08:06 crc kubenswrapper[4791]: I1210 23:08:06.165223 4791 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/77886642-7b06-476f-acb0-7e4df70190a8-operator-scripts\") pod \"neutron-db-create-l5xkj\" (UID: \"77886642-7b06-476f-acb0-7e4df70190a8\") " pod="openstack/neutron-db-create-l5xkj" Dec 10 23:08:06 crc kubenswrapper[4791]: I1210 23:08:06.184538 4791 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-tjvdl\" (UniqueName: \"kubernetes.io/projected/77886642-7b06-476f-acb0-7e4df70190a8-kube-api-access-tjvdl\") pod \"neutron-db-create-l5xkj\" (UID: \"77886642-7b06-476f-acb0-7e4df70190a8\") " pod="openstack/neutron-db-create-l5xkj" Dec 10 23:08:06 crc kubenswrapper[4791]: I1210 23:08:06.269596 4791 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-6rdgm\" (UniqueName: \"kubernetes.io/projected/5ad26083-9c93-4ec4-9159-ef991f116f80-kube-api-access-6rdgm\") pod \"keystone-db-sync-kt8jj\" (UID: \"5ad26083-9c93-4ec4-9159-ef991f116f80\") " pod="openstack/keystone-db-sync-kt8jj" Dec 10 23:08:06 crc kubenswrapper[4791]: I1210 23:08:06.269645 4791 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/5ad26083-9c93-4ec4-9159-ef991f116f80-config-data\") pod \"keystone-db-sync-kt8jj\" (UID: \"5ad26083-9c93-4ec4-9159-ef991f116f80\") " pod="openstack/keystone-db-sync-kt8jj" Dec 10 23:08:06 crc kubenswrapper[4791]: I1210 23:08:06.269687 4791 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/e7649319-d8a5-48d8-b9f8-f236602c72a0-operator-scripts\") pod \"neutron-e982-account-create-update-7fdhc\" (UID: \"e7649319-d8a5-48d8-b9f8-f236602c72a0\") " pod="openstack/neutron-e982-account-create-update-7fdhc" Dec 10 23:08:06 crc kubenswrapper[4791]: I1210 23:08:06.269710 4791 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-k9ps6\" (UniqueName: \"kubernetes.io/projected/e7649319-d8a5-48d8-b9f8-f236602c72a0-kube-api-access-k9ps6\") pod \"neutron-e982-account-create-update-7fdhc\" (UID: \"e7649319-d8a5-48d8-b9f8-f236602c72a0\") " pod="openstack/neutron-e982-account-create-update-7fdhc" Dec 10 23:08:06 crc kubenswrapper[4791]: I1210 23:08:06.269829 4791 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/5ad26083-9c93-4ec4-9159-ef991f116f80-combined-ca-bundle\") pod \"keystone-db-sync-kt8jj\" (UID: \"5ad26083-9c93-4ec4-9159-ef991f116f80\") " pod="openstack/keystone-db-sync-kt8jj" Dec 10 23:08:06 crc kubenswrapper[4791]: I1210 23:08:06.270538 4791 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/e7649319-d8a5-48d8-b9f8-f236602c72a0-operator-scripts\") pod \"neutron-e982-account-create-update-7fdhc\" (UID: \"e7649319-d8a5-48d8-b9f8-f236602c72a0\") " pod="openstack/neutron-e982-account-create-update-7fdhc" Dec 10 23:08:06 crc kubenswrapper[4791]: I1210 23:08:06.293508 4791 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-k9ps6\" (UniqueName: \"kubernetes.io/projected/e7649319-d8a5-48d8-b9f8-f236602c72a0-kube-api-access-k9ps6\") pod \"neutron-e982-account-create-update-7fdhc\" (UID: \"e7649319-d8a5-48d8-b9f8-f236602c72a0\") " pod="openstack/neutron-e982-account-create-update-7fdhc" Dec 10 23:08:06 crc kubenswrapper[4791]: I1210 23:08:06.357517 4791 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-db-create-l5xkj" Dec 10 23:08:06 crc kubenswrapper[4791]: I1210 23:08:06.371080 4791 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-6rdgm\" (UniqueName: \"kubernetes.io/projected/5ad26083-9c93-4ec4-9159-ef991f116f80-kube-api-access-6rdgm\") pod \"keystone-db-sync-kt8jj\" (UID: \"5ad26083-9c93-4ec4-9159-ef991f116f80\") " pod="openstack/keystone-db-sync-kt8jj" Dec 10 23:08:06 crc kubenswrapper[4791]: I1210 23:08:06.371122 4791 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/5ad26083-9c93-4ec4-9159-ef991f116f80-config-data\") pod \"keystone-db-sync-kt8jj\" (UID: \"5ad26083-9c93-4ec4-9159-ef991f116f80\") " pod="openstack/keystone-db-sync-kt8jj" Dec 10 23:08:06 crc kubenswrapper[4791]: I1210 23:08:06.371165 4791 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/5ad26083-9c93-4ec4-9159-ef991f116f80-combined-ca-bundle\") pod \"keystone-db-sync-kt8jj\" (UID: \"5ad26083-9c93-4ec4-9159-ef991f116f80\") " pod="openstack/keystone-db-sync-kt8jj" Dec 10 23:08:06 crc kubenswrapper[4791]: I1210 23:08:06.376210 4791 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/5ad26083-9c93-4ec4-9159-ef991f116f80-config-data\") pod \"keystone-db-sync-kt8jj\" (UID: \"5ad26083-9c93-4ec4-9159-ef991f116f80\") " pod="openstack/keystone-db-sync-kt8jj" Dec 10 23:08:06 crc kubenswrapper[4791]: I1210 23:08:06.376260 4791 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/5ad26083-9c93-4ec4-9159-ef991f116f80-combined-ca-bundle\") pod \"keystone-db-sync-kt8jj\" (UID: \"5ad26083-9c93-4ec4-9159-ef991f116f80\") " pod="openstack/keystone-db-sync-kt8jj" Dec 10 23:08:06 crc kubenswrapper[4791]: I1210 23:08:06.381170 4791 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-e982-account-create-update-7fdhc" Dec 10 23:08:06 crc kubenswrapper[4791]: I1210 23:08:06.393796 4791 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-6rdgm\" (UniqueName: \"kubernetes.io/projected/5ad26083-9c93-4ec4-9159-ef991f116f80-kube-api-access-6rdgm\") pod \"keystone-db-sync-kt8jj\" (UID: \"5ad26083-9c93-4ec4-9159-ef991f116f80\") " pod="openstack/keystone-db-sync-kt8jj" Dec 10 23:08:06 crc kubenswrapper[4791]: I1210 23:08:06.445482 4791 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-db-sync-kt8jj" Dec 10 23:08:06 crc kubenswrapper[4791]: I1210 23:08:06.508645 4791 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-db-create-58mvg"] Dec 10 23:08:06 crc kubenswrapper[4791]: I1210 23:08:06.642791 4791 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/barbican-db-create-w8q8q"] Dec 10 23:08:06 crc kubenswrapper[4791]: I1210 23:08:06.678315 4791 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-db-create-58mvg" event={"ID":"d2eecb39-349b-48d4-8819-aa392095d5d0","Type":"ContainerStarted","Data":"e2ccba20ad7b61bf8f034eaa2c850e0f2bfe5076162c2c6fff7062ccae7cb3c6"} Dec 10 23:08:06 crc kubenswrapper[4791]: I1210 23:08:06.744092 4791 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/barbican-a84b-account-create-update-7ct6w"] Dec 10 23:08:06 crc kubenswrapper[4791]: I1210 23:08:06.862385 4791 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-8c6c-account-create-update-s9h2s"] Dec 10 23:08:06 crc kubenswrapper[4791]: I1210 23:08:06.990386 4791 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/neutron-db-create-l5xkj"] Dec 10 23:08:07 crc kubenswrapper[4791]: W1210 23:08:07.015794 4791 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod77886642_7b06_476f_acb0_7e4df70190a8.slice/crio-448c993671b8c2b72d693d3e7d98608e88c6ed3efe48c0c4eacb89a661bc799d WatchSource:0}: Error finding container 448c993671b8c2b72d693d3e7d98608e88c6ed3efe48c0c4eacb89a661bc799d: Status 404 returned error can't find the container with id 448c993671b8c2b72d693d3e7d98608e88c6ed3efe48c0c4eacb89a661bc799d Dec 10 23:08:07 crc kubenswrapper[4791]: I1210 23:08:07.092695 4791 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/neutron-e982-account-create-update-7fdhc"] Dec 10 23:08:07 crc kubenswrapper[4791]: I1210 23:08:07.110977 4791 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/keystone-db-sync-kt8jj"] Dec 10 23:08:07 crc kubenswrapper[4791]: I1210 23:08:07.689136 4791 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-e982-account-create-update-7fdhc" event={"ID":"e7649319-d8a5-48d8-b9f8-f236602c72a0","Type":"ContainerStarted","Data":"ef2c1e1726ac3fe2b5200e0a1b6289e8846bb2fc6d6fd1dbaa755fddce4936b5"} Dec 10 23:08:07 crc kubenswrapper[4791]: I1210 23:08:07.690765 4791 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-e982-account-create-update-7fdhc" event={"ID":"e7649319-d8a5-48d8-b9f8-f236602c72a0","Type":"ContainerStarted","Data":"a51125ec49b7859f40dafcedb211f7468d71598b5858e13e7cf23c73fe38a383"} Dec 10 23:08:07 crc kubenswrapper[4791]: I1210 23:08:07.693865 4791 generic.go:334] "Generic (PLEG): container finished" podID="c289a9e6-9d03-499b-82f6-f0c898d94f04" containerID="eeb768e1a992b95d7fd4932dcc0d91575705cb14db754bba405b594192a5334d" exitCode=0 Dec 10 23:08:07 crc kubenswrapper[4791]: I1210 23:08:07.693936 4791 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-8c6c-account-create-update-s9h2s" event={"ID":"c289a9e6-9d03-499b-82f6-f0c898d94f04","Type":"ContainerDied","Data":"eeb768e1a992b95d7fd4932dcc0d91575705cb14db754bba405b594192a5334d"} Dec 10 23:08:07 crc kubenswrapper[4791]: I1210 23:08:07.693968 4791 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-8c6c-account-create-update-s9h2s" event={"ID":"c289a9e6-9d03-499b-82f6-f0c898d94f04","Type":"ContainerStarted","Data":"92d9542499a73fe438456a34ab37ad468543662df3ee074ea6eea436b7c35ff8"} Dec 10 23:08:07 crc kubenswrapper[4791]: I1210 23:08:07.697808 4791 generic.go:334] "Generic (PLEG): container finished" podID="77886642-7b06-476f-acb0-7e4df70190a8" containerID="da6698cdf0ff45e48a9d6b08ecbf12721c5b95ee247f5490e3f0ee1861d0c94a" exitCode=0 Dec 10 23:08:07 crc kubenswrapper[4791]: I1210 23:08:07.697835 4791 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-db-create-l5xkj" event={"ID":"77886642-7b06-476f-acb0-7e4df70190a8","Type":"ContainerDied","Data":"da6698cdf0ff45e48a9d6b08ecbf12721c5b95ee247f5490e3f0ee1861d0c94a"} Dec 10 23:08:07 crc kubenswrapper[4791]: I1210 23:08:07.697859 4791 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-db-create-l5xkj" event={"ID":"77886642-7b06-476f-acb0-7e4df70190a8","Type":"ContainerStarted","Data":"448c993671b8c2b72d693d3e7d98608e88c6ed3efe48c0c4eacb89a661bc799d"} Dec 10 23:08:07 crc kubenswrapper[4791]: I1210 23:08:07.702059 4791 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-db-sync-kt8jj" event={"ID":"5ad26083-9c93-4ec4-9159-ef991f116f80","Type":"ContainerStarted","Data":"ed5a216809726f12930fc319273af79fa196ddf4ddc99fb1482154e7f7309231"} Dec 10 23:08:07 crc kubenswrapper[4791]: I1210 23:08:07.704489 4791 generic.go:334] "Generic (PLEG): container finished" podID="ebf742dc-fd16-49ae-a30a-b0f61e4d15fd" containerID="a3eeafd6ea8ce8bb3f96e912604f8a520c9ceb4ac7d92a96d2414c6cfb5c3ec8" exitCode=0 Dec 10 23:08:07 crc kubenswrapper[4791]: I1210 23:08:07.704554 4791 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-db-create-w8q8q" event={"ID":"ebf742dc-fd16-49ae-a30a-b0f61e4d15fd","Type":"ContainerDied","Data":"a3eeafd6ea8ce8bb3f96e912604f8a520c9ceb4ac7d92a96d2414c6cfb5c3ec8"} Dec 10 23:08:07 crc kubenswrapper[4791]: I1210 23:08:07.704581 4791 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-db-create-w8q8q" event={"ID":"ebf742dc-fd16-49ae-a30a-b0f61e4d15fd","Type":"ContainerStarted","Data":"e0244a841fe823a0f534b1facf13177c8298d659a1ac35056c3d09d67533f5b7"} Dec 10 23:08:07 crc kubenswrapper[4791]: I1210 23:08:07.707939 4791 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/neutron-e982-account-create-update-7fdhc" podStartSLOduration=2.707915862 podStartE2EDuration="2.707915862s" podCreationTimestamp="2025-12-10 23:08:05 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-10 23:08:07.705354967 +0000 UTC m=+1122.134972580" watchObservedRunningTime="2025-12-10 23:08:07.707915862 +0000 UTC m=+1122.137533475" Dec 10 23:08:07 crc kubenswrapper[4791]: I1210 23:08:07.709626 4791 generic.go:334] "Generic (PLEG): container finished" podID="25d98ac3-17a9-4885-a64a-2ae593acf177" containerID="88d56927bc9f6ed7c032ad8701f19d0e57298caa95444f144d523b4c8b5ef2bc" exitCode=0 Dec 10 23:08:07 crc kubenswrapper[4791]: I1210 23:08:07.709712 4791 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-a84b-account-create-update-7ct6w" event={"ID":"25d98ac3-17a9-4885-a64a-2ae593acf177","Type":"ContainerDied","Data":"88d56927bc9f6ed7c032ad8701f19d0e57298caa95444f144d523b4c8b5ef2bc"} Dec 10 23:08:07 crc kubenswrapper[4791]: I1210 23:08:07.709745 4791 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-a84b-account-create-update-7ct6w" event={"ID":"25d98ac3-17a9-4885-a64a-2ae593acf177","Type":"ContainerStarted","Data":"790e31a797bfe9a7b2a4c87ed397d0ce615c3bb26b0ad35b21c9d10d6b3e5d7a"} Dec 10 23:08:07 crc kubenswrapper[4791]: I1210 23:08:07.711425 4791 generic.go:334] "Generic (PLEG): container finished" podID="d2eecb39-349b-48d4-8819-aa392095d5d0" containerID="309e845d753d39f9e77682b06f34c9acbc81eb099f4a3042921ed02e899bf5e8" exitCode=0 Dec 10 23:08:07 crc kubenswrapper[4791]: I1210 23:08:07.711461 4791 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-db-create-58mvg" event={"ID":"d2eecb39-349b-48d4-8819-aa392095d5d0","Type":"ContainerDied","Data":"309e845d753d39f9e77682b06f34c9acbc81eb099f4a3042921ed02e899bf5e8"} Dec 10 23:08:08 crc kubenswrapper[4791]: I1210 23:08:08.721903 4791 generic.go:334] "Generic (PLEG): container finished" podID="e7649319-d8a5-48d8-b9f8-f236602c72a0" containerID="ef2c1e1726ac3fe2b5200e0a1b6289e8846bb2fc6d6fd1dbaa755fddce4936b5" exitCode=0 Dec 10 23:08:08 crc kubenswrapper[4791]: I1210 23:08:08.721963 4791 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-e982-account-create-update-7fdhc" event={"ID":"e7649319-d8a5-48d8-b9f8-f236602c72a0","Type":"ContainerDied","Data":"ef2c1e1726ac3fe2b5200e0a1b6289e8846bb2fc6d6fd1dbaa755fddce4936b5"} Dec 10 23:08:09 crc kubenswrapper[4791]: I1210 23:08:09.129401 4791 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-8c6c-account-create-update-s9h2s" Dec 10 23:08:09 crc kubenswrapper[4791]: I1210 23:08:09.270015 4791 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-5tqb8\" (UniqueName: \"kubernetes.io/projected/c289a9e6-9d03-499b-82f6-f0c898d94f04-kube-api-access-5tqb8\") pod \"c289a9e6-9d03-499b-82f6-f0c898d94f04\" (UID: \"c289a9e6-9d03-499b-82f6-f0c898d94f04\") " Dec 10 23:08:09 crc kubenswrapper[4791]: I1210 23:08:09.270282 4791 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/c289a9e6-9d03-499b-82f6-f0c898d94f04-operator-scripts\") pod \"c289a9e6-9d03-499b-82f6-f0c898d94f04\" (UID: \"c289a9e6-9d03-499b-82f6-f0c898d94f04\") " Dec 10 23:08:09 crc kubenswrapper[4791]: I1210 23:08:09.270972 4791 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/c289a9e6-9d03-499b-82f6-f0c898d94f04-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "c289a9e6-9d03-499b-82f6-f0c898d94f04" (UID: "c289a9e6-9d03-499b-82f6-f0c898d94f04"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 10 23:08:09 crc kubenswrapper[4791]: I1210 23:08:09.272404 4791 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/c289a9e6-9d03-499b-82f6-f0c898d94f04-operator-scripts\") on node \"crc\" DevicePath \"\"" Dec 10 23:08:09 crc kubenswrapper[4791]: I1210 23:08:09.273286 4791 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-a84b-account-create-update-7ct6w" Dec 10 23:08:09 crc kubenswrapper[4791]: I1210 23:08:09.278198 4791 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/c289a9e6-9d03-499b-82f6-f0c898d94f04-kube-api-access-5tqb8" (OuterVolumeSpecName: "kube-api-access-5tqb8") pod "c289a9e6-9d03-499b-82f6-f0c898d94f04" (UID: "c289a9e6-9d03-499b-82f6-f0c898d94f04"). InnerVolumeSpecName "kube-api-access-5tqb8". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 10 23:08:09 crc kubenswrapper[4791]: I1210 23:08:09.282788 4791 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-db-create-58mvg" Dec 10 23:08:09 crc kubenswrapper[4791]: I1210 23:08:09.293797 4791 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-db-create-w8q8q" Dec 10 23:08:09 crc kubenswrapper[4791]: I1210 23:08:09.298729 4791 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-db-create-l5xkj" Dec 10 23:08:09 crc kubenswrapper[4791]: I1210 23:08:09.374583 4791 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/25d98ac3-17a9-4885-a64a-2ae593acf177-operator-scripts\") pod \"25d98ac3-17a9-4885-a64a-2ae593acf177\" (UID: \"25d98ac3-17a9-4885-a64a-2ae593acf177\") " Dec 10 23:08:09 crc kubenswrapper[4791]: I1210 23:08:09.374671 4791 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-2w8nc\" (UniqueName: \"kubernetes.io/projected/25d98ac3-17a9-4885-a64a-2ae593acf177-kube-api-access-2w8nc\") pod \"25d98ac3-17a9-4885-a64a-2ae593acf177\" (UID: \"25d98ac3-17a9-4885-a64a-2ae593acf177\") " Dec 10 23:08:09 crc kubenswrapper[4791]: I1210 23:08:09.375080 4791 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-5tqb8\" (UniqueName: \"kubernetes.io/projected/c289a9e6-9d03-499b-82f6-f0c898d94f04-kube-api-access-5tqb8\") on node \"crc\" DevicePath \"\"" Dec 10 23:08:09 crc kubenswrapper[4791]: I1210 23:08:09.376197 4791 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/25d98ac3-17a9-4885-a64a-2ae593acf177-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "25d98ac3-17a9-4885-a64a-2ae593acf177" (UID: "25d98ac3-17a9-4885-a64a-2ae593acf177"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 10 23:08:09 crc kubenswrapper[4791]: I1210 23:08:09.379268 4791 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/25d98ac3-17a9-4885-a64a-2ae593acf177-kube-api-access-2w8nc" (OuterVolumeSpecName: "kube-api-access-2w8nc") pod "25d98ac3-17a9-4885-a64a-2ae593acf177" (UID: "25d98ac3-17a9-4885-a64a-2ae593acf177"). InnerVolumeSpecName "kube-api-access-2w8nc". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 10 23:08:09 crc kubenswrapper[4791]: I1210 23:08:09.476642 4791 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-tjvdl\" (UniqueName: \"kubernetes.io/projected/77886642-7b06-476f-acb0-7e4df70190a8-kube-api-access-tjvdl\") pod \"77886642-7b06-476f-acb0-7e4df70190a8\" (UID: \"77886642-7b06-476f-acb0-7e4df70190a8\") " Dec 10 23:08:09 crc kubenswrapper[4791]: I1210 23:08:09.476726 4791 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-rk79k\" (UniqueName: \"kubernetes.io/projected/d2eecb39-349b-48d4-8819-aa392095d5d0-kube-api-access-rk79k\") pod \"d2eecb39-349b-48d4-8819-aa392095d5d0\" (UID: \"d2eecb39-349b-48d4-8819-aa392095d5d0\") " Dec 10 23:08:09 crc kubenswrapper[4791]: I1210 23:08:09.476753 4791 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-hltjn\" (UniqueName: \"kubernetes.io/projected/ebf742dc-fd16-49ae-a30a-b0f61e4d15fd-kube-api-access-hltjn\") pod \"ebf742dc-fd16-49ae-a30a-b0f61e4d15fd\" (UID: \"ebf742dc-fd16-49ae-a30a-b0f61e4d15fd\") " Dec 10 23:08:09 crc kubenswrapper[4791]: I1210 23:08:09.476807 4791 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/d2eecb39-349b-48d4-8819-aa392095d5d0-operator-scripts\") pod \"d2eecb39-349b-48d4-8819-aa392095d5d0\" (UID: \"d2eecb39-349b-48d4-8819-aa392095d5d0\") " Dec 10 23:08:09 crc kubenswrapper[4791]: I1210 23:08:09.476890 4791 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/77886642-7b06-476f-acb0-7e4df70190a8-operator-scripts\") pod \"77886642-7b06-476f-acb0-7e4df70190a8\" (UID: \"77886642-7b06-476f-acb0-7e4df70190a8\") " Dec 10 23:08:09 crc kubenswrapper[4791]: I1210 23:08:09.476921 4791 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/ebf742dc-fd16-49ae-a30a-b0f61e4d15fd-operator-scripts\") pod \"ebf742dc-fd16-49ae-a30a-b0f61e4d15fd\" (UID: \"ebf742dc-fd16-49ae-a30a-b0f61e4d15fd\") " Dec 10 23:08:09 crc kubenswrapper[4791]: I1210 23:08:09.477272 4791 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-2w8nc\" (UniqueName: \"kubernetes.io/projected/25d98ac3-17a9-4885-a64a-2ae593acf177-kube-api-access-2w8nc\") on node \"crc\" DevicePath \"\"" Dec 10 23:08:09 crc kubenswrapper[4791]: I1210 23:08:09.477285 4791 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/25d98ac3-17a9-4885-a64a-2ae593acf177-operator-scripts\") on node \"crc\" DevicePath \"\"" Dec 10 23:08:09 crc kubenswrapper[4791]: I1210 23:08:09.478454 4791 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/ebf742dc-fd16-49ae-a30a-b0f61e4d15fd-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "ebf742dc-fd16-49ae-a30a-b0f61e4d15fd" (UID: "ebf742dc-fd16-49ae-a30a-b0f61e4d15fd"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 10 23:08:09 crc kubenswrapper[4791]: I1210 23:08:09.478511 4791 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/d2eecb39-349b-48d4-8819-aa392095d5d0-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "d2eecb39-349b-48d4-8819-aa392095d5d0" (UID: "d2eecb39-349b-48d4-8819-aa392095d5d0"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 10 23:08:09 crc kubenswrapper[4791]: I1210 23:08:09.478753 4791 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/77886642-7b06-476f-acb0-7e4df70190a8-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "77886642-7b06-476f-acb0-7e4df70190a8" (UID: "77886642-7b06-476f-acb0-7e4df70190a8"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 10 23:08:09 crc kubenswrapper[4791]: I1210 23:08:09.481144 4791 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/ebf742dc-fd16-49ae-a30a-b0f61e4d15fd-kube-api-access-hltjn" (OuterVolumeSpecName: "kube-api-access-hltjn") pod "ebf742dc-fd16-49ae-a30a-b0f61e4d15fd" (UID: "ebf742dc-fd16-49ae-a30a-b0f61e4d15fd"). InnerVolumeSpecName "kube-api-access-hltjn". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 10 23:08:09 crc kubenswrapper[4791]: I1210 23:08:09.481176 4791 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/d2eecb39-349b-48d4-8819-aa392095d5d0-kube-api-access-rk79k" (OuterVolumeSpecName: "kube-api-access-rk79k") pod "d2eecb39-349b-48d4-8819-aa392095d5d0" (UID: "d2eecb39-349b-48d4-8819-aa392095d5d0"). InnerVolumeSpecName "kube-api-access-rk79k". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 10 23:08:09 crc kubenswrapper[4791]: I1210 23:08:09.481825 4791 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/77886642-7b06-476f-acb0-7e4df70190a8-kube-api-access-tjvdl" (OuterVolumeSpecName: "kube-api-access-tjvdl") pod "77886642-7b06-476f-acb0-7e4df70190a8" (UID: "77886642-7b06-476f-acb0-7e4df70190a8"). InnerVolumeSpecName "kube-api-access-tjvdl". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 10 23:08:09 crc kubenswrapper[4791]: I1210 23:08:09.579058 4791 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/77886642-7b06-476f-acb0-7e4df70190a8-operator-scripts\") on node \"crc\" DevicePath \"\"" Dec 10 23:08:09 crc kubenswrapper[4791]: I1210 23:08:09.579124 4791 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/ebf742dc-fd16-49ae-a30a-b0f61e4d15fd-operator-scripts\") on node \"crc\" DevicePath \"\"" Dec 10 23:08:09 crc kubenswrapper[4791]: I1210 23:08:09.579142 4791 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-tjvdl\" (UniqueName: \"kubernetes.io/projected/77886642-7b06-476f-acb0-7e4df70190a8-kube-api-access-tjvdl\") on node \"crc\" DevicePath \"\"" Dec 10 23:08:09 crc kubenswrapper[4791]: I1210 23:08:09.579164 4791 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-rk79k\" (UniqueName: \"kubernetes.io/projected/d2eecb39-349b-48d4-8819-aa392095d5d0-kube-api-access-rk79k\") on node \"crc\" DevicePath \"\"" Dec 10 23:08:09 crc kubenswrapper[4791]: I1210 23:08:09.579181 4791 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-hltjn\" (UniqueName: \"kubernetes.io/projected/ebf742dc-fd16-49ae-a30a-b0f61e4d15fd-kube-api-access-hltjn\") on node \"crc\" DevicePath \"\"" Dec 10 23:08:09 crc kubenswrapper[4791]: I1210 23:08:09.579198 4791 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/d2eecb39-349b-48d4-8819-aa392095d5d0-operator-scripts\") on node \"crc\" DevicePath \"\"" Dec 10 23:08:09 crc kubenswrapper[4791]: I1210 23:08:09.731469 4791 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-db-create-w8q8q" Dec 10 23:08:09 crc kubenswrapper[4791]: I1210 23:08:09.731488 4791 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-db-create-w8q8q" event={"ID":"ebf742dc-fd16-49ae-a30a-b0f61e4d15fd","Type":"ContainerDied","Data":"e0244a841fe823a0f534b1facf13177c8298d659a1ac35056c3d09d67533f5b7"} Dec 10 23:08:09 crc kubenswrapper[4791]: I1210 23:08:09.731514 4791 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="e0244a841fe823a0f534b1facf13177c8298d659a1ac35056c3d09d67533f5b7" Dec 10 23:08:09 crc kubenswrapper[4791]: I1210 23:08:09.735591 4791 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-a84b-account-create-update-7ct6w" event={"ID":"25d98ac3-17a9-4885-a64a-2ae593acf177","Type":"ContainerDied","Data":"790e31a797bfe9a7b2a4c87ed397d0ce615c3bb26b0ad35b21c9d10d6b3e5d7a"} Dec 10 23:08:09 crc kubenswrapper[4791]: I1210 23:08:09.735619 4791 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-a84b-account-create-update-7ct6w" Dec 10 23:08:09 crc kubenswrapper[4791]: I1210 23:08:09.735625 4791 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="790e31a797bfe9a7b2a4c87ed397d0ce615c3bb26b0ad35b21c9d10d6b3e5d7a" Dec 10 23:08:09 crc kubenswrapper[4791]: I1210 23:08:09.736917 4791 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-8c6c-account-create-update-s9h2s" event={"ID":"c289a9e6-9d03-499b-82f6-f0c898d94f04","Type":"ContainerDied","Data":"92d9542499a73fe438456a34ab37ad468543662df3ee074ea6eea436b7c35ff8"} Dec 10 23:08:09 crc kubenswrapper[4791]: I1210 23:08:09.736943 4791 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="92d9542499a73fe438456a34ab37ad468543662df3ee074ea6eea436b7c35ff8" Dec 10 23:08:09 crc kubenswrapper[4791]: I1210 23:08:09.736924 4791 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-8c6c-account-create-update-s9h2s" Dec 10 23:08:09 crc kubenswrapper[4791]: I1210 23:08:09.737949 4791 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-db-create-58mvg" event={"ID":"d2eecb39-349b-48d4-8819-aa392095d5d0","Type":"ContainerDied","Data":"e2ccba20ad7b61bf8f034eaa2c850e0f2bfe5076162c2c6fff7062ccae7cb3c6"} Dec 10 23:08:09 crc kubenswrapper[4791]: I1210 23:08:09.737971 4791 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="e2ccba20ad7b61bf8f034eaa2c850e0f2bfe5076162c2c6fff7062ccae7cb3c6" Dec 10 23:08:09 crc kubenswrapper[4791]: I1210 23:08:09.738008 4791 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-db-create-58mvg" Dec 10 23:08:09 crc kubenswrapper[4791]: I1210 23:08:09.739294 4791 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-db-create-l5xkj" event={"ID":"77886642-7b06-476f-acb0-7e4df70190a8","Type":"ContainerDied","Data":"448c993671b8c2b72d693d3e7d98608e88c6ed3efe48c0c4eacb89a661bc799d"} Dec 10 23:08:09 crc kubenswrapper[4791]: I1210 23:08:09.739314 4791 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-db-create-l5xkj" Dec 10 23:08:09 crc kubenswrapper[4791]: I1210 23:08:09.739366 4791 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="448c993671b8c2b72d693d3e7d98608e88c6ed3efe48c0c4eacb89a661bc799d" Dec 10 23:08:14 crc kubenswrapper[4791]: I1210 23:08:14.677164 4791 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-e982-account-create-update-7fdhc" Dec 10 23:08:14 crc kubenswrapper[4791]: I1210 23:08:14.727852 4791 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/e7649319-d8a5-48d8-b9f8-f236602c72a0-operator-scripts\") pod \"e7649319-d8a5-48d8-b9f8-f236602c72a0\" (UID: \"e7649319-d8a5-48d8-b9f8-f236602c72a0\") " Dec 10 23:08:14 crc kubenswrapper[4791]: I1210 23:08:14.728057 4791 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-k9ps6\" (UniqueName: \"kubernetes.io/projected/e7649319-d8a5-48d8-b9f8-f236602c72a0-kube-api-access-k9ps6\") pod \"e7649319-d8a5-48d8-b9f8-f236602c72a0\" (UID: \"e7649319-d8a5-48d8-b9f8-f236602c72a0\") " Dec 10 23:08:14 crc kubenswrapper[4791]: I1210 23:08:14.728944 4791 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/e7649319-d8a5-48d8-b9f8-f236602c72a0-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "e7649319-d8a5-48d8-b9f8-f236602c72a0" (UID: "e7649319-d8a5-48d8-b9f8-f236602c72a0"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 10 23:08:14 crc kubenswrapper[4791]: I1210 23:08:14.734270 4791 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/e7649319-d8a5-48d8-b9f8-f236602c72a0-kube-api-access-k9ps6" (OuterVolumeSpecName: "kube-api-access-k9ps6") pod "e7649319-d8a5-48d8-b9f8-f236602c72a0" (UID: "e7649319-d8a5-48d8-b9f8-f236602c72a0"). InnerVolumeSpecName "kube-api-access-k9ps6". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 10 23:08:14 crc kubenswrapper[4791]: I1210 23:08:14.795958 4791 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-e982-account-create-update-7fdhc" event={"ID":"e7649319-d8a5-48d8-b9f8-f236602c72a0","Type":"ContainerDied","Data":"a51125ec49b7859f40dafcedb211f7468d71598b5858e13e7cf23c73fe38a383"} Dec 10 23:08:14 crc kubenswrapper[4791]: I1210 23:08:14.796006 4791 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="a51125ec49b7859f40dafcedb211f7468d71598b5858e13e7cf23c73fe38a383" Dec 10 23:08:14 crc kubenswrapper[4791]: I1210 23:08:14.796062 4791 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-e982-account-create-update-7fdhc" Dec 10 23:08:14 crc kubenswrapper[4791]: I1210 23:08:14.829658 4791 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-k9ps6\" (UniqueName: \"kubernetes.io/projected/e7649319-d8a5-48d8-b9f8-f236602c72a0-kube-api-access-k9ps6\") on node \"crc\" DevicePath \"\"" Dec 10 23:08:14 crc kubenswrapper[4791]: I1210 23:08:14.829700 4791 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/e7649319-d8a5-48d8-b9f8-f236602c72a0-operator-scripts\") on node \"crc\" DevicePath \"\"" Dec 10 23:08:15 crc kubenswrapper[4791]: I1210 23:08:15.804011 4791 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-db-sync-ccl4n" event={"ID":"a2c44c96-70c2-414a-9890-cb95b66f10b2","Type":"ContainerStarted","Data":"1246dddd2dc57e9b79396e6ae705e086d5153488fee1cd18e4dfa7a916a08eed"} Dec 10 23:08:15 crc kubenswrapper[4791]: I1210 23:08:15.805112 4791 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-db-sync-kt8jj" event={"ID":"5ad26083-9c93-4ec4-9159-ef991f116f80","Type":"ContainerStarted","Data":"e728059b66b1ca7c2831961436238782b27e22a524a6ac12d7c043a19d44d59e"} Dec 10 23:08:15 crc kubenswrapper[4791]: I1210 23:08:15.821630 4791 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/glance-db-sync-ccl4n" podStartSLOduration=2.807303364 podStartE2EDuration="38.821610612s" podCreationTimestamp="2025-12-10 23:07:37 +0000 UTC" firstStartedPulling="2025-12-10 23:07:38.661549454 +0000 UTC m=+1093.091167067" lastFinishedPulling="2025-12-10 23:08:14.675856692 +0000 UTC m=+1129.105474315" observedRunningTime="2025-12-10 23:08:15.82014242 +0000 UTC m=+1130.249760033" watchObservedRunningTime="2025-12-10 23:08:15.821610612 +0000 UTC m=+1130.251228225" Dec 10 23:08:15 crc kubenswrapper[4791]: I1210 23:08:15.839702 4791 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/keystone-db-sync-kt8jj" podStartSLOduration=2.263786719 podStartE2EDuration="9.839667095s" podCreationTimestamp="2025-12-10 23:08:06 +0000 UTC" firstStartedPulling="2025-12-10 23:08:07.127255445 +0000 UTC m=+1121.556873058" lastFinishedPulling="2025-12-10 23:08:14.703135781 +0000 UTC m=+1129.132753434" observedRunningTime="2025-12-10 23:08:15.83153882 +0000 UTC m=+1130.261156433" watchObservedRunningTime="2025-12-10 23:08:15.839667095 +0000 UTC m=+1130.269284708" Dec 10 23:08:17 crc kubenswrapper[4791]: I1210 23:08:17.827766 4791 generic.go:334] "Generic (PLEG): container finished" podID="5ad26083-9c93-4ec4-9159-ef991f116f80" containerID="e728059b66b1ca7c2831961436238782b27e22a524a6ac12d7c043a19d44d59e" exitCode=0 Dec 10 23:08:17 crc kubenswrapper[4791]: I1210 23:08:17.827897 4791 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-db-sync-kt8jj" event={"ID":"5ad26083-9c93-4ec4-9159-ef991f116f80","Type":"ContainerDied","Data":"e728059b66b1ca7c2831961436238782b27e22a524a6ac12d7c043a19d44d59e"} Dec 10 23:08:19 crc kubenswrapper[4791]: I1210 23:08:19.102991 4791 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-db-sync-kt8jj" Dec 10 23:08:19 crc kubenswrapper[4791]: I1210 23:08:19.211493 4791 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/5ad26083-9c93-4ec4-9159-ef991f116f80-combined-ca-bundle\") pod \"5ad26083-9c93-4ec4-9159-ef991f116f80\" (UID: \"5ad26083-9c93-4ec4-9159-ef991f116f80\") " Dec 10 23:08:19 crc kubenswrapper[4791]: I1210 23:08:19.211807 4791 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/5ad26083-9c93-4ec4-9159-ef991f116f80-config-data\") pod \"5ad26083-9c93-4ec4-9159-ef991f116f80\" (UID: \"5ad26083-9c93-4ec4-9159-ef991f116f80\") " Dec 10 23:08:19 crc kubenswrapper[4791]: I1210 23:08:19.211877 4791 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-6rdgm\" (UniqueName: \"kubernetes.io/projected/5ad26083-9c93-4ec4-9159-ef991f116f80-kube-api-access-6rdgm\") pod \"5ad26083-9c93-4ec4-9159-ef991f116f80\" (UID: \"5ad26083-9c93-4ec4-9159-ef991f116f80\") " Dec 10 23:08:19 crc kubenswrapper[4791]: I1210 23:08:19.229466 4791 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/5ad26083-9c93-4ec4-9159-ef991f116f80-kube-api-access-6rdgm" (OuterVolumeSpecName: "kube-api-access-6rdgm") pod "5ad26083-9c93-4ec4-9159-ef991f116f80" (UID: "5ad26083-9c93-4ec4-9159-ef991f116f80"). InnerVolumeSpecName "kube-api-access-6rdgm". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 10 23:08:19 crc kubenswrapper[4791]: I1210 23:08:19.234976 4791 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/5ad26083-9c93-4ec4-9159-ef991f116f80-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "5ad26083-9c93-4ec4-9159-ef991f116f80" (UID: "5ad26083-9c93-4ec4-9159-ef991f116f80"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 10 23:08:19 crc kubenswrapper[4791]: I1210 23:08:19.252990 4791 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/5ad26083-9c93-4ec4-9159-ef991f116f80-config-data" (OuterVolumeSpecName: "config-data") pod "5ad26083-9c93-4ec4-9159-ef991f116f80" (UID: "5ad26083-9c93-4ec4-9159-ef991f116f80"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 10 23:08:19 crc kubenswrapper[4791]: I1210 23:08:19.313534 4791 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-6rdgm\" (UniqueName: \"kubernetes.io/projected/5ad26083-9c93-4ec4-9159-ef991f116f80-kube-api-access-6rdgm\") on node \"crc\" DevicePath \"\"" Dec 10 23:08:19 crc kubenswrapper[4791]: I1210 23:08:19.313570 4791 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/5ad26083-9c93-4ec4-9159-ef991f116f80-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 10 23:08:19 crc kubenswrapper[4791]: I1210 23:08:19.313580 4791 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/5ad26083-9c93-4ec4-9159-ef991f116f80-config-data\") on node \"crc\" DevicePath \"\"" Dec 10 23:08:19 crc kubenswrapper[4791]: I1210 23:08:19.844532 4791 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-db-sync-kt8jj" event={"ID":"5ad26083-9c93-4ec4-9159-ef991f116f80","Type":"ContainerDied","Data":"ed5a216809726f12930fc319273af79fa196ddf4ddc99fb1482154e7f7309231"} Dec 10 23:08:19 crc kubenswrapper[4791]: I1210 23:08:19.844583 4791 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-db-sync-kt8jj" Dec 10 23:08:19 crc kubenswrapper[4791]: I1210 23:08:19.844591 4791 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="ed5a216809726f12930fc319273af79fa196ddf4ddc99fb1482154e7f7309231" Dec 10 23:08:20 crc kubenswrapper[4791]: I1210 23:08:20.094429 4791 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-55fff446b9-84xqj"] Dec 10 23:08:20 crc kubenswrapper[4791]: E1210 23:08:20.094802 4791 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="e7649319-d8a5-48d8-b9f8-f236602c72a0" containerName="mariadb-account-create-update" Dec 10 23:08:20 crc kubenswrapper[4791]: I1210 23:08:20.094814 4791 state_mem.go:107] "Deleted CPUSet assignment" podUID="e7649319-d8a5-48d8-b9f8-f236602c72a0" containerName="mariadb-account-create-update" Dec 10 23:08:20 crc kubenswrapper[4791]: E1210 23:08:20.094828 4791 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="d2eecb39-349b-48d4-8819-aa392095d5d0" containerName="mariadb-database-create" Dec 10 23:08:20 crc kubenswrapper[4791]: I1210 23:08:20.094834 4791 state_mem.go:107] "Deleted CPUSet assignment" podUID="d2eecb39-349b-48d4-8819-aa392095d5d0" containerName="mariadb-database-create" Dec 10 23:08:20 crc kubenswrapper[4791]: E1210 23:08:20.094841 4791 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="c289a9e6-9d03-499b-82f6-f0c898d94f04" containerName="mariadb-account-create-update" Dec 10 23:08:20 crc kubenswrapper[4791]: I1210 23:08:20.094847 4791 state_mem.go:107] "Deleted CPUSet assignment" podUID="c289a9e6-9d03-499b-82f6-f0c898d94f04" containerName="mariadb-account-create-update" Dec 10 23:08:20 crc kubenswrapper[4791]: E1210 23:08:20.094869 4791 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="5ad26083-9c93-4ec4-9159-ef991f116f80" containerName="keystone-db-sync" Dec 10 23:08:20 crc kubenswrapper[4791]: I1210 23:08:20.094875 4791 state_mem.go:107] "Deleted CPUSet assignment" podUID="5ad26083-9c93-4ec4-9159-ef991f116f80" containerName="keystone-db-sync" Dec 10 23:08:20 crc kubenswrapper[4791]: E1210 23:08:20.094886 4791 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="77886642-7b06-476f-acb0-7e4df70190a8" containerName="mariadb-database-create" Dec 10 23:08:20 crc kubenswrapper[4791]: I1210 23:08:20.094892 4791 state_mem.go:107] "Deleted CPUSet assignment" podUID="77886642-7b06-476f-acb0-7e4df70190a8" containerName="mariadb-database-create" Dec 10 23:08:20 crc kubenswrapper[4791]: E1210 23:08:20.094900 4791 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="25d98ac3-17a9-4885-a64a-2ae593acf177" containerName="mariadb-account-create-update" Dec 10 23:08:20 crc kubenswrapper[4791]: I1210 23:08:20.094906 4791 state_mem.go:107] "Deleted CPUSet assignment" podUID="25d98ac3-17a9-4885-a64a-2ae593acf177" containerName="mariadb-account-create-update" Dec 10 23:08:20 crc kubenswrapper[4791]: E1210 23:08:20.094920 4791 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="ebf742dc-fd16-49ae-a30a-b0f61e4d15fd" containerName="mariadb-database-create" Dec 10 23:08:20 crc kubenswrapper[4791]: I1210 23:08:20.094925 4791 state_mem.go:107] "Deleted CPUSet assignment" podUID="ebf742dc-fd16-49ae-a30a-b0f61e4d15fd" containerName="mariadb-database-create" Dec 10 23:08:20 crc kubenswrapper[4791]: I1210 23:08:20.095081 4791 memory_manager.go:354] "RemoveStaleState removing state" podUID="d2eecb39-349b-48d4-8819-aa392095d5d0" containerName="mariadb-database-create" Dec 10 23:08:20 crc kubenswrapper[4791]: I1210 23:08:20.095093 4791 memory_manager.go:354] "RemoveStaleState removing state" podUID="c289a9e6-9d03-499b-82f6-f0c898d94f04" containerName="mariadb-account-create-update" Dec 10 23:08:20 crc kubenswrapper[4791]: I1210 23:08:20.095110 4791 memory_manager.go:354] "RemoveStaleState removing state" podUID="5ad26083-9c93-4ec4-9159-ef991f116f80" containerName="keystone-db-sync" Dec 10 23:08:20 crc kubenswrapper[4791]: I1210 23:08:20.095119 4791 memory_manager.go:354] "RemoveStaleState removing state" podUID="ebf742dc-fd16-49ae-a30a-b0f61e4d15fd" containerName="mariadb-database-create" Dec 10 23:08:20 crc kubenswrapper[4791]: I1210 23:08:20.095131 4791 memory_manager.go:354] "RemoveStaleState removing state" podUID="25d98ac3-17a9-4885-a64a-2ae593acf177" containerName="mariadb-account-create-update" Dec 10 23:08:20 crc kubenswrapper[4791]: I1210 23:08:20.095140 4791 memory_manager.go:354] "RemoveStaleState removing state" podUID="77886642-7b06-476f-acb0-7e4df70190a8" containerName="mariadb-database-create" Dec 10 23:08:20 crc kubenswrapper[4791]: I1210 23:08:20.095151 4791 memory_manager.go:354] "RemoveStaleState removing state" podUID="e7649319-d8a5-48d8-b9f8-f236602c72a0" containerName="mariadb-account-create-update" Dec 10 23:08:20 crc kubenswrapper[4791]: I1210 23:08:20.096011 4791 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-55fff446b9-84xqj" Dec 10 23:08:20 crc kubenswrapper[4791]: I1210 23:08:20.119526 4791 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-55fff446b9-84xqj"] Dec 10 23:08:20 crc kubenswrapper[4791]: I1210 23:08:20.125738 4791 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/a9741492-6c12-4225-bab8-f0d6cd39c969-dns-swift-storage-0\") pod \"dnsmasq-dns-55fff446b9-84xqj\" (UID: \"a9741492-6c12-4225-bab8-f0d6cd39c969\") " pod="openstack/dnsmasq-dns-55fff446b9-84xqj" Dec 10 23:08:20 crc kubenswrapper[4791]: I1210 23:08:20.125791 4791 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/a9741492-6c12-4225-bab8-f0d6cd39c969-ovsdbserver-sb\") pod \"dnsmasq-dns-55fff446b9-84xqj\" (UID: \"a9741492-6c12-4225-bab8-f0d6cd39c969\") " pod="openstack/dnsmasq-dns-55fff446b9-84xqj" Dec 10 23:08:20 crc kubenswrapper[4791]: I1210 23:08:20.125856 4791 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-qznw7\" (UniqueName: \"kubernetes.io/projected/a9741492-6c12-4225-bab8-f0d6cd39c969-kube-api-access-qznw7\") pod \"dnsmasq-dns-55fff446b9-84xqj\" (UID: \"a9741492-6c12-4225-bab8-f0d6cd39c969\") " pod="openstack/dnsmasq-dns-55fff446b9-84xqj" Dec 10 23:08:20 crc kubenswrapper[4791]: I1210 23:08:20.125894 4791 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/a9741492-6c12-4225-bab8-f0d6cd39c969-ovsdbserver-nb\") pod \"dnsmasq-dns-55fff446b9-84xqj\" (UID: \"a9741492-6c12-4225-bab8-f0d6cd39c969\") " pod="openstack/dnsmasq-dns-55fff446b9-84xqj" Dec 10 23:08:20 crc kubenswrapper[4791]: I1210 23:08:20.125915 4791 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/a9741492-6c12-4225-bab8-f0d6cd39c969-config\") pod \"dnsmasq-dns-55fff446b9-84xqj\" (UID: \"a9741492-6c12-4225-bab8-f0d6cd39c969\") " pod="openstack/dnsmasq-dns-55fff446b9-84xqj" Dec 10 23:08:20 crc kubenswrapper[4791]: I1210 23:08:20.125935 4791 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/a9741492-6c12-4225-bab8-f0d6cd39c969-dns-svc\") pod \"dnsmasq-dns-55fff446b9-84xqj\" (UID: \"a9741492-6c12-4225-bab8-f0d6cd39c969\") " pod="openstack/dnsmasq-dns-55fff446b9-84xqj" Dec 10 23:08:20 crc kubenswrapper[4791]: I1210 23:08:20.144488 4791 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/keystone-bootstrap-tbpq4"] Dec 10 23:08:20 crc kubenswrapper[4791]: I1210 23:08:20.145640 4791 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-bootstrap-tbpq4" Dec 10 23:08:20 crc kubenswrapper[4791]: I1210 23:08:20.150826 4791 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone-config-data" Dec 10 23:08:20 crc kubenswrapper[4791]: I1210 23:08:20.150988 4791 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone-keystone-dockercfg-t99v8" Dec 10 23:08:20 crc kubenswrapper[4791]: I1210 23:08:20.151888 4791 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone" Dec 10 23:08:20 crc kubenswrapper[4791]: I1210 23:08:20.152083 4791 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"osp-secret" Dec 10 23:08:20 crc kubenswrapper[4791]: I1210 23:08:20.152734 4791 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone-scripts" Dec 10 23:08:20 crc kubenswrapper[4791]: I1210 23:08:20.164666 4791 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/keystone-bootstrap-tbpq4"] Dec 10 23:08:20 crc kubenswrapper[4791]: I1210 23:08:20.227622 4791 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/a9741492-6c12-4225-bab8-f0d6cd39c969-dns-svc\") pod \"dnsmasq-dns-55fff446b9-84xqj\" (UID: \"a9741492-6c12-4225-bab8-f0d6cd39c969\") " pod="openstack/dnsmasq-dns-55fff446b9-84xqj" Dec 10 23:08:20 crc kubenswrapper[4791]: I1210 23:08:20.227701 4791 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/a9741492-6c12-4225-bab8-f0d6cd39c969-dns-swift-storage-0\") pod \"dnsmasq-dns-55fff446b9-84xqj\" (UID: \"a9741492-6c12-4225-bab8-f0d6cd39c969\") " pod="openstack/dnsmasq-dns-55fff446b9-84xqj" Dec 10 23:08:20 crc kubenswrapper[4791]: I1210 23:08:20.227757 4791 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/a9741492-6c12-4225-bab8-f0d6cd39c969-ovsdbserver-sb\") pod \"dnsmasq-dns-55fff446b9-84xqj\" (UID: \"a9741492-6c12-4225-bab8-f0d6cd39c969\") " pod="openstack/dnsmasq-dns-55fff446b9-84xqj" Dec 10 23:08:20 crc kubenswrapper[4791]: I1210 23:08:20.227849 4791 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-qznw7\" (UniqueName: \"kubernetes.io/projected/a9741492-6c12-4225-bab8-f0d6cd39c969-kube-api-access-qznw7\") pod \"dnsmasq-dns-55fff446b9-84xqj\" (UID: \"a9741492-6c12-4225-bab8-f0d6cd39c969\") " pod="openstack/dnsmasq-dns-55fff446b9-84xqj" Dec 10 23:08:20 crc kubenswrapper[4791]: I1210 23:08:20.227916 4791 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/a9741492-6c12-4225-bab8-f0d6cd39c969-ovsdbserver-nb\") pod \"dnsmasq-dns-55fff446b9-84xqj\" (UID: \"a9741492-6c12-4225-bab8-f0d6cd39c969\") " pod="openstack/dnsmasq-dns-55fff446b9-84xqj" Dec 10 23:08:20 crc kubenswrapper[4791]: I1210 23:08:20.227941 4791 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/a9741492-6c12-4225-bab8-f0d6cd39c969-config\") pod \"dnsmasq-dns-55fff446b9-84xqj\" (UID: \"a9741492-6c12-4225-bab8-f0d6cd39c969\") " pod="openstack/dnsmasq-dns-55fff446b9-84xqj" Dec 10 23:08:20 crc kubenswrapper[4791]: I1210 23:08:20.229239 4791 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/a9741492-6c12-4225-bab8-f0d6cd39c969-config\") pod \"dnsmasq-dns-55fff446b9-84xqj\" (UID: \"a9741492-6c12-4225-bab8-f0d6cd39c969\") " pod="openstack/dnsmasq-dns-55fff446b9-84xqj" Dec 10 23:08:20 crc kubenswrapper[4791]: I1210 23:08:20.230041 4791 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/a9741492-6c12-4225-bab8-f0d6cd39c969-ovsdbserver-sb\") pod \"dnsmasq-dns-55fff446b9-84xqj\" (UID: \"a9741492-6c12-4225-bab8-f0d6cd39c969\") " pod="openstack/dnsmasq-dns-55fff446b9-84xqj" Dec 10 23:08:20 crc kubenswrapper[4791]: I1210 23:08:20.231854 4791 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/a9741492-6c12-4225-bab8-f0d6cd39c969-dns-svc\") pod \"dnsmasq-dns-55fff446b9-84xqj\" (UID: \"a9741492-6c12-4225-bab8-f0d6cd39c969\") " pod="openstack/dnsmasq-dns-55fff446b9-84xqj" Dec 10 23:08:20 crc kubenswrapper[4791]: I1210 23:08:20.231914 4791 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/a9741492-6c12-4225-bab8-f0d6cd39c969-ovsdbserver-nb\") pod \"dnsmasq-dns-55fff446b9-84xqj\" (UID: \"a9741492-6c12-4225-bab8-f0d6cd39c969\") " pod="openstack/dnsmasq-dns-55fff446b9-84xqj" Dec 10 23:08:20 crc kubenswrapper[4791]: I1210 23:08:20.233479 4791 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/a9741492-6c12-4225-bab8-f0d6cd39c969-dns-swift-storage-0\") pod \"dnsmasq-dns-55fff446b9-84xqj\" (UID: \"a9741492-6c12-4225-bab8-f0d6cd39c969\") " pod="openstack/dnsmasq-dns-55fff446b9-84xqj" Dec 10 23:08:20 crc kubenswrapper[4791]: I1210 23:08:20.266538 4791 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-qznw7\" (UniqueName: \"kubernetes.io/projected/a9741492-6c12-4225-bab8-f0d6cd39c969-kube-api-access-qznw7\") pod \"dnsmasq-dns-55fff446b9-84xqj\" (UID: \"a9741492-6c12-4225-bab8-f0d6cd39c969\") " pod="openstack/dnsmasq-dns-55fff446b9-84xqj" Dec 10 23:08:20 crc kubenswrapper[4791]: I1210 23:08:20.294847 4791 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/horizon-78fc5d56dc-58cjt"] Dec 10 23:08:20 crc kubenswrapper[4791]: I1210 23:08:20.296439 4791 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/horizon-78fc5d56dc-58cjt" Dec 10 23:08:20 crc kubenswrapper[4791]: I1210 23:08:20.299264 4791 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"horizon-horizon-dockercfg-dz272" Dec 10 23:08:20 crc kubenswrapper[4791]: I1210 23:08:20.299428 4791 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"horizon-scripts" Dec 10 23:08:20 crc kubenswrapper[4791]: I1210 23:08:20.299550 4791 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"horizon-config-data" Dec 10 23:08:20 crc kubenswrapper[4791]: I1210 23:08:20.299651 4791 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"horizon" Dec 10 23:08:20 crc kubenswrapper[4791]: I1210 23:08:20.329291 4791 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/f7db40d5-1f25-4bab-98af-29999f12acf1-config-data\") pod \"keystone-bootstrap-tbpq4\" (UID: \"f7db40d5-1f25-4bab-98af-29999f12acf1\") " pod="openstack/keystone-bootstrap-tbpq4" Dec 10 23:08:20 crc kubenswrapper[4791]: I1210 23:08:20.329397 4791 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f7db40d5-1f25-4bab-98af-29999f12acf1-combined-ca-bundle\") pod \"keystone-bootstrap-tbpq4\" (UID: \"f7db40d5-1f25-4bab-98af-29999f12acf1\") " pod="openstack/keystone-bootstrap-tbpq4" Dec 10 23:08:20 crc kubenswrapper[4791]: I1210 23:08:20.329433 4791 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/f7db40d5-1f25-4bab-98af-29999f12acf1-fernet-keys\") pod \"keystone-bootstrap-tbpq4\" (UID: \"f7db40d5-1f25-4bab-98af-29999f12acf1\") " pod="openstack/keystone-bootstrap-tbpq4" Dec 10 23:08:20 crc kubenswrapper[4791]: I1210 23:08:20.329462 4791 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/f7db40d5-1f25-4bab-98af-29999f12acf1-credential-keys\") pod \"keystone-bootstrap-tbpq4\" (UID: \"f7db40d5-1f25-4bab-98af-29999f12acf1\") " pod="openstack/keystone-bootstrap-tbpq4" Dec 10 23:08:20 crc kubenswrapper[4791]: I1210 23:08:20.329494 4791 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/f7db40d5-1f25-4bab-98af-29999f12acf1-scripts\") pod \"keystone-bootstrap-tbpq4\" (UID: \"f7db40d5-1f25-4bab-98af-29999f12acf1\") " pod="openstack/keystone-bootstrap-tbpq4" Dec 10 23:08:20 crc kubenswrapper[4791]: I1210 23:08:20.329529 4791 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-h44bt\" (UniqueName: \"kubernetes.io/projected/f7db40d5-1f25-4bab-98af-29999f12acf1-kube-api-access-h44bt\") pod \"keystone-bootstrap-tbpq4\" (UID: \"f7db40d5-1f25-4bab-98af-29999f12acf1\") " pod="openstack/keystone-bootstrap-tbpq4" Dec 10 23:08:20 crc kubenswrapper[4791]: I1210 23:08:20.330568 4791 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/horizon-78fc5d56dc-58cjt"] Dec 10 23:08:20 crc kubenswrapper[4791]: I1210 23:08:20.415626 4791 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-55fff446b9-84xqj" Dec 10 23:08:20 crc kubenswrapper[4791]: I1210 23:08:20.438384 4791 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/f7db40d5-1f25-4bab-98af-29999f12acf1-credential-keys\") pod \"keystone-bootstrap-tbpq4\" (UID: \"f7db40d5-1f25-4bab-98af-29999f12acf1\") " pod="openstack/keystone-bootstrap-tbpq4" Dec 10 23:08:20 crc kubenswrapper[4791]: I1210 23:08:20.438463 4791 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/14f46cbc-afe2-40bb-a2a3-5673b5ac53e7-scripts\") pod \"horizon-78fc5d56dc-58cjt\" (UID: \"14f46cbc-afe2-40bb-a2a3-5673b5ac53e7\") " pod="openstack/horizon-78fc5d56dc-58cjt" Dec 10 23:08:20 crc kubenswrapper[4791]: I1210 23:08:20.440179 4791 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/f7db40d5-1f25-4bab-98af-29999f12acf1-scripts\") pod \"keystone-bootstrap-tbpq4\" (UID: \"f7db40d5-1f25-4bab-98af-29999f12acf1\") " pod="openstack/keystone-bootstrap-tbpq4" Dec 10 23:08:20 crc kubenswrapper[4791]: I1210 23:08:20.440241 4791 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-h44bt\" (UniqueName: \"kubernetes.io/projected/f7db40d5-1f25-4bab-98af-29999f12acf1-kube-api-access-h44bt\") pod \"keystone-bootstrap-tbpq4\" (UID: \"f7db40d5-1f25-4bab-98af-29999f12acf1\") " pod="openstack/keystone-bootstrap-tbpq4" Dec 10 23:08:20 crc kubenswrapper[4791]: I1210 23:08:20.440298 4791 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/f7db40d5-1f25-4bab-98af-29999f12acf1-config-data\") pod \"keystone-bootstrap-tbpq4\" (UID: \"f7db40d5-1f25-4bab-98af-29999f12acf1\") " pod="openstack/keystone-bootstrap-tbpq4" Dec 10 23:08:20 crc kubenswrapper[4791]: I1210 23:08:20.440371 4791 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"horizon-secret-key\" (UniqueName: \"kubernetes.io/secret/14f46cbc-afe2-40bb-a2a3-5673b5ac53e7-horizon-secret-key\") pod \"horizon-78fc5d56dc-58cjt\" (UID: \"14f46cbc-afe2-40bb-a2a3-5673b5ac53e7\") " pod="openstack/horizon-78fc5d56dc-58cjt" Dec 10 23:08:20 crc kubenswrapper[4791]: I1210 23:08:20.440420 4791 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/14f46cbc-afe2-40bb-a2a3-5673b5ac53e7-config-data\") pod \"horizon-78fc5d56dc-58cjt\" (UID: \"14f46cbc-afe2-40bb-a2a3-5673b5ac53e7\") " pod="openstack/horizon-78fc5d56dc-58cjt" Dec 10 23:08:20 crc kubenswrapper[4791]: I1210 23:08:20.440450 4791 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-dk8dr\" (UniqueName: \"kubernetes.io/projected/14f46cbc-afe2-40bb-a2a3-5673b5ac53e7-kube-api-access-dk8dr\") pod \"horizon-78fc5d56dc-58cjt\" (UID: \"14f46cbc-afe2-40bb-a2a3-5673b5ac53e7\") " pod="openstack/horizon-78fc5d56dc-58cjt" Dec 10 23:08:20 crc kubenswrapper[4791]: I1210 23:08:20.440501 4791 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f7db40d5-1f25-4bab-98af-29999f12acf1-combined-ca-bundle\") pod \"keystone-bootstrap-tbpq4\" (UID: \"f7db40d5-1f25-4bab-98af-29999f12acf1\") " pod="openstack/keystone-bootstrap-tbpq4" Dec 10 23:08:20 crc kubenswrapper[4791]: I1210 23:08:20.451983 4791 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/f7db40d5-1f25-4bab-98af-29999f12acf1-fernet-keys\") pod \"keystone-bootstrap-tbpq4\" (UID: \"f7db40d5-1f25-4bab-98af-29999f12acf1\") " pod="openstack/keystone-bootstrap-tbpq4" Dec 10 23:08:20 crc kubenswrapper[4791]: I1210 23:08:20.452076 4791 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/14f46cbc-afe2-40bb-a2a3-5673b5ac53e7-logs\") pod \"horizon-78fc5d56dc-58cjt\" (UID: \"14f46cbc-afe2-40bb-a2a3-5673b5ac53e7\") " pod="openstack/horizon-78fc5d56dc-58cjt" Dec 10 23:08:20 crc kubenswrapper[4791]: I1210 23:08:20.455169 4791 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/f7db40d5-1f25-4bab-98af-29999f12acf1-config-data\") pod \"keystone-bootstrap-tbpq4\" (UID: \"f7db40d5-1f25-4bab-98af-29999f12acf1\") " pod="openstack/keystone-bootstrap-tbpq4" Dec 10 23:08:20 crc kubenswrapper[4791]: I1210 23:08:20.456363 4791 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/f7db40d5-1f25-4bab-98af-29999f12acf1-fernet-keys\") pod \"keystone-bootstrap-tbpq4\" (UID: \"f7db40d5-1f25-4bab-98af-29999f12acf1\") " pod="openstack/keystone-bootstrap-tbpq4" Dec 10 23:08:20 crc kubenswrapper[4791]: I1210 23:08:20.468334 4791 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/f7db40d5-1f25-4bab-98af-29999f12acf1-credential-keys\") pod \"keystone-bootstrap-tbpq4\" (UID: \"f7db40d5-1f25-4bab-98af-29999f12acf1\") " pod="openstack/keystone-bootstrap-tbpq4" Dec 10 23:08:20 crc kubenswrapper[4791]: I1210 23:08:20.469594 4791 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/f7db40d5-1f25-4bab-98af-29999f12acf1-scripts\") pod \"keystone-bootstrap-tbpq4\" (UID: \"f7db40d5-1f25-4bab-98af-29999f12acf1\") " pod="openstack/keystone-bootstrap-tbpq4" Dec 10 23:08:20 crc kubenswrapper[4791]: I1210 23:08:20.485289 4791 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f7db40d5-1f25-4bab-98af-29999f12acf1-combined-ca-bundle\") pod \"keystone-bootstrap-tbpq4\" (UID: \"f7db40d5-1f25-4bab-98af-29999f12acf1\") " pod="openstack/keystone-bootstrap-tbpq4" Dec 10 23:08:20 crc kubenswrapper[4791]: I1210 23:08:20.506972 4791 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-h44bt\" (UniqueName: \"kubernetes.io/projected/f7db40d5-1f25-4bab-98af-29999f12acf1-kube-api-access-h44bt\") pod \"keystone-bootstrap-tbpq4\" (UID: \"f7db40d5-1f25-4bab-98af-29999f12acf1\") " pod="openstack/keystone-bootstrap-tbpq4" Dec 10 23:08:20 crc kubenswrapper[4791]: I1210 23:08:20.524854 4791 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ceilometer-0"] Dec 10 23:08:20 crc kubenswrapper[4791]: I1210 23:08:20.527589 4791 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Dec 10 23:08:20 crc kubenswrapper[4791]: I1210 23:08:20.553922 4791 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-config-data" Dec 10 23:08:20 crc kubenswrapper[4791]: I1210 23:08:20.554722 4791 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-scripts" Dec 10 23:08:20 crc kubenswrapper[4791]: I1210 23:08:20.562136 4791 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/14f46cbc-afe2-40bb-a2a3-5673b5ac53e7-logs\") pod \"horizon-78fc5d56dc-58cjt\" (UID: \"14f46cbc-afe2-40bb-a2a3-5673b5ac53e7\") " pod="openstack/horizon-78fc5d56dc-58cjt" Dec 10 23:08:20 crc kubenswrapper[4791]: I1210 23:08:20.562186 4791 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/8eaa1b1d-715f-4059-ad18-fe9632c036a7-log-httpd\") pod \"ceilometer-0\" (UID: \"8eaa1b1d-715f-4059-ad18-fe9632c036a7\") " pod="openstack/ceilometer-0" Dec 10 23:08:20 crc kubenswrapper[4791]: I1210 23:08:20.562239 4791 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/14f46cbc-afe2-40bb-a2a3-5673b5ac53e7-scripts\") pod \"horizon-78fc5d56dc-58cjt\" (UID: \"14f46cbc-afe2-40bb-a2a3-5673b5ac53e7\") " pod="openstack/horizon-78fc5d56dc-58cjt" Dec 10 23:08:20 crc kubenswrapper[4791]: I1210 23:08:20.562269 4791 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/8eaa1b1d-715f-4059-ad18-fe9632c036a7-config-data\") pod \"ceilometer-0\" (UID: \"8eaa1b1d-715f-4059-ad18-fe9632c036a7\") " pod="openstack/ceilometer-0" Dec 10 23:08:20 crc kubenswrapper[4791]: I1210 23:08:20.562291 4791 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/8eaa1b1d-715f-4059-ad18-fe9632c036a7-scripts\") pod \"ceilometer-0\" (UID: \"8eaa1b1d-715f-4059-ad18-fe9632c036a7\") " pod="openstack/ceilometer-0" Dec 10 23:08:20 crc kubenswrapper[4791]: I1210 23:08:20.562309 4791 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/8eaa1b1d-715f-4059-ad18-fe9632c036a7-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"8eaa1b1d-715f-4059-ad18-fe9632c036a7\") " pod="openstack/ceilometer-0" Dec 10 23:08:20 crc kubenswrapper[4791]: I1210 23:08:20.562374 4791 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-7zjdl\" (UniqueName: \"kubernetes.io/projected/8eaa1b1d-715f-4059-ad18-fe9632c036a7-kube-api-access-7zjdl\") pod \"ceilometer-0\" (UID: \"8eaa1b1d-715f-4059-ad18-fe9632c036a7\") " pod="openstack/ceilometer-0" Dec 10 23:08:20 crc kubenswrapper[4791]: I1210 23:08:20.562416 4791 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"horizon-secret-key\" (UniqueName: \"kubernetes.io/secret/14f46cbc-afe2-40bb-a2a3-5673b5ac53e7-horizon-secret-key\") pod \"horizon-78fc5d56dc-58cjt\" (UID: \"14f46cbc-afe2-40bb-a2a3-5673b5ac53e7\") " pod="openstack/horizon-78fc5d56dc-58cjt" Dec 10 23:08:20 crc kubenswrapper[4791]: I1210 23:08:20.562445 4791 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/14f46cbc-afe2-40bb-a2a3-5673b5ac53e7-config-data\") pod \"horizon-78fc5d56dc-58cjt\" (UID: \"14f46cbc-afe2-40bb-a2a3-5673b5ac53e7\") " pod="openstack/horizon-78fc5d56dc-58cjt" Dec 10 23:08:20 crc kubenswrapper[4791]: I1210 23:08:20.562462 4791 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-dk8dr\" (UniqueName: \"kubernetes.io/projected/14f46cbc-afe2-40bb-a2a3-5673b5ac53e7-kube-api-access-dk8dr\") pod \"horizon-78fc5d56dc-58cjt\" (UID: \"14f46cbc-afe2-40bb-a2a3-5673b5ac53e7\") " pod="openstack/horizon-78fc5d56dc-58cjt" Dec 10 23:08:20 crc kubenswrapper[4791]: I1210 23:08:20.562480 4791 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/8eaa1b1d-715f-4059-ad18-fe9632c036a7-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"8eaa1b1d-715f-4059-ad18-fe9632c036a7\") " pod="openstack/ceilometer-0" Dec 10 23:08:20 crc kubenswrapper[4791]: I1210 23:08:20.562505 4791 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/8eaa1b1d-715f-4059-ad18-fe9632c036a7-run-httpd\") pod \"ceilometer-0\" (UID: \"8eaa1b1d-715f-4059-ad18-fe9632c036a7\") " pod="openstack/ceilometer-0" Dec 10 23:08:20 crc kubenswrapper[4791]: I1210 23:08:20.562889 4791 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/14f46cbc-afe2-40bb-a2a3-5673b5ac53e7-logs\") pod \"horizon-78fc5d56dc-58cjt\" (UID: \"14f46cbc-afe2-40bb-a2a3-5673b5ac53e7\") " pod="openstack/horizon-78fc5d56dc-58cjt" Dec 10 23:08:20 crc kubenswrapper[4791]: I1210 23:08:20.563389 4791 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/14f46cbc-afe2-40bb-a2a3-5673b5ac53e7-scripts\") pod \"horizon-78fc5d56dc-58cjt\" (UID: \"14f46cbc-afe2-40bb-a2a3-5673b5ac53e7\") " pod="openstack/horizon-78fc5d56dc-58cjt" Dec 10 23:08:20 crc kubenswrapper[4791]: I1210 23:08:20.564741 4791 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/14f46cbc-afe2-40bb-a2a3-5673b5ac53e7-config-data\") pod \"horizon-78fc5d56dc-58cjt\" (UID: \"14f46cbc-afe2-40bb-a2a3-5673b5ac53e7\") " pod="openstack/horizon-78fc5d56dc-58cjt" Dec 10 23:08:20 crc kubenswrapper[4791]: I1210 23:08:20.568714 4791 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"horizon-secret-key\" (UniqueName: \"kubernetes.io/secret/14f46cbc-afe2-40bb-a2a3-5673b5ac53e7-horizon-secret-key\") pod \"horizon-78fc5d56dc-58cjt\" (UID: \"14f46cbc-afe2-40bb-a2a3-5673b5ac53e7\") " pod="openstack/horizon-78fc5d56dc-58cjt" Dec 10 23:08:20 crc kubenswrapper[4791]: I1210 23:08:20.582608 4791 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/cinder-db-sync-g4s5p"] Dec 10 23:08:20 crc kubenswrapper[4791]: I1210 23:08:20.583700 4791 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-db-sync-g4s5p" Dec 10 23:08:20 crc kubenswrapper[4791]: I1210 23:08:20.598545 4791 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceilometer-0"] Dec 10 23:08:20 crc kubenswrapper[4791]: I1210 23:08:20.598836 4791 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cinder-scripts" Dec 10 23:08:20 crc kubenswrapper[4791]: I1210 23:08:20.599195 4791 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cinder-cinder-dockercfg-lxnf7" Dec 10 23:08:20 crc kubenswrapper[4791]: I1210 23:08:20.599397 4791 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cinder-config-data" Dec 10 23:08:20 crc kubenswrapper[4791]: I1210 23:08:20.625306 4791 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-dk8dr\" (UniqueName: \"kubernetes.io/projected/14f46cbc-afe2-40bb-a2a3-5673b5ac53e7-kube-api-access-dk8dr\") pod \"horizon-78fc5d56dc-58cjt\" (UID: \"14f46cbc-afe2-40bb-a2a3-5673b5ac53e7\") " pod="openstack/horizon-78fc5d56dc-58cjt" Dec 10 23:08:20 crc kubenswrapper[4791]: I1210 23:08:20.630004 4791 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-db-sync-g4s5p"] Dec 10 23:08:20 crc kubenswrapper[4791]: I1210 23:08:20.641631 4791 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/horizon-78fc5d56dc-58cjt" Dec 10 23:08:20 crc kubenswrapper[4791]: I1210 23:08:20.647442 4791 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/neutron-db-sync-kg4vv"] Dec 10 23:08:20 crc kubenswrapper[4791]: I1210 23:08:20.649096 4791 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-db-sync-kg4vv" Dec 10 23:08:20 crc kubenswrapper[4791]: I1210 23:08:20.658006 4791 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"neutron-httpd-config" Dec 10 23:08:20 crc kubenswrapper[4791]: I1210 23:08:20.658417 4791 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"neutron-neutron-dockercfg-fblml" Dec 10 23:08:20 crc kubenswrapper[4791]: I1210 23:08:20.658539 4791 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"neutron-config" Dec 10 23:08:20 crc kubenswrapper[4791]: I1210 23:08:20.659032 4791 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/neutron-db-sync-kg4vv"] Dec 10 23:08:20 crc kubenswrapper[4791]: I1210 23:08:20.664284 4791 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/8eaa1b1d-715f-4059-ad18-fe9632c036a7-log-httpd\") pod \"ceilometer-0\" (UID: \"8eaa1b1d-715f-4059-ad18-fe9632c036a7\") " pod="openstack/ceilometer-0" Dec 10 23:08:20 crc kubenswrapper[4791]: I1210 23:08:20.664354 4791 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/8eaa1b1d-715f-4059-ad18-fe9632c036a7-config-data\") pod \"ceilometer-0\" (UID: \"8eaa1b1d-715f-4059-ad18-fe9632c036a7\") " pod="openstack/ceilometer-0" Dec 10 23:08:20 crc kubenswrapper[4791]: I1210 23:08:20.664375 4791 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/a85df2ed-be84-41b3-8993-c3d963daef08-etc-machine-id\") pod \"cinder-db-sync-g4s5p\" (UID: \"a85df2ed-be84-41b3-8993-c3d963daef08\") " pod="openstack/cinder-db-sync-g4s5p" Dec 10 23:08:20 crc kubenswrapper[4791]: I1210 23:08:20.664394 4791 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/a85df2ed-be84-41b3-8993-c3d963daef08-config-data\") pod \"cinder-db-sync-g4s5p\" (UID: \"a85df2ed-be84-41b3-8993-c3d963daef08\") " pod="openstack/cinder-db-sync-g4s5p" Dec 10 23:08:20 crc kubenswrapper[4791]: I1210 23:08:20.664412 4791 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/8eaa1b1d-715f-4059-ad18-fe9632c036a7-scripts\") pod \"ceilometer-0\" (UID: \"8eaa1b1d-715f-4059-ad18-fe9632c036a7\") " pod="openstack/ceilometer-0" Dec 10 23:08:20 crc kubenswrapper[4791]: I1210 23:08:20.664432 4791 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/8eaa1b1d-715f-4059-ad18-fe9632c036a7-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"8eaa1b1d-715f-4059-ad18-fe9632c036a7\") " pod="openstack/ceilometer-0" Dec 10 23:08:20 crc kubenswrapper[4791]: I1210 23:08:20.664457 4791 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-7zjdl\" (UniqueName: \"kubernetes.io/projected/8eaa1b1d-715f-4059-ad18-fe9632c036a7-kube-api-access-7zjdl\") pod \"ceilometer-0\" (UID: \"8eaa1b1d-715f-4059-ad18-fe9632c036a7\") " pod="openstack/ceilometer-0" Dec 10 23:08:20 crc kubenswrapper[4791]: I1210 23:08:20.664503 4791 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/a85df2ed-be84-41b3-8993-c3d963daef08-db-sync-config-data\") pod \"cinder-db-sync-g4s5p\" (UID: \"a85df2ed-be84-41b3-8993-c3d963daef08\") " pod="openstack/cinder-db-sync-g4s5p" Dec 10 23:08:20 crc kubenswrapper[4791]: I1210 23:08:20.664525 4791 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/a85df2ed-be84-41b3-8993-c3d963daef08-scripts\") pod \"cinder-db-sync-g4s5p\" (UID: \"a85df2ed-be84-41b3-8993-c3d963daef08\") " pod="openstack/cinder-db-sync-g4s5p" Dec 10 23:08:20 crc kubenswrapper[4791]: I1210 23:08:20.664544 4791 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-5fw8x\" (UniqueName: \"kubernetes.io/projected/a85df2ed-be84-41b3-8993-c3d963daef08-kube-api-access-5fw8x\") pod \"cinder-db-sync-g4s5p\" (UID: \"a85df2ed-be84-41b3-8993-c3d963daef08\") " pod="openstack/cinder-db-sync-g4s5p" Dec 10 23:08:20 crc kubenswrapper[4791]: I1210 23:08:20.664560 4791 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/8eaa1b1d-715f-4059-ad18-fe9632c036a7-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"8eaa1b1d-715f-4059-ad18-fe9632c036a7\") " pod="openstack/ceilometer-0" Dec 10 23:08:20 crc kubenswrapper[4791]: I1210 23:08:20.664584 4791 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/8eaa1b1d-715f-4059-ad18-fe9632c036a7-run-httpd\") pod \"ceilometer-0\" (UID: \"8eaa1b1d-715f-4059-ad18-fe9632c036a7\") " pod="openstack/ceilometer-0" Dec 10 23:08:20 crc kubenswrapper[4791]: I1210 23:08:20.664616 4791 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a85df2ed-be84-41b3-8993-c3d963daef08-combined-ca-bundle\") pod \"cinder-db-sync-g4s5p\" (UID: \"a85df2ed-be84-41b3-8993-c3d963daef08\") " pod="openstack/cinder-db-sync-g4s5p" Dec 10 23:08:20 crc kubenswrapper[4791]: I1210 23:08:20.664838 4791 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/8eaa1b1d-715f-4059-ad18-fe9632c036a7-log-httpd\") pod \"ceilometer-0\" (UID: \"8eaa1b1d-715f-4059-ad18-fe9632c036a7\") " pod="openstack/ceilometer-0" Dec 10 23:08:20 crc kubenswrapper[4791]: I1210 23:08:20.666490 4791 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/8eaa1b1d-715f-4059-ad18-fe9632c036a7-run-httpd\") pod \"ceilometer-0\" (UID: \"8eaa1b1d-715f-4059-ad18-fe9632c036a7\") " pod="openstack/ceilometer-0" Dec 10 23:08:20 crc kubenswrapper[4791]: I1210 23:08:20.668765 4791 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/8eaa1b1d-715f-4059-ad18-fe9632c036a7-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"8eaa1b1d-715f-4059-ad18-fe9632c036a7\") " pod="openstack/ceilometer-0" Dec 10 23:08:20 crc kubenswrapper[4791]: I1210 23:08:20.671722 4791 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/placement-db-sync-svgbr"] Dec 10 23:08:20 crc kubenswrapper[4791]: I1210 23:08:20.675016 4791 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/8eaa1b1d-715f-4059-ad18-fe9632c036a7-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"8eaa1b1d-715f-4059-ad18-fe9632c036a7\") " pod="openstack/ceilometer-0" Dec 10 23:08:20 crc kubenswrapper[4791]: I1210 23:08:20.675428 4791 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/8eaa1b1d-715f-4059-ad18-fe9632c036a7-scripts\") pod \"ceilometer-0\" (UID: \"8eaa1b1d-715f-4059-ad18-fe9632c036a7\") " pod="openstack/ceilometer-0" Dec 10 23:08:20 crc kubenswrapper[4791]: I1210 23:08:20.676921 4791 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/8eaa1b1d-715f-4059-ad18-fe9632c036a7-config-data\") pod \"ceilometer-0\" (UID: \"8eaa1b1d-715f-4059-ad18-fe9632c036a7\") " pod="openstack/ceilometer-0" Dec 10 23:08:20 crc kubenswrapper[4791]: I1210 23:08:20.677242 4791 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/placement-db-sync-svgbr" Dec 10 23:08:20 crc kubenswrapper[4791]: I1210 23:08:20.684367 4791 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"placement-config-data" Dec 10 23:08:20 crc kubenswrapper[4791]: I1210 23:08:20.685202 4791 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"placement-scripts" Dec 10 23:08:20 crc kubenswrapper[4791]: I1210 23:08:20.690020 4791 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"placement-placement-dockercfg-wklr7" Dec 10 23:08:20 crc kubenswrapper[4791]: I1210 23:08:20.693798 4791 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-7zjdl\" (UniqueName: \"kubernetes.io/projected/8eaa1b1d-715f-4059-ad18-fe9632c036a7-kube-api-access-7zjdl\") pod \"ceilometer-0\" (UID: \"8eaa1b1d-715f-4059-ad18-fe9632c036a7\") " pod="openstack/ceilometer-0" Dec 10 23:08:20 crc kubenswrapper[4791]: I1210 23:08:20.705470 4791 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-55fff446b9-84xqj"] Dec 10 23:08:20 crc kubenswrapper[4791]: I1210 23:08:20.716025 4791 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/placement-db-sync-svgbr"] Dec 10 23:08:20 crc kubenswrapper[4791]: I1210 23:08:20.726726 4791 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-76fcf4b695-r7wvb"] Dec 10 23:08:20 crc kubenswrapper[4791]: I1210 23:08:20.730574 4791 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-76fcf4b695-r7wvb" Dec 10 23:08:20 crc kubenswrapper[4791]: I1210 23:08:20.738974 4791 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/barbican-db-sync-zllvm"] Dec 10 23:08:20 crc kubenswrapper[4791]: I1210 23:08:20.741255 4791 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-db-sync-zllvm" Dec 10 23:08:20 crc kubenswrapper[4791]: I1210 23:08:20.753937 4791 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/horizon-5f54ccbf57-wwpfk"] Dec 10 23:08:20 crc kubenswrapper[4791]: I1210 23:08:20.755348 4791 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/horizon-5f54ccbf57-wwpfk" Dec 10 23:08:20 crc kubenswrapper[4791]: I1210 23:08:20.761104 4791 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"barbican-barbican-dockercfg-rkxgd" Dec 10 23:08:20 crc kubenswrapper[4791]: I1210 23:08:20.761446 4791 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"barbican-config-data" Dec 10 23:08:20 crc kubenswrapper[4791]: I1210 23:08:20.765709 4791 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/851556d7-fa38-4eec-ae74-de989d66d27b-combined-ca-bundle\") pod \"barbican-db-sync-zllvm\" (UID: \"851556d7-fa38-4eec-ae74-de989d66d27b\") " pod="openstack/barbican-db-sync-zllvm" Dec 10 23:08:20 crc kubenswrapper[4791]: I1210 23:08:20.765786 4791 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/6076a930-1583-41b7-b5ae-bc6c61086e48-combined-ca-bundle\") pod \"neutron-db-sync-kg4vv\" (UID: \"6076a930-1583-41b7-b5ae-bc6c61086e48\") " pod="openstack/neutron-db-sync-kg4vv" Dec 10 23:08:20 crc kubenswrapper[4791]: I1210 23:08:20.765813 4791 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/b9efeadd-d54b-4f78-a0ad-85cc62321b84-scripts\") pod \"placement-db-sync-svgbr\" (UID: \"b9efeadd-d54b-4f78-a0ad-85cc62321b84\") " pod="openstack/placement-db-sync-svgbr" Dec 10 23:08:20 crc kubenswrapper[4791]: I1210 23:08:20.765832 4791 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/851556d7-fa38-4eec-ae74-de989d66d27b-db-sync-config-data\") pod \"barbican-db-sync-zllvm\" (UID: \"851556d7-fa38-4eec-ae74-de989d66d27b\") " pod="openstack/barbican-db-sync-zllvm" Dec 10 23:08:20 crc kubenswrapper[4791]: I1210 23:08:20.765853 4791 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/secret/6076a930-1583-41b7-b5ae-bc6c61086e48-config\") pod \"neutron-db-sync-kg4vv\" (UID: \"6076a930-1583-41b7-b5ae-bc6c61086e48\") " pod="openstack/neutron-db-sync-kg4vv" Dec 10 23:08:20 crc kubenswrapper[4791]: I1210 23:08:20.765873 4791 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/a85df2ed-be84-41b3-8993-c3d963daef08-db-sync-config-data\") pod \"cinder-db-sync-g4s5p\" (UID: \"a85df2ed-be84-41b3-8993-c3d963daef08\") " pod="openstack/cinder-db-sync-g4s5p" Dec 10 23:08:20 crc kubenswrapper[4791]: I1210 23:08:20.765896 4791 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/a85df2ed-be84-41b3-8993-c3d963daef08-scripts\") pod \"cinder-db-sync-g4s5p\" (UID: \"a85df2ed-be84-41b3-8993-c3d963daef08\") " pod="openstack/cinder-db-sync-g4s5p" Dec 10 23:08:20 crc kubenswrapper[4791]: I1210 23:08:20.765916 4791 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-66x4z\" (UniqueName: \"kubernetes.io/projected/6076a930-1583-41b7-b5ae-bc6c61086e48-kube-api-access-66x4z\") pod \"neutron-db-sync-kg4vv\" (UID: \"6076a930-1583-41b7-b5ae-bc6c61086e48\") " pod="openstack/neutron-db-sync-kg4vv" Dec 10 23:08:20 crc kubenswrapper[4791]: I1210 23:08:20.765932 4791 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-crjrd\" (UniqueName: \"kubernetes.io/projected/851556d7-fa38-4eec-ae74-de989d66d27b-kube-api-access-crjrd\") pod \"barbican-db-sync-zllvm\" (UID: \"851556d7-fa38-4eec-ae74-de989d66d27b\") " pod="openstack/barbican-db-sync-zllvm" Dec 10 23:08:20 crc kubenswrapper[4791]: I1210 23:08:20.765951 4791 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-5fw8x\" (UniqueName: \"kubernetes.io/projected/a85df2ed-be84-41b3-8993-c3d963daef08-kube-api-access-5fw8x\") pod \"cinder-db-sync-g4s5p\" (UID: \"a85df2ed-be84-41b3-8993-c3d963daef08\") " pod="openstack/cinder-db-sync-g4s5p" Dec 10 23:08:20 crc kubenswrapper[4791]: I1210 23:08:20.765978 4791 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/b9efeadd-d54b-4f78-a0ad-85cc62321b84-logs\") pod \"placement-db-sync-svgbr\" (UID: \"b9efeadd-d54b-4f78-a0ad-85cc62321b84\") " pod="openstack/placement-db-sync-svgbr" Dec 10 23:08:20 crc kubenswrapper[4791]: I1210 23:08:20.766003 4791 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-x949f\" (UniqueName: \"kubernetes.io/projected/00320ef6-9362-427b-a2eb-12329d652a0e-kube-api-access-x949f\") pod \"dnsmasq-dns-76fcf4b695-r7wvb\" (UID: \"00320ef6-9362-427b-a2eb-12329d652a0e\") " pod="openstack/dnsmasq-dns-76fcf4b695-r7wvb" Dec 10 23:08:20 crc kubenswrapper[4791]: I1210 23:08:20.766019 4791 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-q27r9\" (UniqueName: \"kubernetes.io/projected/b9efeadd-d54b-4f78-a0ad-85cc62321b84-kube-api-access-q27r9\") pod \"placement-db-sync-svgbr\" (UID: \"b9efeadd-d54b-4f78-a0ad-85cc62321b84\") " pod="openstack/placement-db-sync-svgbr" Dec 10 23:08:20 crc kubenswrapper[4791]: I1210 23:08:20.766042 4791 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/00320ef6-9362-427b-a2eb-12329d652a0e-ovsdbserver-sb\") pod \"dnsmasq-dns-76fcf4b695-r7wvb\" (UID: \"00320ef6-9362-427b-a2eb-12329d652a0e\") " pod="openstack/dnsmasq-dns-76fcf4b695-r7wvb" Dec 10 23:08:20 crc kubenswrapper[4791]: I1210 23:08:20.766070 4791 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a85df2ed-be84-41b3-8993-c3d963daef08-combined-ca-bundle\") pod \"cinder-db-sync-g4s5p\" (UID: \"a85df2ed-be84-41b3-8993-c3d963daef08\") " pod="openstack/cinder-db-sync-g4s5p" Dec 10 23:08:20 crc kubenswrapper[4791]: I1210 23:08:20.766097 4791 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/b9efeadd-d54b-4f78-a0ad-85cc62321b84-config-data\") pod \"placement-db-sync-svgbr\" (UID: \"b9efeadd-d54b-4f78-a0ad-85cc62321b84\") " pod="openstack/placement-db-sync-svgbr" Dec 10 23:08:20 crc kubenswrapper[4791]: I1210 23:08:20.766117 4791 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/00320ef6-9362-427b-a2eb-12329d652a0e-ovsdbserver-nb\") pod \"dnsmasq-dns-76fcf4b695-r7wvb\" (UID: \"00320ef6-9362-427b-a2eb-12329d652a0e\") " pod="openstack/dnsmasq-dns-76fcf4b695-r7wvb" Dec 10 23:08:20 crc kubenswrapper[4791]: I1210 23:08:20.766137 4791 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/b9efeadd-d54b-4f78-a0ad-85cc62321b84-combined-ca-bundle\") pod \"placement-db-sync-svgbr\" (UID: \"b9efeadd-d54b-4f78-a0ad-85cc62321b84\") " pod="openstack/placement-db-sync-svgbr" Dec 10 23:08:20 crc kubenswrapper[4791]: I1210 23:08:20.766152 4791 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/00320ef6-9362-427b-a2eb-12329d652a0e-dns-swift-storage-0\") pod \"dnsmasq-dns-76fcf4b695-r7wvb\" (UID: \"00320ef6-9362-427b-a2eb-12329d652a0e\") " pod="openstack/dnsmasq-dns-76fcf4b695-r7wvb" Dec 10 23:08:20 crc kubenswrapper[4791]: I1210 23:08:20.766180 4791 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/a85df2ed-be84-41b3-8993-c3d963daef08-etc-machine-id\") pod \"cinder-db-sync-g4s5p\" (UID: \"a85df2ed-be84-41b3-8993-c3d963daef08\") " pod="openstack/cinder-db-sync-g4s5p" Dec 10 23:08:20 crc kubenswrapper[4791]: I1210 23:08:20.766198 4791 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/00320ef6-9362-427b-a2eb-12329d652a0e-dns-svc\") pod \"dnsmasq-dns-76fcf4b695-r7wvb\" (UID: \"00320ef6-9362-427b-a2eb-12329d652a0e\") " pod="openstack/dnsmasq-dns-76fcf4b695-r7wvb" Dec 10 23:08:20 crc kubenswrapper[4791]: I1210 23:08:20.766224 4791 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/a85df2ed-be84-41b3-8993-c3d963daef08-config-data\") pod \"cinder-db-sync-g4s5p\" (UID: \"a85df2ed-be84-41b3-8993-c3d963daef08\") " pod="openstack/cinder-db-sync-g4s5p" Dec 10 23:08:20 crc kubenswrapper[4791]: I1210 23:08:20.766252 4791 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/00320ef6-9362-427b-a2eb-12329d652a0e-config\") pod \"dnsmasq-dns-76fcf4b695-r7wvb\" (UID: \"00320ef6-9362-427b-a2eb-12329d652a0e\") " pod="openstack/dnsmasq-dns-76fcf4b695-r7wvb" Dec 10 23:08:20 crc kubenswrapper[4791]: I1210 23:08:20.771061 4791 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/a85df2ed-be84-41b3-8993-c3d963daef08-etc-machine-id\") pod \"cinder-db-sync-g4s5p\" (UID: \"a85df2ed-be84-41b3-8993-c3d963daef08\") " pod="openstack/cinder-db-sync-g4s5p" Dec 10 23:08:20 crc kubenswrapper[4791]: I1210 23:08:20.776324 4791 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a85df2ed-be84-41b3-8993-c3d963daef08-combined-ca-bundle\") pod \"cinder-db-sync-g4s5p\" (UID: \"a85df2ed-be84-41b3-8993-c3d963daef08\") " pod="openstack/cinder-db-sync-g4s5p" Dec 10 23:08:20 crc kubenswrapper[4791]: I1210 23:08:20.776438 4791 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-bootstrap-tbpq4" Dec 10 23:08:20 crc kubenswrapper[4791]: I1210 23:08:20.778967 4791 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/a85df2ed-be84-41b3-8993-c3d963daef08-scripts\") pod \"cinder-db-sync-g4s5p\" (UID: \"a85df2ed-be84-41b3-8993-c3d963daef08\") " pod="openstack/cinder-db-sync-g4s5p" Dec 10 23:08:20 crc kubenswrapper[4791]: I1210 23:08:20.783554 4791 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/a85df2ed-be84-41b3-8993-c3d963daef08-db-sync-config-data\") pod \"cinder-db-sync-g4s5p\" (UID: \"a85df2ed-be84-41b3-8993-c3d963daef08\") " pod="openstack/cinder-db-sync-g4s5p" Dec 10 23:08:20 crc kubenswrapper[4791]: I1210 23:08:20.788262 4791 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/a85df2ed-be84-41b3-8993-c3d963daef08-config-data\") pod \"cinder-db-sync-g4s5p\" (UID: \"a85df2ed-be84-41b3-8993-c3d963daef08\") " pod="openstack/cinder-db-sync-g4s5p" Dec 10 23:08:20 crc kubenswrapper[4791]: I1210 23:08:20.798855 4791 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/horizon-5f54ccbf57-wwpfk"] Dec 10 23:08:20 crc kubenswrapper[4791]: I1210 23:08:20.798997 4791 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-5fw8x\" (UniqueName: \"kubernetes.io/projected/a85df2ed-be84-41b3-8993-c3d963daef08-kube-api-access-5fw8x\") pod \"cinder-db-sync-g4s5p\" (UID: \"a85df2ed-be84-41b3-8993-c3d963daef08\") " pod="openstack/cinder-db-sync-g4s5p" Dec 10 23:08:20 crc kubenswrapper[4791]: I1210 23:08:20.807015 4791 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-76fcf4b695-r7wvb"] Dec 10 23:08:20 crc kubenswrapper[4791]: I1210 23:08:20.818730 4791 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/barbican-db-sync-zllvm"] Dec 10 23:08:20 crc kubenswrapper[4791]: I1210 23:08:20.868682 4791 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/00320ef6-9362-427b-a2eb-12329d652a0e-ovsdbserver-nb\") pod \"dnsmasq-dns-76fcf4b695-r7wvb\" (UID: \"00320ef6-9362-427b-a2eb-12329d652a0e\") " pod="openstack/dnsmasq-dns-76fcf4b695-r7wvb" Dec 10 23:08:20 crc kubenswrapper[4791]: I1210 23:08:20.868791 4791 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/b9efeadd-d54b-4f78-a0ad-85cc62321b84-combined-ca-bundle\") pod \"placement-db-sync-svgbr\" (UID: \"b9efeadd-d54b-4f78-a0ad-85cc62321b84\") " pod="openstack/placement-db-sync-svgbr" Dec 10 23:08:20 crc kubenswrapper[4791]: I1210 23:08:20.869626 4791 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/00320ef6-9362-427b-a2eb-12329d652a0e-ovsdbserver-nb\") pod \"dnsmasq-dns-76fcf4b695-r7wvb\" (UID: \"00320ef6-9362-427b-a2eb-12329d652a0e\") " pod="openstack/dnsmasq-dns-76fcf4b695-r7wvb" Dec 10 23:08:20 crc kubenswrapper[4791]: I1210 23:08:20.869633 4791 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/00320ef6-9362-427b-a2eb-12329d652a0e-dns-swift-storage-0\") pod \"dnsmasq-dns-76fcf4b695-r7wvb\" (UID: \"00320ef6-9362-427b-a2eb-12329d652a0e\") " pod="openstack/dnsmasq-dns-76fcf4b695-r7wvb" Dec 10 23:08:20 crc kubenswrapper[4791]: I1210 23:08:20.869699 4791 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/aa3a63ab-6092-46ce-aaf6-40360aa060ac-logs\") pod \"horizon-5f54ccbf57-wwpfk\" (UID: \"aa3a63ab-6092-46ce-aaf6-40360aa060ac\") " pod="openstack/horizon-5f54ccbf57-wwpfk" Dec 10 23:08:20 crc kubenswrapper[4791]: I1210 23:08:20.870483 4791 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/00320ef6-9362-427b-a2eb-12329d652a0e-dns-swift-storage-0\") pod \"dnsmasq-dns-76fcf4b695-r7wvb\" (UID: \"00320ef6-9362-427b-a2eb-12329d652a0e\") " pod="openstack/dnsmasq-dns-76fcf4b695-r7wvb" Dec 10 23:08:20 crc kubenswrapper[4791]: I1210 23:08:20.871199 4791 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-85n5v\" (UniqueName: \"kubernetes.io/projected/aa3a63ab-6092-46ce-aaf6-40360aa060ac-kube-api-access-85n5v\") pod \"horizon-5f54ccbf57-wwpfk\" (UID: \"aa3a63ab-6092-46ce-aaf6-40360aa060ac\") " pod="openstack/horizon-5f54ccbf57-wwpfk" Dec 10 23:08:20 crc kubenswrapper[4791]: I1210 23:08:20.871254 4791 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/00320ef6-9362-427b-a2eb-12329d652a0e-dns-svc\") pod \"dnsmasq-dns-76fcf4b695-r7wvb\" (UID: \"00320ef6-9362-427b-a2eb-12329d652a0e\") " pod="openstack/dnsmasq-dns-76fcf4b695-r7wvb" Dec 10 23:08:20 crc kubenswrapper[4791]: I1210 23:08:20.871293 4791 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/00320ef6-9362-427b-a2eb-12329d652a0e-config\") pod \"dnsmasq-dns-76fcf4b695-r7wvb\" (UID: \"00320ef6-9362-427b-a2eb-12329d652a0e\") " pod="openstack/dnsmasq-dns-76fcf4b695-r7wvb" Dec 10 23:08:20 crc kubenswrapper[4791]: I1210 23:08:20.871323 4791 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/851556d7-fa38-4eec-ae74-de989d66d27b-combined-ca-bundle\") pod \"barbican-db-sync-zllvm\" (UID: \"851556d7-fa38-4eec-ae74-de989d66d27b\") " pod="openstack/barbican-db-sync-zllvm" Dec 10 23:08:20 crc kubenswrapper[4791]: I1210 23:08:20.871404 4791 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/aa3a63ab-6092-46ce-aaf6-40360aa060ac-scripts\") pod \"horizon-5f54ccbf57-wwpfk\" (UID: \"aa3a63ab-6092-46ce-aaf6-40360aa060ac\") " pod="openstack/horizon-5f54ccbf57-wwpfk" Dec 10 23:08:20 crc kubenswrapper[4791]: I1210 23:08:20.871454 4791 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/6076a930-1583-41b7-b5ae-bc6c61086e48-combined-ca-bundle\") pod \"neutron-db-sync-kg4vv\" (UID: \"6076a930-1583-41b7-b5ae-bc6c61086e48\") " pod="openstack/neutron-db-sync-kg4vv" Dec 10 23:08:20 crc kubenswrapper[4791]: I1210 23:08:20.871492 4791 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/b9efeadd-d54b-4f78-a0ad-85cc62321b84-scripts\") pod \"placement-db-sync-svgbr\" (UID: \"b9efeadd-d54b-4f78-a0ad-85cc62321b84\") " pod="openstack/placement-db-sync-svgbr" Dec 10 23:08:20 crc kubenswrapper[4791]: I1210 23:08:20.871516 4791 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/851556d7-fa38-4eec-ae74-de989d66d27b-db-sync-config-data\") pod \"barbican-db-sync-zllvm\" (UID: \"851556d7-fa38-4eec-ae74-de989d66d27b\") " pod="openstack/barbican-db-sync-zllvm" Dec 10 23:08:20 crc kubenswrapper[4791]: I1210 23:08:20.871537 4791 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/secret/6076a930-1583-41b7-b5ae-bc6c61086e48-config\") pod \"neutron-db-sync-kg4vv\" (UID: \"6076a930-1583-41b7-b5ae-bc6c61086e48\") " pod="openstack/neutron-db-sync-kg4vv" Dec 10 23:08:20 crc kubenswrapper[4791]: I1210 23:08:20.871571 4791 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-66x4z\" (UniqueName: \"kubernetes.io/projected/6076a930-1583-41b7-b5ae-bc6c61086e48-kube-api-access-66x4z\") pod \"neutron-db-sync-kg4vv\" (UID: \"6076a930-1583-41b7-b5ae-bc6c61086e48\") " pod="openstack/neutron-db-sync-kg4vv" Dec 10 23:08:20 crc kubenswrapper[4791]: I1210 23:08:20.871588 4791 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-crjrd\" (UniqueName: \"kubernetes.io/projected/851556d7-fa38-4eec-ae74-de989d66d27b-kube-api-access-crjrd\") pod \"barbican-db-sync-zllvm\" (UID: \"851556d7-fa38-4eec-ae74-de989d66d27b\") " pod="openstack/barbican-db-sync-zllvm" Dec 10 23:08:20 crc kubenswrapper[4791]: I1210 23:08:20.871625 4791 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/b9efeadd-d54b-4f78-a0ad-85cc62321b84-logs\") pod \"placement-db-sync-svgbr\" (UID: \"b9efeadd-d54b-4f78-a0ad-85cc62321b84\") " pod="openstack/placement-db-sync-svgbr" Dec 10 23:08:20 crc kubenswrapper[4791]: I1210 23:08:20.871670 4791 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-x949f\" (UniqueName: \"kubernetes.io/projected/00320ef6-9362-427b-a2eb-12329d652a0e-kube-api-access-x949f\") pod \"dnsmasq-dns-76fcf4b695-r7wvb\" (UID: \"00320ef6-9362-427b-a2eb-12329d652a0e\") " pod="openstack/dnsmasq-dns-76fcf4b695-r7wvb" Dec 10 23:08:20 crc kubenswrapper[4791]: I1210 23:08:20.871685 4791 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-q27r9\" (UniqueName: \"kubernetes.io/projected/b9efeadd-d54b-4f78-a0ad-85cc62321b84-kube-api-access-q27r9\") pod \"placement-db-sync-svgbr\" (UID: \"b9efeadd-d54b-4f78-a0ad-85cc62321b84\") " pod="openstack/placement-db-sync-svgbr" Dec 10 23:08:20 crc kubenswrapper[4791]: I1210 23:08:20.871717 4791 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/00320ef6-9362-427b-a2eb-12329d652a0e-ovsdbserver-sb\") pod \"dnsmasq-dns-76fcf4b695-r7wvb\" (UID: \"00320ef6-9362-427b-a2eb-12329d652a0e\") " pod="openstack/dnsmasq-dns-76fcf4b695-r7wvb" Dec 10 23:08:20 crc kubenswrapper[4791]: I1210 23:08:20.871751 4791 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"horizon-secret-key\" (UniqueName: \"kubernetes.io/secret/aa3a63ab-6092-46ce-aaf6-40360aa060ac-horizon-secret-key\") pod \"horizon-5f54ccbf57-wwpfk\" (UID: \"aa3a63ab-6092-46ce-aaf6-40360aa060ac\") " pod="openstack/horizon-5f54ccbf57-wwpfk" Dec 10 23:08:20 crc kubenswrapper[4791]: I1210 23:08:20.871828 4791 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/aa3a63ab-6092-46ce-aaf6-40360aa060ac-config-data\") pod \"horizon-5f54ccbf57-wwpfk\" (UID: \"aa3a63ab-6092-46ce-aaf6-40360aa060ac\") " pod="openstack/horizon-5f54ccbf57-wwpfk" Dec 10 23:08:20 crc kubenswrapper[4791]: I1210 23:08:20.871871 4791 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/b9efeadd-d54b-4f78-a0ad-85cc62321b84-config-data\") pod \"placement-db-sync-svgbr\" (UID: \"b9efeadd-d54b-4f78-a0ad-85cc62321b84\") " pod="openstack/placement-db-sync-svgbr" Dec 10 23:08:20 crc kubenswrapper[4791]: I1210 23:08:20.873002 4791 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/b9efeadd-d54b-4f78-a0ad-85cc62321b84-logs\") pod \"placement-db-sync-svgbr\" (UID: \"b9efeadd-d54b-4f78-a0ad-85cc62321b84\") " pod="openstack/placement-db-sync-svgbr" Dec 10 23:08:20 crc kubenswrapper[4791]: I1210 23:08:20.873019 4791 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/00320ef6-9362-427b-a2eb-12329d652a0e-ovsdbserver-sb\") pod \"dnsmasq-dns-76fcf4b695-r7wvb\" (UID: \"00320ef6-9362-427b-a2eb-12329d652a0e\") " pod="openstack/dnsmasq-dns-76fcf4b695-r7wvb" Dec 10 23:08:20 crc kubenswrapper[4791]: I1210 23:08:20.873647 4791 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/00320ef6-9362-427b-a2eb-12329d652a0e-config\") pod \"dnsmasq-dns-76fcf4b695-r7wvb\" (UID: \"00320ef6-9362-427b-a2eb-12329d652a0e\") " pod="openstack/dnsmasq-dns-76fcf4b695-r7wvb" Dec 10 23:08:20 crc kubenswrapper[4791]: I1210 23:08:20.873834 4791 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/00320ef6-9362-427b-a2eb-12329d652a0e-dns-svc\") pod \"dnsmasq-dns-76fcf4b695-r7wvb\" (UID: \"00320ef6-9362-427b-a2eb-12329d652a0e\") " pod="openstack/dnsmasq-dns-76fcf4b695-r7wvb" Dec 10 23:08:20 crc kubenswrapper[4791]: I1210 23:08:20.875457 4791 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/b9efeadd-d54b-4f78-a0ad-85cc62321b84-combined-ca-bundle\") pod \"placement-db-sync-svgbr\" (UID: \"b9efeadd-d54b-4f78-a0ad-85cc62321b84\") " pod="openstack/placement-db-sync-svgbr" Dec 10 23:08:20 crc kubenswrapper[4791]: I1210 23:08:20.877691 4791 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/b9efeadd-d54b-4f78-a0ad-85cc62321b84-config-data\") pod \"placement-db-sync-svgbr\" (UID: \"b9efeadd-d54b-4f78-a0ad-85cc62321b84\") " pod="openstack/placement-db-sync-svgbr" Dec 10 23:08:20 crc kubenswrapper[4791]: I1210 23:08:20.879098 4791 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/b9efeadd-d54b-4f78-a0ad-85cc62321b84-scripts\") pod \"placement-db-sync-svgbr\" (UID: \"b9efeadd-d54b-4f78-a0ad-85cc62321b84\") " pod="openstack/placement-db-sync-svgbr" Dec 10 23:08:20 crc kubenswrapper[4791]: I1210 23:08:20.879750 4791 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/6076a930-1583-41b7-b5ae-bc6c61086e48-combined-ca-bundle\") pod \"neutron-db-sync-kg4vv\" (UID: \"6076a930-1583-41b7-b5ae-bc6c61086e48\") " pod="openstack/neutron-db-sync-kg4vv" Dec 10 23:08:20 crc kubenswrapper[4791]: I1210 23:08:20.880109 4791 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/851556d7-fa38-4eec-ae74-de989d66d27b-combined-ca-bundle\") pod \"barbican-db-sync-zllvm\" (UID: \"851556d7-fa38-4eec-ae74-de989d66d27b\") " pod="openstack/barbican-db-sync-zllvm" Dec 10 23:08:20 crc kubenswrapper[4791]: I1210 23:08:20.885748 4791 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/851556d7-fa38-4eec-ae74-de989d66d27b-db-sync-config-data\") pod \"barbican-db-sync-zllvm\" (UID: \"851556d7-fa38-4eec-ae74-de989d66d27b\") " pod="openstack/barbican-db-sync-zllvm" Dec 10 23:08:20 crc kubenswrapper[4791]: I1210 23:08:20.886580 4791 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/secret/6076a930-1583-41b7-b5ae-bc6c61086e48-config\") pod \"neutron-db-sync-kg4vv\" (UID: \"6076a930-1583-41b7-b5ae-bc6c61086e48\") " pod="openstack/neutron-db-sync-kg4vv" Dec 10 23:08:20 crc kubenswrapper[4791]: I1210 23:08:20.889706 4791 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-66x4z\" (UniqueName: \"kubernetes.io/projected/6076a930-1583-41b7-b5ae-bc6c61086e48-kube-api-access-66x4z\") pod \"neutron-db-sync-kg4vv\" (UID: \"6076a930-1583-41b7-b5ae-bc6c61086e48\") " pod="openstack/neutron-db-sync-kg4vv" Dec 10 23:08:20 crc kubenswrapper[4791]: I1210 23:08:20.889933 4791 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-x949f\" (UniqueName: \"kubernetes.io/projected/00320ef6-9362-427b-a2eb-12329d652a0e-kube-api-access-x949f\") pod \"dnsmasq-dns-76fcf4b695-r7wvb\" (UID: \"00320ef6-9362-427b-a2eb-12329d652a0e\") " pod="openstack/dnsmasq-dns-76fcf4b695-r7wvb" Dec 10 23:08:20 crc kubenswrapper[4791]: I1210 23:08:20.892824 4791 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-crjrd\" (UniqueName: \"kubernetes.io/projected/851556d7-fa38-4eec-ae74-de989d66d27b-kube-api-access-crjrd\") pod \"barbican-db-sync-zllvm\" (UID: \"851556d7-fa38-4eec-ae74-de989d66d27b\") " pod="openstack/barbican-db-sync-zllvm" Dec 10 23:08:20 crc kubenswrapper[4791]: I1210 23:08:20.895322 4791 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-q27r9\" (UniqueName: \"kubernetes.io/projected/b9efeadd-d54b-4f78-a0ad-85cc62321b84-kube-api-access-q27r9\") pod \"placement-db-sync-svgbr\" (UID: \"b9efeadd-d54b-4f78-a0ad-85cc62321b84\") " pod="openstack/placement-db-sync-svgbr" Dec 10 23:08:20 crc kubenswrapper[4791]: I1210 23:08:20.960977 4791 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Dec 10 23:08:20 crc kubenswrapper[4791]: I1210 23:08:20.975116 4791 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/aa3a63ab-6092-46ce-aaf6-40360aa060ac-scripts\") pod \"horizon-5f54ccbf57-wwpfk\" (UID: \"aa3a63ab-6092-46ce-aaf6-40360aa060ac\") " pod="openstack/horizon-5f54ccbf57-wwpfk" Dec 10 23:08:20 crc kubenswrapper[4791]: I1210 23:08:20.975613 4791 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"horizon-secret-key\" (UniqueName: \"kubernetes.io/secret/aa3a63ab-6092-46ce-aaf6-40360aa060ac-horizon-secret-key\") pod \"horizon-5f54ccbf57-wwpfk\" (UID: \"aa3a63ab-6092-46ce-aaf6-40360aa060ac\") " pod="openstack/horizon-5f54ccbf57-wwpfk" Dec 10 23:08:20 crc kubenswrapper[4791]: I1210 23:08:20.975659 4791 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/aa3a63ab-6092-46ce-aaf6-40360aa060ac-config-data\") pod \"horizon-5f54ccbf57-wwpfk\" (UID: \"aa3a63ab-6092-46ce-aaf6-40360aa060ac\") " pod="openstack/horizon-5f54ccbf57-wwpfk" Dec 10 23:08:20 crc kubenswrapper[4791]: I1210 23:08:20.975737 4791 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/aa3a63ab-6092-46ce-aaf6-40360aa060ac-logs\") pod \"horizon-5f54ccbf57-wwpfk\" (UID: \"aa3a63ab-6092-46ce-aaf6-40360aa060ac\") " pod="openstack/horizon-5f54ccbf57-wwpfk" Dec 10 23:08:20 crc kubenswrapper[4791]: I1210 23:08:20.975762 4791 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-85n5v\" (UniqueName: \"kubernetes.io/projected/aa3a63ab-6092-46ce-aaf6-40360aa060ac-kube-api-access-85n5v\") pod \"horizon-5f54ccbf57-wwpfk\" (UID: \"aa3a63ab-6092-46ce-aaf6-40360aa060ac\") " pod="openstack/horizon-5f54ccbf57-wwpfk" Dec 10 23:08:20 crc kubenswrapper[4791]: I1210 23:08:20.979161 4791 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/aa3a63ab-6092-46ce-aaf6-40360aa060ac-scripts\") pod \"horizon-5f54ccbf57-wwpfk\" (UID: \"aa3a63ab-6092-46ce-aaf6-40360aa060ac\") " pod="openstack/horizon-5f54ccbf57-wwpfk" Dec 10 23:08:20 crc kubenswrapper[4791]: I1210 23:08:20.979596 4791 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/aa3a63ab-6092-46ce-aaf6-40360aa060ac-logs\") pod \"horizon-5f54ccbf57-wwpfk\" (UID: \"aa3a63ab-6092-46ce-aaf6-40360aa060ac\") " pod="openstack/horizon-5f54ccbf57-wwpfk" Dec 10 23:08:20 crc kubenswrapper[4791]: I1210 23:08:20.980719 4791 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/aa3a63ab-6092-46ce-aaf6-40360aa060ac-config-data\") pod \"horizon-5f54ccbf57-wwpfk\" (UID: \"aa3a63ab-6092-46ce-aaf6-40360aa060ac\") " pod="openstack/horizon-5f54ccbf57-wwpfk" Dec 10 23:08:20 crc kubenswrapper[4791]: I1210 23:08:20.983396 4791 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"horizon-secret-key\" (UniqueName: \"kubernetes.io/secret/aa3a63ab-6092-46ce-aaf6-40360aa060ac-horizon-secret-key\") pod \"horizon-5f54ccbf57-wwpfk\" (UID: \"aa3a63ab-6092-46ce-aaf6-40360aa060ac\") " pod="openstack/horizon-5f54ccbf57-wwpfk" Dec 10 23:08:20 crc kubenswrapper[4791]: I1210 23:08:20.994189 4791 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-db-sync-g4s5p" Dec 10 23:08:20 crc kubenswrapper[4791]: I1210 23:08:20.996669 4791 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-85n5v\" (UniqueName: \"kubernetes.io/projected/aa3a63ab-6092-46ce-aaf6-40360aa060ac-kube-api-access-85n5v\") pod \"horizon-5f54ccbf57-wwpfk\" (UID: \"aa3a63ab-6092-46ce-aaf6-40360aa060ac\") " pod="openstack/horizon-5f54ccbf57-wwpfk" Dec 10 23:08:21 crc kubenswrapper[4791]: I1210 23:08:21.012750 4791 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-db-sync-kg4vv" Dec 10 23:08:21 crc kubenswrapper[4791]: I1210 23:08:21.037411 4791 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/placement-db-sync-svgbr" Dec 10 23:08:21 crc kubenswrapper[4791]: I1210 23:08:21.070867 4791 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-76fcf4b695-r7wvb" Dec 10 23:08:21 crc kubenswrapper[4791]: I1210 23:08:21.077872 4791 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-db-sync-zllvm" Dec 10 23:08:21 crc kubenswrapper[4791]: I1210 23:08:21.098035 4791 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/horizon-5f54ccbf57-wwpfk" Dec 10 23:08:21 crc kubenswrapper[4791]: I1210 23:08:21.167453 4791 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-55fff446b9-84xqj"] Dec 10 23:08:21 crc kubenswrapper[4791]: W1210 23:08:21.176608 4791 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-poda9741492_6c12_4225_bab8_f0d6cd39c969.slice/crio-af98316e7186debd5dbb69da4dbea2f45f8dba4f6dd2ab14fbc92d9c5e221c4f WatchSource:0}: Error finding container af98316e7186debd5dbb69da4dbea2f45f8dba4f6dd2ab14fbc92d9c5e221c4f: Status 404 returned error can't find the container with id af98316e7186debd5dbb69da4dbea2f45f8dba4f6dd2ab14fbc92d9c5e221c4f Dec 10 23:08:21 crc kubenswrapper[4791]: I1210 23:08:21.225845 4791 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/horizon-78fc5d56dc-58cjt"] Dec 10 23:08:21 crc kubenswrapper[4791]: I1210 23:08:21.402966 4791 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/keystone-bootstrap-tbpq4"] Dec 10 23:08:21 crc kubenswrapper[4791]: W1210 23:08:21.716403 4791 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod8eaa1b1d_715f_4059_ad18_fe9632c036a7.slice/crio-759f432e76faae97192a0bd18f244848d3975025239dc1ed6f67864d6c26fd4a WatchSource:0}: Error finding container 759f432e76faae97192a0bd18f244848d3975025239dc1ed6f67864d6c26fd4a: Status 404 returned error can't find the container with id 759f432e76faae97192a0bd18f244848d3975025239dc1ed6f67864d6c26fd4a Dec 10 23:08:21 crc kubenswrapper[4791]: I1210 23:08:21.725370 4791 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceilometer-0"] Dec 10 23:08:21 crc kubenswrapper[4791]: I1210 23:08:21.866409 4791 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/horizon-78fc5d56dc-58cjt" event={"ID":"14f46cbc-afe2-40bb-a2a3-5673b5ac53e7","Type":"ContainerStarted","Data":"24570b6eae27ba7121cf1d2b73b3c0e94c1598dfd6a06e134fc597203116ee98"} Dec 10 23:08:21 crc kubenswrapper[4791]: I1210 23:08:21.871407 4791 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"8eaa1b1d-715f-4059-ad18-fe9632c036a7","Type":"ContainerStarted","Data":"759f432e76faae97192a0bd18f244848d3975025239dc1ed6f67864d6c26fd4a"} Dec 10 23:08:21 crc kubenswrapper[4791]: I1210 23:08:21.874095 4791 generic.go:334] "Generic (PLEG): container finished" podID="a9741492-6c12-4225-bab8-f0d6cd39c969" containerID="206268dbd38be5a86301e5ea2ba3c086665b81f005bc0a3d41a4d3cda6e4a174" exitCode=0 Dec 10 23:08:21 crc kubenswrapper[4791]: I1210 23:08:21.874166 4791 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-55fff446b9-84xqj" event={"ID":"a9741492-6c12-4225-bab8-f0d6cd39c969","Type":"ContainerDied","Data":"206268dbd38be5a86301e5ea2ba3c086665b81f005bc0a3d41a4d3cda6e4a174"} Dec 10 23:08:21 crc kubenswrapper[4791]: I1210 23:08:21.874193 4791 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-55fff446b9-84xqj" event={"ID":"a9741492-6c12-4225-bab8-f0d6cd39c969","Type":"ContainerStarted","Data":"af98316e7186debd5dbb69da4dbea2f45f8dba4f6dd2ab14fbc92d9c5e221c4f"} Dec 10 23:08:21 crc kubenswrapper[4791]: I1210 23:08:21.876159 4791 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-bootstrap-tbpq4" event={"ID":"f7db40d5-1f25-4bab-98af-29999f12acf1","Type":"ContainerStarted","Data":"07fb0486c1e3fbfccc4819aa21c7ffea14c6bbcfde3f6c560ded3b5e60ee75a1"} Dec 10 23:08:21 crc kubenswrapper[4791]: I1210 23:08:21.876209 4791 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-bootstrap-tbpq4" event={"ID":"f7db40d5-1f25-4bab-98af-29999f12acf1","Type":"ContainerStarted","Data":"e53c596e23188bd58e688e924d170d746c36b6faae11999ae74ff5b8c804f1b8"} Dec 10 23:08:21 crc kubenswrapper[4791]: I1210 23:08:21.935958 4791 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/keystone-bootstrap-tbpq4" podStartSLOduration=1.9359370569999999 podStartE2EDuration="1.935937057s" podCreationTimestamp="2025-12-10 23:08:20 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-10 23:08:21.922110886 +0000 UTC m=+1136.351728499" watchObservedRunningTime="2025-12-10 23:08:21.935937057 +0000 UTC m=+1136.365554670" Dec 10 23:08:21 crc kubenswrapper[4791]: I1210 23:08:21.959898 4791 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/neutron-db-sync-kg4vv"] Dec 10 23:08:21 crc kubenswrapper[4791]: W1210 23:08:21.965157 4791 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod6076a930_1583_41b7_b5ae_bc6c61086e48.slice/crio-8d4d2124883e0dca88bed5ae049d5f29c1818ec35294b04b7dd019fa938fea46 WatchSource:0}: Error finding container 8d4d2124883e0dca88bed5ae049d5f29c1818ec35294b04b7dd019fa938fea46: Status 404 returned error can't find the container with id 8d4d2124883e0dca88bed5ae049d5f29c1818ec35294b04b7dd019fa938fea46 Dec 10 23:08:21 crc kubenswrapper[4791]: I1210 23:08:21.976540 4791 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-db-sync-g4s5p"] Dec 10 23:08:21 crc kubenswrapper[4791]: I1210 23:08:21.987573 4791 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/barbican-db-sync-zllvm"] Dec 10 23:08:22 crc kubenswrapper[4791]: I1210 23:08:22.010694 4791 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/placement-db-sync-svgbr"] Dec 10 23:08:22 crc kubenswrapper[4791]: I1210 23:08:22.123009 4791 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/horizon-5f54ccbf57-wwpfk"] Dec 10 23:08:22 crc kubenswrapper[4791]: I1210 23:08:22.137457 4791 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-76fcf4b695-r7wvb"] Dec 10 23:08:22 crc kubenswrapper[4791]: I1210 23:08:22.244669 4791 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-55fff446b9-84xqj" Dec 10 23:08:22 crc kubenswrapper[4791]: I1210 23:08:22.320151 4791 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/horizon-5f54ccbf57-wwpfk"] Dec 10 23:08:22 crc kubenswrapper[4791]: I1210 23:08:22.333796 4791 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/a9741492-6c12-4225-bab8-f0d6cd39c969-ovsdbserver-nb\") pod \"a9741492-6c12-4225-bab8-f0d6cd39c969\" (UID: \"a9741492-6c12-4225-bab8-f0d6cd39c969\") " Dec 10 23:08:22 crc kubenswrapper[4791]: I1210 23:08:22.333889 4791 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/a9741492-6c12-4225-bab8-f0d6cd39c969-dns-svc\") pod \"a9741492-6c12-4225-bab8-f0d6cd39c969\" (UID: \"a9741492-6c12-4225-bab8-f0d6cd39c969\") " Dec 10 23:08:22 crc kubenswrapper[4791]: I1210 23:08:22.333990 4791 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/a9741492-6c12-4225-bab8-f0d6cd39c969-dns-swift-storage-0\") pod \"a9741492-6c12-4225-bab8-f0d6cd39c969\" (UID: \"a9741492-6c12-4225-bab8-f0d6cd39c969\") " Dec 10 23:08:22 crc kubenswrapper[4791]: I1210 23:08:22.334272 4791 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/a9741492-6c12-4225-bab8-f0d6cd39c969-config\") pod \"a9741492-6c12-4225-bab8-f0d6cd39c969\" (UID: \"a9741492-6c12-4225-bab8-f0d6cd39c969\") " Dec 10 23:08:22 crc kubenswrapper[4791]: I1210 23:08:22.334324 4791 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-qznw7\" (UniqueName: \"kubernetes.io/projected/a9741492-6c12-4225-bab8-f0d6cd39c969-kube-api-access-qznw7\") pod \"a9741492-6c12-4225-bab8-f0d6cd39c969\" (UID: \"a9741492-6c12-4225-bab8-f0d6cd39c969\") " Dec 10 23:08:22 crc kubenswrapper[4791]: I1210 23:08:22.334387 4791 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/a9741492-6c12-4225-bab8-f0d6cd39c969-ovsdbserver-sb\") pod \"a9741492-6c12-4225-bab8-f0d6cd39c969\" (UID: \"a9741492-6c12-4225-bab8-f0d6cd39c969\") " Dec 10 23:08:22 crc kubenswrapper[4791]: I1210 23:08:22.363348 4791 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/a9741492-6c12-4225-bab8-f0d6cd39c969-kube-api-access-qznw7" (OuterVolumeSpecName: "kube-api-access-qznw7") pod "a9741492-6c12-4225-bab8-f0d6cd39c969" (UID: "a9741492-6c12-4225-bab8-f0d6cd39c969"). InnerVolumeSpecName "kube-api-access-qznw7". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 10 23:08:22 crc kubenswrapper[4791]: I1210 23:08:22.364067 4791 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/a9741492-6c12-4225-bab8-f0d6cd39c969-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "a9741492-6c12-4225-bab8-f0d6cd39c969" (UID: "a9741492-6c12-4225-bab8-f0d6cd39c969"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 10 23:08:22 crc kubenswrapper[4791]: I1210 23:08:22.364145 4791 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ceilometer-0"] Dec 10 23:08:22 crc kubenswrapper[4791]: I1210 23:08:22.383154 4791 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/a9741492-6c12-4225-bab8-f0d6cd39c969-dns-swift-storage-0" (OuterVolumeSpecName: "dns-swift-storage-0") pod "a9741492-6c12-4225-bab8-f0d6cd39c969" (UID: "a9741492-6c12-4225-bab8-f0d6cd39c969"). InnerVolumeSpecName "dns-swift-storage-0". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 10 23:08:22 crc kubenswrapper[4791]: I1210 23:08:22.388294 4791 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/horizon-5cc98b55bf-7xpnj"] Dec 10 23:08:22 crc kubenswrapper[4791]: E1210 23:08:22.410025 4791 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="a9741492-6c12-4225-bab8-f0d6cd39c969" containerName="init" Dec 10 23:08:22 crc kubenswrapper[4791]: I1210 23:08:22.410190 4791 state_mem.go:107] "Deleted CPUSet assignment" podUID="a9741492-6c12-4225-bab8-f0d6cd39c969" containerName="init" Dec 10 23:08:22 crc kubenswrapper[4791]: I1210 23:08:22.410943 4791 memory_manager.go:354] "RemoveStaleState removing state" podUID="a9741492-6c12-4225-bab8-f0d6cd39c969" containerName="init" Dec 10 23:08:22 crc kubenswrapper[4791]: I1210 23:08:22.415765 4791 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/a9741492-6c12-4225-bab8-f0d6cd39c969-config" (OuterVolumeSpecName: "config") pod "a9741492-6c12-4225-bab8-f0d6cd39c969" (UID: "a9741492-6c12-4225-bab8-f0d6cd39c969"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 10 23:08:22 crc kubenswrapper[4791]: I1210 23:08:22.421236 4791 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/a9741492-6c12-4225-bab8-f0d6cd39c969-ovsdbserver-sb" (OuterVolumeSpecName: "ovsdbserver-sb") pod "a9741492-6c12-4225-bab8-f0d6cd39c969" (UID: "a9741492-6c12-4225-bab8-f0d6cd39c969"). InnerVolumeSpecName "ovsdbserver-sb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 10 23:08:22 crc kubenswrapper[4791]: I1210 23:08:22.423488 4791 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/horizon-5cc98b55bf-7xpnj" Dec 10 23:08:22 crc kubenswrapper[4791]: I1210 23:08:22.435034 4791 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/a9741492-6c12-4225-bab8-f0d6cd39c969-ovsdbserver-nb" (OuterVolumeSpecName: "ovsdbserver-nb") pod "a9741492-6c12-4225-bab8-f0d6cd39c969" (UID: "a9741492-6c12-4225-bab8-f0d6cd39c969"). InnerVolumeSpecName "ovsdbserver-nb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 10 23:08:22 crc kubenswrapper[4791]: I1210 23:08:22.443066 4791 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-qznw7\" (UniqueName: \"kubernetes.io/projected/a9741492-6c12-4225-bab8-f0d6cd39c969-kube-api-access-qznw7\") on node \"crc\" DevicePath \"\"" Dec 10 23:08:22 crc kubenswrapper[4791]: I1210 23:08:22.443100 4791 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/a9741492-6c12-4225-bab8-f0d6cd39c969-ovsdbserver-sb\") on node \"crc\" DevicePath \"\"" Dec 10 23:08:22 crc kubenswrapper[4791]: I1210 23:08:22.443109 4791 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/a9741492-6c12-4225-bab8-f0d6cd39c969-ovsdbserver-nb\") on node \"crc\" DevicePath \"\"" Dec 10 23:08:22 crc kubenswrapper[4791]: I1210 23:08:22.443118 4791 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/a9741492-6c12-4225-bab8-f0d6cd39c969-dns-svc\") on node \"crc\" DevicePath \"\"" Dec 10 23:08:22 crc kubenswrapper[4791]: I1210 23:08:22.443128 4791 reconciler_common.go:293] "Volume detached for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/a9741492-6c12-4225-bab8-f0d6cd39c969-dns-swift-storage-0\") on node \"crc\" DevicePath \"\"" Dec 10 23:08:22 crc kubenswrapper[4791]: I1210 23:08:22.443136 4791 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/a9741492-6c12-4225-bab8-f0d6cd39c969-config\") on node \"crc\" DevicePath \"\"" Dec 10 23:08:22 crc kubenswrapper[4791]: I1210 23:08:22.450110 4791 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/horizon-5cc98b55bf-7xpnj"] Dec 10 23:08:22 crc kubenswrapper[4791]: I1210 23:08:22.544987 4791 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/9e9c1419-c6f3-4947-8788-4915f48677d7-logs\") pod \"horizon-5cc98b55bf-7xpnj\" (UID: \"9e9c1419-c6f3-4947-8788-4915f48677d7\") " pod="openstack/horizon-5cc98b55bf-7xpnj" Dec 10 23:08:22 crc kubenswrapper[4791]: I1210 23:08:22.545450 4791 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-v97s8\" (UniqueName: \"kubernetes.io/projected/9e9c1419-c6f3-4947-8788-4915f48677d7-kube-api-access-v97s8\") pod \"horizon-5cc98b55bf-7xpnj\" (UID: \"9e9c1419-c6f3-4947-8788-4915f48677d7\") " pod="openstack/horizon-5cc98b55bf-7xpnj" Dec 10 23:08:22 crc kubenswrapper[4791]: I1210 23:08:22.545503 4791 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/9e9c1419-c6f3-4947-8788-4915f48677d7-scripts\") pod \"horizon-5cc98b55bf-7xpnj\" (UID: \"9e9c1419-c6f3-4947-8788-4915f48677d7\") " pod="openstack/horizon-5cc98b55bf-7xpnj" Dec 10 23:08:22 crc kubenswrapper[4791]: I1210 23:08:22.545534 4791 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"horizon-secret-key\" (UniqueName: \"kubernetes.io/secret/9e9c1419-c6f3-4947-8788-4915f48677d7-horizon-secret-key\") pod \"horizon-5cc98b55bf-7xpnj\" (UID: \"9e9c1419-c6f3-4947-8788-4915f48677d7\") " pod="openstack/horizon-5cc98b55bf-7xpnj" Dec 10 23:08:22 crc kubenswrapper[4791]: I1210 23:08:22.545576 4791 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/9e9c1419-c6f3-4947-8788-4915f48677d7-config-data\") pod \"horizon-5cc98b55bf-7xpnj\" (UID: \"9e9c1419-c6f3-4947-8788-4915f48677d7\") " pod="openstack/horizon-5cc98b55bf-7xpnj" Dec 10 23:08:22 crc kubenswrapper[4791]: I1210 23:08:22.647643 4791 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/9e9c1419-c6f3-4947-8788-4915f48677d7-scripts\") pod \"horizon-5cc98b55bf-7xpnj\" (UID: \"9e9c1419-c6f3-4947-8788-4915f48677d7\") " pod="openstack/horizon-5cc98b55bf-7xpnj" Dec 10 23:08:22 crc kubenswrapper[4791]: I1210 23:08:22.647706 4791 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"horizon-secret-key\" (UniqueName: \"kubernetes.io/secret/9e9c1419-c6f3-4947-8788-4915f48677d7-horizon-secret-key\") pod \"horizon-5cc98b55bf-7xpnj\" (UID: \"9e9c1419-c6f3-4947-8788-4915f48677d7\") " pod="openstack/horizon-5cc98b55bf-7xpnj" Dec 10 23:08:22 crc kubenswrapper[4791]: I1210 23:08:22.648497 4791 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/9e9c1419-c6f3-4947-8788-4915f48677d7-scripts\") pod \"horizon-5cc98b55bf-7xpnj\" (UID: \"9e9c1419-c6f3-4947-8788-4915f48677d7\") " pod="openstack/horizon-5cc98b55bf-7xpnj" Dec 10 23:08:22 crc kubenswrapper[4791]: I1210 23:08:22.648538 4791 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/9e9c1419-c6f3-4947-8788-4915f48677d7-config-data\") pod \"horizon-5cc98b55bf-7xpnj\" (UID: \"9e9c1419-c6f3-4947-8788-4915f48677d7\") " pod="openstack/horizon-5cc98b55bf-7xpnj" Dec 10 23:08:22 crc kubenswrapper[4791]: I1210 23:08:22.648604 4791 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/9e9c1419-c6f3-4947-8788-4915f48677d7-logs\") pod \"horizon-5cc98b55bf-7xpnj\" (UID: \"9e9c1419-c6f3-4947-8788-4915f48677d7\") " pod="openstack/horizon-5cc98b55bf-7xpnj" Dec 10 23:08:22 crc kubenswrapper[4791]: I1210 23:08:22.648652 4791 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-v97s8\" (UniqueName: \"kubernetes.io/projected/9e9c1419-c6f3-4947-8788-4915f48677d7-kube-api-access-v97s8\") pod \"horizon-5cc98b55bf-7xpnj\" (UID: \"9e9c1419-c6f3-4947-8788-4915f48677d7\") " pod="openstack/horizon-5cc98b55bf-7xpnj" Dec 10 23:08:22 crc kubenswrapper[4791]: I1210 23:08:22.649148 4791 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/9e9c1419-c6f3-4947-8788-4915f48677d7-logs\") pod \"horizon-5cc98b55bf-7xpnj\" (UID: \"9e9c1419-c6f3-4947-8788-4915f48677d7\") " pod="openstack/horizon-5cc98b55bf-7xpnj" Dec 10 23:08:22 crc kubenswrapper[4791]: I1210 23:08:22.649567 4791 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/9e9c1419-c6f3-4947-8788-4915f48677d7-config-data\") pod \"horizon-5cc98b55bf-7xpnj\" (UID: \"9e9c1419-c6f3-4947-8788-4915f48677d7\") " pod="openstack/horizon-5cc98b55bf-7xpnj" Dec 10 23:08:22 crc kubenswrapper[4791]: I1210 23:08:22.654326 4791 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"horizon-secret-key\" (UniqueName: \"kubernetes.io/secret/9e9c1419-c6f3-4947-8788-4915f48677d7-horizon-secret-key\") pod \"horizon-5cc98b55bf-7xpnj\" (UID: \"9e9c1419-c6f3-4947-8788-4915f48677d7\") " pod="openstack/horizon-5cc98b55bf-7xpnj" Dec 10 23:08:22 crc kubenswrapper[4791]: I1210 23:08:22.672402 4791 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-v97s8\" (UniqueName: \"kubernetes.io/projected/9e9c1419-c6f3-4947-8788-4915f48677d7-kube-api-access-v97s8\") pod \"horizon-5cc98b55bf-7xpnj\" (UID: \"9e9c1419-c6f3-4947-8788-4915f48677d7\") " pod="openstack/horizon-5cc98b55bf-7xpnj" Dec 10 23:08:22 crc kubenswrapper[4791]: I1210 23:08:22.833837 4791 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/horizon-5cc98b55bf-7xpnj" Dec 10 23:08:22 crc kubenswrapper[4791]: I1210 23:08:22.902224 4791 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-db-sync-svgbr" event={"ID":"b9efeadd-d54b-4f78-a0ad-85cc62321b84","Type":"ContainerStarted","Data":"539bc24935d590728b85d4263f0e96b5faf6637750a2619fa9490a40c7dbf5c6"} Dec 10 23:08:22 crc kubenswrapper[4791]: I1210 23:08:22.909159 4791 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-db-sync-zllvm" event={"ID":"851556d7-fa38-4eec-ae74-de989d66d27b","Type":"ContainerStarted","Data":"85645cc296d872c3ce2efa9f113d3205f9f481a970d015419a478a6a5c20d362"} Dec 10 23:08:22 crc kubenswrapper[4791]: I1210 23:08:22.913298 4791 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-db-sync-kg4vv" event={"ID":"6076a930-1583-41b7-b5ae-bc6c61086e48","Type":"ContainerStarted","Data":"b409fe2d3a34ee8ca490698ce05841ae136b5e63e5b04879ee4a783bcca8809e"} Dec 10 23:08:22 crc kubenswrapper[4791]: I1210 23:08:22.913428 4791 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-db-sync-kg4vv" event={"ID":"6076a930-1583-41b7-b5ae-bc6c61086e48","Type":"ContainerStarted","Data":"8d4d2124883e0dca88bed5ae049d5f29c1818ec35294b04b7dd019fa938fea46"} Dec 10 23:08:22 crc kubenswrapper[4791]: I1210 23:08:22.915860 4791 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-55fff446b9-84xqj" Dec 10 23:08:22 crc kubenswrapper[4791]: I1210 23:08:22.915855 4791 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-55fff446b9-84xqj" event={"ID":"a9741492-6c12-4225-bab8-f0d6cd39c969","Type":"ContainerDied","Data":"af98316e7186debd5dbb69da4dbea2f45f8dba4f6dd2ab14fbc92d9c5e221c4f"} Dec 10 23:08:22 crc kubenswrapper[4791]: I1210 23:08:22.916036 4791 scope.go:117] "RemoveContainer" containerID="206268dbd38be5a86301e5ea2ba3c086665b81f005bc0a3d41a4d3cda6e4a174" Dec 10 23:08:22 crc kubenswrapper[4791]: I1210 23:08:22.917327 4791 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/horizon-5f54ccbf57-wwpfk" event={"ID":"aa3a63ab-6092-46ce-aaf6-40360aa060ac","Type":"ContainerStarted","Data":"cd3a5fe7eb6d3ca058066e281d4e2debdd7d84c3d08f93c34d63063dbabf1a7d"} Dec 10 23:08:22 crc kubenswrapper[4791]: I1210 23:08:22.920062 4791 generic.go:334] "Generic (PLEG): container finished" podID="00320ef6-9362-427b-a2eb-12329d652a0e" containerID="b4b22d6561fd9eb3dbbf460fde86eaf3110f3e0633b12d60b315543205b7c7a1" exitCode=0 Dec 10 23:08:22 crc kubenswrapper[4791]: I1210 23:08:22.920125 4791 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-76fcf4b695-r7wvb" event={"ID":"00320ef6-9362-427b-a2eb-12329d652a0e","Type":"ContainerDied","Data":"b4b22d6561fd9eb3dbbf460fde86eaf3110f3e0633b12d60b315543205b7c7a1"} Dec 10 23:08:22 crc kubenswrapper[4791]: I1210 23:08:22.920153 4791 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-76fcf4b695-r7wvb" event={"ID":"00320ef6-9362-427b-a2eb-12329d652a0e","Type":"ContainerStarted","Data":"89a0427f33bd75ec7105ab124933fbf02951e2bcaf48b56a70dfc6beac2d88c3"} Dec 10 23:08:22 crc kubenswrapper[4791]: I1210 23:08:22.922571 4791 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-db-sync-g4s5p" event={"ID":"a85df2ed-be84-41b3-8993-c3d963daef08","Type":"ContainerStarted","Data":"31c7a8174d2b3adae98a2a223b58b992d2148dcb534d2f097625927e1a32586e"} Dec 10 23:08:22 crc kubenswrapper[4791]: I1210 23:08:22.937930 4791 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/neutron-db-sync-kg4vv" podStartSLOduration=2.937911955 podStartE2EDuration="2.937911955s" podCreationTimestamp="2025-12-10 23:08:20 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-10 23:08:22.933253031 +0000 UTC m=+1137.362870644" watchObservedRunningTime="2025-12-10 23:08:22.937911955 +0000 UTC m=+1137.367529568" Dec 10 23:08:23 crc kubenswrapper[4791]: I1210 23:08:23.033597 4791 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-55fff446b9-84xqj"] Dec 10 23:08:23 crc kubenswrapper[4791]: I1210 23:08:23.064479 4791 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-55fff446b9-84xqj"] Dec 10 23:08:23 crc kubenswrapper[4791]: I1210 23:08:23.392632 4791 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/horizon-5cc98b55bf-7xpnj"] Dec 10 23:08:23 crc kubenswrapper[4791]: W1210 23:08:23.408024 4791 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod9e9c1419_c6f3_4947_8788_4915f48677d7.slice/crio-4a159644255a6732d7d58d79d7ad6c6e1295b53c7457c9ca2759ec3d5712484b WatchSource:0}: Error finding container 4a159644255a6732d7d58d79d7ad6c6e1295b53c7457c9ca2759ec3d5712484b: Status 404 returned error can't find the container with id 4a159644255a6732d7d58d79d7ad6c6e1295b53c7457c9ca2759ec3d5712484b Dec 10 23:08:23 crc kubenswrapper[4791]: I1210 23:08:23.899045 4791 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="a9741492-6c12-4225-bab8-f0d6cd39c969" path="/var/lib/kubelet/pods/a9741492-6c12-4225-bab8-f0d6cd39c969/volumes" Dec 10 23:08:23 crc kubenswrapper[4791]: I1210 23:08:23.941464 4791 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-76fcf4b695-r7wvb" event={"ID":"00320ef6-9362-427b-a2eb-12329d652a0e","Type":"ContainerStarted","Data":"f6ba39c429e2e3771ef8f6a8d0b8099d2c828d474f01bd5a46f1fd73a72e07ab"} Dec 10 23:08:23 crc kubenswrapper[4791]: I1210 23:08:23.944110 4791 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/dnsmasq-dns-76fcf4b695-r7wvb" Dec 10 23:08:23 crc kubenswrapper[4791]: I1210 23:08:23.951972 4791 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/horizon-5cc98b55bf-7xpnj" event={"ID":"9e9c1419-c6f3-4947-8788-4915f48677d7","Type":"ContainerStarted","Data":"4a159644255a6732d7d58d79d7ad6c6e1295b53c7457c9ca2759ec3d5712484b"} Dec 10 23:08:23 crc kubenswrapper[4791]: I1210 23:08:23.966901 4791 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/dnsmasq-dns-76fcf4b695-r7wvb" podStartSLOduration=3.966885347 podStartE2EDuration="3.966885347s" podCreationTimestamp="2025-12-10 23:08:20 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-10 23:08:23.965820576 +0000 UTC m=+1138.395438209" watchObservedRunningTime="2025-12-10 23:08:23.966885347 +0000 UTC m=+1138.396502960" Dec 10 23:08:24 crc kubenswrapper[4791]: I1210 23:08:24.972098 4791 generic.go:334] "Generic (PLEG): container finished" podID="a2c44c96-70c2-414a-9890-cb95b66f10b2" containerID="1246dddd2dc57e9b79396e6ae705e086d5153488fee1cd18e4dfa7a916a08eed" exitCode=0 Dec 10 23:08:24 crc kubenswrapper[4791]: I1210 23:08:24.972224 4791 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-db-sync-ccl4n" event={"ID":"a2c44c96-70c2-414a-9890-cb95b66f10b2","Type":"ContainerDied","Data":"1246dddd2dc57e9b79396e6ae705e086d5153488fee1cd18e4dfa7a916a08eed"} Dec 10 23:08:25 crc kubenswrapper[4791]: I1210 23:08:25.038214 4791 patch_prober.go:28] interesting pod/machine-config-daemon-5rb5l container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 10 23:08:25 crc kubenswrapper[4791]: I1210 23:08:25.038286 4791 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-5rb5l" podUID="ba35653c-6e06-4cee-a4d6-137764090d18" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 10 23:08:27 crc kubenswrapper[4791]: I1210 23:08:27.023507 4791 generic.go:334] "Generic (PLEG): container finished" podID="f7db40d5-1f25-4bab-98af-29999f12acf1" containerID="07fb0486c1e3fbfccc4819aa21c7ffea14c6bbcfde3f6c560ded3b5e60ee75a1" exitCode=0 Dec 10 23:08:27 crc kubenswrapper[4791]: I1210 23:08:27.023723 4791 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-bootstrap-tbpq4" event={"ID":"f7db40d5-1f25-4bab-98af-29999f12acf1","Type":"ContainerDied","Data":"07fb0486c1e3fbfccc4819aa21c7ffea14c6bbcfde3f6c560ded3b5e60ee75a1"} Dec 10 23:08:29 crc kubenswrapper[4791]: I1210 23:08:29.042270 4791 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/horizon-78fc5d56dc-58cjt"] Dec 10 23:08:29 crc kubenswrapper[4791]: I1210 23:08:29.075177 4791 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/horizon-7cc66567bd-pwb2l"] Dec 10 23:08:29 crc kubenswrapper[4791]: I1210 23:08:29.076987 4791 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/horizon-7cc66567bd-pwb2l" Dec 10 23:08:29 crc kubenswrapper[4791]: I1210 23:08:29.081698 4791 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-horizon-svc" Dec 10 23:08:29 crc kubenswrapper[4791]: I1210 23:08:29.127401 4791 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/horizon-7cc66567bd-pwb2l"] Dec 10 23:08:29 crc kubenswrapper[4791]: I1210 23:08:29.143114 4791 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/horizon-5cc98b55bf-7xpnj"] Dec 10 23:08:29 crc kubenswrapper[4791]: I1210 23:08:29.164571 4791 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/horizon-6999c5bf5b-7cr9t"] Dec 10 23:08:29 crc kubenswrapper[4791]: I1210 23:08:29.166966 4791 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/horizon-6999c5bf5b-7cr9t" Dec 10 23:08:29 crc kubenswrapper[4791]: I1210 23:08:29.172630 4791 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"horizon-tls-certs\" (UniqueName: \"kubernetes.io/secret/55139c7c-e9fa-4539-b7b2-b2be7ee43b44-horizon-tls-certs\") pod \"horizon-7cc66567bd-pwb2l\" (UID: \"55139c7c-e9fa-4539-b7b2-b2be7ee43b44\") " pod="openstack/horizon-7cc66567bd-pwb2l" Dec 10 23:08:29 crc kubenswrapper[4791]: I1210 23:08:29.172690 4791 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/0fab3b41-ed57-4536-ade0-f5f6ccde40e2-logs\") pod \"horizon-6999c5bf5b-7cr9t\" (UID: \"0fab3b41-ed57-4536-ade0-f5f6ccde40e2\") " pod="openstack/horizon-6999c5bf5b-7cr9t" Dec 10 23:08:29 crc kubenswrapper[4791]: I1210 23:08:29.172748 4791 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/0fab3b41-ed57-4536-ade0-f5f6ccde40e2-scripts\") pod \"horizon-6999c5bf5b-7cr9t\" (UID: \"0fab3b41-ed57-4536-ade0-f5f6ccde40e2\") " pod="openstack/horizon-6999c5bf5b-7cr9t" Dec 10 23:08:29 crc kubenswrapper[4791]: I1210 23:08:29.172789 4791 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/55139c7c-e9fa-4539-b7b2-b2be7ee43b44-config-data\") pod \"horizon-7cc66567bd-pwb2l\" (UID: \"55139c7c-e9fa-4539-b7b2-b2be7ee43b44\") " pod="openstack/horizon-7cc66567bd-pwb2l" Dec 10 23:08:29 crc kubenswrapper[4791]: I1210 23:08:29.172822 4791 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-qmqzj\" (UniqueName: \"kubernetes.io/projected/55139c7c-e9fa-4539-b7b2-b2be7ee43b44-kube-api-access-qmqzj\") pod \"horizon-7cc66567bd-pwb2l\" (UID: \"55139c7c-e9fa-4539-b7b2-b2be7ee43b44\") " pod="openstack/horizon-7cc66567bd-pwb2l" Dec 10 23:08:29 crc kubenswrapper[4791]: I1210 23:08:29.172896 4791 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/0fab3b41-ed57-4536-ade0-f5f6ccde40e2-combined-ca-bundle\") pod \"horizon-6999c5bf5b-7cr9t\" (UID: \"0fab3b41-ed57-4536-ade0-f5f6ccde40e2\") " pod="openstack/horizon-6999c5bf5b-7cr9t" Dec 10 23:08:29 crc kubenswrapper[4791]: I1210 23:08:29.172926 4791 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/0fab3b41-ed57-4536-ade0-f5f6ccde40e2-config-data\") pod \"horizon-6999c5bf5b-7cr9t\" (UID: \"0fab3b41-ed57-4536-ade0-f5f6ccde40e2\") " pod="openstack/horizon-6999c5bf5b-7cr9t" Dec 10 23:08:29 crc kubenswrapper[4791]: I1210 23:08:29.172983 4791 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/55139c7c-e9fa-4539-b7b2-b2be7ee43b44-scripts\") pod \"horizon-7cc66567bd-pwb2l\" (UID: \"55139c7c-e9fa-4539-b7b2-b2be7ee43b44\") " pod="openstack/horizon-7cc66567bd-pwb2l" Dec 10 23:08:29 crc kubenswrapper[4791]: I1210 23:08:29.173006 4791 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/55139c7c-e9fa-4539-b7b2-b2be7ee43b44-combined-ca-bundle\") pod \"horizon-7cc66567bd-pwb2l\" (UID: \"55139c7c-e9fa-4539-b7b2-b2be7ee43b44\") " pod="openstack/horizon-7cc66567bd-pwb2l" Dec 10 23:08:29 crc kubenswrapper[4791]: I1210 23:08:29.173031 4791 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"horizon-secret-key\" (UniqueName: \"kubernetes.io/secret/55139c7c-e9fa-4539-b7b2-b2be7ee43b44-horizon-secret-key\") pod \"horizon-7cc66567bd-pwb2l\" (UID: \"55139c7c-e9fa-4539-b7b2-b2be7ee43b44\") " pod="openstack/horizon-7cc66567bd-pwb2l" Dec 10 23:08:29 crc kubenswrapper[4791]: I1210 23:08:29.173068 4791 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-z4vql\" (UniqueName: \"kubernetes.io/projected/0fab3b41-ed57-4536-ade0-f5f6ccde40e2-kube-api-access-z4vql\") pod \"horizon-6999c5bf5b-7cr9t\" (UID: \"0fab3b41-ed57-4536-ade0-f5f6ccde40e2\") " pod="openstack/horizon-6999c5bf5b-7cr9t" Dec 10 23:08:29 crc kubenswrapper[4791]: I1210 23:08:29.173094 4791 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/55139c7c-e9fa-4539-b7b2-b2be7ee43b44-logs\") pod \"horizon-7cc66567bd-pwb2l\" (UID: \"55139c7c-e9fa-4539-b7b2-b2be7ee43b44\") " pod="openstack/horizon-7cc66567bd-pwb2l" Dec 10 23:08:29 crc kubenswrapper[4791]: I1210 23:08:29.173118 4791 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"horizon-tls-certs\" (UniqueName: \"kubernetes.io/secret/0fab3b41-ed57-4536-ade0-f5f6ccde40e2-horizon-tls-certs\") pod \"horizon-6999c5bf5b-7cr9t\" (UID: \"0fab3b41-ed57-4536-ade0-f5f6ccde40e2\") " pod="openstack/horizon-6999c5bf5b-7cr9t" Dec 10 23:08:29 crc kubenswrapper[4791]: I1210 23:08:29.173156 4791 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"horizon-secret-key\" (UniqueName: \"kubernetes.io/secret/0fab3b41-ed57-4536-ade0-f5f6ccde40e2-horizon-secret-key\") pod \"horizon-6999c5bf5b-7cr9t\" (UID: \"0fab3b41-ed57-4536-ade0-f5f6ccde40e2\") " pod="openstack/horizon-6999c5bf5b-7cr9t" Dec 10 23:08:29 crc kubenswrapper[4791]: I1210 23:08:29.194774 4791 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/horizon-6999c5bf5b-7cr9t"] Dec 10 23:08:29 crc kubenswrapper[4791]: I1210 23:08:29.274711 4791 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/55139c7c-e9fa-4539-b7b2-b2be7ee43b44-scripts\") pod \"horizon-7cc66567bd-pwb2l\" (UID: \"55139c7c-e9fa-4539-b7b2-b2be7ee43b44\") " pod="openstack/horizon-7cc66567bd-pwb2l" Dec 10 23:08:29 crc kubenswrapper[4791]: I1210 23:08:29.274784 4791 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"horizon-secret-key\" (UniqueName: \"kubernetes.io/secret/55139c7c-e9fa-4539-b7b2-b2be7ee43b44-horizon-secret-key\") pod \"horizon-7cc66567bd-pwb2l\" (UID: \"55139c7c-e9fa-4539-b7b2-b2be7ee43b44\") " pod="openstack/horizon-7cc66567bd-pwb2l" Dec 10 23:08:29 crc kubenswrapper[4791]: I1210 23:08:29.274806 4791 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/55139c7c-e9fa-4539-b7b2-b2be7ee43b44-combined-ca-bundle\") pod \"horizon-7cc66567bd-pwb2l\" (UID: \"55139c7c-e9fa-4539-b7b2-b2be7ee43b44\") " pod="openstack/horizon-7cc66567bd-pwb2l" Dec 10 23:08:29 crc kubenswrapper[4791]: I1210 23:08:29.274863 4791 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-z4vql\" (UniqueName: \"kubernetes.io/projected/0fab3b41-ed57-4536-ade0-f5f6ccde40e2-kube-api-access-z4vql\") pod \"horizon-6999c5bf5b-7cr9t\" (UID: \"0fab3b41-ed57-4536-ade0-f5f6ccde40e2\") " pod="openstack/horizon-6999c5bf5b-7cr9t" Dec 10 23:08:29 crc kubenswrapper[4791]: I1210 23:08:29.274887 4791 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/55139c7c-e9fa-4539-b7b2-b2be7ee43b44-logs\") pod \"horizon-7cc66567bd-pwb2l\" (UID: \"55139c7c-e9fa-4539-b7b2-b2be7ee43b44\") " pod="openstack/horizon-7cc66567bd-pwb2l" Dec 10 23:08:29 crc kubenswrapper[4791]: I1210 23:08:29.274927 4791 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"horizon-tls-certs\" (UniqueName: \"kubernetes.io/secret/0fab3b41-ed57-4536-ade0-f5f6ccde40e2-horizon-tls-certs\") pod \"horizon-6999c5bf5b-7cr9t\" (UID: \"0fab3b41-ed57-4536-ade0-f5f6ccde40e2\") " pod="openstack/horizon-6999c5bf5b-7cr9t" Dec 10 23:08:29 crc kubenswrapper[4791]: I1210 23:08:29.274954 4791 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"horizon-secret-key\" (UniqueName: \"kubernetes.io/secret/0fab3b41-ed57-4536-ade0-f5f6ccde40e2-horizon-secret-key\") pod \"horizon-6999c5bf5b-7cr9t\" (UID: \"0fab3b41-ed57-4536-ade0-f5f6ccde40e2\") " pod="openstack/horizon-6999c5bf5b-7cr9t" Dec 10 23:08:29 crc kubenswrapper[4791]: I1210 23:08:29.275008 4791 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"horizon-tls-certs\" (UniqueName: \"kubernetes.io/secret/55139c7c-e9fa-4539-b7b2-b2be7ee43b44-horizon-tls-certs\") pod \"horizon-7cc66567bd-pwb2l\" (UID: \"55139c7c-e9fa-4539-b7b2-b2be7ee43b44\") " pod="openstack/horizon-7cc66567bd-pwb2l" Dec 10 23:08:29 crc kubenswrapper[4791]: I1210 23:08:29.275024 4791 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/0fab3b41-ed57-4536-ade0-f5f6ccde40e2-logs\") pod \"horizon-6999c5bf5b-7cr9t\" (UID: \"0fab3b41-ed57-4536-ade0-f5f6ccde40e2\") " pod="openstack/horizon-6999c5bf5b-7cr9t" Dec 10 23:08:29 crc kubenswrapper[4791]: I1210 23:08:29.275091 4791 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/0fab3b41-ed57-4536-ade0-f5f6ccde40e2-scripts\") pod \"horizon-6999c5bf5b-7cr9t\" (UID: \"0fab3b41-ed57-4536-ade0-f5f6ccde40e2\") " pod="openstack/horizon-6999c5bf5b-7cr9t" Dec 10 23:08:29 crc kubenswrapper[4791]: I1210 23:08:29.275194 4791 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/55139c7c-e9fa-4539-b7b2-b2be7ee43b44-config-data\") pod \"horizon-7cc66567bd-pwb2l\" (UID: \"55139c7c-e9fa-4539-b7b2-b2be7ee43b44\") " pod="openstack/horizon-7cc66567bd-pwb2l" Dec 10 23:08:29 crc kubenswrapper[4791]: I1210 23:08:29.275220 4791 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-qmqzj\" (UniqueName: \"kubernetes.io/projected/55139c7c-e9fa-4539-b7b2-b2be7ee43b44-kube-api-access-qmqzj\") pod \"horizon-7cc66567bd-pwb2l\" (UID: \"55139c7c-e9fa-4539-b7b2-b2be7ee43b44\") " pod="openstack/horizon-7cc66567bd-pwb2l" Dec 10 23:08:29 crc kubenswrapper[4791]: I1210 23:08:29.275424 4791 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/0fab3b41-ed57-4536-ade0-f5f6ccde40e2-combined-ca-bundle\") pod \"horizon-6999c5bf5b-7cr9t\" (UID: \"0fab3b41-ed57-4536-ade0-f5f6ccde40e2\") " pod="openstack/horizon-6999c5bf5b-7cr9t" Dec 10 23:08:29 crc kubenswrapper[4791]: I1210 23:08:29.275506 4791 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/0fab3b41-ed57-4536-ade0-f5f6ccde40e2-config-data\") pod \"horizon-6999c5bf5b-7cr9t\" (UID: \"0fab3b41-ed57-4536-ade0-f5f6ccde40e2\") " pod="openstack/horizon-6999c5bf5b-7cr9t" Dec 10 23:08:29 crc kubenswrapper[4791]: I1210 23:08:29.275525 4791 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/55139c7c-e9fa-4539-b7b2-b2be7ee43b44-scripts\") pod \"horizon-7cc66567bd-pwb2l\" (UID: \"55139c7c-e9fa-4539-b7b2-b2be7ee43b44\") " pod="openstack/horizon-7cc66567bd-pwb2l" Dec 10 23:08:29 crc kubenswrapper[4791]: I1210 23:08:29.275815 4791 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/0fab3b41-ed57-4536-ade0-f5f6ccde40e2-logs\") pod \"horizon-6999c5bf5b-7cr9t\" (UID: \"0fab3b41-ed57-4536-ade0-f5f6ccde40e2\") " pod="openstack/horizon-6999c5bf5b-7cr9t" Dec 10 23:08:29 crc kubenswrapper[4791]: I1210 23:08:29.276739 4791 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/55139c7c-e9fa-4539-b7b2-b2be7ee43b44-logs\") pod \"horizon-7cc66567bd-pwb2l\" (UID: \"55139c7c-e9fa-4539-b7b2-b2be7ee43b44\") " pod="openstack/horizon-7cc66567bd-pwb2l" Dec 10 23:08:29 crc kubenswrapper[4791]: I1210 23:08:29.277154 4791 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/0fab3b41-ed57-4536-ade0-f5f6ccde40e2-scripts\") pod \"horizon-6999c5bf5b-7cr9t\" (UID: \"0fab3b41-ed57-4536-ade0-f5f6ccde40e2\") " pod="openstack/horizon-6999c5bf5b-7cr9t" Dec 10 23:08:29 crc kubenswrapper[4791]: I1210 23:08:29.278272 4791 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/0fab3b41-ed57-4536-ade0-f5f6ccde40e2-config-data\") pod \"horizon-6999c5bf5b-7cr9t\" (UID: \"0fab3b41-ed57-4536-ade0-f5f6ccde40e2\") " pod="openstack/horizon-6999c5bf5b-7cr9t" Dec 10 23:08:29 crc kubenswrapper[4791]: I1210 23:08:29.281757 4791 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/55139c7c-e9fa-4539-b7b2-b2be7ee43b44-config-data\") pod \"horizon-7cc66567bd-pwb2l\" (UID: \"55139c7c-e9fa-4539-b7b2-b2be7ee43b44\") " pod="openstack/horizon-7cc66567bd-pwb2l" Dec 10 23:08:29 crc kubenswrapper[4791]: I1210 23:08:29.283694 4791 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/0fab3b41-ed57-4536-ade0-f5f6ccde40e2-combined-ca-bundle\") pod \"horizon-6999c5bf5b-7cr9t\" (UID: \"0fab3b41-ed57-4536-ade0-f5f6ccde40e2\") " pod="openstack/horizon-6999c5bf5b-7cr9t" Dec 10 23:08:29 crc kubenswrapper[4791]: I1210 23:08:29.285062 4791 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"horizon-secret-key\" (UniqueName: \"kubernetes.io/secret/0fab3b41-ed57-4536-ade0-f5f6ccde40e2-horizon-secret-key\") pod \"horizon-6999c5bf5b-7cr9t\" (UID: \"0fab3b41-ed57-4536-ade0-f5f6ccde40e2\") " pod="openstack/horizon-6999c5bf5b-7cr9t" Dec 10 23:08:29 crc kubenswrapper[4791]: I1210 23:08:29.285085 4791 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"horizon-tls-certs\" (UniqueName: \"kubernetes.io/secret/55139c7c-e9fa-4539-b7b2-b2be7ee43b44-horizon-tls-certs\") pod \"horizon-7cc66567bd-pwb2l\" (UID: \"55139c7c-e9fa-4539-b7b2-b2be7ee43b44\") " pod="openstack/horizon-7cc66567bd-pwb2l" Dec 10 23:08:29 crc kubenswrapper[4791]: I1210 23:08:29.286784 4791 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"horizon-tls-certs\" (UniqueName: \"kubernetes.io/secret/0fab3b41-ed57-4536-ade0-f5f6ccde40e2-horizon-tls-certs\") pod \"horizon-6999c5bf5b-7cr9t\" (UID: \"0fab3b41-ed57-4536-ade0-f5f6ccde40e2\") " pod="openstack/horizon-6999c5bf5b-7cr9t" Dec 10 23:08:29 crc kubenswrapper[4791]: I1210 23:08:29.286945 4791 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/55139c7c-e9fa-4539-b7b2-b2be7ee43b44-combined-ca-bundle\") pod \"horizon-7cc66567bd-pwb2l\" (UID: \"55139c7c-e9fa-4539-b7b2-b2be7ee43b44\") " pod="openstack/horizon-7cc66567bd-pwb2l" Dec 10 23:08:29 crc kubenswrapper[4791]: I1210 23:08:29.295172 4791 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"horizon-secret-key\" (UniqueName: \"kubernetes.io/secret/55139c7c-e9fa-4539-b7b2-b2be7ee43b44-horizon-secret-key\") pod \"horizon-7cc66567bd-pwb2l\" (UID: \"55139c7c-e9fa-4539-b7b2-b2be7ee43b44\") " pod="openstack/horizon-7cc66567bd-pwb2l" Dec 10 23:08:29 crc kubenswrapper[4791]: I1210 23:08:29.295765 4791 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-z4vql\" (UniqueName: \"kubernetes.io/projected/0fab3b41-ed57-4536-ade0-f5f6ccde40e2-kube-api-access-z4vql\") pod \"horizon-6999c5bf5b-7cr9t\" (UID: \"0fab3b41-ed57-4536-ade0-f5f6ccde40e2\") " pod="openstack/horizon-6999c5bf5b-7cr9t" Dec 10 23:08:29 crc kubenswrapper[4791]: I1210 23:08:29.299437 4791 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-qmqzj\" (UniqueName: \"kubernetes.io/projected/55139c7c-e9fa-4539-b7b2-b2be7ee43b44-kube-api-access-qmqzj\") pod \"horizon-7cc66567bd-pwb2l\" (UID: \"55139c7c-e9fa-4539-b7b2-b2be7ee43b44\") " pod="openstack/horizon-7cc66567bd-pwb2l" Dec 10 23:08:29 crc kubenswrapper[4791]: I1210 23:08:29.409216 4791 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/horizon-7cc66567bd-pwb2l" Dec 10 23:08:29 crc kubenswrapper[4791]: I1210 23:08:29.489017 4791 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/horizon-6999c5bf5b-7cr9t" Dec 10 23:08:31 crc kubenswrapper[4791]: I1210 23:08:31.072766 4791 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/dnsmasq-dns-76fcf4b695-r7wvb" Dec 10 23:08:31 crc kubenswrapper[4791]: I1210 23:08:31.146003 4791 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-77585f5f8c-drnrs"] Dec 10 23:08:31 crc kubenswrapper[4791]: I1210 23:08:31.146813 4791 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/dnsmasq-dns-77585f5f8c-drnrs" podUID="defe4ef9-9037-4a53-a860-ce6713ee219c" containerName="dnsmasq-dns" containerID="cri-o://799a9c08f13a56a6860ed5b1f58366c2cfea680c89ada5e8e0d0083357e1d1ed" gracePeriod=10 Dec 10 23:08:31 crc kubenswrapper[4791]: E1210 23:08:31.627020 4791 cadvisor_stats_provider.go:516] "Partial failure issuing cadvisor.ContainerInfoV2" err="partial failures: [\"/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-poddefe4ef9_9037_4a53_a860_ce6713ee219c.slice/crio-conmon-799a9c08f13a56a6860ed5b1f58366c2cfea680c89ada5e8e0d0083357e1d1ed.scope\": RecentStats: unable to find data in memory cache]" Dec 10 23:08:32 crc kubenswrapper[4791]: I1210 23:08:32.088712 4791 generic.go:334] "Generic (PLEG): container finished" podID="defe4ef9-9037-4a53-a860-ce6713ee219c" containerID="799a9c08f13a56a6860ed5b1f58366c2cfea680c89ada5e8e0d0083357e1d1ed" exitCode=0 Dec 10 23:08:32 crc kubenswrapper[4791]: I1210 23:08:32.088774 4791 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-77585f5f8c-drnrs" event={"ID":"defe4ef9-9037-4a53-a860-ce6713ee219c","Type":"ContainerDied","Data":"799a9c08f13a56a6860ed5b1f58366c2cfea680c89ada5e8e0d0083357e1d1ed"} Dec 10 23:08:34 crc kubenswrapper[4791]: I1210 23:08:34.313561 4791 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-bootstrap-tbpq4" Dec 10 23:08:34 crc kubenswrapper[4791]: I1210 23:08:34.488941 4791 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/f7db40d5-1f25-4bab-98af-29999f12acf1-credential-keys\") pod \"f7db40d5-1f25-4bab-98af-29999f12acf1\" (UID: \"f7db40d5-1f25-4bab-98af-29999f12acf1\") " Dec 10 23:08:34 crc kubenswrapper[4791]: I1210 23:08:34.489015 4791 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/f7db40d5-1f25-4bab-98af-29999f12acf1-scripts\") pod \"f7db40d5-1f25-4bab-98af-29999f12acf1\" (UID: \"f7db40d5-1f25-4bab-98af-29999f12acf1\") " Dec 10 23:08:34 crc kubenswrapper[4791]: I1210 23:08:34.489058 4791 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f7db40d5-1f25-4bab-98af-29999f12acf1-combined-ca-bundle\") pod \"f7db40d5-1f25-4bab-98af-29999f12acf1\" (UID: \"f7db40d5-1f25-4bab-98af-29999f12acf1\") " Dec 10 23:08:34 crc kubenswrapper[4791]: I1210 23:08:34.489145 4791 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/f7db40d5-1f25-4bab-98af-29999f12acf1-fernet-keys\") pod \"f7db40d5-1f25-4bab-98af-29999f12acf1\" (UID: \"f7db40d5-1f25-4bab-98af-29999f12acf1\") " Dec 10 23:08:34 crc kubenswrapper[4791]: I1210 23:08:34.489246 4791 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-h44bt\" (UniqueName: \"kubernetes.io/projected/f7db40d5-1f25-4bab-98af-29999f12acf1-kube-api-access-h44bt\") pod \"f7db40d5-1f25-4bab-98af-29999f12acf1\" (UID: \"f7db40d5-1f25-4bab-98af-29999f12acf1\") " Dec 10 23:08:34 crc kubenswrapper[4791]: I1210 23:08:34.489279 4791 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/f7db40d5-1f25-4bab-98af-29999f12acf1-config-data\") pod \"f7db40d5-1f25-4bab-98af-29999f12acf1\" (UID: \"f7db40d5-1f25-4bab-98af-29999f12acf1\") " Dec 10 23:08:34 crc kubenswrapper[4791]: I1210 23:08:34.496485 4791 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/f7db40d5-1f25-4bab-98af-29999f12acf1-credential-keys" (OuterVolumeSpecName: "credential-keys") pod "f7db40d5-1f25-4bab-98af-29999f12acf1" (UID: "f7db40d5-1f25-4bab-98af-29999f12acf1"). InnerVolumeSpecName "credential-keys". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 10 23:08:34 crc kubenswrapper[4791]: I1210 23:08:34.496528 4791 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/f7db40d5-1f25-4bab-98af-29999f12acf1-scripts" (OuterVolumeSpecName: "scripts") pod "f7db40d5-1f25-4bab-98af-29999f12acf1" (UID: "f7db40d5-1f25-4bab-98af-29999f12acf1"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 10 23:08:34 crc kubenswrapper[4791]: I1210 23:08:34.496544 4791 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/f7db40d5-1f25-4bab-98af-29999f12acf1-kube-api-access-h44bt" (OuterVolumeSpecName: "kube-api-access-h44bt") pod "f7db40d5-1f25-4bab-98af-29999f12acf1" (UID: "f7db40d5-1f25-4bab-98af-29999f12acf1"). InnerVolumeSpecName "kube-api-access-h44bt". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 10 23:08:34 crc kubenswrapper[4791]: I1210 23:08:34.507676 4791 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/f7db40d5-1f25-4bab-98af-29999f12acf1-fernet-keys" (OuterVolumeSpecName: "fernet-keys") pod "f7db40d5-1f25-4bab-98af-29999f12acf1" (UID: "f7db40d5-1f25-4bab-98af-29999f12acf1"). InnerVolumeSpecName "fernet-keys". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 10 23:08:34 crc kubenswrapper[4791]: I1210 23:08:34.517029 4791 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/f7db40d5-1f25-4bab-98af-29999f12acf1-config-data" (OuterVolumeSpecName: "config-data") pod "f7db40d5-1f25-4bab-98af-29999f12acf1" (UID: "f7db40d5-1f25-4bab-98af-29999f12acf1"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 10 23:08:34 crc kubenswrapper[4791]: I1210 23:08:34.522132 4791 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/f7db40d5-1f25-4bab-98af-29999f12acf1-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "f7db40d5-1f25-4bab-98af-29999f12acf1" (UID: "f7db40d5-1f25-4bab-98af-29999f12acf1"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 10 23:08:34 crc kubenswrapper[4791]: I1210 23:08:34.594607 4791 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/f7db40d5-1f25-4bab-98af-29999f12acf1-scripts\") on node \"crc\" DevicePath \"\"" Dec 10 23:08:34 crc kubenswrapper[4791]: I1210 23:08:34.594670 4791 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f7db40d5-1f25-4bab-98af-29999f12acf1-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 10 23:08:34 crc kubenswrapper[4791]: I1210 23:08:34.594691 4791 reconciler_common.go:293] "Volume detached for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/f7db40d5-1f25-4bab-98af-29999f12acf1-fernet-keys\") on node \"crc\" DevicePath \"\"" Dec 10 23:08:34 crc kubenswrapper[4791]: I1210 23:08:34.594707 4791 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-h44bt\" (UniqueName: \"kubernetes.io/projected/f7db40d5-1f25-4bab-98af-29999f12acf1-kube-api-access-h44bt\") on node \"crc\" DevicePath \"\"" Dec 10 23:08:34 crc kubenswrapper[4791]: I1210 23:08:34.594722 4791 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/f7db40d5-1f25-4bab-98af-29999f12acf1-config-data\") on node \"crc\" DevicePath \"\"" Dec 10 23:08:34 crc kubenswrapper[4791]: I1210 23:08:34.594735 4791 reconciler_common.go:293] "Volume detached for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/f7db40d5-1f25-4bab-98af-29999f12acf1-credential-keys\") on node \"crc\" DevicePath \"\"" Dec 10 23:08:35 crc kubenswrapper[4791]: I1210 23:08:35.131504 4791 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-bootstrap-tbpq4" event={"ID":"f7db40d5-1f25-4bab-98af-29999f12acf1","Type":"ContainerDied","Data":"e53c596e23188bd58e688e924d170d746c36b6faae11999ae74ff5b8c804f1b8"} Dec 10 23:08:35 crc kubenswrapper[4791]: I1210 23:08:35.131828 4791 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="e53c596e23188bd58e688e924d170d746c36b6faae11999ae74ff5b8c804f1b8" Dec 10 23:08:35 crc kubenswrapper[4791]: I1210 23:08:35.131592 4791 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-bootstrap-tbpq4" Dec 10 23:08:35 crc kubenswrapper[4791]: I1210 23:08:35.394932 4791 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/keystone-bootstrap-tbpq4"] Dec 10 23:08:35 crc kubenswrapper[4791]: I1210 23:08:35.408709 4791 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/keystone-bootstrap-tbpq4"] Dec 10 23:08:35 crc kubenswrapper[4791]: I1210 23:08:35.483959 4791 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/keystone-bootstrap-mwqkg"] Dec 10 23:08:35 crc kubenswrapper[4791]: E1210 23:08:35.484329 4791 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f7db40d5-1f25-4bab-98af-29999f12acf1" containerName="keystone-bootstrap" Dec 10 23:08:35 crc kubenswrapper[4791]: I1210 23:08:35.484359 4791 state_mem.go:107] "Deleted CPUSet assignment" podUID="f7db40d5-1f25-4bab-98af-29999f12acf1" containerName="keystone-bootstrap" Dec 10 23:08:35 crc kubenswrapper[4791]: I1210 23:08:35.484584 4791 memory_manager.go:354] "RemoveStaleState removing state" podUID="f7db40d5-1f25-4bab-98af-29999f12acf1" containerName="keystone-bootstrap" Dec 10 23:08:35 crc kubenswrapper[4791]: I1210 23:08:35.485109 4791 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-bootstrap-mwqkg" Dec 10 23:08:35 crc kubenswrapper[4791]: I1210 23:08:35.487937 4791 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone-config-data" Dec 10 23:08:35 crc kubenswrapper[4791]: I1210 23:08:35.489124 4791 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"osp-secret" Dec 10 23:08:35 crc kubenswrapper[4791]: I1210 23:08:35.489318 4791 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone-scripts" Dec 10 23:08:35 crc kubenswrapper[4791]: I1210 23:08:35.491784 4791 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone" Dec 10 23:08:35 crc kubenswrapper[4791]: I1210 23:08:35.493882 4791 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone-keystone-dockercfg-t99v8" Dec 10 23:08:35 crc kubenswrapper[4791]: I1210 23:08:35.500484 4791 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/keystone-bootstrap-mwqkg"] Dec 10 23:08:35 crc kubenswrapper[4791]: I1210 23:08:35.615476 4791 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-xzlsz\" (UniqueName: \"kubernetes.io/projected/7c65a05d-c2cc-45c4-b469-e6a271e01856-kube-api-access-xzlsz\") pod \"keystone-bootstrap-mwqkg\" (UID: \"7c65a05d-c2cc-45c4-b469-e6a271e01856\") " pod="openstack/keystone-bootstrap-mwqkg" Dec 10 23:08:35 crc kubenswrapper[4791]: I1210 23:08:35.615558 4791 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/7c65a05d-c2cc-45c4-b469-e6a271e01856-config-data\") pod \"keystone-bootstrap-mwqkg\" (UID: \"7c65a05d-c2cc-45c4-b469-e6a271e01856\") " pod="openstack/keystone-bootstrap-mwqkg" Dec 10 23:08:35 crc kubenswrapper[4791]: I1210 23:08:35.615627 4791 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/7c65a05d-c2cc-45c4-b469-e6a271e01856-fernet-keys\") pod \"keystone-bootstrap-mwqkg\" (UID: \"7c65a05d-c2cc-45c4-b469-e6a271e01856\") " pod="openstack/keystone-bootstrap-mwqkg" Dec 10 23:08:35 crc kubenswrapper[4791]: I1210 23:08:35.615740 4791 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/7c65a05d-c2cc-45c4-b469-e6a271e01856-scripts\") pod \"keystone-bootstrap-mwqkg\" (UID: \"7c65a05d-c2cc-45c4-b469-e6a271e01856\") " pod="openstack/keystone-bootstrap-mwqkg" Dec 10 23:08:35 crc kubenswrapper[4791]: I1210 23:08:35.616026 4791 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/7c65a05d-c2cc-45c4-b469-e6a271e01856-credential-keys\") pod \"keystone-bootstrap-mwqkg\" (UID: \"7c65a05d-c2cc-45c4-b469-e6a271e01856\") " pod="openstack/keystone-bootstrap-mwqkg" Dec 10 23:08:35 crc kubenswrapper[4791]: I1210 23:08:35.616289 4791 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/7c65a05d-c2cc-45c4-b469-e6a271e01856-combined-ca-bundle\") pod \"keystone-bootstrap-mwqkg\" (UID: \"7c65a05d-c2cc-45c4-b469-e6a271e01856\") " pod="openstack/keystone-bootstrap-mwqkg" Dec 10 23:08:35 crc kubenswrapper[4791]: I1210 23:08:35.718194 4791 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/7c65a05d-c2cc-45c4-b469-e6a271e01856-credential-keys\") pod \"keystone-bootstrap-mwqkg\" (UID: \"7c65a05d-c2cc-45c4-b469-e6a271e01856\") " pod="openstack/keystone-bootstrap-mwqkg" Dec 10 23:08:35 crc kubenswrapper[4791]: I1210 23:08:35.718293 4791 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/7c65a05d-c2cc-45c4-b469-e6a271e01856-combined-ca-bundle\") pod \"keystone-bootstrap-mwqkg\" (UID: \"7c65a05d-c2cc-45c4-b469-e6a271e01856\") " pod="openstack/keystone-bootstrap-mwqkg" Dec 10 23:08:35 crc kubenswrapper[4791]: I1210 23:08:35.718359 4791 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-xzlsz\" (UniqueName: \"kubernetes.io/projected/7c65a05d-c2cc-45c4-b469-e6a271e01856-kube-api-access-xzlsz\") pod \"keystone-bootstrap-mwqkg\" (UID: \"7c65a05d-c2cc-45c4-b469-e6a271e01856\") " pod="openstack/keystone-bootstrap-mwqkg" Dec 10 23:08:35 crc kubenswrapper[4791]: I1210 23:08:35.718414 4791 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/7c65a05d-c2cc-45c4-b469-e6a271e01856-config-data\") pod \"keystone-bootstrap-mwqkg\" (UID: \"7c65a05d-c2cc-45c4-b469-e6a271e01856\") " pod="openstack/keystone-bootstrap-mwqkg" Dec 10 23:08:35 crc kubenswrapper[4791]: I1210 23:08:35.718432 4791 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/7c65a05d-c2cc-45c4-b469-e6a271e01856-fernet-keys\") pod \"keystone-bootstrap-mwqkg\" (UID: \"7c65a05d-c2cc-45c4-b469-e6a271e01856\") " pod="openstack/keystone-bootstrap-mwqkg" Dec 10 23:08:35 crc kubenswrapper[4791]: I1210 23:08:35.718449 4791 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/7c65a05d-c2cc-45c4-b469-e6a271e01856-scripts\") pod \"keystone-bootstrap-mwqkg\" (UID: \"7c65a05d-c2cc-45c4-b469-e6a271e01856\") " pod="openstack/keystone-bootstrap-mwqkg" Dec 10 23:08:35 crc kubenswrapper[4791]: I1210 23:08:35.722694 4791 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/7c65a05d-c2cc-45c4-b469-e6a271e01856-combined-ca-bundle\") pod \"keystone-bootstrap-mwqkg\" (UID: \"7c65a05d-c2cc-45c4-b469-e6a271e01856\") " pod="openstack/keystone-bootstrap-mwqkg" Dec 10 23:08:35 crc kubenswrapper[4791]: I1210 23:08:35.723132 4791 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/7c65a05d-c2cc-45c4-b469-e6a271e01856-config-data\") pod \"keystone-bootstrap-mwqkg\" (UID: \"7c65a05d-c2cc-45c4-b469-e6a271e01856\") " pod="openstack/keystone-bootstrap-mwqkg" Dec 10 23:08:35 crc kubenswrapper[4791]: I1210 23:08:35.729927 4791 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/7c65a05d-c2cc-45c4-b469-e6a271e01856-credential-keys\") pod \"keystone-bootstrap-mwqkg\" (UID: \"7c65a05d-c2cc-45c4-b469-e6a271e01856\") " pod="openstack/keystone-bootstrap-mwqkg" Dec 10 23:08:35 crc kubenswrapper[4791]: I1210 23:08:35.730110 4791 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/7c65a05d-c2cc-45c4-b469-e6a271e01856-fernet-keys\") pod \"keystone-bootstrap-mwqkg\" (UID: \"7c65a05d-c2cc-45c4-b469-e6a271e01856\") " pod="openstack/keystone-bootstrap-mwqkg" Dec 10 23:08:35 crc kubenswrapper[4791]: I1210 23:08:35.734642 4791 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/7c65a05d-c2cc-45c4-b469-e6a271e01856-scripts\") pod \"keystone-bootstrap-mwqkg\" (UID: \"7c65a05d-c2cc-45c4-b469-e6a271e01856\") " pod="openstack/keystone-bootstrap-mwqkg" Dec 10 23:08:35 crc kubenswrapper[4791]: I1210 23:08:35.735678 4791 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-xzlsz\" (UniqueName: \"kubernetes.io/projected/7c65a05d-c2cc-45c4-b469-e6a271e01856-kube-api-access-xzlsz\") pod \"keystone-bootstrap-mwqkg\" (UID: \"7c65a05d-c2cc-45c4-b469-e6a271e01856\") " pod="openstack/keystone-bootstrap-mwqkg" Dec 10 23:08:35 crc kubenswrapper[4791]: I1210 23:08:35.838544 4791 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-bootstrap-mwqkg" Dec 10 23:08:35 crc kubenswrapper[4791]: I1210 23:08:35.897476 4791 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="f7db40d5-1f25-4bab-98af-29999f12acf1" path="/var/lib/kubelet/pods/f7db40d5-1f25-4bab-98af-29999f12acf1/volumes" Dec 10 23:08:36 crc kubenswrapper[4791]: E1210 23:08:36.474448 4791 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="quay.io/podified-antelope-centos9/openstack-placement-api:current-podified" Dec 10 23:08:36 crc kubenswrapper[4791]: E1210 23:08:36.474827 4791 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:placement-db-sync,Image:quay.io/podified-antelope-centos9/openstack-placement-api:current-podified,Command:[/bin/bash],Args:[-c /usr/local/bin/kolla_start],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:KOLLA_BOOTSTRAP,Value:true,ValueFrom:nil,},EnvVar{Name:KOLLA_CONFIG_STRATEGY,Value:COPY_ALWAYS,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:scripts,ReadOnly:true,MountPath:/usr/local/bin/container-scripts,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:logs,ReadOnly:false,MountPath:/var/log/placement,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:config-data,ReadOnly:false,MountPath:/var/lib/openstack/config,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:config-data,ReadOnly:true,MountPath:/var/lib/kolla/config_files/config.json,SubPath:placement-dbsync-config.json,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:combined-ca-bundle,ReadOnly:true,MountPath:/etc/pki/ca-trust/extracted/pem/tls-ca-bundle.pem,SubPath:tls-ca-bundle.pem,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-q27r9,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*42482,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:nil,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod placement-db-sync-svgbr_openstack(b9efeadd-d54b-4f78-a0ad-85cc62321b84): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError" Dec 10 23:08:36 crc kubenswrapper[4791]: E1210 23:08:36.476240 4791 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"placement-db-sync\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\"" pod="openstack/placement-db-sync-svgbr" podUID="b9efeadd-d54b-4f78-a0ad-85cc62321b84" Dec 10 23:08:36 crc kubenswrapper[4791]: I1210 23:08:36.603733 4791 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glance-db-sync-ccl4n" Dec 10 23:08:36 crc kubenswrapper[4791]: I1210 23:08:36.626850 4791 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-77585f5f8c-drnrs" Dec 10 23:08:36 crc kubenswrapper[4791]: I1210 23:08:36.742776 4791 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/a2c44c96-70c2-414a-9890-cb95b66f10b2-db-sync-config-data\") pod \"a2c44c96-70c2-414a-9890-cb95b66f10b2\" (UID: \"a2c44c96-70c2-414a-9890-cb95b66f10b2\") " Dec 10 23:08:36 crc kubenswrapper[4791]: I1210 23:08:36.742858 4791 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/defe4ef9-9037-4a53-a860-ce6713ee219c-dns-swift-storage-0\") pod \"defe4ef9-9037-4a53-a860-ce6713ee219c\" (UID: \"defe4ef9-9037-4a53-a860-ce6713ee219c\") " Dec 10 23:08:36 crc kubenswrapper[4791]: I1210 23:08:36.742922 4791 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/defe4ef9-9037-4a53-a860-ce6713ee219c-ovsdbserver-sb\") pod \"defe4ef9-9037-4a53-a860-ce6713ee219c\" (UID: \"defe4ef9-9037-4a53-a860-ce6713ee219c\") " Dec 10 23:08:36 crc kubenswrapper[4791]: I1210 23:08:36.742985 4791 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/defe4ef9-9037-4a53-a860-ce6713ee219c-dns-svc\") pod \"defe4ef9-9037-4a53-a860-ce6713ee219c\" (UID: \"defe4ef9-9037-4a53-a860-ce6713ee219c\") " Dec 10 23:08:36 crc kubenswrapper[4791]: I1210 23:08:36.743016 4791 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/defe4ef9-9037-4a53-a860-ce6713ee219c-ovsdbserver-nb\") pod \"defe4ef9-9037-4a53-a860-ce6713ee219c\" (UID: \"defe4ef9-9037-4a53-a860-ce6713ee219c\") " Dec 10 23:08:36 crc kubenswrapper[4791]: I1210 23:08:36.743057 4791 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a2c44c96-70c2-414a-9890-cb95b66f10b2-combined-ca-bundle\") pod \"a2c44c96-70c2-414a-9890-cb95b66f10b2\" (UID: \"a2c44c96-70c2-414a-9890-cb95b66f10b2\") " Dec 10 23:08:36 crc kubenswrapper[4791]: I1210 23:08:36.743140 4791 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/a2c44c96-70c2-414a-9890-cb95b66f10b2-config-data\") pod \"a2c44c96-70c2-414a-9890-cb95b66f10b2\" (UID: \"a2c44c96-70c2-414a-9890-cb95b66f10b2\") " Dec 10 23:08:36 crc kubenswrapper[4791]: I1210 23:08:36.743182 4791 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-v7lq8\" (UniqueName: \"kubernetes.io/projected/a2c44c96-70c2-414a-9890-cb95b66f10b2-kube-api-access-v7lq8\") pod \"a2c44c96-70c2-414a-9890-cb95b66f10b2\" (UID: \"a2c44c96-70c2-414a-9890-cb95b66f10b2\") " Dec 10 23:08:36 crc kubenswrapper[4791]: I1210 23:08:36.743249 4791 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/defe4ef9-9037-4a53-a860-ce6713ee219c-config\") pod \"defe4ef9-9037-4a53-a860-ce6713ee219c\" (UID: \"defe4ef9-9037-4a53-a860-ce6713ee219c\") " Dec 10 23:08:36 crc kubenswrapper[4791]: I1210 23:08:36.743399 4791 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-9s79j\" (UniqueName: \"kubernetes.io/projected/defe4ef9-9037-4a53-a860-ce6713ee219c-kube-api-access-9s79j\") pod \"defe4ef9-9037-4a53-a860-ce6713ee219c\" (UID: \"defe4ef9-9037-4a53-a860-ce6713ee219c\") " Dec 10 23:08:36 crc kubenswrapper[4791]: I1210 23:08:36.752122 4791 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/a2c44c96-70c2-414a-9890-cb95b66f10b2-kube-api-access-v7lq8" (OuterVolumeSpecName: "kube-api-access-v7lq8") pod "a2c44c96-70c2-414a-9890-cb95b66f10b2" (UID: "a2c44c96-70c2-414a-9890-cb95b66f10b2"). InnerVolumeSpecName "kube-api-access-v7lq8". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 10 23:08:36 crc kubenswrapper[4791]: I1210 23:08:36.760408 4791 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/defe4ef9-9037-4a53-a860-ce6713ee219c-kube-api-access-9s79j" (OuterVolumeSpecName: "kube-api-access-9s79j") pod "defe4ef9-9037-4a53-a860-ce6713ee219c" (UID: "defe4ef9-9037-4a53-a860-ce6713ee219c"). InnerVolumeSpecName "kube-api-access-9s79j". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 10 23:08:36 crc kubenswrapper[4791]: I1210 23:08:36.787877 4791 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/a2c44c96-70c2-414a-9890-cb95b66f10b2-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "a2c44c96-70c2-414a-9890-cb95b66f10b2" (UID: "a2c44c96-70c2-414a-9890-cb95b66f10b2"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 10 23:08:36 crc kubenswrapper[4791]: I1210 23:08:36.799481 4791 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/defe4ef9-9037-4a53-a860-ce6713ee219c-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "defe4ef9-9037-4a53-a860-ce6713ee219c" (UID: "defe4ef9-9037-4a53-a860-ce6713ee219c"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 10 23:08:36 crc kubenswrapper[4791]: I1210 23:08:36.815488 4791 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/defe4ef9-9037-4a53-a860-ce6713ee219c-config" (OuterVolumeSpecName: "config") pod "defe4ef9-9037-4a53-a860-ce6713ee219c" (UID: "defe4ef9-9037-4a53-a860-ce6713ee219c"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 10 23:08:36 crc kubenswrapper[4791]: I1210 23:08:36.816351 4791 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/defe4ef9-9037-4a53-a860-ce6713ee219c-dns-swift-storage-0" (OuterVolumeSpecName: "dns-swift-storage-0") pod "defe4ef9-9037-4a53-a860-ce6713ee219c" (UID: "defe4ef9-9037-4a53-a860-ce6713ee219c"). InnerVolumeSpecName "dns-swift-storage-0". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 10 23:08:36 crc kubenswrapper[4791]: I1210 23:08:36.817649 4791 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/a2c44c96-70c2-414a-9890-cb95b66f10b2-db-sync-config-data" (OuterVolumeSpecName: "db-sync-config-data") pod "a2c44c96-70c2-414a-9890-cb95b66f10b2" (UID: "a2c44c96-70c2-414a-9890-cb95b66f10b2"). InnerVolumeSpecName "db-sync-config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 10 23:08:36 crc kubenswrapper[4791]: I1210 23:08:36.824887 4791 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/a2c44c96-70c2-414a-9890-cb95b66f10b2-config-data" (OuterVolumeSpecName: "config-data") pod "a2c44c96-70c2-414a-9890-cb95b66f10b2" (UID: "a2c44c96-70c2-414a-9890-cb95b66f10b2"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 10 23:08:36 crc kubenswrapper[4791]: I1210 23:08:36.840133 4791 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/defe4ef9-9037-4a53-a860-ce6713ee219c-ovsdbserver-nb" (OuterVolumeSpecName: "ovsdbserver-nb") pod "defe4ef9-9037-4a53-a860-ce6713ee219c" (UID: "defe4ef9-9037-4a53-a860-ce6713ee219c"). InnerVolumeSpecName "ovsdbserver-nb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 10 23:08:36 crc kubenswrapper[4791]: I1210 23:08:36.844904 4791 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/defe4ef9-9037-4a53-a860-ce6713ee219c-ovsdbserver-sb" (OuterVolumeSpecName: "ovsdbserver-sb") pod "defe4ef9-9037-4a53-a860-ce6713ee219c" (UID: "defe4ef9-9037-4a53-a860-ce6713ee219c"). InnerVolumeSpecName "ovsdbserver-sb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 10 23:08:36 crc kubenswrapper[4791]: I1210 23:08:36.846147 4791 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/defe4ef9-9037-4a53-a860-ce6713ee219c-config\") on node \"crc\" DevicePath \"\"" Dec 10 23:08:36 crc kubenswrapper[4791]: I1210 23:08:36.846190 4791 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-9s79j\" (UniqueName: \"kubernetes.io/projected/defe4ef9-9037-4a53-a860-ce6713ee219c-kube-api-access-9s79j\") on node \"crc\" DevicePath \"\"" Dec 10 23:08:36 crc kubenswrapper[4791]: I1210 23:08:36.846237 4791 reconciler_common.go:293] "Volume detached for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/a2c44c96-70c2-414a-9890-cb95b66f10b2-db-sync-config-data\") on node \"crc\" DevicePath \"\"" Dec 10 23:08:36 crc kubenswrapper[4791]: I1210 23:08:36.846246 4791 reconciler_common.go:293] "Volume detached for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/defe4ef9-9037-4a53-a860-ce6713ee219c-dns-swift-storage-0\") on node \"crc\" DevicePath \"\"" Dec 10 23:08:36 crc kubenswrapper[4791]: I1210 23:08:36.846256 4791 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/defe4ef9-9037-4a53-a860-ce6713ee219c-ovsdbserver-sb\") on node \"crc\" DevicePath \"\"" Dec 10 23:08:36 crc kubenswrapper[4791]: I1210 23:08:36.846264 4791 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/defe4ef9-9037-4a53-a860-ce6713ee219c-dns-svc\") on node \"crc\" DevicePath \"\"" Dec 10 23:08:36 crc kubenswrapper[4791]: I1210 23:08:36.846273 4791 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/defe4ef9-9037-4a53-a860-ce6713ee219c-ovsdbserver-nb\") on node \"crc\" DevicePath \"\"" Dec 10 23:08:36 crc kubenswrapper[4791]: I1210 23:08:36.846283 4791 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a2c44c96-70c2-414a-9890-cb95b66f10b2-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 10 23:08:36 crc kubenswrapper[4791]: I1210 23:08:36.846298 4791 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/a2c44c96-70c2-414a-9890-cb95b66f10b2-config-data\") on node \"crc\" DevicePath \"\"" Dec 10 23:08:36 crc kubenswrapper[4791]: I1210 23:08:36.846308 4791 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-v7lq8\" (UniqueName: \"kubernetes.io/projected/a2c44c96-70c2-414a-9890-cb95b66f10b2-kube-api-access-v7lq8\") on node \"crc\" DevicePath \"\"" Dec 10 23:08:37 crc kubenswrapper[4791]: I1210 23:08:37.148097 4791 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-db-sync-ccl4n" event={"ID":"a2c44c96-70c2-414a-9890-cb95b66f10b2","Type":"ContainerDied","Data":"c32c4e4e6da6ca2408460931b7bc5e7926a5bf56f0c2b87506e93a5fb695ed2c"} Dec 10 23:08:37 crc kubenswrapper[4791]: I1210 23:08:37.148348 4791 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="c32c4e4e6da6ca2408460931b7bc5e7926a5bf56f0c2b87506e93a5fb695ed2c" Dec 10 23:08:37 crc kubenswrapper[4791]: I1210 23:08:37.148135 4791 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glance-db-sync-ccl4n" Dec 10 23:08:37 crc kubenswrapper[4791]: I1210 23:08:37.150421 4791 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-77585f5f8c-drnrs" event={"ID":"defe4ef9-9037-4a53-a860-ce6713ee219c","Type":"ContainerDied","Data":"7dade463ebac697fb2c29c5b384910908465c1fb5f97888b67133f0c327c79b4"} Dec 10 23:08:37 crc kubenswrapper[4791]: I1210 23:08:37.150444 4791 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-77585f5f8c-drnrs" Dec 10 23:08:37 crc kubenswrapper[4791]: I1210 23:08:37.150480 4791 scope.go:117] "RemoveContainer" containerID="799a9c08f13a56a6860ed5b1f58366c2cfea680c89ada5e8e0d0083357e1d1ed" Dec 10 23:08:37 crc kubenswrapper[4791]: E1210 23:08:37.152171 4791 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"placement-db-sync\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/podified-antelope-centos9/openstack-placement-api:current-podified\\\"\"" pod="openstack/placement-db-sync-svgbr" podUID="b9efeadd-d54b-4f78-a0ad-85cc62321b84" Dec 10 23:08:37 crc kubenswrapper[4791]: I1210 23:08:37.197212 4791 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-77585f5f8c-drnrs"] Dec 10 23:08:37 crc kubenswrapper[4791]: I1210 23:08:37.206476 4791 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-77585f5f8c-drnrs"] Dec 10 23:08:37 crc kubenswrapper[4791]: I1210 23:08:37.907289 4791 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="defe4ef9-9037-4a53-a860-ce6713ee219c" path="/var/lib/kubelet/pods/defe4ef9-9037-4a53-a860-ce6713ee219c/volumes" Dec 10 23:08:38 crc kubenswrapper[4791]: I1210 23:08:38.074697 4791 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-8b5c85b87-d7p92"] Dec 10 23:08:38 crc kubenswrapper[4791]: E1210 23:08:38.075360 4791 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="a2c44c96-70c2-414a-9890-cb95b66f10b2" containerName="glance-db-sync" Dec 10 23:08:38 crc kubenswrapper[4791]: I1210 23:08:38.075377 4791 state_mem.go:107] "Deleted CPUSet assignment" podUID="a2c44c96-70c2-414a-9890-cb95b66f10b2" containerName="glance-db-sync" Dec 10 23:08:38 crc kubenswrapper[4791]: E1210 23:08:38.075403 4791 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="defe4ef9-9037-4a53-a860-ce6713ee219c" containerName="dnsmasq-dns" Dec 10 23:08:38 crc kubenswrapper[4791]: I1210 23:08:38.075410 4791 state_mem.go:107] "Deleted CPUSet assignment" podUID="defe4ef9-9037-4a53-a860-ce6713ee219c" containerName="dnsmasq-dns" Dec 10 23:08:38 crc kubenswrapper[4791]: E1210 23:08:38.075428 4791 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="defe4ef9-9037-4a53-a860-ce6713ee219c" containerName="init" Dec 10 23:08:38 crc kubenswrapper[4791]: I1210 23:08:38.075433 4791 state_mem.go:107] "Deleted CPUSet assignment" podUID="defe4ef9-9037-4a53-a860-ce6713ee219c" containerName="init" Dec 10 23:08:38 crc kubenswrapper[4791]: I1210 23:08:38.075607 4791 memory_manager.go:354] "RemoveStaleState removing state" podUID="a2c44c96-70c2-414a-9890-cb95b66f10b2" containerName="glance-db-sync" Dec 10 23:08:38 crc kubenswrapper[4791]: I1210 23:08:38.075621 4791 memory_manager.go:354] "RemoveStaleState removing state" podUID="defe4ef9-9037-4a53-a860-ce6713ee219c" containerName="dnsmasq-dns" Dec 10 23:08:38 crc kubenswrapper[4791]: I1210 23:08:38.081896 4791 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-8b5c85b87-d7p92" Dec 10 23:08:38 crc kubenswrapper[4791]: I1210 23:08:38.096799 4791 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-8b5c85b87-d7p92"] Dec 10 23:08:38 crc kubenswrapper[4791]: I1210 23:08:38.168845 4791 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/7b395104-85b9-4ab1-8349-00fb9cb52b64-ovsdbserver-sb\") pod \"dnsmasq-dns-8b5c85b87-d7p92\" (UID: \"7b395104-85b9-4ab1-8349-00fb9cb52b64\") " pod="openstack/dnsmasq-dns-8b5c85b87-d7p92" Dec 10 23:08:38 crc kubenswrapper[4791]: I1210 23:08:38.174575 4791 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/7b395104-85b9-4ab1-8349-00fb9cb52b64-dns-svc\") pod \"dnsmasq-dns-8b5c85b87-d7p92\" (UID: \"7b395104-85b9-4ab1-8349-00fb9cb52b64\") " pod="openstack/dnsmasq-dns-8b5c85b87-d7p92" Dec 10 23:08:38 crc kubenswrapper[4791]: I1210 23:08:38.174684 4791 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-4k249\" (UniqueName: \"kubernetes.io/projected/7b395104-85b9-4ab1-8349-00fb9cb52b64-kube-api-access-4k249\") pod \"dnsmasq-dns-8b5c85b87-d7p92\" (UID: \"7b395104-85b9-4ab1-8349-00fb9cb52b64\") " pod="openstack/dnsmasq-dns-8b5c85b87-d7p92" Dec 10 23:08:38 crc kubenswrapper[4791]: I1210 23:08:38.174792 4791 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/7b395104-85b9-4ab1-8349-00fb9cb52b64-dns-swift-storage-0\") pod \"dnsmasq-dns-8b5c85b87-d7p92\" (UID: \"7b395104-85b9-4ab1-8349-00fb9cb52b64\") " pod="openstack/dnsmasq-dns-8b5c85b87-d7p92" Dec 10 23:08:38 crc kubenswrapper[4791]: I1210 23:08:38.174851 4791 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/7b395104-85b9-4ab1-8349-00fb9cb52b64-ovsdbserver-nb\") pod \"dnsmasq-dns-8b5c85b87-d7p92\" (UID: \"7b395104-85b9-4ab1-8349-00fb9cb52b64\") " pod="openstack/dnsmasq-dns-8b5c85b87-d7p92" Dec 10 23:08:38 crc kubenswrapper[4791]: I1210 23:08:38.174900 4791 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/7b395104-85b9-4ab1-8349-00fb9cb52b64-config\") pod \"dnsmasq-dns-8b5c85b87-d7p92\" (UID: \"7b395104-85b9-4ab1-8349-00fb9cb52b64\") " pod="openstack/dnsmasq-dns-8b5c85b87-d7p92" Dec 10 23:08:38 crc kubenswrapper[4791]: I1210 23:08:38.276611 4791 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/7b395104-85b9-4ab1-8349-00fb9cb52b64-config\") pod \"dnsmasq-dns-8b5c85b87-d7p92\" (UID: \"7b395104-85b9-4ab1-8349-00fb9cb52b64\") " pod="openstack/dnsmasq-dns-8b5c85b87-d7p92" Dec 10 23:08:38 crc kubenswrapper[4791]: I1210 23:08:38.276721 4791 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/7b395104-85b9-4ab1-8349-00fb9cb52b64-ovsdbserver-sb\") pod \"dnsmasq-dns-8b5c85b87-d7p92\" (UID: \"7b395104-85b9-4ab1-8349-00fb9cb52b64\") " pod="openstack/dnsmasq-dns-8b5c85b87-d7p92" Dec 10 23:08:38 crc kubenswrapper[4791]: I1210 23:08:38.276790 4791 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/7b395104-85b9-4ab1-8349-00fb9cb52b64-dns-svc\") pod \"dnsmasq-dns-8b5c85b87-d7p92\" (UID: \"7b395104-85b9-4ab1-8349-00fb9cb52b64\") " pod="openstack/dnsmasq-dns-8b5c85b87-d7p92" Dec 10 23:08:38 crc kubenswrapper[4791]: I1210 23:08:38.276819 4791 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-4k249\" (UniqueName: \"kubernetes.io/projected/7b395104-85b9-4ab1-8349-00fb9cb52b64-kube-api-access-4k249\") pod \"dnsmasq-dns-8b5c85b87-d7p92\" (UID: \"7b395104-85b9-4ab1-8349-00fb9cb52b64\") " pod="openstack/dnsmasq-dns-8b5c85b87-d7p92" Dec 10 23:08:38 crc kubenswrapper[4791]: I1210 23:08:38.276845 4791 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/7b395104-85b9-4ab1-8349-00fb9cb52b64-dns-swift-storage-0\") pod \"dnsmasq-dns-8b5c85b87-d7p92\" (UID: \"7b395104-85b9-4ab1-8349-00fb9cb52b64\") " pod="openstack/dnsmasq-dns-8b5c85b87-d7p92" Dec 10 23:08:38 crc kubenswrapper[4791]: I1210 23:08:38.276865 4791 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/7b395104-85b9-4ab1-8349-00fb9cb52b64-ovsdbserver-nb\") pod \"dnsmasq-dns-8b5c85b87-d7p92\" (UID: \"7b395104-85b9-4ab1-8349-00fb9cb52b64\") " pod="openstack/dnsmasq-dns-8b5c85b87-d7p92" Dec 10 23:08:38 crc kubenswrapper[4791]: I1210 23:08:38.277774 4791 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/7b395104-85b9-4ab1-8349-00fb9cb52b64-config\") pod \"dnsmasq-dns-8b5c85b87-d7p92\" (UID: \"7b395104-85b9-4ab1-8349-00fb9cb52b64\") " pod="openstack/dnsmasq-dns-8b5c85b87-d7p92" Dec 10 23:08:38 crc kubenswrapper[4791]: I1210 23:08:38.277789 4791 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/7b395104-85b9-4ab1-8349-00fb9cb52b64-ovsdbserver-sb\") pod \"dnsmasq-dns-8b5c85b87-d7p92\" (UID: \"7b395104-85b9-4ab1-8349-00fb9cb52b64\") " pod="openstack/dnsmasq-dns-8b5c85b87-d7p92" Dec 10 23:08:38 crc kubenswrapper[4791]: I1210 23:08:38.278521 4791 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/7b395104-85b9-4ab1-8349-00fb9cb52b64-dns-swift-storage-0\") pod \"dnsmasq-dns-8b5c85b87-d7p92\" (UID: \"7b395104-85b9-4ab1-8349-00fb9cb52b64\") " pod="openstack/dnsmasq-dns-8b5c85b87-d7p92" Dec 10 23:08:38 crc kubenswrapper[4791]: I1210 23:08:38.278598 4791 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/7b395104-85b9-4ab1-8349-00fb9cb52b64-ovsdbserver-nb\") pod \"dnsmasq-dns-8b5c85b87-d7p92\" (UID: \"7b395104-85b9-4ab1-8349-00fb9cb52b64\") " pod="openstack/dnsmasq-dns-8b5c85b87-d7p92" Dec 10 23:08:38 crc kubenswrapper[4791]: I1210 23:08:38.278892 4791 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/7b395104-85b9-4ab1-8349-00fb9cb52b64-dns-svc\") pod \"dnsmasq-dns-8b5c85b87-d7p92\" (UID: \"7b395104-85b9-4ab1-8349-00fb9cb52b64\") " pod="openstack/dnsmasq-dns-8b5c85b87-d7p92" Dec 10 23:08:38 crc kubenswrapper[4791]: I1210 23:08:38.300765 4791 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-4k249\" (UniqueName: \"kubernetes.io/projected/7b395104-85b9-4ab1-8349-00fb9cb52b64-kube-api-access-4k249\") pod \"dnsmasq-dns-8b5c85b87-d7p92\" (UID: \"7b395104-85b9-4ab1-8349-00fb9cb52b64\") " pod="openstack/dnsmasq-dns-8b5c85b87-d7p92" Dec 10 23:08:38 crc kubenswrapper[4791]: I1210 23:08:38.420172 4791 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-8b5c85b87-d7p92" Dec 10 23:08:39 crc kubenswrapper[4791]: I1210 23:08:39.030290 4791 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/glance-default-external-api-0"] Dec 10 23:08:39 crc kubenswrapper[4791]: I1210 23:08:39.033375 4791 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-external-api-0" Dec 10 23:08:39 crc kubenswrapper[4791]: I1210 23:08:39.035965 4791 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"glance-glance-dockercfg-rsl4w" Dec 10 23:08:39 crc kubenswrapper[4791]: I1210 23:08:39.036092 4791 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"glance-scripts" Dec 10 23:08:39 crc kubenswrapper[4791]: I1210 23:08:39.036156 4791 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"glance-default-external-config-data" Dec 10 23:08:39 crc kubenswrapper[4791]: I1210 23:08:39.053854 4791 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-default-external-api-0"] Dec 10 23:08:39 crc kubenswrapper[4791]: I1210 23:08:39.099014 4791 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/glance-default-internal-api-0"] Dec 10 23:08:39 crc kubenswrapper[4791]: I1210 23:08:39.100533 4791 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-internal-api-0" Dec 10 23:08:39 crc kubenswrapper[4791]: I1210 23:08:39.102795 4791 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"glance-default-internal-config-data" Dec 10 23:08:39 crc kubenswrapper[4791]: I1210 23:08:39.118072 4791 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-default-internal-api-0"] Dec 10 23:08:39 crc kubenswrapper[4791]: I1210 23:08:39.194244 4791 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/51437220-85e7-4c11-a567-d1bb94871e81-combined-ca-bundle\") pod \"glance-default-external-api-0\" (UID: \"51437220-85e7-4c11-a567-d1bb94871e81\") " pod="openstack/glance-default-external-api-0" Dec 10 23:08:39 crc kubenswrapper[4791]: I1210 23:08:39.194286 4791 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-hdz5s\" (UniqueName: \"kubernetes.io/projected/51437220-85e7-4c11-a567-d1bb94871e81-kube-api-access-hdz5s\") pod \"glance-default-external-api-0\" (UID: \"51437220-85e7-4c11-a567-d1bb94871e81\") " pod="openstack/glance-default-external-api-0" Dec 10 23:08:39 crc kubenswrapper[4791]: I1210 23:08:39.194356 4791 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/abd5935b-9f5b-4f25-9eec-eeedcb95bc16-combined-ca-bundle\") pod \"glance-default-internal-api-0\" (UID: \"abd5935b-9f5b-4f25-9eec-eeedcb95bc16\") " pod="openstack/glance-default-internal-api-0" Dec 10 23:08:39 crc kubenswrapper[4791]: I1210 23:08:39.194431 4791 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-knxlc\" (UniqueName: \"kubernetes.io/projected/abd5935b-9f5b-4f25-9eec-eeedcb95bc16-kube-api-access-knxlc\") pod \"glance-default-internal-api-0\" (UID: \"abd5935b-9f5b-4f25-9eec-eeedcb95bc16\") " pod="openstack/glance-default-internal-api-0" Dec 10 23:08:39 crc kubenswrapper[4791]: I1210 23:08:39.194529 4791 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/51437220-85e7-4c11-a567-d1bb94871e81-config-data\") pod \"glance-default-external-api-0\" (UID: \"51437220-85e7-4c11-a567-d1bb94871e81\") " pod="openstack/glance-default-external-api-0" Dec 10 23:08:39 crc kubenswrapper[4791]: I1210 23:08:39.194566 4791 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/51437220-85e7-4c11-a567-d1bb94871e81-httpd-run\") pod \"glance-default-external-api-0\" (UID: \"51437220-85e7-4c11-a567-d1bb94871e81\") " pod="openstack/glance-default-external-api-0" Dec 10 23:08:39 crc kubenswrapper[4791]: I1210 23:08:39.194605 4791 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/abd5935b-9f5b-4f25-9eec-eeedcb95bc16-config-data\") pod \"glance-default-internal-api-0\" (UID: \"abd5935b-9f5b-4f25-9eec-eeedcb95bc16\") " pod="openstack/glance-default-internal-api-0" Dec 10 23:08:39 crc kubenswrapper[4791]: I1210 23:08:39.194670 4791 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/51437220-85e7-4c11-a567-d1bb94871e81-logs\") pod \"glance-default-external-api-0\" (UID: \"51437220-85e7-4c11-a567-d1bb94871e81\") " pod="openstack/glance-default-external-api-0" Dec 10 23:08:39 crc kubenswrapper[4791]: I1210 23:08:39.194765 4791 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/abd5935b-9f5b-4f25-9eec-eeedcb95bc16-logs\") pod \"glance-default-internal-api-0\" (UID: \"abd5935b-9f5b-4f25-9eec-eeedcb95bc16\") " pod="openstack/glance-default-internal-api-0" Dec 10 23:08:39 crc kubenswrapper[4791]: I1210 23:08:39.194805 4791 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/abd5935b-9f5b-4f25-9eec-eeedcb95bc16-httpd-run\") pod \"glance-default-internal-api-0\" (UID: \"abd5935b-9f5b-4f25-9eec-eeedcb95bc16\") " pod="openstack/glance-default-internal-api-0" Dec 10 23:08:39 crc kubenswrapper[4791]: I1210 23:08:39.194846 4791 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/abd5935b-9f5b-4f25-9eec-eeedcb95bc16-scripts\") pod \"glance-default-internal-api-0\" (UID: \"abd5935b-9f5b-4f25-9eec-eeedcb95bc16\") " pod="openstack/glance-default-internal-api-0" Dec 10 23:08:39 crc kubenswrapper[4791]: I1210 23:08:39.194912 4791 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage06-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage06-crc\") pod \"glance-default-external-api-0\" (UID: \"51437220-85e7-4c11-a567-d1bb94871e81\") " pod="openstack/glance-default-external-api-0" Dec 10 23:08:39 crc kubenswrapper[4791]: I1210 23:08:39.194989 4791 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/51437220-85e7-4c11-a567-d1bb94871e81-scripts\") pod \"glance-default-external-api-0\" (UID: \"51437220-85e7-4c11-a567-d1bb94871e81\") " pod="openstack/glance-default-external-api-0" Dec 10 23:08:39 crc kubenswrapper[4791]: I1210 23:08:39.195009 4791 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage03-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage03-crc\") pod \"glance-default-internal-api-0\" (UID: \"abd5935b-9f5b-4f25-9eec-eeedcb95bc16\") " pod="openstack/glance-default-internal-api-0" Dec 10 23:08:39 crc kubenswrapper[4791]: I1210 23:08:39.296857 4791 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/abd5935b-9f5b-4f25-9eec-eeedcb95bc16-logs\") pod \"glance-default-internal-api-0\" (UID: \"abd5935b-9f5b-4f25-9eec-eeedcb95bc16\") " pod="openstack/glance-default-internal-api-0" Dec 10 23:08:39 crc kubenswrapper[4791]: I1210 23:08:39.296924 4791 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/abd5935b-9f5b-4f25-9eec-eeedcb95bc16-httpd-run\") pod \"glance-default-internal-api-0\" (UID: \"abd5935b-9f5b-4f25-9eec-eeedcb95bc16\") " pod="openstack/glance-default-internal-api-0" Dec 10 23:08:39 crc kubenswrapper[4791]: I1210 23:08:39.296963 4791 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/abd5935b-9f5b-4f25-9eec-eeedcb95bc16-scripts\") pod \"glance-default-internal-api-0\" (UID: \"abd5935b-9f5b-4f25-9eec-eeedcb95bc16\") " pod="openstack/glance-default-internal-api-0" Dec 10 23:08:39 crc kubenswrapper[4791]: I1210 23:08:39.296991 4791 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage06-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage06-crc\") pod \"glance-default-external-api-0\" (UID: \"51437220-85e7-4c11-a567-d1bb94871e81\") " pod="openstack/glance-default-external-api-0" Dec 10 23:08:39 crc kubenswrapper[4791]: I1210 23:08:39.297030 4791 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/51437220-85e7-4c11-a567-d1bb94871e81-scripts\") pod \"glance-default-external-api-0\" (UID: \"51437220-85e7-4c11-a567-d1bb94871e81\") " pod="openstack/glance-default-external-api-0" Dec 10 23:08:39 crc kubenswrapper[4791]: I1210 23:08:39.297053 4791 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage03-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage03-crc\") pod \"glance-default-internal-api-0\" (UID: \"abd5935b-9f5b-4f25-9eec-eeedcb95bc16\") " pod="openstack/glance-default-internal-api-0" Dec 10 23:08:39 crc kubenswrapper[4791]: I1210 23:08:39.297098 4791 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/51437220-85e7-4c11-a567-d1bb94871e81-combined-ca-bundle\") pod \"glance-default-external-api-0\" (UID: \"51437220-85e7-4c11-a567-d1bb94871e81\") " pod="openstack/glance-default-external-api-0" Dec 10 23:08:39 crc kubenswrapper[4791]: I1210 23:08:39.297121 4791 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-hdz5s\" (UniqueName: \"kubernetes.io/projected/51437220-85e7-4c11-a567-d1bb94871e81-kube-api-access-hdz5s\") pod \"glance-default-external-api-0\" (UID: \"51437220-85e7-4c11-a567-d1bb94871e81\") " pod="openstack/glance-default-external-api-0" Dec 10 23:08:39 crc kubenswrapper[4791]: I1210 23:08:39.297177 4791 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/abd5935b-9f5b-4f25-9eec-eeedcb95bc16-combined-ca-bundle\") pod \"glance-default-internal-api-0\" (UID: \"abd5935b-9f5b-4f25-9eec-eeedcb95bc16\") " pod="openstack/glance-default-internal-api-0" Dec 10 23:08:39 crc kubenswrapper[4791]: I1210 23:08:39.297200 4791 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-knxlc\" (UniqueName: \"kubernetes.io/projected/abd5935b-9f5b-4f25-9eec-eeedcb95bc16-kube-api-access-knxlc\") pod \"glance-default-internal-api-0\" (UID: \"abd5935b-9f5b-4f25-9eec-eeedcb95bc16\") " pod="openstack/glance-default-internal-api-0" Dec 10 23:08:39 crc kubenswrapper[4791]: I1210 23:08:39.297227 4791 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/51437220-85e7-4c11-a567-d1bb94871e81-config-data\") pod \"glance-default-external-api-0\" (UID: \"51437220-85e7-4c11-a567-d1bb94871e81\") " pod="openstack/glance-default-external-api-0" Dec 10 23:08:39 crc kubenswrapper[4791]: I1210 23:08:39.297251 4791 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/51437220-85e7-4c11-a567-d1bb94871e81-httpd-run\") pod \"glance-default-external-api-0\" (UID: \"51437220-85e7-4c11-a567-d1bb94871e81\") " pod="openstack/glance-default-external-api-0" Dec 10 23:08:39 crc kubenswrapper[4791]: I1210 23:08:39.297273 4791 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/abd5935b-9f5b-4f25-9eec-eeedcb95bc16-config-data\") pod \"glance-default-internal-api-0\" (UID: \"abd5935b-9f5b-4f25-9eec-eeedcb95bc16\") " pod="openstack/glance-default-internal-api-0" Dec 10 23:08:39 crc kubenswrapper[4791]: I1210 23:08:39.297306 4791 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/51437220-85e7-4c11-a567-d1bb94871e81-logs\") pod \"glance-default-external-api-0\" (UID: \"51437220-85e7-4c11-a567-d1bb94871e81\") " pod="openstack/glance-default-external-api-0" Dec 10 23:08:39 crc kubenswrapper[4791]: I1210 23:08:39.297793 4791 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/51437220-85e7-4c11-a567-d1bb94871e81-logs\") pod \"glance-default-external-api-0\" (UID: \"51437220-85e7-4c11-a567-d1bb94871e81\") " pod="openstack/glance-default-external-api-0" Dec 10 23:08:39 crc kubenswrapper[4791]: I1210 23:08:39.298132 4791 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/abd5935b-9f5b-4f25-9eec-eeedcb95bc16-logs\") pod \"glance-default-internal-api-0\" (UID: \"abd5935b-9f5b-4f25-9eec-eeedcb95bc16\") " pod="openstack/glance-default-internal-api-0" Dec 10 23:08:39 crc kubenswrapper[4791]: I1210 23:08:39.298411 4791 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/abd5935b-9f5b-4f25-9eec-eeedcb95bc16-httpd-run\") pod \"glance-default-internal-api-0\" (UID: \"abd5935b-9f5b-4f25-9eec-eeedcb95bc16\") " pod="openstack/glance-default-internal-api-0" Dec 10 23:08:39 crc kubenswrapper[4791]: I1210 23:08:39.299370 4791 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/51437220-85e7-4c11-a567-d1bb94871e81-httpd-run\") pod \"glance-default-external-api-0\" (UID: \"51437220-85e7-4c11-a567-d1bb94871e81\") " pod="openstack/glance-default-external-api-0" Dec 10 23:08:39 crc kubenswrapper[4791]: I1210 23:08:39.299556 4791 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage06-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage06-crc\") pod \"glance-default-external-api-0\" (UID: \"51437220-85e7-4c11-a567-d1bb94871e81\") device mount path \"/mnt/openstack/pv06\"" pod="openstack/glance-default-external-api-0" Dec 10 23:08:39 crc kubenswrapper[4791]: I1210 23:08:39.299658 4791 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage03-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage03-crc\") pod \"glance-default-internal-api-0\" (UID: \"abd5935b-9f5b-4f25-9eec-eeedcb95bc16\") device mount path \"/mnt/openstack/pv03\"" pod="openstack/glance-default-internal-api-0" Dec 10 23:08:39 crc kubenswrapper[4791]: I1210 23:08:39.303165 4791 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/abd5935b-9f5b-4f25-9eec-eeedcb95bc16-scripts\") pod \"glance-default-internal-api-0\" (UID: \"abd5935b-9f5b-4f25-9eec-eeedcb95bc16\") " pod="openstack/glance-default-internal-api-0" Dec 10 23:08:39 crc kubenswrapper[4791]: I1210 23:08:39.304615 4791 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/51437220-85e7-4c11-a567-d1bb94871e81-config-data\") pod \"glance-default-external-api-0\" (UID: \"51437220-85e7-4c11-a567-d1bb94871e81\") " pod="openstack/glance-default-external-api-0" Dec 10 23:08:39 crc kubenswrapper[4791]: I1210 23:08:39.304731 4791 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/51437220-85e7-4c11-a567-d1bb94871e81-combined-ca-bundle\") pod \"glance-default-external-api-0\" (UID: \"51437220-85e7-4c11-a567-d1bb94871e81\") " pod="openstack/glance-default-external-api-0" Dec 10 23:08:39 crc kubenswrapper[4791]: I1210 23:08:39.309855 4791 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/abd5935b-9f5b-4f25-9eec-eeedcb95bc16-config-data\") pod \"glance-default-internal-api-0\" (UID: \"abd5935b-9f5b-4f25-9eec-eeedcb95bc16\") " pod="openstack/glance-default-internal-api-0" Dec 10 23:08:39 crc kubenswrapper[4791]: I1210 23:08:39.313006 4791 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/51437220-85e7-4c11-a567-d1bb94871e81-scripts\") pod \"glance-default-external-api-0\" (UID: \"51437220-85e7-4c11-a567-d1bb94871e81\") " pod="openstack/glance-default-external-api-0" Dec 10 23:08:39 crc kubenswrapper[4791]: I1210 23:08:39.321998 4791 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-knxlc\" (UniqueName: \"kubernetes.io/projected/abd5935b-9f5b-4f25-9eec-eeedcb95bc16-kube-api-access-knxlc\") pod \"glance-default-internal-api-0\" (UID: \"abd5935b-9f5b-4f25-9eec-eeedcb95bc16\") " pod="openstack/glance-default-internal-api-0" Dec 10 23:08:39 crc kubenswrapper[4791]: I1210 23:08:39.333369 4791 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/abd5935b-9f5b-4f25-9eec-eeedcb95bc16-combined-ca-bundle\") pod \"glance-default-internal-api-0\" (UID: \"abd5935b-9f5b-4f25-9eec-eeedcb95bc16\") " pod="openstack/glance-default-internal-api-0" Dec 10 23:08:39 crc kubenswrapper[4791]: I1210 23:08:39.346143 4791 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-hdz5s\" (UniqueName: \"kubernetes.io/projected/51437220-85e7-4c11-a567-d1bb94871e81-kube-api-access-hdz5s\") pod \"glance-default-external-api-0\" (UID: \"51437220-85e7-4c11-a567-d1bb94871e81\") " pod="openstack/glance-default-external-api-0" Dec 10 23:08:39 crc kubenswrapper[4791]: I1210 23:08:39.367025 4791 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage03-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage03-crc\") pod \"glance-default-internal-api-0\" (UID: \"abd5935b-9f5b-4f25-9eec-eeedcb95bc16\") " pod="openstack/glance-default-internal-api-0" Dec 10 23:08:39 crc kubenswrapper[4791]: I1210 23:08:39.377973 4791 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage06-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage06-crc\") pod \"glance-default-external-api-0\" (UID: \"51437220-85e7-4c11-a567-d1bb94871e81\") " pod="openstack/glance-default-external-api-0" Dec 10 23:08:39 crc kubenswrapper[4791]: I1210 23:08:39.426354 4791 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-internal-api-0" Dec 10 23:08:39 crc kubenswrapper[4791]: I1210 23:08:39.677248 4791 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-external-api-0" Dec 10 23:08:40 crc kubenswrapper[4791]: I1210 23:08:40.089837 4791 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/dnsmasq-dns-77585f5f8c-drnrs" podUID="defe4ef9-9037-4a53-a860-ce6713ee219c" containerName="dnsmasq-dns" probeResult="failure" output="dial tcp 10.217.0.125:5353: i/o timeout" Dec 10 23:08:40 crc kubenswrapper[4791]: I1210 23:08:40.734718 4791 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/glance-default-external-api-0"] Dec 10 23:08:40 crc kubenswrapper[4791]: I1210 23:08:40.808825 4791 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/glance-default-internal-api-0"] Dec 10 23:08:47 crc kubenswrapper[4791]: E1210 23:08:47.839742 4791 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="quay.io/podified-antelope-centos9/openstack-cinder-api:current-podified" Dec 10 23:08:47 crc kubenswrapper[4791]: E1210 23:08:47.840556 4791 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:cinder-db-sync,Image:quay.io/podified-antelope-centos9/openstack-cinder-api:current-podified,Command:[/bin/bash],Args:[-c /usr/local/bin/kolla_set_configs && /usr/local/bin/kolla_start],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:KOLLA_BOOTSTRAP,Value:TRUE,ValueFrom:nil,},EnvVar{Name:KOLLA_CONFIG_STRATEGY,Value:COPY_ALWAYS,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:etc-machine-id,ReadOnly:true,MountPath:/etc/machine-id,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:scripts,ReadOnly:true,MountPath:/usr/local/bin/container-scripts,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:config-data,ReadOnly:true,MountPath:/var/lib/config-data/merged,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:config-data,ReadOnly:true,MountPath:/etc/my.cnf,SubPath:my.cnf,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:db-sync-config-data,ReadOnly:true,MountPath:/etc/cinder/cinder.conf.d,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:config-data,ReadOnly:true,MountPath:/var/lib/kolla/config_files/config.json,SubPath:db-sync-config.json,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:combined-ca-bundle,ReadOnly:true,MountPath:/etc/pki/ca-trust/extracted/pem/tls-ca-bundle.pem,SubPath:tls-ca-bundle.pem,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-5fw8x,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:nil,Privileged:nil,SELinuxOptions:nil,RunAsUser:*0,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:nil,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod cinder-db-sync-g4s5p_openstack(a85df2ed-be84-41b3-8993-c3d963daef08): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError" Dec 10 23:08:47 crc kubenswrapper[4791]: E1210 23:08:47.841722 4791 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"cinder-db-sync\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\"" pod="openstack/cinder-db-sync-g4s5p" podUID="a85df2ed-be84-41b3-8993-c3d963daef08" Dec 10 23:08:48 crc kubenswrapper[4791]: E1210 23:08:48.248881 4791 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"cinder-db-sync\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/podified-antelope-centos9/openstack-cinder-api:current-podified\\\"\"" pod="openstack/cinder-db-sync-g4s5p" podUID="a85df2ed-be84-41b3-8993-c3d963daef08" Dec 10 23:08:49 crc kubenswrapper[4791]: I1210 23:08:49.256391 4791 generic.go:334] "Generic (PLEG): container finished" podID="6076a930-1583-41b7-b5ae-bc6c61086e48" containerID="b409fe2d3a34ee8ca490698ce05841ae136b5e63e5b04879ee4a783bcca8809e" exitCode=0 Dec 10 23:08:49 crc kubenswrapper[4791]: I1210 23:08:49.256535 4791 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-db-sync-kg4vv" event={"ID":"6076a930-1583-41b7-b5ae-bc6c61086e48","Type":"ContainerDied","Data":"b409fe2d3a34ee8ca490698ce05841ae136b5e63e5b04879ee4a783bcca8809e"} Dec 10 23:08:49 crc kubenswrapper[4791]: I1210 23:08:49.833675 4791 scope.go:117] "RemoveContainer" containerID="b677977e766e127dbf26f7ec4252e1db8771f1088e3cfb922d4d6f5e770a4a8c" Dec 10 23:08:50 crc kubenswrapper[4791]: E1210 23:08:50.315169 4791 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="quay.io/podified-antelope-centos9/openstack-barbican-api:current-podified" Dec 10 23:08:50 crc kubenswrapper[4791]: E1210 23:08:50.315333 4791 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:barbican-db-sync,Image:quay.io/podified-antelope-centos9/openstack-barbican-api:current-podified,Command:[/bin/bash],Args:[-c barbican-manage db upgrade],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:KOLLA_BOOTSTRAP,Value:TRUE,ValueFrom:nil,},EnvVar{Name:KOLLA_CONFIG_STRATEGY,Value:COPY_ALWAYS,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:db-sync-config-data,ReadOnly:true,MountPath:/etc/barbican/barbican.conf.d,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:combined-ca-bundle,ReadOnly:true,MountPath:/etc/pki/ca-trust/extracted/pem/tls-ca-bundle.pem,SubPath:tls-ca-bundle.pem,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-crjrd,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*42403,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:nil,RunAsGroup:*42403,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod barbican-db-sync-zllvm_openstack(851556d7-fa38-4eec-ae74-de989d66d27b): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError" Dec 10 23:08:50 crc kubenswrapper[4791]: E1210 23:08:50.316813 4791 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"barbican-db-sync\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\"" pod="openstack/barbican-db-sync-zllvm" podUID="851556d7-fa38-4eec-ae74-de989d66d27b" Dec 10 23:08:50 crc kubenswrapper[4791]: I1210 23:08:50.642623 4791 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-db-sync-kg4vv" Dec 10 23:08:50 crc kubenswrapper[4791]: I1210 23:08:50.766592 4791 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-66x4z\" (UniqueName: \"kubernetes.io/projected/6076a930-1583-41b7-b5ae-bc6c61086e48-kube-api-access-66x4z\") pod \"6076a930-1583-41b7-b5ae-bc6c61086e48\" (UID: \"6076a930-1583-41b7-b5ae-bc6c61086e48\") " Dec 10 23:08:50 crc kubenswrapper[4791]: I1210 23:08:50.767095 4791 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/6076a930-1583-41b7-b5ae-bc6c61086e48-combined-ca-bundle\") pod \"6076a930-1583-41b7-b5ae-bc6c61086e48\" (UID: \"6076a930-1583-41b7-b5ae-bc6c61086e48\") " Dec 10 23:08:50 crc kubenswrapper[4791]: I1210 23:08:50.767318 4791 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/secret/6076a930-1583-41b7-b5ae-bc6c61086e48-config\") pod \"6076a930-1583-41b7-b5ae-bc6c61086e48\" (UID: \"6076a930-1583-41b7-b5ae-bc6c61086e48\") " Dec 10 23:08:50 crc kubenswrapper[4791]: I1210 23:08:50.770257 4791 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/6076a930-1583-41b7-b5ae-bc6c61086e48-kube-api-access-66x4z" (OuterVolumeSpecName: "kube-api-access-66x4z") pod "6076a930-1583-41b7-b5ae-bc6c61086e48" (UID: "6076a930-1583-41b7-b5ae-bc6c61086e48"). InnerVolumeSpecName "kube-api-access-66x4z". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 10 23:08:50 crc kubenswrapper[4791]: I1210 23:08:50.806466 4791 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/6076a930-1583-41b7-b5ae-bc6c61086e48-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "6076a930-1583-41b7-b5ae-bc6c61086e48" (UID: "6076a930-1583-41b7-b5ae-bc6c61086e48"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 10 23:08:50 crc kubenswrapper[4791]: I1210 23:08:50.821676 4791 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/6076a930-1583-41b7-b5ae-bc6c61086e48-config" (OuterVolumeSpecName: "config") pod "6076a930-1583-41b7-b5ae-bc6c61086e48" (UID: "6076a930-1583-41b7-b5ae-bc6c61086e48"). InnerVolumeSpecName "config". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 10 23:08:50 crc kubenswrapper[4791]: I1210 23:08:50.840974 4791 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/horizon-6999c5bf5b-7cr9t"] Dec 10 23:08:50 crc kubenswrapper[4791]: I1210 23:08:50.850566 4791 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/horizon-7cc66567bd-pwb2l"] Dec 10 23:08:50 crc kubenswrapper[4791]: W1210 23:08:50.855360 4791 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod55139c7c_e9fa_4539_b7b2_b2be7ee43b44.slice/crio-d2883b26349a02a4c60df2f542814b6fa9be4acda940054df2fa1a9938da6439 WatchSource:0}: Error finding container d2883b26349a02a4c60df2f542814b6fa9be4acda940054df2fa1a9938da6439: Status 404 returned error can't find the container with id d2883b26349a02a4c60df2f542814b6fa9be4acda940054df2fa1a9938da6439 Dec 10 23:08:50 crc kubenswrapper[4791]: I1210 23:08:50.869796 4791 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/6076a930-1583-41b7-b5ae-bc6c61086e48-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 10 23:08:50 crc kubenswrapper[4791]: I1210 23:08:50.869829 4791 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/secret/6076a930-1583-41b7-b5ae-bc6c61086e48-config\") on node \"crc\" DevicePath \"\"" Dec 10 23:08:50 crc kubenswrapper[4791]: I1210 23:08:50.869839 4791 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-66x4z\" (UniqueName: \"kubernetes.io/projected/6076a930-1583-41b7-b5ae-bc6c61086e48-kube-api-access-66x4z\") on node \"crc\" DevicePath \"\"" Dec 10 23:08:51 crc kubenswrapper[4791]: I1210 23:08:51.052560 4791 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/keystone-bootstrap-mwqkg"] Dec 10 23:08:51 crc kubenswrapper[4791]: I1210 23:08:51.161271 4791 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/glance-default-internal-api-0"] Dec 10 23:08:51 crc kubenswrapper[4791]: I1210 23:08:51.189525 4791 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-8b5c85b87-d7p92"] Dec 10 23:08:51 crc kubenswrapper[4791]: W1210 23:08:51.198025 4791 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod7b395104_85b9_4ab1_8349_00fb9cb52b64.slice/crio-37f5ca22ddd27fe1dd921f99f9afe6eed6231af68c44250bd66acf7a89b3b193 WatchSource:0}: Error finding container 37f5ca22ddd27fe1dd921f99f9afe6eed6231af68c44250bd66acf7a89b3b193: Status 404 returned error can't find the container with id 37f5ca22ddd27fe1dd921f99f9afe6eed6231af68c44250bd66acf7a89b3b193 Dec 10 23:08:51 crc kubenswrapper[4791]: I1210 23:08:51.284847 4791 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"8eaa1b1d-715f-4059-ad18-fe9632c036a7","Type":"ContainerStarted","Data":"5ae0fb7182370bec008f1c24c5d245fd6cb390d075a883adda305056ffab03b5"} Dec 10 23:08:51 crc kubenswrapper[4791]: I1210 23:08:51.287440 4791 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/horizon-7cc66567bd-pwb2l" event={"ID":"55139c7c-e9fa-4539-b7b2-b2be7ee43b44","Type":"ContainerStarted","Data":"7b1ada7642cab048e60fe308d453e086b82053b893d130082d56f6aaf472608a"} Dec 10 23:08:51 crc kubenswrapper[4791]: I1210 23:08:51.287487 4791 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/horizon-7cc66567bd-pwb2l" event={"ID":"55139c7c-e9fa-4539-b7b2-b2be7ee43b44","Type":"ContainerStarted","Data":"d2883b26349a02a4c60df2f542814b6fa9be4acda940054df2fa1a9938da6439"} Dec 10 23:08:51 crc kubenswrapper[4791]: I1210 23:08:51.289418 4791 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-db-sync-kg4vv" event={"ID":"6076a930-1583-41b7-b5ae-bc6c61086e48","Type":"ContainerDied","Data":"8d4d2124883e0dca88bed5ae049d5f29c1818ec35294b04b7dd019fa938fea46"} Dec 10 23:08:51 crc kubenswrapper[4791]: I1210 23:08:51.289456 4791 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="8d4d2124883e0dca88bed5ae049d5f29c1818ec35294b04b7dd019fa938fea46" Dec 10 23:08:51 crc kubenswrapper[4791]: I1210 23:08:51.289541 4791 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-db-sync-kg4vv" Dec 10 23:08:51 crc kubenswrapper[4791]: I1210 23:08:51.289996 4791 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/glance-default-external-api-0"] Dec 10 23:08:51 crc kubenswrapper[4791]: I1210 23:08:51.296458 4791 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/horizon-78fc5d56dc-58cjt" event={"ID":"14f46cbc-afe2-40bb-a2a3-5673b5ac53e7","Type":"ContainerStarted","Data":"1bf24a4c4c7be94d64b9919fd150ec64d3f744be3bf1b6f17541d5c0b7a27a30"} Dec 10 23:08:51 crc kubenswrapper[4791]: I1210 23:08:51.296510 4791 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/horizon-78fc5d56dc-58cjt" event={"ID":"14f46cbc-afe2-40bb-a2a3-5673b5ac53e7","Type":"ContainerStarted","Data":"d5bed1e64f4f7d6b0cbb4910d9037ebbea49eb4ac927509fe80353feed0ed97d"} Dec 10 23:08:51 crc kubenswrapper[4791]: I1210 23:08:51.296646 4791 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/horizon-78fc5d56dc-58cjt" podUID="14f46cbc-afe2-40bb-a2a3-5673b5ac53e7" containerName="horizon-log" containerID="cri-o://d5bed1e64f4f7d6b0cbb4910d9037ebbea49eb4ac927509fe80353feed0ed97d" gracePeriod=30 Dec 10 23:08:51 crc kubenswrapper[4791]: I1210 23:08:51.298252 4791 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/horizon-78fc5d56dc-58cjt" podUID="14f46cbc-afe2-40bb-a2a3-5673b5ac53e7" containerName="horizon" containerID="cri-o://1bf24a4c4c7be94d64b9919fd150ec64d3f744be3bf1b6f17541d5c0b7a27a30" gracePeriod=30 Dec 10 23:08:51 crc kubenswrapper[4791]: I1210 23:08:51.341630 4791 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-bootstrap-mwqkg" event={"ID":"7c65a05d-c2cc-45c4-b469-e6a271e01856","Type":"ContainerStarted","Data":"3e22738048c400928a86f51ef7f5d400be8b23990d8caa8f9edd0625131b8147"} Dec 10 23:08:51 crc kubenswrapper[4791]: I1210 23:08:51.348072 4791 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/horizon-5cc98b55bf-7xpnj" podUID="9e9c1419-c6f3-4947-8788-4915f48677d7" containerName="horizon-log" containerID="cri-o://9369387266f0bf17a3578a2973371e39479e41aeaf91821d2bd23efe9fbbe09e" gracePeriod=30 Dec 10 23:08:51 crc kubenswrapper[4791]: I1210 23:08:51.348319 4791 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/horizon-5cc98b55bf-7xpnj" event={"ID":"9e9c1419-c6f3-4947-8788-4915f48677d7","Type":"ContainerStarted","Data":"e8a0fecaed859f1a8a83f5c36eabd8013be451668e6363d31a5ebdbc8e9a0863"} Dec 10 23:08:51 crc kubenswrapper[4791]: I1210 23:08:51.348369 4791 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/horizon-5cc98b55bf-7xpnj" event={"ID":"9e9c1419-c6f3-4947-8788-4915f48677d7","Type":"ContainerStarted","Data":"9369387266f0bf17a3578a2973371e39479e41aeaf91821d2bd23efe9fbbe09e"} Dec 10 23:08:51 crc kubenswrapper[4791]: I1210 23:08:51.348404 4791 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/horizon-5cc98b55bf-7xpnj" podUID="9e9c1419-c6f3-4947-8788-4915f48677d7" containerName="horizon" containerID="cri-o://e8a0fecaed859f1a8a83f5c36eabd8013be451668e6363d31a5ebdbc8e9a0863" gracePeriod=30 Dec 10 23:08:51 crc kubenswrapper[4791]: I1210 23:08:51.373275 4791 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/horizon-6999c5bf5b-7cr9t" event={"ID":"0fab3b41-ed57-4536-ade0-f5f6ccde40e2","Type":"ContainerStarted","Data":"222024b7a1e9af424dc501d8c3cc8325bfeaccce56c6c04425fd958735119e44"} Dec 10 23:08:51 crc kubenswrapper[4791]: I1210 23:08:51.373326 4791 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/horizon-6999c5bf5b-7cr9t" event={"ID":"0fab3b41-ed57-4536-ade0-f5f6ccde40e2","Type":"ContainerStarted","Data":"590a8d39a453289055a426f8a1236fdfcb2f842798e8cd98c595ab6ff4e2e5c5"} Dec 10 23:08:51 crc kubenswrapper[4791]: I1210 23:08:51.374643 4791 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/horizon-78fc5d56dc-58cjt" podStartSLOduration=2.35934816 podStartE2EDuration="31.374624869s" podCreationTimestamp="2025-12-10 23:08:20 +0000 UTC" firstStartedPulling="2025-12-10 23:08:21.309626649 +0000 UTC m=+1135.739244262" lastFinishedPulling="2025-12-10 23:08:50.324903348 +0000 UTC m=+1164.754520971" observedRunningTime="2025-12-10 23:08:51.356729881 +0000 UTC m=+1165.786347514" watchObservedRunningTime="2025-12-10 23:08:51.374624869 +0000 UTC m=+1165.804242482" Dec 10 23:08:51 crc kubenswrapper[4791]: I1210 23:08:51.389443 4791 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/horizon-5f54ccbf57-wwpfk" event={"ID":"aa3a63ab-6092-46ce-aaf6-40360aa060ac","Type":"ContainerStarted","Data":"51262b7b3d283cb5b5b3f2953648269ed55c1d6a470bbdf743fc7e6a314d9407"} Dec 10 23:08:51 crc kubenswrapper[4791]: I1210 23:08:51.395353 4791 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/horizon-5f54ccbf57-wwpfk" podUID="aa3a63ab-6092-46ce-aaf6-40360aa060ac" containerName="horizon-log" containerID="cri-o://35844ea7a50bd278dafc5819fb34072df86afbe70e63e900d65a4f0f56c5cfc8" gracePeriod=30 Dec 10 23:08:51 crc kubenswrapper[4791]: I1210 23:08:51.402308 4791 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/horizon-5f54ccbf57-wwpfk" event={"ID":"aa3a63ab-6092-46ce-aaf6-40360aa060ac","Type":"ContainerStarted","Data":"35844ea7a50bd278dafc5819fb34072df86afbe70e63e900d65a4f0f56c5cfc8"} Dec 10 23:08:51 crc kubenswrapper[4791]: I1210 23:08:51.395440 4791 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/horizon-5f54ccbf57-wwpfk" podUID="aa3a63ab-6092-46ce-aaf6-40360aa060ac" containerName="horizon" containerID="cri-o://51262b7b3d283cb5b5b3f2953648269ed55c1d6a470bbdf743fc7e6a314d9407" gracePeriod=30 Dec 10 23:08:51 crc kubenswrapper[4791]: I1210 23:08:51.407185 4791 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/horizon-5cc98b55bf-7xpnj" podStartSLOduration=2.520057856 podStartE2EDuration="29.40716226s" podCreationTimestamp="2025-12-10 23:08:22 +0000 UTC" firstStartedPulling="2025-12-10 23:08:23.415445977 +0000 UTC m=+1137.845063590" lastFinishedPulling="2025-12-10 23:08:50.302550381 +0000 UTC m=+1164.732167994" observedRunningTime="2025-12-10 23:08:51.394813913 +0000 UTC m=+1165.824431536" watchObservedRunningTime="2025-12-10 23:08:51.40716226 +0000 UTC m=+1165.836779873" Dec 10 23:08:51 crc kubenswrapper[4791]: I1210 23:08:51.412421 4791 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-internal-api-0" event={"ID":"abd5935b-9f5b-4f25-9eec-eeedcb95bc16","Type":"ContainerStarted","Data":"4b04d4980fab42d0d10f3d85865d46ed6e9222985e5ebd06945fc2460e8de169"} Dec 10 23:08:51 crc kubenswrapper[4791]: I1210 23:08:51.429766 4791 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-8b5c85b87-d7p92" event={"ID":"7b395104-85b9-4ab1-8349-00fb9cb52b64","Type":"ContainerStarted","Data":"37f5ca22ddd27fe1dd921f99f9afe6eed6231af68c44250bd66acf7a89b3b193"} Dec 10 23:08:51 crc kubenswrapper[4791]: E1210 23:08:51.435844 4791 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"barbican-db-sync\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/podified-antelope-centos9/openstack-barbican-api:current-podified\\\"\"" pod="openstack/barbican-db-sync-zllvm" podUID="851556d7-fa38-4eec-ae74-de989d66d27b" Dec 10 23:08:51 crc kubenswrapper[4791]: I1210 23:08:51.474861 4791 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/horizon-5f54ccbf57-wwpfk" podStartSLOduration=5.831934947 podStartE2EDuration="31.474841309s" podCreationTimestamp="2025-12-10 23:08:20 +0000 UTC" firstStartedPulling="2025-12-10 23:08:22.145854492 +0000 UTC m=+1136.575472105" lastFinishedPulling="2025-12-10 23:08:47.788760854 +0000 UTC m=+1162.218378467" observedRunningTime="2025-12-10 23:08:51.450461844 +0000 UTC m=+1165.880079457" watchObservedRunningTime="2025-12-10 23:08:51.474841309 +0000 UTC m=+1165.904458922" Dec 10 23:08:51 crc kubenswrapper[4791]: I1210 23:08:51.522963 4791 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-8b5c85b87-d7p92"] Dec 10 23:08:51 crc kubenswrapper[4791]: I1210 23:08:51.552496 4791 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/neutron-85794c6bd6-ht8sz"] Dec 10 23:08:51 crc kubenswrapper[4791]: E1210 23:08:51.553181 4791 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="6076a930-1583-41b7-b5ae-bc6c61086e48" containerName="neutron-db-sync" Dec 10 23:08:51 crc kubenswrapper[4791]: I1210 23:08:51.553198 4791 state_mem.go:107] "Deleted CPUSet assignment" podUID="6076a930-1583-41b7-b5ae-bc6c61086e48" containerName="neutron-db-sync" Dec 10 23:08:51 crc kubenswrapper[4791]: I1210 23:08:51.553396 4791 memory_manager.go:354] "RemoveStaleState removing state" podUID="6076a930-1583-41b7-b5ae-bc6c61086e48" containerName="neutron-db-sync" Dec 10 23:08:51 crc kubenswrapper[4791]: I1210 23:08:51.556910 4791 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-85794c6bd6-ht8sz" Dec 10 23:08:51 crc kubenswrapper[4791]: I1210 23:08:51.563416 4791 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-neutron-ovndbs" Dec 10 23:08:51 crc kubenswrapper[4791]: I1210 23:08:51.563690 4791 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"neutron-httpd-config" Dec 10 23:08:51 crc kubenswrapper[4791]: I1210 23:08:51.563793 4791 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"neutron-config" Dec 10 23:08:51 crc kubenswrapper[4791]: I1210 23:08:51.563890 4791 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"neutron-neutron-dockercfg-fblml" Dec 10 23:08:51 crc kubenswrapper[4791]: I1210 23:08:51.568415 4791 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-84b966f6c9-tz266"] Dec 10 23:08:51 crc kubenswrapper[4791]: I1210 23:08:51.571086 4791 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-84b966f6c9-tz266" Dec 10 23:08:51 crc kubenswrapper[4791]: I1210 23:08:51.609673 4791 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/neutron-85794c6bd6-ht8sz"] Dec 10 23:08:51 crc kubenswrapper[4791]: I1210 23:08:51.626482 4791 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-84b966f6c9-tz266"] Dec 10 23:08:51 crc kubenswrapper[4791]: I1210 23:08:51.697375 4791 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/secret/75eb971a-04fc-40f5-9890-d5a0b4daccd1-config\") pod \"neutron-85794c6bd6-ht8sz\" (UID: \"75eb971a-04fc-40f5-9890-d5a0b4daccd1\") " pod="openstack/neutron-85794c6bd6-ht8sz" Dec 10 23:08:51 crc kubenswrapper[4791]: I1210 23:08:51.697417 4791 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/3a4fcb67-0b78-43d6-a6a1-8d319bf19184-config\") pod \"dnsmasq-dns-84b966f6c9-tz266\" (UID: \"3a4fcb67-0b78-43d6-a6a1-8d319bf19184\") " pod="openstack/dnsmasq-dns-84b966f6c9-tz266" Dec 10 23:08:51 crc kubenswrapper[4791]: I1210 23:08:51.697468 4791 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/3a4fcb67-0b78-43d6-a6a1-8d319bf19184-ovsdbserver-nb\") pod \"dnsmasq-dns-84b966f6c9-tz266\" (UID: \"3a4fcb67-0b78-43d6-a6a1-8d319bf19184\") " pod="openstack/dnsmasq-dns-84b966f6c9-tz266" Dec 10 23:08:51 crc kubenswrapper[4791]: I1210 23:08:51.697758 4791 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/3a4fcb67-0b78-43d6-a6a1-8d319bf19184-dns-swift-storage-0\") pod \"dnsmasq-dns-84b966f6c9-tz266\" (UID: \"3a4fcb67-0b78-43d6-a6a1-8d319bf19184\") " pod="openstack/dnsmasq-dns-84b966f6c9-tz266" Dec 10 23:08:51 crc kubenswrapper[4791]: I1210 23:08:51.697797 4791 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/75eb971a-04fc-40f5-9890-d5a0b4daccd1-combined-ca-bundle\") pod \"neutron-85794c6bd6-ht8sz\" (UID: \"75eb971a-04fc-40f5-9890-d5a0b4daccd1\") " pod="openstack/neutron-85794c6bd6-ht8sz" Dec 10 23:08:51 crc kubenswrapper[4791]: I1210 23:08:51.697817 4791 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-9cgk8\" (UniqueName: \"kubernetes.io/projected/75eb971a-04fc-40f5-9890-d5a0b4daccd1-kube-api-access-9cgk8\") pod \"neutron-85794c6bd6-ht8sz\" (UID: \"75eb971a-04fc-40f5-9890-d5a0b4daccd1\") " pod="openstack/neutron-85794c6bd6-ht8sz" Dec 10 23:08:51 crc kubenswrapper[4791]: I1210 23:08:51.697834 4791 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/3a4fcb67-0b78-43d6-a6a1-8d319bf19184-dns-svc\") pod \"dnsmasq-dns-84b966f6c9-tz266\" (UID: \"3a4fcb67-0b78-43d6-a6a1-8d319bf19184\") " pod="openstack/dnsmasq-dns-84b966f6c9-tz266" Dec 10 23:08:51 crc kubenswrapper[4791]: I1210 23:08:51.697859 4791 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/3a4fcb67-0b78-43d6-a6a1-8d319bf19184-ovsdbserver-sb\") pod \"dnsmasq-dns-84b966f6c9-tz266\" (UID: \"3a4fcb67-0b78-43d6-a6a1-8d319bf19184\") " pod="openstack/dnsmasq-dns-84b966f6c9-tz266" Dec 10 23:08:51 crc kubenswrapper[4791]: I1210 23:08:51.697981 4791 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovndb-tls-certs\" (UniqueName: \"kubernetes.io/secret/75eb971a-04fc-40f5-9890-d5a0b4daccd1-ovndb-tls-certs\") pod \"neutron-85794c6bd6-ht8sz\" (UID: \"75eb971a-04fc-40f5-9890-d5a0b4daccd1\") " pod="openstack/neutron-85794c6bd6-ht8sz" Dec 10 23:08:51 crc kubenswrapper[4791]: I1210 23:08:51.698013 4791 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-pg94c\" (UniqueName: \"kubernetes.io/projected/3a4fcb67-0b78-43d6-a6a1-8d319bf19184-kube-api-access-pg94c\") pod \"dnsmasq-dns-84b966f6c9-tz266\" (UID: \"3a4fcb67-0b78-43d6-a6a1-8d319bf19184\") " pod="openstack/dnsmasq-dns-84b966f6c9-tz266" Dec 10 23:08:51 crc kubenswrapper[4791]: I1210 23:08:51.698031 4791 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"httpd-config\" (UniqueName: \"kubernetes.io/secret/75eb971a-04fc-40f5-9890-d5a0b4daccd1-httpd-config\") pod \"neutron-85794c6bd6-ht8sz\" (UID: \"75eb971a-04fc-40f5-9890-d5a0b4daccd1\") " pod="openstack/neutron-85794c6bd6-ht8sz" Dec 10 23:08:51 crc kubenswrapper[4791]: I1210 23:08:51.799089 4791 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/secret/75eb971a-04fc-40f5-9890-d5a0b4daccd1-config\") pod \"neutron-85794c6bd6-ht8sz\" (UID: \"75eb971a-04fc-40f5-9890-d5a0b4daccd1\") " pod="openstack/neutron-85794c6bd6-ht8sz" Dec 10 23:08:51 crc kubenswrapper[4791]: I1210 23:08:51.799134 4791 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/3a4fcb67-0b78-43d6-a6a1-8d319bf19184-config\") pod \"dnsmasq-dns-84b966f6c9-tz266\" (UID: \"3a4fcb67-0b78-43d6-a6a1-8d319bf19184\") " pod="openstack/dnsmasq-dns-84b966f6c9-tz266" Dec 10 23:08:51 crc kubenswrapper[4791]: I1210 23:08:51.799185 4791 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/3a4fcb67-0b78-43d6-a6a1-8d319bf19184-ovsdbserver-nb\") pod \"dnsmasq-dns-84b966f6c9-tz266\" (UID: \"3a4fcb67-0b78-43d6-a6a1-8d319bf19184\") " pod="openstack/dnsmasq-dns-84b966f6c9-tz266" Dec 10 23:08:51 crc kubenswrapper[4791]: I1210 23:08:51.799211 4791 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/3a4fcb67-0b78-43d6-a6a1-8d319bf19184-dns-swift-storage-0\") pod \"dnsmasq-dns-84b966f6c9-tz266\" (UID: \"3a4fcb67-0b78-43d6-a6a1-8d319bf19184\") " pod="openstack/dnsmasq-dns-84b966f6c9-tz266" Dec 10 23:08:51 crc kubenswrapper[4791]: I1210 23:08:51.799231 4791 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/75eb971a-04fc-40f5-9890-d5a0b4daccd1-combined-ca-bundle\") pod \"neutron-85794c6bd6-ht8sz\" (UID: \"75eb971a-04fc-40f5-9890-d5a0b4daccd1\") " pod="openstack/neutron-85794c6bd6-ht8sz" Dec 10 23:08:51 crc kubenswrapper[4791]: I1210 23:08:51.799249 4791 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-9cgk8\" (UniqueName: \"kubernetes.io/projected/75eb971a-04fc-40f5-9890-d5a0b4daccd1-kube-api-access-9cgk8\") pod \"neutron-85794c6bd6-ht8sz\" (UID: \"75eb971a-04fc-40f5-9890-d5a0b4daccd1\") " pod="openstack/neutron-85794c6bd6-ht8sz" Dec 10 23:08:51 crc kubenswrapper[4791]: I1210 23:08:51.799266 4791 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/3a4fcb67-0b78-43d6-a6a1-8d319bf19184-dns-svc\") pod \"dnsmasq-dns-84b966f6c9-tz266\" (UID: \"3a4fcb67-0b78-43d6-a6a1-8d319bf19184\") " pod="openstack/dnsmasq-dns-84b966f6c9-tz266" Dec 10 23:08:51 crc kubenswrapper[4791]: I1210 23:08:51.799289 4791 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/3a4fcb67-0b78-43d6-a6a1-8d319bf19184-ovsdbserver-sb\") pod \"dnsmasq-dns-84b966f6c9-tz266\" (UID: \"3a4fcb67-0b78-43d6-a6a1-8d319bf19184\") " pod="openstack/dnsmasq-dns-84b966f6c9-tz266" Dec 10 23:08:51 crc kubenswrapper[4791]: I1210 23:08:51.799383 4791 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovndb-tls-certs\" (UniqueName: \"kubernetes.io/secret/75eb971a-04fc-40f5-9890-d5a0b4daccd1-ovndb-tls-certs\") pod \"neutron-85794c6bd6-ht8sz\" (UID: \"75eb971a-04fc-40f5-9890-d5a0b4daccd1\") " pod="openstack/neutron-85794c6bd6-ht8sz" Dec 10 23:08:51 crc kubenswrapper[4791]: I1210 23:08:51.799412 4791 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-pg94c\" (UniqueName: \"kubernetes.io/projected/3a4fcb67-0b78-43d6-a6a1-8d319bf19184-kube-api-access-pg94c\") pod \"dnsmasq-dns-84b966f6c9-tz266\" (UID: \"3a4fcb67-0b78-43d6-a6a1-8d319bf19184\") " pod="openstack/dnsmasq-dns-84b966f6c9-tz266" Dec 10 23:08:51 crc kubenswrapper[4791]: I1210 23:08:51.799428 4791 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"httpd-config\" (UniqueName: \"kubernetes.io/secret/75eb971a-04fc-40f5-9890-d5a0b4daccd1-httpd-config\") pod \"neutron-85794c6bd6-ht8sz\" (UID: \"75eb971a-04fc-40f5-9890-d5a0b4daccd1\") " pod="openstack/neutron-85794c6bd6-ht8sz" Dec 10 23:08:51 crc kubenswrapper[4791]: I1210 23:08:51.801294 4791 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/3a4fcb67-0b78-43d6-a6a1-8d319bf19184-ovsdbserver-sb\") pod \"dnsmasq-dns-84b966f6c9-tz266\" (UID: \"3a4fcb67-0b78-43d6-a6a1-8d319bf19184\") " pod="openstack/dnsmasq-dns-84b966f6c9-tz266" Dec 10 23:08:51 crc kubenswrapper[4791]: I1210 23:08:51.802304 4791 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/3a4fcb67-0b78-43d6-a6a1-8d319bf19184-dns-svc\") pod \"dnsmasq-dns-84b966f6c9-tz266\" (UID: \"3a4fcb67-0b78-43d6-a6a1-8d319bf19184\") " pod="openstack/dnsmasq-dns-84b966f6c9-tz266" Dec 10 23:08:51 crc kubenswrapper[4791]: I1210 23:08:51.802715 4791 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/3a4fcb67-0b78-43d6-a6a1-8d319bf19184-config\") pod \"dnsmasq-dns-84b966f6c9-tz266\" (UID: \"3a4fcb67-0b78-43d6-a6a1-8d319bf19184\") " pod="openstack/dnsmasq-dns-84b966f6c9-tz266" Dec 10 23:08:51 crc kubenswrapper[4791]: I1210 23:08:51.803792 4791 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/3a4fcb67-0b78-43d6-a6a1-8d319bf19184-dns-swift-storage-0\") pod \"dnsmasq-dns-84b966f6c9-tz266\" (UID: \"3a4fcb67-0b78-43d6-a6a1-8d319bf19184\") " pod="openstack/dnsmasq-dns-84b966f6c9-tz266" Dec 10 23:08:51 crc kubenswrapper[4791]: I1210 23:08:51.815592 4791 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/3a4fcb67-0b78-43d6-a6a1-8d319bf19184-ovsdbserver-nb\") pod \"dnsmasq-dns-84b966f6c9-tz266\" (UID: \"3a4fcb67-0b78-43d6-a6a1-8d319bf19184\") " pod="openstack/dnsmasq-dns-84b966f6c9-tz266" Dec 10 23:08:51 crc kubenswrapper[4791]: I1210 23:08:51.819274 4791 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"httpd-config\" (UniqueName: \"kubernetes.io/secret/75eb971a-04fc-40f5-9890-d5a0b4daccd1-httpd-config\") pod \"neutron-85794c6bd6-ht8sz\" (UID: \"75eb971a-04fc-40f5-9890-d5a0b4daccd1\") " pod="openstack/neutron-85794c6bd6-ht8sz" Dec 10 23:08:51 crc kubenswrapper[4791]: I1210 23:08:51.831737 4791 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/secret/75eb971a-04fc-40f5-9890-d5a0b4daccd1-config\") pod \"neutron-85794c6bd6-ht8sz\" (UID: \"75eb971a-04fc-40f5-9890-d5a0b4daccd1\") " pod="openstack/neutron-85794c6bd6-ht8sz" Dec 10 23:08:51 crc kubenswrapper[4791]: I1210 23:08:51.854430 4791 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/75eb971a-04fc-40f5-9890-d5a0b4daccd1-combined-ca-bundle\") pod \"neutron-85794c6bd6-ht8sz\" (UID: \"75eb971a-04fc-40f5-9890-d5a0b4daccd1\") " pod="openstack/neutron-85794c6bd6-ht8sz" Dec 10 23:08:51 crc kubenswrapper[4791]: I1210 23:08:51.855158 4791 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovndb-tls-certs\" (UniqueName: \"kubernetes.io/secret/75eb971a-04fc-40f5-9890-d5a0b4daccd1-ovndb-tls-certs\") pod \"neutron-85794c6bd6-ht8sz\" (UID: \"75eb971a-04fc-40f5-9890-d5a0b4daccd1\") " pod="openstack/neutron-85794c6bd6-ht8sz" Dec 10 23:08:51 crc kubenswrapper[4791]: I1210 23:08:51.855910 4791 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-9cgk8\" (UniqueName: \"kubernetes.io/projected/75eb971a-04fc-40f5-9890-d5a0b4daccd1-kube-api-access-9cgk8\") pod \"neutron-85794c6bd6-ht8sz\" (UID: \"75eb971a-04fc-40f5-9890-d5a0b4daccd1\") " pod="openstack/neutron-85794c6bd6-ht8sz" Dec 10 23:08:51 crc kubenswrapper[4791]: I1210 23:08:51.868498 4791 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-pg94c\" (UniqueName: \"kubernetes.io/projected/3a4fcb67-0b78-43d6-a6a1-8d319bf19184-kube-api-access-pg94c\") pod \"dnsmasq-dns-84b966f6c9-tz266\" (UID: \"3a4fcb67-0b78-43d6-a6a1-8d319bf19184\") " pod="openstack/dnsmasq-dns-84b966f6c9-tz266" Dec 10 23:08:51 crc kubenswrapper[4791]: I1210 23:08:51.888626 4791 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-85794c6bd6-ht8sz" Dec 10 23:08:51 crc kubenswrapper[4791]: I1210 23:08:51.909773 4791 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-84b966f6c9-tz266" Dec 10 23:08:52 crc kubenswrapper[4791]: I1210 23:08:52.446926 4791 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"51437220-85e7-4c11-a567-d1bb94871e81","Type":"ContainerStarted","Data":"b8ce0e374219a112d8cb5058ba1f707e84b6d757ed98ce0a13df0f32a5294422"} Dec 10 23:08:52 crc kubenswrapper[4791]: I1210 23:08:52.457628 4791 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-84b966f6c9-tz266"] Dec 10 23:08:52 crc kubenswrapper[4791]: I1210 23:08:52.468677 4791 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/horizon-6999c5bf5b-7cr9t" event={"ID":"0fab3b41-ed57-4536-ade0-f5f6ccde40e2","Type":"ContainerStarted","Data":"54e6b83fa4ce4b9c4c31ce3aa15fc9d74a39ed65397050b4667c1bcda2c6ac0e"} Dec 10 23:08:52 crc kubenswrapper[4791]: I1210 23:08:52.482752 4791 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/horizon-7cc66567bd-pwb2l" event={"ID":"55139c7c-e9fa-4539-b7b2-b2be7ee43b44","Type":"ContainerStarted","Data":"a441e9dd03ea293d73ed83e0ce82cd1f391a27dfc4d4c0f86b23bba57c39c2c5"} Dec 10 23:08:52 crc kubenswrapper[4791]: E1210 23:08:52.482883 4791 cadvisor_stats_provider.go:516] "Partial failure issuing cadvisor.ContainerInfoV2" err="partial failures: [\"/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod7b395104_85b9_4ab1_8349_00fb9cb52b64.slice/crio-de48999e170ebe660dbbd7bf91bdbab79f2d9bdbefc589037cbc53ec4d79bd79.scope\": RecentStats: unable to find data in memory cache], [\"/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod7b395104_85b9_4ab1_8349_00fb9cb52b64.slice/crio-conmon-de48999e170ebe660dbbd7bf91bdbab79f2d9bdbefc589037cbc53ec4d79bd79.scope\": RecentStats: unable to find data in memory cache]" Dec 10 23:08:52 crc kubenswrapper[4791]: I1210 23:08:52.497509 4791 generic.go:334] "Generic (PLEG): container finished" podID="7b395104-85b9-4ab1-8349-00fb9cb52b64" containerID="de48999e170ebe660dbbd7bf91bdbab79f2d9bdbefc589037cbc53ec4d79bd79" exitCode=0 Dec 10 23:08:52 crc kubenswrapper[4791]: I1210 23:08:52.497594 4791 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-8b5c85b87-d7p92" event={"ID":"7b395104-85b9-4ab1-8349-00fb9cb52b64","Type":"ContainerDied","Data":"de48999e170ebe660dbbd7bf91bdbab79f2d9bdbefc589037cbc53ec4d79bd79"} Dec 10 23:08:52 crc kubenswrapper[4791]: I1210 23:08:52.500487 4791 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/horizon-6999c5bf5b-7cr9t" podStartSLOduration=23.500468603 podStartE2EDuration="23.500468603s" podCreationTimestamp="2025-12-10 23:08:29 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-10 23:08:52.498010412 +0000 UTC m=+1166.927628025" watchObservedRunningTime="2025-12-10 23:08:52.500468603 +0000 UTC m=+1166.930086216" Dec 10 23:08:52 crc kubenswrapper[4791]: I1210 23:08:52.505938 4791 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-bootstrap-mwqkg" event={"ID":"7c65a05d-c2cc-45c4-b469-e6a271e01856","Type":"ContainerStarted","Data":"f0eb0592a4a0cdde8425fc8222755e9b651cc41049d8057c83c99f2ae97e73a5"} Dec 10 23:08:52 crc kubenswrapper[4791]: I1210 23:08:52.553203 4791 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/horizon-7cc66567bd-pwb2l" podStartSLOduration=23.553184654 podStartE2EDuration="23.553184654s" podCreationTimestamp="2025-12-10 23:08:29 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-10 23:08:52.523801596 +0000 UTC m=+1166.953419199" watchObservedRunningTime="2025-12-10 23:08:52.553184654 +0000 UTC m=+1166.982802267" Dec 10 23:08:52 crc kubenswrapper[4791]: I1210 23:08:52.607900 4791 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/keystone-bootstrap-mwqkg" podStartSLOduration=17.607884363 podStartE2EDuration="17.607884363s" podCreationTimestamp="2025-12-10 23:08:35 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-10 23:08:52.564406108 +0000 UTC m=+1166.994023721" watchObservedRunningTime="2025-12-10 23:08:52.607884363 +0000 UTC m=+1167.037501976" Dec 10 23:08:52 crc kubenswrapper[4791]: I1210 23:08:52.656733 4791 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/neutron-85794c6bd6-ht8sz"] Dec 10 23:08:52 crc kubenswrapper[4791]: W1210 23:08:52.662446 4791 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod75eb971a_04fc_40f5_9890_d5a0b4daccd1.slice/crio-dd2f20735a4b4cd7f75dfe9937d9cb51e2dc61038563f55ba640cae578951ee2 WatchSource:0}: Error finding container dd2f20735a4b4cd7f75dfe9937d9cb51e2dc61038563f55ba640cae578951ee2: Status 404 returned error can't find the container with id dd2f20735a4b4cd7f75dfe9937d9cb51e2dc61038563f55ba640cae578951ee2 Dec 10 23:08:52 crc kubenswrapper[4791]: I1210 23:08:52.837463 4791 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/horizon-5cc98b55bf-7xpnj" Dec 10 23:08:53 crc kubenswrapper[4791]: I1210 23:08:53.207691 4791 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-8b5c85b87-d7p92" Dec 10 23:08:53 crc kubenswrapper[4791]: I1210 23:08:53.349886 4791 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/7b395104-85b9-4ab1-8349-00fb9cb52b64-config\") pod \"7b395104-85b9-4ab1-8349-00fb9cb52b64\" (UID: \"7b395104-85b9-4ab1-8349-00fb9cb52b64\") " Dec 10 23:08:53 crc kubenswrapper[4791]: I1210 23:08:53.350196 4791 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/7b395104-85b9-4ab1-8349-00fb9cb52b64-dns-swift-storage-0\") pod \"7b395104-85b9-4ab1-8349-00fb9cb52b64\" (UID: \"7b395104-85b9-4ab1-8349-00fb9cb52b64\") " Dec 10 23:08:53 crc kubenswrapper[4791]: I1210 23:08:53.350295 4791 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/7b395104-85b9-4ab1-8349-00fb9cb52b64-dns-svc\") pod \"7b395104-85b9-4ab1-8349-00fb9cb52b64\" (UID: \"7b395104-85b9-4ab1-8349-00fb9cb52b64\") " Dec 10 23:08:53 crc kubenswrapper[4791]: I1210 23:08:53.350426 4791 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-4k249\" (UniqueName: \"kubernetes.io/projected/7b395104-85b9-4ab1-8349-00fb9cb52b64-kube-api-access-4k249\") pod \"7b395104-85b9-4ab1-8349-00fb9cb52b64\" (UID: \"7b395104-85b9-4ab1-8349-00fb9cb52b64\") " Dec 10 23:08:53 crc kubenswrapper[4791]: I1210 23:08:53.350463 4791 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/7b395104-85b9-4ab1-8349-00fb9cb52b64-ovsdbserver-sb\") pod \"7b395104-85b9-4ab1-8349-00fb9cb52b64\" (UID: \"7b395104-85b9-4ab1-8349-00fb9cb52b64\") " Dec 10 23:08:53 crc kubenswrapper[4791]: I1210 23:08:53.350517 4791 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/7b395104-85b9-4ab1-8349-00fb9cb52b64-ovsdbserver-nb\") pod \"7b395104-85b9-4ab1-8349-00fb9cb52b64\" (UID: \"7b395104-85b9-4ab1-8349-00fb9cb52b64\") " Dec 10 23:08:53 crc kubenswrapper[4791]: I1210 23:08:53.383773 4791 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/7b395104-85b9-4ab1-8349-00fb9cb52b64-config" (OuterVolumeSpecName: "config") pod "7b395104-85b9-4ab1-8349-00fb9cb52b64" (UID: "7b395104-85b9-4ab1-8349-00fb9cb52b64"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 10 23:08:53 crc kubenswrapper[4791]: I1210 23:08:53.383901 4791 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/7b395104-85b9-4ab1-8349-00fb9cb52b64-kube-api-access-4k249" (OuterVolumeSpecName: "kube-api-access-4k249") pod "7b395104-85b9-4ab1-8349-00fb9cb52b64" (UID: "7b395104-85b9-4ab1-8349-00fb9cb52b64"). InnerVolumeSpecName "kube-api-access-4k249". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 10 23:08:53 crc kubenswrapper[4791]: I1210 23:08:53.434521 4791 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/7b395104-85b9-4ab1-8349-00fb9cb52b64-ovsdbserver-nb" (OuterVolumeSpecName: "ovsdbserver-nb") pod "7b395104-85b9-4ab1-8349-00fb9cb52b64" (UID: "7b395104-85b9-4ab1-8349-00fb9cb52b64"). InnerVolumeSpecName "ovsdbserver-nb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 10 23:08:53 crc kubenswrapper[4791]: I1210 23:08:53.453280 4791 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-4k249\" (UniqueName: \"kubernetes.io/projected/7b395104-85b9-4ab1-8349-00fb9cb52b64-kube-api-access-4k249\") on node \"crc\" DevicePath \"\"" Dec 10 23:08:53 crc kubenswrapper[4791]: I1210 23:08:53.453333 4791 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/7b395104-85b9-4ab1-8349-00fb9cb52b64-ovsdbserver-nb\") on node \"crc\" DevicePath \"\"" Dec 10 23:08:53 crc kubenswrapper[4791]: I1210 23:08:53.453366 4791 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/7b395104-85b9-4ab1-8349-00fb9cb52b64-config\") on node \"crc\" DevicePath \"\"" Dec 10 23:08:53 crc kubenswrapper[4791]: I1210 23:08:53.525231 4791 generic.go:334] "Generic (PLEG): container finished" podID="3a4fcb67-0b78-43d6-a6a1-8d319bf19184" containerID="e3d0d1d6e6602ea537ea9b3c0b7772d80cba882bb829676e2912ef4ae01fb09f" exitCode=0 Dec 10 23:08:53 crc kubenswrapper[4791]: I1210 23:08:53.525711 4791 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-84b966f6c9-tz266" event={"ID":"3a4fcb67-0b78-43d6-a6a1-8d319bf19184","Type":"ContainerDied","Data":"e3d0d1d6e6602ea537ea9b3c0b7772d80cba882bb829676e2912ef4ae01fb09f"} Dec 10 23:08:53 crc kubenswrapper[4791]: I1210 23:08:53.525739 4791 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-84b966f6c9-tz266" event={"ID":"3a4fcb67-0b78-43d6-a6a1-8d319bf19184","Type":"ContainerStarted","Data":"47471e5879b32ac699bed8b99af0a4c42ce82327f2054dd5c2ccb69ce6de9494"} Dec 10 23:08:53 crc kubenswrapper[4791]: I1210 23:08:53.549003 4791 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"51437220-85e7-4c11-a567-d1bb94871e81","Type":"ContainerStarted","Data":"b3200bc2f8d202f18e9123084c35cf7170a1703fc8f14d712cf5afdc12502263"} Dec 10 23:08:53 crc kubenswrapper[4791]: I1210 23:08:53.559608 4791 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-85794c6bd6-ht8sz" event={"ID":"75eb971a-04fc-40f5-9890-d5a0b4daccd1","Type":"ContainerStarted","Data":"a1c20ac32bb42bda941e83eca0be34d9e4905b4b8eff8dab079f24798af60c6b"} Dec 10 23:08:53 crc kubenswrapper[4791]: I1210 23:08:53.559654 4791 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-85794c6bd6-ht8sz" event={"ID":"75eb971a-04fc-40f5-9890-d5a0b4daccd1","Type":"ContainerStarted","Data":"dd2fb7dddb229ce25296aefb965623011367c1640da7f68c5f11bd17f39afbdc"} Dec 10 23:08:53 crc kubenswrapper[4791]: I1210 23:08:53.559666 4791 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-85794c6bd6-ht8sz" event={"ID":"75eb971a-04fc-40f5-9890-d5a0b4daccd1","Type":"ContainerStarted","Data":"dd2f20735a4b4cd7f75dfe9937d9cb51e2dc61038563f55ba640cae578951ee2"} Dec 10 23:08:53 crc kubenswrapper[4791]: I1210 23:08:53.559708 4791 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/neutron-85794c6bd6-ht8sz" Dec 10 23:08:53 crc kubenswrapper[4791]: I1210 23:08:53.565596 4791 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-db-sync-svgbr" event={"ID":"b9efeadd-d54b-4f78-a0ad-85cc62321b84","Type":"ContainerStarted","Data":"c8c6f523ef5f1d59b753e569f79692fbe8d9bc4939c81ccec02f0720f1fb8447"} Dec 10 23:08:53 crc kubenswrapper[4791]: I1210 23:08:53.575094 4791 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/7b395104-85b9-4ab1-8349-00fb9cb52b64-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "7b395104-85b9-4ab1-8349-00fb9cb52b64" (UID: "7b395104-85b9-4ab1-8349-00fb9cb52b64"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 10 23:08:53 crc kubenswrapper[4791]: I1210 23:08:53.579556 4791 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-internal-api-0" event={"ID":"abd5935b-9f5b-4f25-9eec-eeedcb95bc16","Type":"ContainerStarted","Data":"1c77bc5e67a79a02182644e651dc347dfb3e1f1a0fac9d42fa8544f476591532"} Dec 10 23:08:53 crc kubenswrapper[4791]: I1210 23:08:53.633927 4791 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-8b5c85b87-d7p92" event={"ID":"7b395104-85b9-4ab1-8349-00fb9cb52b64","Type":"ContainerDied","Data":"37f5ca22ddd27fe1dd921f99f9afe6eed6231af68c44250bd66acf7a89b3b193"} Dec 10 23:08:53 crc kubenswrapper[4791]: I1210 23:08:53.634147 4791 scope.go:117] "RemoveContainer" containerID="de48999e170ebe660dbbd7bf91bdbab79f2d9bdbefc589037cbc53ec4d79bd79" Dec 10 23:08:53 crc kubenswrapper[4791]: I1210 23:08:53.634572 4791 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-8b5c85b87-d7p92" Dec 10 23:08:53 crc kubenswrapper[4791]: I1210 23:08:53.647469 4791 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/neutron-85794c6bd6-ht8sz" podStartSLOduration=2.647438983 podStartE2EDuration="2.647438983s" podCreationTimestamp="2025-12-10 23:08:51 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-10 23:08:53.589664235 +0000 UTC m=+1168.019281868" watchObservedRunningTime="2025-12-10 23:08:53.647438983 +0000 UTC m=+1168.077056596" Dec 10 23:08:53 crc kubenswrapper[4791]: I1210 23:08:53.648317 4791 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/7b395104-85b9-4ab1-8349-00fb9cb52b64-ovsdbserver-sb" (OuterVolumeSpecName: "ovsdbserver-sb") pod "7b395104-85b9-4ab1-8349-00fb9cb52b64" (UID: "7b395104-85b9-4ab1-8349-00fb9cb52b64"). InnerVolumeSpecName "ovsdbserver-sb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 10 23:08:53 crc kubenswrapper[4791]: I1210 23:08:53.658482 4791 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/placement-db-sync-svgbr" podStartSLOduration=2.950911123 podStartE2EDuration="33.658461551s" podCreationTimestamp="2025-12-10 23:08:20 +0000 UTC" firstStartedPulling="2025-12-10 23:08:22.030920355 +0000 UTC m=+1136.460537958" lastFinishedPulling="2025-12-10 23:08:52.738470773 +0000 UTC m=+1167.168088386" observedRunningTime="2025-12-10 23:08:53.648465412 +0000 UTC m=+1168.078083025" watchObservedRunningTime="2025-12-10 23:08:53.658461551 +0000 UTC m=+1168.088079174" Dec 10 23:08:53 crc kubenswrapper[4791]: I1210 23:08:53.666730 4791 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/7b395104-85b9-4ab1-8349-00fb9cb52b64-dns-swift-storage-0" (OuterVolumeSpecName: "dns-swift-storage-0") pod "7b395104-85b9-4ab1-8349-00fb9cb52b64" (UID: "7b395104-85b9-4ab1-8349-00fb9cb52b64"). InnerVolumeSpecName "dns-swift-storage-0". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 10 23:08:53 crc kubenswrapper[4791]: I1210 23:08:53.680474 4791 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/7b395104-85b9-4ab1-8349-00fb9cb52b64-ovsdbserver-sb\") on node \"crc\" DevicePath \"\"" Dec 10 23:08:53 crc kubenswrapper[4791]: I1210 23:08:53.680516 4791 reconciler_common.go:293] "Volume detached for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/7b395104-85b9-4ab1-8349-00fb9cb52b64-dns-swift-storage-0\") on node \"crc\" DevicePath \"\"" Dec 10 23:08:53 crc kubenswrapper[4791]: I1210 23:08:53.680526 4791 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/7b395104-85b9-4ab1-8349-00fb9cb52b64-dns-svc\") on node \"crc\" DevicePath \"\"" Dec 10 23:08:54 crc kubenswrapper[4791]: I1210 23:08:54.004858 4791 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-8b5c85b87-d7p92"] Dec 10 23:08:54 crc kubenswrapper[4791]: I1210 23:08:54.048631 4791 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-8b5c85b87-d7p92"] Dec 10 23:08:54 crc kubenswrapper[4791]: I1210 23:08:54.416024 4791 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/neutron-856d8bb659-x5kwg"] Dec 10 23:08:54 crc kubenswrapper[4791]: E1210 23:08:54.416750 4791 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="7b395104-85b9-4ab1-8349-00fb9cb52b64" containerName="init" Dec 10 23:08:54 crc kubenswrapper[4791]: I1210 23:08:54.416775 4791 state_mem.go:107] "Deleted CPUSet assignment" podUID="7b395104-85b9-4ab1-8349-00fb9cb52b64" containerName="init" Dec 10 23:08:54 crc kubenswrapper[4791]: I1210 23:08:54.417058 4791 memory_manager.go:354] "RemoveStaleState removing state" podUID="7b395104-85b9-4ab1-8349-00fb9cb52b64" containerName="init" Dec 10 23:08:54 crc kubenswrapper[4791]: I1210 23:08:54.418246 4791 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-856d8bb659-x5kwg" Dec 10 23:08:54 crc kubenswrapper[4791]: I1210 23:08:54.420646 4791 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-neutron-internal-svc" Dec 10 23:08:54 crc kubenswrapper[4791]: I1210 23:08:54.425063 4791 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-neutron-public-svc" Dec 10 23:08:54 crc kubenswrapper[4791]: I1210 23:08:54.451313 4791 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/neutron-856d8bb659-x5kwg"] Dec 10 23:08:54 crc kubenswrapper[4791]: I1210 23:08:54.608906 4791 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"httpd-config\" (UniqueName: \"kubernetes.io/secret/60652b01-72de-481d-b2ea-8c349e9aee52-httpd-config\") pod \"neutron-856d8bb659-x5kwg\" (UID: \"60652b01-72de-481d-b2ea-8c349e9aee52\") " pod="openstack/neutron-856d8bb659-x5kwg" Dec 10 23:08:54 crc kubenswrapper[4791]: I1210 23:08:54.608969 4791 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/60652b01-72de-481d-b2ea-8c349e9aee52-public-tls-certs\") pod \"neutron-856d8bb659-x5kwg\" (UID: \"60652b01-72de-481d-b2ea-8c349e9aee52\") " pod="openstack/neutron-856d8bb659-x5kwg" Dec 10 23:08:54 crc kubenswrapper[4791]: I1210 23:08:54.609029 4791 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/60652b01-72de-481d-b2ea-8c349e9aee52-internal-tls-certs\") pod \"neutron-856d8bb659-x5kwg\" (UID: \"60652b01-72de-481d-b2ea-8c349e9aee52\") " pod="openstack/neutron-856d8bb659-x5kwg" Dec 10 23:08:54 crc kubenswrapper[4791]: I1210 23:08:54.609196 4791 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/60652b01-72de-481d-b2ea-8c349e9aee52-combined-ca-bundle\") pod \"neutron-856d8bb659-x5kwg\" (UID: \"60652b01-72de-481d-b2ea-8c349e9aee52\") " pod="openstack/neutron-856d8bb659-x5kwg" Dec 10 23:08:54 crc kubenswrapper[4791]: I1210 23:08:54.609324 4791 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/secret/60652b01-72de-481d-b2ea-8c349e9aee52-config\") pod \"neutron-856d8bb659-x5kwg\" (UID: \"60652b01-72de-481d-b2ea-8c349e9aee52\") " pod="openstack/neutron-856d8bb659-x5kwg" Dec 10 23:08:54 crc kubenswrapper[4791]: I1210 23:08:54.609563 4791 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-2srmj\" (UniqueName: \"kubernetes.io/projected/60652b01-72de-481d-b2ea-8c349e9aee52-kube-api-access-2srmj\") pod \"neutron-856d8bb659-x5kwg\" (UID: \"60652b01-72de-481d-b2ea-8c349e9aee52\") " pod="openstack/neutron-856d8bb659-x5kwg" Dec 10 23:08:54 crc kubenswrapper[4791]: I1210 23:08:54.609610 4791 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovndb-tls-certs\" (UniqueName: \"kubernetes.io/secret/60652b01-72de-481d-b2ea-8c349e9aee52-ovndb-tls-certs\") pod \"neutron-856d8bb659-x5kwg\" (UID: \"60652b01-72de-481d-b2ea-8c349e9aee52\") " pod="openstack/neutron-856d8bb659-x5kwg" Dec 10 23:08:54 crc kubenswrapper[4791]: I1210 23:08:54.643679 4791 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-internal-api-0" event={"ID":"abd5935b-9f5b-4f25-9eec-eeedcb95bc16","Type":"ContainerStarted","Data":"c41db6dc374c530ec370e8b68d2cf357805394432183a75efdc95665439b1c5a"} Dec 10 23:08:54 crc kubenswrapper[4791]: I1210 23:08:54.643826 4791 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/glance-default-internal-api-0" podUID="abd5935b-9f5b-4f25-9eec-eeedcb95bc16" containerName="glance-log" containerID="cri-o://1c77bc5e67a79a02182644e651dc347dfb3e1f1a0fac9d42fa8544f476591532" gracePeriod=30 Dec 10 23:08:54 crc kubenswrapper[4791]: I1210 23:08:54.644268 4791 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/glance-default-internal-api-0" podUID="abd5935b-9f5b-4f25-9eec-eeedcb95bc16" containerName="glance-httpd" containerID="cri-o://c41db6dc374c530ec370e8b68d2cf357805394432183a75efdc95665439b1c5a" gracePeriod=30 Dec 10 23:08:54 crc kubenswrapper[4791]: I1210 23:08:54.679702 4791 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-84b966f6c9-tz266" event={"ID":"3a4fcb67-0b78-43d6-a6a1-8d319bf19184","Type":"ContainerStarted","Data":"4294aa9ba3a7a9440f41cfc9d38f97e9bcc1fa80423767e0df5c6d4016638bf9"} Dec 10 23:08:54 crc kubenswrapper[4791]: I1210 23:08:54.680568 4791 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/dnsmasq-dns-84b966f6c9-tz266" Dec 10 23:08:54 crc kubenswrapper[4791]: I1210 23:08:54.689839 4791 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/glance-default-internal-api-0" podStartSLOduration=16.689818524 podStartE2EDuration="16.689818524s" podCreationTimestamp="2025-12-10 23:08:38 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-10 23:08:54.685217401 +0000 UTC m=+1169.114835014" watchObservedRunningTime="2025-12-10 23:08:54.689818524 +0000 UTC m=+1169.119436137" Dec 10 23:08:54 crc kubenswrapper[4791]: I1210 23:08:54.694123 4791 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/glance-default-external-api-0" podUID="51437220-85e7-4c11-a567-d1bb94871e81" containerName="glance-log" containerID="cri-o://b3200bc2f8d202f18e9123084c35cf7170a1703fc8f14d712cf5afdc12502263" gracePeriod=30 Dec 10 23:08:54 crc kubenswrapper[4791]: I1210 23:08:54.694332 4791 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"51437220-85e7-4c11-a567-d1bb94871e81","Type":"ContainerStarted","Data":"d87c65da14b64634d54e9c9920efe50091ebd48562861281d0b21e5537c4ed60"} Dec 10 23:08:54 crc kubenswrapper[4791]: I1210 23:08:54.694402 4791 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/glance-default-external-api-0" podUID="51437220-85e7-4c11-a567-d1bb94871e81" containerName="glance-httpd" containerID="cri-o://d87c65da14b64634d54e9c9920efe50091ebd48562861281d0b21e5537c4ed60" gracePeriod=30 Dec 10 23:08:54 crc kubenswrapper[4791]: I1210 23:08:54.711395 4791 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/secret/60652b01-72de-481d-b2ea-8c349e9aee52-config\") pod \"neutron-856d8bb659-x5kwg\" (UID: \"60652b01-72de-481d-b2ea-8c349e9aee52\") " pod="openstack/neutron-856d8bb659-x5kwg" Dec 10 23:08:54 crc kubenswrapper[4791]: I1210 23:08:54.711475 4791 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-2srmj\" (UniqueName: \"kubernetes.io/projected/60652b01-72de-481d-b2ea-8c349e9aee52-kube-api-access-2srmj\") pod \"neutron-856d8bb659-x5kwg\" (UID: \"60652b01-72de-481d-b2ea-8c349e9aee52\") " pod="openstack/neutron-856d8bb659-x5kwg" Dec 10 23:08:54 crc kubenswrapper[4791]: I1210 23:08:54.711499 4791 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovndb-tls-certs\" (UniqueName: \"kubernetes.io/secret/60652b01-72de-481d-b2ea-8c349e9aee52-ovndb-tls-certs\") pod \"neutron-856d8bb659-x5kwg\" (UID: \"60652b01-72de-481d-b2ea-8c349e9aee52\") " pod="openstack/neutron-856d8bb659-x5kwg" Dec 10 23:08:54 crc kubenswrapper[4791]: I1210 23:08:54.711590 4791 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"httpd-config\" (UniqueName: \"kubernetes.io/secret/60652b01-72de-481d-b2ea-8c349e9aee52-httpd-config\") pod \"neutron-856d8bb659-x5kwg\" (UID: \"60652b01-72de-481d-b2ea-8c349e9aee52\") " pod="openstack/neutron-856d8bb659-x5kwg" Dec 10 23:08:54 crc kubenswrapper[4791]: I1210 23:08:54.711612 4791 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/60652b01-72de-481d-b2ea-8c349e9aee52-public-tls-certs\") pod \"neutron-856d8bb659-x5kwg\" (UID: \"60652b01-72de-481d-b2ea-8c349e9aee52\") " pod="openstack/neutron-856d8bb659-x5kwg" Dec 10 23:08:54 crc kubenswrapper[4791]: I1210 23:08:54.711649 4791 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/60652b01-72de-481d-b2ea-8c349e9aee52-internal-tls-certs\") pod \"neutron-856d8bb659-x5kwg\" (UID: \"60652b01-72de-481d-b2ea-8c349e9aee52\") " pod="openstack/neutron-856d8bb659-x5kwg" Dec 10 23:08:54 crc kubenswrapper[4791]: I1210 23:08:54.711693 4791 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/60652b01-72de-481d-b2ea-8c349e9aee52-combined-ca-bundle\") pod \"neutron-856d8bb659-x5kwg\" (UID: \"60652b01-72de-481d-b2ea-8c349e9aee52\") " pod="openstack/neutron-856d8bb659-x5kwg" Dec 10 23:08:54 crc kubenswrapper[4791]: I1210 23:08:54.726116 4791 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovndb-tls-certs\" (UniqueName: \"kubernetes.io/secret/60652b01-72de-481d-b2ea-8c349e9aee52-ovndb-tls-certs\") pod \"neutron-856d8bb659-x5kwg\" (UID: \"60652b01-72de-481d-b2ea-8c349e9aee52\") " pod="openstack/neutron-856d8bb659-x5kwg" Dec 10 23:08:54 crc kubenswrapper[4791]: I1210 23:08:54.726119 4791 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"httpd-config\" (UniqueName: \"kubernetes.io/secret/60652b01-72de-481d-b2ea-8c349e9aee52-httpd-config\") pod \"neutron-856d8bb659-x5kwg\" (UID: \"60652b01-72de-481d-b2ea-8c349e9aee52\") " pod="openstack/neutron-856d8bb659-x5kwg" Dec 10 23:08:54 crc kubenswrapper[4791]: I1210 23:08:54.726596 4791 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/60652b01-72de-481d-b2ea-8c349e9aee52-public-tls-certs\") pod \"neutron-856d8bb659-x5kwg\" (UID: \"60652b01-72de-481d-b2ea-8c349e9aee52\") " pod="openstack/neutron-856d8bb659-x5kwg" Dec 10 23:08:54 crc kubenswrapper[4791]: I1210 23:08:54.729601 4791 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/secret/60652b01-72de-481d-b2ea-8c349e9aee52-config\") pod \"neutron-856d8bb659-x5kwg\" (UID: \"60652b01-72de-481d-b2ea-8c349e9aee52\") " pod="openstack/neutron-856d8bb659-x5kwg" Dec 10 23:08:54 crc kubenswrapper[4791]: I1210 23:08:54.730139 4791 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/60652b01-72de-481d-b2ea-8c349e9aee52-combined-ca-bundle\") pod \"neutron-856d8bb659-x5kwg\" (UID: \"60652b01-72de-481d-b2ea-8c349e9aee52\") " pod="openstack/neutron-856d8bb659-x5kwg" Dec 10 23:08:54 crc kubenswrapper[4791]: I1210 23:08:54.737368 4791 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/dnsmasq-dns-84b966f6c9-tz266" podStartSLOduration=3.737324575 podStartE2EDuration="3.737324575s" podCreationTimestamp="2025-12-10 23:08:51 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-10 23:08:54.730503228 +0000 UTC m=+1169.160120841" watchObservedRunningTime="2025-12-10 23:08:54.737324575 +0000 UTC m=+1169.166942188" Dec 10 23:08:54 crc kubenswrapper[4791]: I1210 23:08:54.739388 4791 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/60652b01-72de-481d-b2ea-8c349e9aee52-internal-tls-certs\") pod \"neutron-856d8bb659-x5kwg\" (UID: \"60652b01-72de-481d-b2ea-8c349e9aee52\") " pod="openstack/neutron-856d8bb659-x5kwg" Dec 10 23:08:54 crc kubenswrapper[4791]: I1210 23:08:54.762131 4791 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-2srmj\" (UniqueName: \"kubernetes.io/projected/60652b01-72de-481d-b2ea-8c349e9aee52-kube-api-access-2srmj\") pod \"neutron-856d8bb659-x5kwg\" (UID: \"60652b01-72de-481d-b2ea-8c349e9aee52\") " pod="openstack/neutron-856d8bb659-x5kwg" Dec 10 23:08:54 crc kubenswrapper[4791]: I1210 23:08:54.781165 4791 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/glance-default-external-api-0" podStartSLOduration=17.781145749 podStartE2EDuration="17.781145749s" podCreationTimestamp="2025-12-10 23:08:37 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-10 23:08:54.776087383 +0000 UTC m=+1169.205704996" watchObservedRunningTime="2025-12-10 23:08:54.781145749 +0000 UTC m=+1169.210763362" Dec 10 23:08:55 crc kubenswrapper[4791]: I1210 23:08:55.037724 4791 patch_prober.go:28] interesting pod/machine-config-daemon-5rb5l container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 10 23:08:55 crc kubenswrapper[4791]: I1210 23:08:55.038086 4791 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-5rb5l" podUID="ba35653c-6e06-4cee-a4d6-137764090d18" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 10 23:08:55 crc kubenswrapper[4791]: I1210 23:08:55.037815 4791 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-856d8bb659-x5kwg" Dec 10 23:08:55 crc kubenswrapper[4791]: I1210 23:08:55.718328 4791 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"8eaa1b1d-715f-4059-ad18-fe9632c036a7","Type":"ContainerStarted","Data":"a6fb5e073d90b22471fee522d523321fa4497ebdf36b79f38fd1ee68b1a79815"} Dec 10 23:08:55 crc kubenswrapper[4791]: I1210 23:08:55.723569 4791 generic.go:334] "Generic (PLEG): container finished" podID="51437220-85e7-4c11-a567-d1bb94871e81" containerID="d87c65da14b64634d54e9c9920efe50091ebd48562861281d0b21e5537c4ed60" exitCode=0 Dec 10 23:08:55 crc kubenswrapper[4791]: I1210 23:08:55.723599 4791 generic.go:334] "Generic (PLEG): container finished" podID="51437220-85e7-4c11-a567-d1bb94871e81" containerID="b3200bc2f8d202f18e9123084c35cf7170a1703fc8f14d712cf5afdc12502263" exitCode=143 Dec 10 23:08:55 crc kubenswrapper[4791]: I1210 23:08:55.723643 4791 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"51437220-85e7-4c11-a567-d1bb94871e81","Type":"ContainerDied","Data":"d87c65da14b64634d54e9c9920efe50091ebd48562861281d0b21e5537c4ed60"} Dec 10 23:08:55 crc kubenswrapper[4791]: I1210 23:08:55.723668 4791 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"51437220-85e7-4c11-a567-d1bb94871e81","Type":"ContainerDied","Data":"b3200bc2f8d202f18e9123084c35cf7170a1703fc8f14d712cf5afdc12502263"} Dec 10 23:08:55 crc kubenswrapper[4791]: I1210 23:08:55.734800 4791 generic.go:334] "Generic (PLEG): container finished" podID="abd5935b-9f5b-4f25-9eec-eeedcb95bc16" containerID="c41db6dc374c530ec370e8b68d2cf357805394432183a75efdc95665439b1c5a" exitCode=0 Dec 10 23:08:55 crc kubenswrapper[4791]: I1210 23:08:55.734842 4791 generic.go:334] "Generic (PLEG): container finished" podID="abd5935b-9f5b-4f25-9eec-eeedcb95bc16" containerID="1c77bc5e67a79a02182644e651dc347dfb3e1f1a0fac9d42fa8544f476591532" exitCode=143 Dec 10 23:08:55 crc kubenswrapper[4791]: I1210 23:08:55.736119 4791 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-internal-api-0" event={"ID":"abd5935b-9f5b-4f25-9eec-eeedcb95bc16","Type":"ContainerDied","Data":"c41db6dc374c530ec370e8b68d2cf357805394432183a75efdc95665439b1c5a"} Dec 10 23:08:55 crc kubenswrapper[4791]: I1210 23:08:55.736153 4791 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-internal-api-0" event={"ID":"abd5935b-9f5b-4f25-9eec-eeedcb95bc16","Type":"ContainerDied","Data":"1c77bc5e67a79a02182644e651dc347dfb3e1f1a0fac9d42fa8544f476591532"} Dec 10 23:08:55 crc kubenswrapper[4791]: I1210 23:08:55.862975 4791 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-internal-api-0" Dec 10 23:08:55 crc kubenswrapper[4791]: I1210 23:08:55.922885 4791 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="7b395104-85b9-4ab1-8349-00fb9cb52b64" path="/var/lib/kubelet/pods/7b395104-85b9-4ab1-8349-00fb9cb52b64/volumes" Dec 10 23:08:55 crc kubenswrapper[4791]: I1210 23:08:55.941034 4791 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"glance\" (UniqueName: \"kubernetes.io/local-volume/local-storage03-crc\") pod \"abd5935b-9f5b-4f25-9eec-eeedcb95bc16\" (UID: \"abd5935b-9f5b-4f25-9eec-eeedcb95bc16\") " Dec 10 23:08:55 crc kubenswrapper[4791]: I1210 23:08:55.941131 4791 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/abd5935b-9f5b-4f25-9eec-eeedcb95bc16-logs\") pod \"abd5935b-9f5b-4f25-9eec-eeedcb95bc16\" (UID: \"abd5935b-9f5b-4f25-9eec-eeedcb95bc16\") " Dec 10 23:08:55 crc kubenswrapper[4791]: I1210 23:08:55.941172 4791 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/abd5935b-9f5b-4f25-9eec-eeedcb95bc16-httpd-run\") pod \"abd5935b-9f5b-4f25-9eec-eeedcb95bc16\" (UID: \"abd5935b-9f5b-4f25-9eec-eeedcb95bc16\") " Dec 10 23:08:55 crc kubenswrapper[4791]: I1210 23:08:55.941206 4791 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/abd5935b-9f5b-4f25-9eec-eeedcb95bc16-combined-ca-bundle\") pod \"abd5935b-9f5b-4f25-9eec-eeedcb95bc16\" (UID: \"abd5935b-9f5b-4f25-9eec-eeedcb95bc16\") " Dec 10 23:08:55 crc kubenswrapper[4791]: I1210 23:08:55.941284 4791 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-knxlc\" (UniqueName: \"kubernetes.io/projected/abd5935b-9f5b-4f25-9eec-eeedcb95bc16-kube-api-access-knxlc\") pod \"abd5935b-9f5b-4f25-9eec-eeedcb95bc16\" (UID: \"abd5935b-9f5b-4f25-9eec-eeedcb95bc16\") " Dec 10 23:08:55 crc kubenswrapper[4791]: I1210 23:08:55.941314 4791 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/abd5935b-9f5b-4f25-9eec-eeedcb95bc16-scripts\") pod \"abd5935b-9f5b-4f25-9eec-eeedcb95bc16\" (UID: \"abd5935b-9f5b-4f25-9eec-eeedcb95bc16\") " Dec 10 23:08:55 crc kubenswrapper[4791]: I1210 23:08:55.941424 4791 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/abd5935b-9f5b-4f25-9eec-eeedcb95bc16-config-data\") pod \"abd5935b-9f5b-4f25-9eec-eeedcb95bc16\" (UID: \"abd5935b-9f5b-4f25-9eec-eeedcb95bc16\") " Dec 10 23:08:55 crc kubenswrapper[4791]: I1210 23:08:55.942663 4791 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/abd5935b-9f5b-4f25-9eec-eeedcb95bc16-logs" (OuterVolumeSpecName: "logs") pod "abd5935b-9f5b-4f25-9eec-eeedcb95bc16" (UID: "abd5935b-9f5b-4f25-9eec-eeedcb95bc16"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 10 23:08:55 crc kubenswrapper[4791]: I1210 23:08:55.943139 4791 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/abd5935b-9f5b-4f25-9eec-eeedcb95bc16-httpd-run" (OuterVolumeSpecName: "httpd-run") pod "abd5935b-9f5b-4f25-9eec-eeedcb95bc16" (UID: "abd5935b-9f5b-4f25-9eec-eeedcb95bc16"). InnerVolumeSpecName "httpd-run". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 10 23:08:55 crc kubenswrapper[4791]: I1210 23:08:55.958677 4791 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/local-volume/local-storage03-crc" (OuterVolumeSpecName: "glance") pod "abd5935b-9f5b-4f25-9eec-eeedcb95bc16" (UID: "abd5935b-9f5b-4f25-9eec-eeedcb95bc16"). InnerVolumeSpecName "local-storage03-crc". PluginName "kubernetes.io/local-volume", VolumeGidValue "" Dec 10 23:08:55 crc kubenswrapper[4791]: I1210 23:08:55.960549 4791 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/abd5935b-9f5b-4f25-9eec-eeedcb95bc16-scripts" (OuterVolumeSpecName: "scripts") pod "abd5935b-9f5b-4f25-9eec-eeedcb95bc16" (UID: "abd5935b-9f5b-4f25-9eec-eeedcb95bc16"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 10 23:08:55 crc kubenswrapper[4791]: I1210 23:08:55.975034 4791 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/abd5935b-9f5b-4f25-9eec-eeedcb95bc16-kube-api-access-knxlc" (OuterVolumeSpecName: "kube-api-access-knxlc") pod "abd5935b-9f5b-4f25-9eec-eeedcb95bc16" (UID: "abd5935b-9f5b-4f25-9eec-eeedcb95bc16"). InnerVolumeSpecName "kube-api-access-knxlc". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 10 23:08:56 crc kubenswrapper[4791]: I1210 23:08:56.007577 4791 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/abd5935b-9f5b-4f25-9eec-eeedcb95bc16-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "abd5935b-9f5b-4f25-9eec-eeedcb95bc16" (UID: "abd5935b-9f5b-4f25-9eec-eeedcb95bc16"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 10 23:08:56 crc kubenswrapper[4791]: I1210 23:08:56.021508 4791 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/abd5935b-9f5b-4f25-9eec-eeedcb95bc16-config-data" (OuterVolumeSpecName: "config-data") pod "abd5935b-9f5b-4f25-9eec-eeedcb95bc16" (UID: "abd5935b-9f5b-4f25-9eec-eeedcb95bc16"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 10 23:08:56 crc kubenswrapper[4791]: I1210 23:08:56.044072 4791 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-knxlc\" (UniqueName: \"kubernetes.io/projected/abd5935b-9f5b-4f25-9eec-eeedcb95bc16-kube-api-access-knxlc\") on node \"crc\" DevicePath \"\"" Dec 10 23:08:56 crc kubenswrapper[4791]: I1210 23:08:56.044119 4791 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/abd5935b-9f5b-4f25-9eec-eeedcb95bc16-scripts\") on node \"crc\" DevicePath \"\"" Dec 10 23:08:56 crc kubenswrapper[4791]: I1210 23:08:56.044131 4791 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/abd5935b-9f5b-4f25-9eec-eeedcb95bc16-config-data\") on node \"crc\" DevicePath \"\"" Dec 10 23:08:56 crc kubenswrapper[4791]: I1210 23:08:56.044157 4791 reconciler_common.go:286] "operationExecutor.UnmountDevice started for volume \"local-storage03-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage03-crc\") on node \"crc\" " Dec 10 23:08:56 crc kubenswrapper[4791]: I1210 23:08:56.044168 4791 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/abd5935b-9f5b-4f25-9eec-eeedcb95bc16-logs\") on node \"crc\" DevicePath \"\"" Dec 10 23:08:56 crc kubenswrapper[4791]: I1210 23:08:56.044178 4791 reconciler_common.go:293] "Volume detached for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/abd5935b-9f5b-4f25-9eec-eeedcb95bc16-httpd-run\") on node \"crc\" DevicePath \"\"" Dec 10 23:08:56 crc kubenswrapper[4791]: I1210 23:08:56.044189 4791 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/abd5935b-9f5b-4f25-9eec-eeedcb95bc16-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 10 23:08:56 crc kubenswrapper[4791]: I1210 23:08:56.060764 4791 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-external-api-0" Dec 10 23:08:56 crc kubenswrapper[4791]: I1210 23:08:56.073557 4791 operation_generator.go:917] UnmountDevice succeeded for volume "local-storage03-crc" (UniqueName: "kubernetes.io/local-volume/local-storage03-crc") on node "crc" Dec 10 23:08:56 crc kubenswrapper[4791]: I1210 23:08:56.115131 4791 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/neutron-856d8bb659-x5kwg"] Dec 10 23:08:56 crc kubenswrapper[4791]: I1210 23:08:56.149374 4791 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/51437220-85e7-4c11-a567-d1bb94871e81-combined-ca-bundle\") pod \"51437220-85e7-4c11-a567-d1bb94871e81\" (UID: \"51437220-85e7-4c11-a567-d1bb94871e81\") " Dec 10 23:08:56 crc kubenswrapper[4791]: I1210 23:08:56.149484 4791 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"glance\" (UniqueName: \"kubernetes.io/local-volume/local-storage06-crc\") pod \"51437220-85e7-4c11-a567-d1bb94871e81\" (UID: \"51437220-85e7-4c11-a567-d1bb94871e81\") " Dec 10 23:08:56 crc kubenswrapper[4791]: I1210 23:08:56.149536 4791 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/51437220-85e7-4c11-a567-d1bb94871e81-config-data\") pod \"51437220-85e7-4c11-a567-d1bb94871e81\" (UID: \"51437220-85e7-4c11-a567-d1bb94871e81\") " Dec 10 23:08:56 crc kubenswrapper[4791]: I1210 23:08:56.149845 4791 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/51437220-85e7-4c11-a567-d1bb94871e81-logs\") pod \"51437220-85e7-4c11-a567-d1bb94871e81\" (UID: \"51437220-85e7-4c11-a567-d1bb94871e81\") " Dec 10 23:08:56 crc kubenswrapper[4791]: I1210 23:08:56.149914 4791 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/51437220-85e7-4c11-a567-d1bb94871e81-scripts\") pod \"51437220-85e7-4c11-a567-d1bb94871e81\" (UID: \"51437220-85e7-4c11-a567-d1bb94871e81\") " Dec 10 23:08:56 crc kubenswrapper[4791]: I1210 23:08:56.149943 4791 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/51437220-85e7-4c11-a567-d1bb94871e81-httpd-run\") pod \"51437220-85e7-4c11-a567-d1bb94871e81\" (UID: \"51437220-85e7-4c11-a567-d1bb94871e81\") " Dec 10 23:08:56 crc kubenswrapper[4791]: I1210 23:08:56.149985 4791 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-hdz5s\" (UniqueName: \"kubernetes.io/projected/51437220-85e7-4c11-a567-d1bb94871e81-kube-api-access-hdz5s\") pod \"51437220-85e7-4c11-a567-d1bb94871e81\" (UID: \"51437220-85e7-4c11-a567-d1bb94871e81\") " Dec 10 23:08:56 crc kubenswrapper[4791]: I1210 23:08:56.151827 4791 reconciler_common.go:293] "Volume detached for volume \"local-storage03-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage03-crc\") on node \"crc\" DevicePath \"\"" Dec 10 23:08:56 crc kubenswrapper[4791]: I1210 23:08:56.155894 4791 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/51437220-85e7-4c11-a567-d1bb94871e81-logs" (OuterVolumeSpecName: "logs") pod "51437220-85e7-4c11-a567-d1bb94871e81" (UID: "51437220-85e7-4c11-a567-d1bb94871e81"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 10 23:08:56 crc kubenswrapper[4791]: I1210 23:08:56.156191 4791 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/51437220-85e7-4c11-a567-d1bb94871e81-httpd-run" (OuterVolumeSpecName: "httpd-run") pod "51437220-85e7-4c11-a567-d1bb94871e81" (UID: "51437220-85e7-4c11-a567-d1bb94871e81"). InnerVolumeSpecName "httpd-run". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 10 23:08:56 crc kubenswrapper[4791]: I1210 23:08:56.156208 4791 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/51437220-85e7-4c11-a567-d1bb94871e81-kube-api-access-hdz5s" (OuterVolumeSpecName: "kube-api-access-hdz5s") pod "51437220-85e7-4c11-a567-d1bb94871e81" (UID: "51437220-85e7-4c11-a567-d1bb94871e81"). InnerVolumeSpecName "kube-api-access-hdz5s". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 10 23:08:56 crc kubenswrapper[4791]: I1210 23:08:56.169815 4791 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/local-volume/local-storage06-crc" (OuterVolumeSpecName: "glance") pod "51437220-85e7-4c11-a567-d1bb94871e81" (UID: "51437220-85e7-4c11-a567-d1bb94871e81"). InnerVolumeSpecName "local-storage06-crc". PluginName "kubernetes.io/local-volume", VolumeGidValue "" Dec 10 23:08:56 crc kubenswrapper[4791]: I1210 23:08:56.183403 4791 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/51437220-85e7-4c11-a567-d1bb94871e81-scripts" (OuterVolumeSpecName: "scripts") pod "51437220-85e7-4c11-a567-d1bb94871e81" (UID: "51437220-85e7-4c11-a567-d1bb94871e81"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 10 23:08:56 crc kubenswrapper[4791]: I1210 23:08:56.234100 4791 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/51437220-85e7-4c11-a567-d1bb94871e81-config-data" (OuterVolumeSpecName: "config-data") pod "51437220-85e7-4c11-a567-d1bb94871e81" (UID: "51437220-85e7-4c11-a567-d1bb94871e81"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 10 23:08:56 crc kubenswrapper[4791]: I1210 23:08:56.234326 4791 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/51437220-85e7-4c11-a567-d1bb94871e81-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "51437220-85e7-4c11-a567-d1bb94871e81" (UID: "51437220-85e7-4c11-a567-d1bb94871e81"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 10 23:08:56 crc kubenswrapper[4791]: I1210 23:08:56.253927 4791 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/51437220-85e7-4c11-a567-d1bb94871e81-logs\") on node \"crc\" DevicePath \"\"" Dec 10 23:08:56 crc kubenswrapper[4791]: I1210 23:08:56.253962 4791 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/51437220-85e7-4c11-a567-d1bb94871e81-scripts\") on node \"crc\" DevicePath \"\"" Dec 10 23:08:56 crc kubenswrapper[4791]: I1210 23:08:56.253975 4791 reconciler_common.go:293] "Volume detached for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/51437220-85e7-4c11-a567-d1bb94871e81-httpd-run\") on node \"crc\" DevicePath \"\"" Dec 10 23:08:56 crc kubenswrapper[4791]: I1210 23:08:56.254012 4791 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-hdz5s\" (UniqueName: \"kubernetes.io/projected/51437220-85e7-4c11-a567-d1bb94871e81-kube-api-access-hdz5s\") on node \"crc\" DevicePath \"\"" Dec 10 23:08:56 crc kubenswrapper[4791]: I1210 23:08:56.254027 4791 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/51437220-85e7-4c11-a567-d1bb94871e81-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 10 23:08:56 crc kubenswrapper[4791]: I1210 23:08:56.254071 4791 reconciler_common.go:286] "operationExecutor.UnmountDevice started for volume \"local-storage06-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage06-crc\") on node \"crc\" " Dec 10 23:08:56 crc kubenswrapper[4791]: I1210 23:08:56.254087 4791 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/51437220-85e7-4c11-a567-d1bb94871e81-config-data\") on node \"crc\" DevicePath \"\"" Dec 10 23:08:56 crc kubenswrapper[4791]: I1210 23:08:56.275077 4791 operation_generator.go:917] UnmountDevice succeeded for volume "local-storage06-crc" (UniqueName: "kubernetes.io/local-volume/local-storage06-crc") on node "crc" Dec 10 23:08:56 crc kubenswrapper[4791]: I1210 23:08:56.356054 4791 reconciler_common.go:293] "Volume detached for volume \"local-storage06-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage06-crc\") on node \"crc\" DevicePath \"\"" Dec 10 23:08:56 crc kubenswrapper[4791]: I1210 23:08:56.758005 4791 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"51437220-85e7-4c11-a567-d1bb94871e81","Type":"ContainerDied","Data":"b8ce0e374219a112d8cb5058ba1f707e84b6d757ed98ce0a13df0f32a5294422"} Dec 10 23:08:56 crc kubenswrapper[4791]: I1210 23:08:56.758071 4791 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-external-api-0" Dec 10 23:08:56 crc kubenswrapper[4791]: I1210 23:08:56.758100 4791 scope.go:117] "RemoveContainer" containerID="d87c65da14b64634d54e9c9920efe50091ebd48562861281d0b21e5537c4ed60" Dec 10 23:08:56 crc kubenswrapper[4791]: I1210 23:08:56.759627 4791 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-856d8bb659-x5kwg" event={"ID":"60652b01-72de-481d-b2ea-8c349e9aee52","Type":"ContainerStarted","Data":"d16b70ccb879eecdc077e88fb83baf286ae999fe11ae37fa2b8190eb7cff77f8"} Dec 10 23:08:56 crc kubenswrapper[4791]: I1210 23:08:56.773932 4791 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-internal-api-0" Dec 10 23:08:56 crc kubenswrapper[4791]: I1210 23:08:56.774390 4791 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-internal-api-0" event={"ID":"abd5935b-9f5b-4f25-9eec-eeedcb95bc16","Type":"ContainerDied","Data":"4b04d4980fab42d0d10f3d85865d46ed6e9222985e5ebd06945fc2460e8de169"} Dec 10 23:08:56 crc kubenswrapper[4791]: I1210 23:08:56.879599 4791 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/glance-default-external-api-0"] Dec 10 23:08:56 crc kubenswrapper[4791]: I1210 23:08:56.912000 4791 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/glance-default-external-api-0"] Dec 10 23:08:56 crc kubenswrapper[4791]: I1210 23:08:56.925732 4791 scope.go:117] "RemoveContainer" containerID="b3200bc2f8d202f18e9123084c35cf7170a1703fc8f14d712cf5afdc12502263" Dec 10 23:08:56 crc kubenswrapper[4791]: I1210 23:08:56.931454 4791 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/glance-default-external-api-0"] Dec 10 23:08:56 crc kubenswrapper[4791]: E1210 23:08:56.931813 4791 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="51437220-85e7-4c11-a567-d1bb94871e81" containerName="glance-httpd" Dec 10 23:08:56 crc kubenswrapper[4791]: I1210 23:08:56.931831 4791 state_mem.go:107] "Deleted CPUSet assignment" podUID="51437220-85e7-4c11-a567-d1bb94871e81" containerName="glance-httpd" Dec 10 23:08:56 crc kubenswrapper[4791]: E1210 23:08:56.931842 4791 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="51437220-85e7-4c11-a567-d1bb94871e81" containerName="glance-log" Dec 10 23:08:56 crc kubenswrapper[4791]: I1210 23:08:56.931848 4791 state_mem.go:107] "Deleted CPUSet assignment" podUID="51437220-85e7-4c11-a567-d1bb94871e81" containerName="glance-log" Dec 10 23:08:56 crc kubenswrapper[4791]: E1210 23:08:56.931864 4791 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="abd5935b-9f5b-4f25-9eec-eeedcb95bc16" containerName="glance-httpd" Dec 10 23:08:56 crc kubenswrapper[4791]: I1210 23:08:56.931871 4791 state_mem.go:107] "Deleted CPUSet assignment" podUID="abd5935b-9f5b-4f25-9eec-eeedcb95bc16" containerName="glance-httpd" Dec 10 23:08:56 crc kubenswrapper[4791]: E1210 23:08:56.931905 4791 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="abd5935b-9f5b-4f25-9eec-eeedcb95bc16" containerName="glance-log" Dec 10 23:08:56 crc kubenswrapper[4791]: I1210 23:08:56.931911 4791 state_mem.go:107] "Deleted CPUSet assignment" podUID="abd5935b-9f5b-4f25-9eec-eeedcb95bc16" containerName="glance-log" Dec 10 23:08:56 crc kubenswrapper[4791]: I1210 23:08:56.933220 4791 memory_manager.go:354] "RemoveStaleState removing state" podUID="abd5935b-9f5b-4f25-9eec-eeedcb95bc16" containerName="glance-log" Dec 10 23:08:56 crc kubenswrapper[4791]: I1210 23:08:56.933245 4791 memory_manager.go:354] "RemoveStaleState removing state" podUID="51437220-85e7-4c11-a567-d1bb94871e81" containerName="glance-log" Dec 10 23:08:56 crc kubenswrapper[4791]: I1210 23:08:56.933264 4791 memory_manager.go:354] "RemoveStaleState removing state" podUID="abd5935b-9f5b-4f25-9eec-eeedcb95bc16" containerName="glance-httpd" Dec 10 23:08:56 crc kubenswrapper[4791]: I1210 23:08:56.933275 4791 memory_manager.go:354] "RemoveStaleState removing state" podUID="51437220-85e7-4c11-a567-d1bb94871e81" containerName="glance-httpd" Dec 10 23:08:56 crc kubenswrapper[4791]: I1210 23:08:56.934290 4791 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-external-api-0" Dec 10 23:08:56 crc kubenswrapper[4791]: I1210 23:08:56.936921 4791 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"glance-glance-dockercfg-rsl4w" Dec 10 23:08:56 crc kubenswrapper[4791]: I1210 23:08:56.944804 4791 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"glance-scripts" Dec 10 23:08:56 crc kubenswrapper[4791]: I1210 23:08:56.945513 4791 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-glance-default-public-svc" Dec 10 23:08:56 crc kubenswrapper[4791]: I1210 23:08:56.945622 4791 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"glance-default-external-config-data" Dec 10 23:08:56 crc kubenswrapper[4791]: I1210 23:08:56.945761 4791 scope.go:117] "RemoveContainer" containerID="c41db6dc374c530ec370e8b68d2cf357805394432183a75efdc95665439b1c5a" Dec 10 23:08:56 crc kubenswrapper[4791]: I1210 23:08:56.966224 4791 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/glance-default-internal-api-0"] Dec 10 23:08:56 crc kubenswrapper[4791]: I1210 23:08:56.989060 4791 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/glance-default-internal-api-0"] Dec 10 23:08:56 crc kubenswrapper[4791]: I1210 23:08:56.994428 4791 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-default-external-api-0"] Dec 10 23:08:57 crc kubenswrapper[4791]: I1210 23:08:57.001286 4791 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/glance-default-internal-api-0"] Dec 10 23:08:57 crc kubenswrapper[4791]: I1210 23:08:57.002857 4791 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-internal-api-0" Dec 10 23:08:57 crc kubenswrapper[4791]: I1210 23:08:57.016415 4791 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-default-internal-api-0"] Dec 10 23:08:57 crc kubenswrapper[4791]: I1210 23:08:57.030043 4791 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-glance-default-internal-svc" Dec 10 23:08:57 crc kubenswrapper[4791]: I1210 23:08:57.030792 4791 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"glance-default-internal-config-data" Dec 10 23:08:57 crc kubenswrapper[4791]: I1210 23:08:57.049706 4791 scope.go:117] "RemoveContainer" containerID="1c77bc5e67a79a02182644e651dc347dfb3e1f1a0fac9d42fa8544f476591532" Dec 10 23:08:57 crc kubenswrapper[4791]: I1210 23:08:57.079876 4791 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/e51944e5-b91a-4310-9693-5333a680b9ff-logs\") pod \"glance-default-external-api-0\" (UID: \"e51944e5-b91a-4310-9693-5333a680b9ff\") " pod="openstack/glance-default-external-api-0" Dec 10 23:08:57 crc kubenswrapper[4791]: I1210 23:08:57.079923 4791 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-wsh77\" (UniqueName: \"kubernetes.io/projected/e51944e5-b91a-4310-9693-5333a680b9ff-kube-api-access-wsh77\") pod \"glance-default-external-api-0\" (UID: \"e51944e5-b91a-4310-9693-5333a680b9ff\") " pod="openstack/glance-default-external-api-0" Dec 10 23:08:57 crc kubenswrapper[4791]: I1210 23:08:57.079967 4791 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage06-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage06-crc\") pod \"glance-default-external-api-0\" (UID: \"e51944e5-b91a-4310-9693-5333a680b9ff\") " pod="openstack/glance-default-external-api-0" Dec 10 23:08:57 crc kubenswrapper[4791]: I1210 23:08:57.080003 4791 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/e51944e5-b91a-4310-9693-5333a680b9ff-httpd-run\") pod \"glance-default-external-api-0\" (UID: \"e51944e5-b91a-4310-9693-5333a680b9ff\") " pod="openstack/glance-default-external-api-0" Dec 10 23:08:57 crc kubenswrapper[4791]: I1210 23:08:57.080019 4791 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/e51944e5-b91a-4310-9693-5333a680b9ff-public-tls-certs\") pod \"glance-default-external-api-0\" (UID: \"e51944e5-b91a-4310-9693-5333a680b9ff\") " pod="openstack/glance-default-external-api-0" Dec 10 23:08:57 crc kubenswrapper[4791]: I1210 23:08:57.080054 4791 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/e51944e5-b91a-4310-9693-5333a680b9ff-config-data\") pod \"glance-default-external-api-0\" (UID: \"e51944e5-b91a-4310-9693-5333a680b9ff\") " pod="openstack/glance-default-external-api-0" Dec 10 23:08:57 crc kubenswrapper[4791]: I1210 23:08:57.080092 4791 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/e51944e5-b91a-4310-9693-5333a680b9ff-scripts\") pod \"glance-default-external-api-0\" (UID: \"e51944e5-b91a-4310-9693-5333a680b9ff\") " pod="openstack/glance-default-external-api-0" Dec 10 23:08:57 crc kubenswrapper[4791]: I1210 23:08:57.080140 4791 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/e51944e5-b91a-4310-9693-5333a680b9ff-combined-ca-bundle\") pod \"glance-default-external-api-0\" (UID: \"e51944e5-b91a-4310-9693-5333a680b9ff\") " pod="openstack/glance-default-external-api-0" Dec 10 23:08:57 crc kubenswrapper[4791]: I1210 23:08:57.181445 4791 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-9v6tz\" (UniqueName: \"kubernetes.io/projected/9460828b-2713-404b-8c14-ca25566937f4-kube-api-access-9v6tz\") pod \"glance-default-internal-api-0\" (UID: \"9460828b-2713-404b-8c14-ca25566937f4\") " pod="openstack/glance-default-internal-api-0" Dec 10 23:08:57 crc kubenswrapper[4791]: I1210 23:08:57.181514 4791 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/9460828b-2713-404b-8c14-ca25566937f4-scripts\") pod \"glance-default-internal-api-0\" (UID: \"9460828b-2713-404b-8c14-ca25566937f4\") " pod="openstack/glance-default-internal-api-0" Dec 10 23:08:57 crc kubenswrapper[4791]: I1210 23:08:57.181539 4791 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/9460828b-2713-404b-8c14-ca25566937f4-combined-ca-bundle\") pod \"glance-default-internal-api-0\" (UID: \"9460828b-2713-404b-8c14-ca25566937f4\") " pod="openstack/glance-default-internal-api-0" Dec 10 23:08:57 crc kubenswrapper[4791]: I1210 23:08:57.181583 4791 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/e51944e5-b91a-4310-9693-5333a680b9ff-logs\") pod \"glance-default-external-api-0\" (UID: \"e51944e5-b91a-4310-9693-5333a680b9ff\") " pod="openstack/glance-default-external-api-0" Dec 10 23:08:57 crc kubenswrapper[4791]: I1210 23:08:57.181616 4791 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-wsh77\" (UniqueName: \"kubernetes.io/projected/e51944e5-b91a-4310-9693-5333a680b9ff-kube-api-access-wsh77\") pod \"glance-default-external-api-0\" (UID: \"e51944e5-b91a-4310-9693-5333a680b9ff\") " pod="openstack/glance-default-external-api-0" Dec 10 23:08:57 crc kubenswrapper[4791]: I1210 23:08:57.181736 4791 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/9460828b-2713-404b-8c14-ca25566937f4-config-data\") pod \"glance-default-internal-api-0\" (UID: \"9460828b-2713-404b-8c14-ca25566937f4\") " pod="openstack/glance-default-internal-api-0" Dec 10 23:08:57 crc kubenswrapper[4791]: I1210 23:08:57.181830 4791 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage06-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage06-crc\") pod \"glance-default-external-api-0\" (UID: \"e51944e5-b91a-4310-9693-5333a680b9ff\") " pod="openstack/glance-default-external-api-0" Dec 10 23:08:57 crc kubenswrapper[4791]: I1210 23:08:57.181889 4791 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/9460828b-2713-404b-8c14-ca25566937f4-internal-tls-certs\") pod \"glance-default-internal-api-0\" (UID: \"9460828b-2713-404b-8c14-ca25566937f4\") " pod="openstack/glance-default-internal-api-0" Dec 10 23:08:57 crc kubenswrapper[4791]: I1210 23:08:57.181945 4791 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/9460828b-2713-404b-8c14-ca25566937f4-logs\") pod \"glance-default-internal-api-0\" (UID: \"9460828b-2713-404b-8c14-ca25566937f4\") " pod="openstack/glance-default-internal-api-0" Dec 10 23:08:57 crc kubenswrapper[4791]: I1210 23:08:57.181986 4791 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/e51944e5-b91a-4310-9693-5333a680b9ff-public-tls-certs\") pod \"glance-default-external-api-0\" (UID: \"e51944e5-b91a-4310-9693-5333a680b9ff\") " pod="openstack/glance-default-external-api-0" Dec 10 23:08:57 crc kubenswrapper[4791]: I1210 23:08:57.182010 4791 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage03-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage03-crc\") pod \"glance-default-internal-api-0\" (UID: \"9460828b-2713-404b-8c14-ca25566937f4\") " pod="openstack/glance-default-internal-api-0" Dec 10 23:08:57 crc kubenswrapper[4791]: I1210 23:08:57.182034 4791 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/e51944e5-b91a-4310-9693-5333a680b9ff-httpd-run\") pod \"glance-default-external-api-0\" (UID: \"e51944e5-b91a-4310-9693-5333a680b9ff\") " pod="openstack/glance-default-external-api-0" Dec 10 23:08:57 crc kubenswrapper[4791]: I1210 23:08:57.182068 4791 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage06-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage06-crc\") pod \"glance-default-external-api-0\" (UID: \"e51944e5-b91a-4310-9693-5333a680b9ff\") device mount path \"/mnt/openstack/pv06\"" pod="openstack/glance-default-external-api-0" Dec 10 23:08:57 crc kubenswrapper[4791]: I1210 23:08:57.182068 4791 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/e51944e5-b91a-4310-9693-5333a680b9ff-logs\") pod \"glance-default-external-api-0\" (UID: \"e51944e5-b91a-4310-9693-5333a680b9ff\") " pod="openstack/glance-default-external-api-0" Dec 10 23:08:57 crc kubenswrapper[4791]: I1210 23:08:57.182192 4791 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/e51944e5-b91a-4310-9693-5333a680b9ff-config-data\") pod \"glance-default-external-api-0\" (UID: \"e51944e5-b91a-4310-9693-5333a680b9ff\") " pod="openstack/glance-default-external-api-0" Dec 10 23:08:57 crc kubenswrapper[4791]: I1210 23:08:57.182244 4791 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/9460828b-2713-404b-8c14-ca25566937f4-httpd-run\") pod \"glance-default-internal-api-0\" (UID: \"9460828b-2713-404b-8c14-ca25566937f4\") " pod="openstack/glance-default-internal-api-0" Dec 10 23:08:57 crc kubenswrapper[4791]: I1210 23:08:57.182270 4791 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/e51944e5-b91a-4310-9693-5333a680b9ff-scripts\") pod \"glance-default-external-api-0\" (UID: \"e51944e5-b91a-4310-9693-5333a680b9ff\") " pod="openstack/glance-default-external-api-0" Dec 10 23:08:57 crc kubenswrapper[4791]: I1210 23:08:57.182354 4791 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/e51944e5-b91a-4310-9693-5333a680b9ff-combined-ca-bundle\") pod \"glance-default-external-api-0\" (UID: \"e51944e5-b91a-4310-9693-5333a680b9ff\") " pod="openstack/glance-default-external-api-0" Dec 10 23:08:57 crc kubenswrapper[4791]: I1210 23:08:57.182472 4791 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/e51944e5-b91a-4310-9693-5333a680b9ff-httpd-run\") pod \"glance-default-external-api-0\" (UID: \"e51944e5-b91a-4310-9693-5333a680b9ff\") " pod="openstack/glance-default-external-api-0" Dec 10 23:08:57 crc kubenswrapper[4791]: I1210 23:08:57.189208 4791 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/e51944e5-b91a-4310-9693-5333a680b9ff-scripts\") pod \"glance-default-external-api-0\" (UID: \"e51944e5-b91a-4310-9693-5333a680b9ff\") " pod="openstack/glance-default-external-api-0" Dec 10 23:08:57 crc kubenswrapper[4791]: I1210 23:08:57.190927 4791 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/e51944e5-b91a-4310-9693-5333a680b9ff-combined-ca-bundle\") pod \"glance-default-external-api-0\" (UID: \"e51944e5-b91a-4310-9693-5333a680b9ff\") " pod="openstack/glance-default-external-api-0" Dec 10 23:08:57 crc kubenswrapper[4791]: I1210 23:08:57.199080 4791 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/e51944e5-b91a-4310-9693-5333a680b9ff-public-tls-certs\") pod \"glance-default-external-api-0\" (UID: \"e51944e5-b91a-4310-9693-5333a680b9ff\") " pod="openstack/glance-default-external-api-0" Dec 10 23:08:57 crc kubenswrapper[4791]: I1210 23:08:57.207229 4791 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/e51944e5-b91a-4310-9693-5333a680b9ff-config-data\") pod \"glance-default-external-api-0\" (UID: \"e51944e5-b91a-4310-9693-5333a680b9ff\") " pod="openstack/glance-default-external-api-0" Dec 10 23:08:57 crc kubenswrapper[4791]: I1210 23:08:57.216155 4791 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-wsh77\" (UniqueName: \"kubernetes.io/projected/e51944e5-b91a-4310-9693-5333a680b9ff-kube-api-access-wsh77\") pod \"glance-default-external-api-0\" (UID: \"e51944e5-b91a-4310-9693-5333a680b9ff\") " pod="openstack/glance-default-external-api-0" Dec 10 23:08:57 crc kubenswrapper[4791]: I1210 23:08:57.241946 4791 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage06-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage06-crc\") pod \"glance-default-external-api-0\" (UID: \"e51944e5-b91a-4310-9693-5333a680b9ff\") " pod="openstack/glance-default-external-api-0" Dec 10 23:08:57 crc kubenswrapper[4791]: I1210 23:08:57.286442 4791 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/9460828b-2713-404b-8c14-ca25566937f4-internal-tls-certs\") pod \"glance-default-internal-api-0\" (UID: \"9460828b-2713-404b-8c14-ca25566937f4\") " pod="openstack/glance-default-internal-api-0" Dec 10 23:08:57 crc kubenswrapper[4791]: I1210 23:08:57.286513 4791 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/9460828b-2713-404b-8c14-ca25566937f4-logs\") pod \"glance-default-internal-api-0\" (UID: \"9460828b-2713-404b-8c14-ca25566937f4\") " pod="openstack/glance-default-internal-api-0" Dec 10 23:08:57 crc kubenswrapper[4791]: I1210 23:08:57.286548 4791 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage03-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage03-crc\") pod \"glance-default-internal-api-0\" (UID: \"9460828b-2713-404b-8c14-ca25566937f4\") " pod="openstack/glance-default-internal-api-0" Dec 10 23:08:57 crc kubenswrapper[4791]: I1210 23:08:57.286620 4791 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/9460828b-2713-404b-8c14-ca25566937f4-httpd-run\") pod \"glance-default-internal-api-0\" (UID: \"9460828b-2713-404b-8c14-ca25566937f4\") " pod="openstack/glance-default-internal-api-0" Dec 10 23:08:57 crc kubenswrapper[4791]: I1210 23:08:57.286707 4791 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-9v6tz\" (UniqueName: \"kubernetes.io/projected/9460828b-2713-404b-8c14-ca25566937f4-kube-api-access-9v6tz\") pod \"glance-default-internal-api-0\" (UID: \"9460828b-2713-404b-8c14-ca25566937f4\") " pod="openstack/glance-default-internal-api-0" Dec 10 23:08:57 crc kubenswrapper[4791]: I1210 23:08:57.286735 4791 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/9460828b-2713-404b-8c14-ca25566937f4-scripts\") pod \"glance-default-internal-api-0\" (UID: \"9460828b-2713-404b-8c14-ca25566937f4\") " pod="openstack/glance-default-internal-api-0" Dec 10 23:08:57 crc kubenswrapper[4791]: I1210 23:08:57.286775 4791 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/9460828b-2713-404b-8c14-ca25566937f4-combined-ca-bundle\") pod \"glance-default-internal-api-0\" (UID: \"9460828b-2713-404b-8c14-ca25566937f4\") " pod="openstack/glance-default-internal-api-0" Dec 10 23:08:57 crc kubenswrapper[4791]: I1210 23:08:57.286815 4791 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/9460828b-2713-404b-8c14-ca25566937f4-config-data\") pod \"glance-default-internal-api-0\" (UID: \"9460828b-2713-404b-8c14-ca25566937f4\") " pod="openstack/glance-default-internal-api-0" Dec 10 23:08:57 crc kubenswrapper[4791]: I1210 23:08:57.287691 4791 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/9460828b-2713-404b-8c14-ca25566937f4-httpd-run\") pod \"glance-default-internal-api-0\" (UID: \"9460828b-2713-404b-8c14-ca25566937f4\") " pod="openstack/glance-default-internal-api-0" Dec 10 23:08:57 crc kubenswrapper[4791]: I1210 23:08:57.288888 4791 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/9460828b-2713-404b-8c14-ca25566937f4-logs\") pod \"glance-default-internal-api-0\" (UID: \"9460828b-2713-404b-8c14-ca25566937f4\") " pod="openstack/glance-default-internal-api-0" Dec 10 23:08:57 crc kubenswrapper[4791]: I1210 23:08:57.289031 4791 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage03-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage03-crc\") pod \"glance-default-internal-api-0\" (UID: \"9460828b-2713-404b-8c14-ca25566937f4\") device mount path \"/mnt/openstack/pv03\"" pod="openstack/glance-default-internal-api-0" Dec 10 23:08:57 crc kubenswrapper[4791]: I1210 23:08:57.292231 4791 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/9460828b-2713-404b-8c14-ca25566937f4-config-data\") pod \"glance-default-internal-api-0\" (UID: \"9460828b-2713-404b-8c14-ca25566937f4\") " pod="openstack/glance-default-internal-api-0" Dec 10 23:08:57 crc kubenswrapper[4791]: I1210 23:08:57.296584 4791 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-external-api-0" Dec 10 23:08:57 crc kubenswrapper[4791]: I1210 23:08:57.297227 4791 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/9460828b-2713-404b-8c14-ca25566937f4-combined-ca-bundle\") pod \"glance-default-internal-api-0\" (UID: \"9460828b-2713-404b-8c14-ca25566937f4\") " pod="openstack/glance-default-internal-api-0" Dec 10 23:08:57 crc kubenswrapper[4791]: I1210 23:08:57.298770 4791 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/9460828b-2713-404b-8c14-ca25566937f4-internal-tls-certs\") pod \"glance-default-internal-api-0\" (UID: \"9460828b-2713-404b-8c14-ca25566937f4\") " pod="openstack/glance-default-internal-api-0" Dec 10 23:08:57 crc kubenswrapper[4791]: I1210 23:08:57.302016 4791 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/9460828b-2713-404b-8c14-ca25566937f4-scripts\") pod \"glance-default-internal-api-0\" (UID: \"9460828b-2713-404b-8c14-ca25566937f4\") " pod="openstack/glance-default-internal-api-0" Dec 10 23:08:57 crc kubenswrapper[4791]: I1210 23:08:57.317085 4791 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-9v6tz\" (UniqueName: \"kubernetes.io/projected/9460828b-2713-404b-8c14-ca25566937f4-kube-api-access-9v6tz\") pod \"glance-default-internal-api-0\" (UID: \"9460828b-2713-404b-8c14-ca25566937f4\") " pod="openstack/glance-default-internal-api-0" Dec 10 23:08:57 crc kubenswrapper[4791]: I1210 23:08:57.369693 4791 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage03-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage03-crc\") pod \"glance-default-internal-api-0\" (UID: \"9460828b-2713-404b-8c14-ca25566937f4\") " pod="openstack/glance-default-internal-api-0" Dec 10 23:08:57 crc kubenswrapper[4791]: I1210 23:08:57.658281 4791 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-internal-api-0" Dec 10 23:08:57 crc kubenswrapper[4791]: I1210 23:08:57.798928 4791 generic.go:334] "Generic (PLEG): container finished" podID="7c65a05d-c2cc-45c4-b469-e6a271e01856" containerID="f0eb0592a4a0cdde8425fc8222755e9b651cc41049d8057c83c99f2ae97e73a5" exitCode=0 Dec 10 23:08:57 crc kubenswrapper[4791]: I1210 23:08:57.799057 4791 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-bootstrap-mwqkg" event={"ID":"7c65a05d-c2cc-45c4-b469-e6a271e01856","Type":"ContainerDied","Data":"f0eb0592a4a0cdde8425fc8222755e9b651cc41049d8057c83c99f2ae97e73a5"} Dec 10 23:08:57 crc kubenswrapper[4791]: I1210 23:08:57.817947 4791 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-856d8bb659-x5kwg" event={"ID":"60652b01-72de-481d-b2ea-8c349e9aee52","Type":"ContainerStarted","Data":"d18a560dcceb6a5ded619d71ad3633b0464b3edfbfcce229712c984d40d39f87"} Dec 10 23:08:57 crc kubenswrapper[4791]: I1210 23:08:57.817990 4791 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-856d8bb659-x5kwg" event={"ID":"60652b01-72de-481d-b2ea-8c349e9aee52","Type":"ContainerStarted","Data":"27441d329ae77fc53456abeff253c9fe2a09749a4809ca870531bcf1df39ecea"} Dec 10 23:08:57 crc kubenswrapper[4791]: I1210 23:08:57.818804 4791 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/neutron-856d8bb659-x5kwg" Dec 10 23:08:57 crc kubenswrapper[4791]: I1210 23:08:57.877662 4791 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/neutron-856d8bb659-x5kwg" podStartSLOduration=3.877638717 podStartE2EDuration="3.877638717s" podCreationTimestamp="2025-12-10 23:08:54 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-10 23:08:57.852641646 +0000 UTC m=+1172.282259259" watchObservedRunningTime="2025-12-10 23:08:57.877638717 +0000 UTC m=+1172.307256330" Dec 10 23:08:57 crc kubenswrapper[4791]: I1210 23:08:57.948382 4791 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="51437220-85e7-4c11-a567-d1bb94871e81" path="/var/lib/kubelet/pods/51437220-85e7-4c11-a567-d1bb94871e81/volumes" Dec 10 23:08:57 crc kubenswrapper[4791]: I1210 23:08:57.952750 4791 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="abd5935b-9f5b-4f25-9eec-eeedcb95bc16" path="/var/lib/kubelet/pods/abd5935b-9f5b-4f25-9eec-eeedcb95bc16/volumes" Dec 10 23:08:58 crc kubenswrapper[4791]: I1210 23:08:58.003091 4791 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-default-external-api-0"] Dec 10 23:08:58 crc kubenswrapper[4791]: I1210 23:08:58.339216 4791 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-default-internal-api-0"] Dec 10 23:08:58 crc kubenswrapper[4791]: W1210 23:08:58.343842 4791 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod9460828b_2713_404b_8c14_ca25566937f4.slice/crio-49081a7073620d4096b556ad3cf34629657b4fce3626e6ec53cb34cca27a990c WatchSource:0}: Error finding container 49081a7073620d4096b556ad3cf34629657b4fce3626e6ec53cb34cca27a990c: Status 404 returned error can't find the container with id 49081a7073620d4096b556ad3cf34629657b4fce3626e6ec53cb34cca27a990c Dec 10 23:08:58 crc kubenswrapper[4791]: I1210 23:08:58.858739 4791 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"e51944e5-b91a-4310-9693-5333a680b9ff","Type":"ContainerStarted","Data":"0c8b58f66eadcfcb89b26e96d631131882ba4ffa15fe1e84d9225a05f54bbd28"} Dec 10 23:08:58 crc kubenswrapper[4791]: I1210 23:08:58.864862 4791 generic.go:334] "Generic (PLEG): container finished" podID="b9efeadd-d54b-4f78-a0ad-85cc62321b84" containerID="c8c6f523ef5f1d59b753e569f79692fbe8d9bc4939c81ccec02f0720f1fb8447" exitCode=0 Dec 10 23:08:58 crc kubenswrapper[4791]: I1210 23:08:58.864928 4791 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-db-sync-svgbr" event={"ID":"b9efeadd-d54b-4f78-a0ad-85cc62321b84","Type":"ContainerDied","Data":"c8c6f523ef5f1d59b753e569f79692fbe8d9bc4939c81ccec02f0720f1fb8447"} Dec 10 23:08:58 crc kubenswrapper[4791]: I1210 23:08:58.866258 4791 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-internal-api-0" event={"ID":"9460828b-2713-404b-8c14-ca25566937f4","Type":"ContainerStarted","Data":"49081a7073620d4096b556ad3cf34629657b4fce3626e6ec53cb34cca27a990c"} Dec 10 23:08:59 crc kubenswrapper[4791]: I1210 23:08:59.339700 4791 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-bootstrap-mwqkg" Dec 10 23:08:59 crc kubenswrapper[4791]: I1210 23:08:59.410277 4791 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/horizon-7cc66567bd-pwb2l" Dec 10 23:08:59 crc kubenswrapper[4791]: I1210 23:08:59.410318 4791 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/horizon-7cc66567bd-pwb2l" Dec 10 23:08:59 crc kubenswrapper[4791]: I1210 23:08:59.442587 4791 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/7c65a05d-c2cc-45c4-b469-e6a271e01856-config-data\") pod \"7c65a05d-c2cc-45c4-b469-e6a271e01856\" (UID: \"7c65a05d-c2cc-45c4-b469-e6a271e01856\") " Dec 10 23:08:59 crc kubenswrapper[4791]: I1210 23:08:59.442673 4791 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/7c65a05d-c2cc-45c4-b469-e6a271e01856-fernet-keys\") pod \"7c65a05d-c2cc-45c4-b469-e6a271e01856\" (UID: \"7c65a05d-c2cc-45c4-b469-e6a271e01856\") " Dec 10 23:08:59 crc kubenswrapper[4791]: I1210 23:08:59.442820 4791 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-xzlsz\" (UniqueName: \"kubernetes.io/projected/7c65a05d-c2cc-45c4-b469-e6a271e01856-kube-api-access-xzlsz\") pod \"7c65a05d-c2cc-45c4-b469-e6a271e01856\" (UID: \"7c65a05d-c2cc-45c4-b469-e6a271e01856\") " Dec 10 23:08:59 crc kubenswrapper[4791]: I1210 23:08:59.442874 4791 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/7c65a05d-c2cc-45c4-b469-e6a271e01856-credential-keys\") pod \"7c65a05d-c2cc-45c4-b469-e6a271e01856\" (UID: \"7c65a05d-c2cc-45c4-b469-e6a271e01856\") " Dec 10 23:08:59 crc kubenswrapper[4791]: I1210 23:08:59.442893 4791 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/7c65a05d-c2cc-45c4-b469-e6a271e01856-combined-ca-bundle\") pod \"7c65a05d-c2cc-45c4-b469-e6a271e01856\" (UID: \"7c65a05d-c2cc-45c4-b469-e6a271e01856\") " Dec 10 23:08:59 crc kubenswrapper[4791]: I1210 23:08:59.443377 4791 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/7c65a05d-c2cc-45c4-b469-e6a271e01856-scripts\") pod \"7c65a05d-c2cc-45c4-b469-e6a271e01856\" (UID: \"7c65a05d-c2cc-45c4-b469-e6a271e01856\") " Dec 10 23:08:59 crc kubenswrapper[4791]: I1210 23:08:59.457557 4791 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/7c65a05d-c2cc-45c4-b469-e6a271e01856-credential-keys" (OuterVolumeSpecName: "credential-keys") pod "7c65a05d-c2cc-45c4-b469-e6a271e01856" (UID: "7c65a05d-c2cc-45c4-b469-e6a271e01856"). InnerVolumeSpecName "credential-keys". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 10 23:08:59 crc kubenswrapper[4791]: I1210 23:08:59.457588 4791 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/7c65a05d-c2cc-45c4-b469-e6a271e01856-kube-api-access-xzlsz" (OuterVolumeSpecName: "kube-api-access-xzlsz") pod "7c65a05d-c2cc-45c4-b469-e6a271e01856" (UID: "7c65a05d-c2cc-45c4-b469-e6a271e01856"). InnerVolumeSpecName "kube-api-access-xzlsz". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 10 23:08:59 crc kubenswrapper[4791]: I1210 23:08:59.459423 4791 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/7c65a05d-c2cc-45c4-b469-e6a271e01856-fernet-keys" (OuterVolumeSpecName: "fernet-keys") pod "7c65a05d-c2cc-45c4-b469-e6a271e01856" (UID: "7c65a05d-c2cc-45c4-b469-e6a271e01856"). InnerVolumeSpecName "fernet-keys". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 10 23:08:59 crc kubenswrapper[4791]: I1210 23:08:59.460910 4791 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/7c65a05d-c2cc-45c4-b469-e6a271e01856-scripts" (OuterVolumeSpecName: "scripts") pod "7c65a05d-c2cc-45c4-b469-e6a271e01856" (UID: "7c65a05d-c2cc-45c4-b469-e6a271e01856"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 10 23:08:59 crc kubenswrapper[4791]: I1210 23:08:59.491780 4791 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/horizon-6999c5bf5b-7cr9t" Dec 10 23:08:59 crc kubenswrapper[4791]: I1210 23:08:59.492374 4791 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/horizon-6999c5bf5b-7cr9t" Dec 10 23:08:59 crc kubenswrapper[4791]: I1210 23:08:59.495762 4791 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/7c65a05d-c2cc-45c4-b469-e6a271e01856-config-data" (OuterVolumeSpecName: "config-data") pod "7c65a05d-c2cc-45c4-b469-e6a271e01856" (UID: "7c65a05d-c2cc-45c4-b469-e6a271e01856"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 10 23:08:59 crc kubenswrapper[4791]: I1210 23:08:59.545670 4791 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/7c65a05d-c2cc-45c4-b469-e6a271e01856-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "7c65a05d-c2cc-45c4-b469-e6a271e01856" (UID: "7c65a05d-c2cc-45c4-b469-e6a271e01856"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 10 23:08:59 crc kubenswrapper[4791]: I1210 23:08:59.550046 4791 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-xzlsz\" (UniqueName: \"kubernetes.io/projected/7c65a05d-c2cc-45c4-b469-e6a271e01856-kube-api-access-xzlsz\") on node \"crc\" DevicePath \"\"" Dec 10 23:08:59 crc kubenswrapper[4791]: I1210 23:08:59.550105 4791 reconciler_common.go:293] "Volume detached for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/7c65a05d-c2cc-45c4-b469-e6a271e01856-credential-keys\") on node \"crc\" DevicePath \"\"" Dec 10 23:08:59 crc kubenswrapper[4791]: I1210 23:08:59.550118 4791 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/7c65a05d-c2cc-45c4-b469-e6a271e01856-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 10 23:08:59 crc kubenswrapper[4791]: I1210 23:08:59.550133 4791 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/7c65a05d-c2cc-45c4-b469-e6a271e01856-scripts\") on node \"crc\" DevicePath \"\"" Dec 10 23:08:59 crc kubenswrapper[4791]: I1210 23:08:59.550145 4791 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/7c65a05d-c2cc-45c4-b469-e6a271e01856-config-data\") on node \"crc\" DevicePath \"\"" Dec 10 23:08:59 crc kubenswrapper[4791]: I1210 23:08:59.550156 4791 reconciler_common.go:293] "Volume detached for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/7c65a05d-c2cc-45c4-b469-e6a271e01856-fernet-keys\") on node \"crc\" DevicePath \"\"" Dec 10 23:08:59 crc kubenswrapper[4791]: I1210 23:08:59.881313 4791 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-bootstrap-mwqkg" event={"ID":"7c65a05d-c2cc-45c4-b469-e6a271e01856","Type":"ContainerDied","Data":"3e22738048c400928a86f51ef7f5d400be8b23990d8caa8f9edd0625131b8147"} Dec 10 23:08:59 crc kubenswrapper[4791]: I1210 23:08:59.881467 4791 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-bootstrap-mwqkg" Dec 10 23:08:59 crc kubenswrapper[4791]: I1210 23:08:59.883695 4791 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="3e22738048c400928a86f51ef7f5d400be8b23990d8caa8f9edd0625131b8147" Dec 10 23:09:00 crc kubenswrapper[4791]: I1210 23:09:00.045236 4791 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/keystone-54b7f6dc79-9klwj"] Dec 10 23:09:00 crc kubenswrapper[4791]: E1210 23:09:00.045955 4791 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="7c65a05d-c2cc-45c4-b469-e6a271e01856" containerName="keystone-bootstrap" Dec 10 23:09:00 crc kubenswrapper[4791]: I1210 23:09:00.045974 4791 state_mem.go:107] "Deleted CPUSet assignment" podUID="7c65a05d-c2cc-45c4-b469-e6a271e01856" containerName="keystone-bootstrap" Dec 10 23:09:00 crc kubenswrapper[4791]: I1210 23:09:00.046235 4791 memory_manager.go:354] "RemoveStaleState removing state" podUID="7c65a05d-c2cc-45c4-b469-e6a271e01856" containerName="keystone-bootstrap" Dec 10 23:09:00 crc kubenswrapper[4791]: I1210 23:09:00.046924 4791 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-54b7f6dc79-9klwj" Dec 10 23:09:00 crc kubenswrapper[4791]: I1210 23:09:00.051659 4791 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone" Dec 10 23:09:00 crc kubenswrapper[4791]: I1210 23:09:00.051842 4791 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone-scripts" Dec 10 23:09:00 crc kubenswrapper[4791]: I1210 23:09:00.052050 4791 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone-keystone-dockercfg-t99v8" Dec 10 23:09:00 crc kubenswrapper[4791]: I1210 23:09:00.052165 4791 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-keystone-public-svc" Dec 10 23:09:00 crc kubenswrapper[4791]: I1210 23:09:00.052271 4791 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-keystone-internal-svc" Dec 10 23:09:00 crc kubenswrapper[4791]: I1210 23:09:00.052396 4791 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone-config-data" Dec 10 23:09:00 crc kubenswrapper[4791]: I1210 23:09:00.070118 4791 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/5a5460d8-5939-40d3-8453-a487b3c4b58f-fernet-keys\") pod \"keystone-54b7f6dc79-9klwj\" (UID: \"5a5460d8-5939-40d3-8453-a487b3c4b58f\") " pod="openstack/keystone-54b7f6dc79-9klwj" Dec 10 23:09:00 crc kubenswrapper[4791]: I1210 23:09:00.070164 4791 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/5a5460d8-5939-40d3-8453-a487b3c4b58f-public-tls-certs\") pod \"keystone-54b7f6dc79-9klwj\" (UID: \"5a5460d8-5939-40d3-8453-a487b3c4b58f\") " pod="openstack/keystone-54b7f6dc79-9klwj" Dec 10 23:09:00 crc kubenswrapper[4791]: I1210 23:09:00.070188 4791 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/5a5460d8-5939-40d3-8453-a487b3c4b58f-internal-tls-certs\") pod \"keystone-54b7f6dc79-9klwj\" (UID: \"5a5460d8-5939-40d3-8453-a487b3c4b58f\") " pod="openstack/keystone-54b7f6dc79-9klwj" Dec 10 23:09:00 crc kubenswrapper[4791]: I1210 23:09:00.070234 4791 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-4vczv\" (UniqueName: \"kubernetes.io/projected/5a5460d8-5939-40d3-8453-a487b3c4b58f-kube-api-access-4vczv\") pod \"keystone-54b7f6dc79-9klwj\" (UID: \"5a5460d8-5939-40d3-8453-a487b3c4b58f\") " pod="openstack/keystone-54b7f6dc79-9klwj" Dec 10 23:09:00 crc kubenswrapper[4791]: I1210 23:09:00.070257 4791 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/5a5460d8-5939-40d3-8453-a487b3c4b58f-credential-keys\") pod \"keystone-54b7f6dc79-9klwj\" (UID: \"5a5460d8-5939-40d3-8453-a487b3c4b58f\") " pod="openstack/keystone-54b7f6dc79-9klwj" Dec 10 23:09:00 crc kubenswrapper[4791]: I1210 23:09:00.070281 4791 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/5a5460d8-5939-40d3-8453-a487b3c4b58f-scripts\") pod \"keystone-54b7f6dc79-9klwj\" (UID: \"5a5460d8-5939-40d3-8453-a487b3c4b58f\") " pod="openstack/keystone-54b7f6dc79-9klwj" Dec 10 23:09:00 crc kubenswrapper[4791]: I1210 23:09:00.070315 4791 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/5a5460d8-5939-40d3-8453-a487b3c4b58f-combined-ca-bundle\") pod \"keystone-54b7f6dc79-9klwj\" (UID: \"5a5460d8-5939-40d3-8453-a487b3c4b58f\") " pod="openstack/keystone-54b7f6dc79-9klwj" Dec 10 23:09:00 crc kubenswrapper[4791]: I1210 23:09:00.070353 4791 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/5a5460d8-5939-40d3-8453-a487b3c4b58f-config-data\") pod \"keystone-54b7f6dc79-9klwj\" (UID: \"5a5460d8-5939-40d3-8453-a487b3c4b58f\") " pod="openstack/keystone-54b7f6dc79-9klwj" Dec 10 23:09:00 crc kubenswrapper[4791]: I1210 23:09:00.092857 4791 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/keystone-54b7f6dc79-9klwj"] Dec 10 23:09:00 crc kubenswrapper[4791]: I1210 23:09:00.204797 4791 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/5a5460d8-5939-40d3-8453-a487b3c4b58f-scripts\") pod \"keystone-54b7f6dc79-9klwj\" (UID: \"5a5460d8-5939-40d3-8453-a487b3c4b58f\") " pod="openstack/keystone-54b7f6dc79-9klwj" Dec 10 23:09:00 crc kubenswrapper[4791]: I1210 23:09:00.204865 4791 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/5a5460d8-5939-40d3-8453-a487b3c4b58f-combined-ca-bundle\") pod \"keystone-54b7f6dc79-9klwj\" (UID: \"5a5460d8-5939-40d3-8453-a487b3c4b58f\") " pod="openstack/keystone-54b7f6dc79-9klwj" Dec 10 23:09:00 crc kubenswrapper[4791]: I1210 23:09:00.204902 4791 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/5a5460d8-5939-40d3-8453-a487b3c4b58f-config-data\") pod \"keystone-54b7f6dc79-9klwj\" (UID: \"5a5460d8-5939-40d3-8453-a487b3c4b58f\") " pod="openstack/keystone-54b7f6dc79-9klwj" Dec 10 23:09:00 crc kubenswrapper[4791]: I1210 23:09:00.204979 4791 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/5a5460d8-5939-40d3-8453-a487b3c4b58f-fernet-keys\") pod \"keystone-54b7f6dc79-9klwj\" (UID: \"5a5460d8-5939-40d3-8453-a487b3c4b58f\") " pod="openstack/keystone-54b7f6dc79-9klwj" Dec 10 23:09:00 crc kubenswrapper[4791]: I1210 23:09:00.205000 4791 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/5a5460d8-5939-40d3-8453-a487b3c4b58f-public-tls-certs\") pod \"keystone-54b7f6dc79-9klwj\" (UID: \"5a5460d8-5939-40d3-8453-a487b3c4b58f\") " pod="openstack/keystone-54b7f6dc79-9klwj" Dec 10 23:09:00 crc kubenswrapper[4791]: I1210 23:09:00.205022 4791 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/5a5460d8-5939-40d3-8453-a487b3c4b58f-internal-tls-certs\") pod \"keystone-54b7f6dc79-9klwj\" (UID: \"5a5460d8-5939-40d3-8453-a487b3c4b58f\") " pod="openstack/keystone-54b7f6dc79-9klwj" Dec 10 23:09:00 crc kubenswrapper[4791]: I1210 23:09:00.205060 4791 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-4vczv\" (UniqueName: \"kubernetes.io/projected/5a5460d8-5939-40d3-8453-a487b3c4b58f-kube-api-access-4vczv\") pod \"keystone-54b7f6dc79-9klwj\" (UID: \"5a5460d8-5939-40d3-8453-a487b3c4b58f\") " pod="openstack/keystone-54b7f6dc79-9klwj" Dec 10 23:09:00 crc kubenswrapper[4791]: I1210 23:09:00.205082 4791 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/5a5460d8-5939-40d3-8453-a487b3c4b58f-credential-keys\") pod \"keystone-54b7f6dc79-9klwj\" (UID: \"5a5460d8-5939-40d3-8453-a487b3c4b58f\") " pod="openstack/keystone-54b7f6dc79-9klwj" Dec 10 23:09:00 crc kubenswrapper[4791]: I1210 23:09:00.212805 4791 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/5a5460d8-5939-40d3-8453-a487b3c4b58f-scripts\") pod \"keystone-54b7f6dc79-9klwj\" (UID: \"5a5460d8-5939-40d3-8453-a487b3c4b58f\") " pod="openstack/keystone-54b7f6dc79-9klwj" Dec 10 23:09:00 crc kubenswrapper[4791]: I1210 23:09:00.214609 4791 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/5a5460d8-5939-40d3-8453-a487b3c4b58f-config-data\") pod \"keystone-54b7f6dc79-9klwj\" (UID: \"5a5460d8-5939-40d3-8453-a487b3c4b58f\") " pod="openstack/keystone-54b7f6dc79-9klwj" Dec 10 23:09:00 crc kubenswrapper[4791]: I1210 23:09:00.216253 4791 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/5a5460d8-5939-40d3-8453-a487b3c4b58f-credential-keys\") pod \"keystone-54b7f6dc79-9klwj\" (UID: \"5a5460d8-5939-40d3-8453-a487b3c4b58f\") " pod="openstack/keystone-54b7f6dc79-9klwj" Dec 10 23:09:00 crc kubenswrapper[4791]: I1210 23:09:00.221234 4791 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/5a5460d8-5939-40d3-8453-a487b3c4b58f-combined-ca-bundle\") pod \"keystone-54b7f6dc79-9klwj\" (UID: \"5a5460d8-5939-40d3-8453-a487b3c4b58f\") " pod="openstack/keystone-54b7f6dc79-9klwj" Dec 10 23:09:00 crc kubenswrapper[4791]: I1210 23:09:00.228004 4791 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/5a5460d8-5939-40d3-8453-a487b3c4b58f-internal-tls-certs\") pod \"keystone-54b7f6dc79-9klwj\" (UID: \"5a5460d8-5939-40d3-8453-a487b3c4b58f\") " pod="openstack/keystone-54b7f6dc79-9klwj" Dec 10 23:09:00 crc kubenswrapper[4791]: I1210 23:09:00.231097 4791 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/5a5460d8-5939-40d3-8453-a487b3c4b58f-fernet-keys\") pod \"keystone-54b7f6dc79-9klwj\" (UID: \"5a5460d8-5939-40d3-8453-a487b3c4b58f\") " pod="openstack/keystone-54b7f6dc79-9klwj" Dec 10 23:09:00 crc kubenswrapper[4791]: I1210 23:09:00.243394 4791 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-4vczv\" (UniqueName: \"kubernetes.io/projected/5a5460d8-5939-40d3-8453-a487b3c4b58f-kube-api-access-4vczv\") pod \"keystone-54b7f6dc79-9klwj\" (UID: \"5a5460d8-5939-40d3-8453-a487b3c4b58f\") " pod="openstack/keystone-54b7f6dc79-9klwj" Dec 10 23:09:00 crc kubenswrapper[4791]: I1210 23:09:00.249883 4791 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/5a5460d8-5939-40d3-8453-a487b3c4b58f-public-tls-certs\") pod \"keystone-54b7f6dc79-9klwj\" (UID: \"5a5460d8-5939-40d3-8453-a487b3c4b58f\") " pod="openstack/keystone-54b7f6dc79-9klwj" Dec 10 23:09:00 crc kubenswrapper[4791]: I1210 23:09:00.388131 4791 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-54b7f6dc79-9klwj" Dec 10 23:09:00 crc kubenswrapper[4791]: I1210 23:09:00.642392 4791 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/horizon-78fc5d56dc-58cjt" Dec 10 23:09:00 crc kubenswrapper[4791]: I1210 23:09:00.702037 4791 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/placement-db-sync-svgbr" Dec 10 23:09:00 crc kubenswrapper[4791]: I1210 23:09:00.837624 4791 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/b9efeadd-d54b-4f78-a0ad-85cc62321b84-logs\") pod \"b9efeadd-d54b-4f78-a0ad-85cc62321b84\" (UID: \"b9efeadd-d54b-4f78-a0ad-85cc62321b84\") " Dec 10 23:09:00 crc kubenswrapper[4791]: I1210 23:09:00.837992 4791 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/b9efeadd-d54b-4f78-a0ad-85cc62321b84-scripts\") pod \"b9efeadd-d54b-4f78-a0ad-85cc62321b84\" (UID: \"b9efeadd-d54b-4f78-a0ad-85cc62321b84\") " Dec 10 23:09:00 crc kubenswrapper[4791]: I1210 23:09:00.838018 4791 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-q27r9\" (UniqueName: \"kubernetes.io/projected/b9efeadd-d54b-4f78-a0ad-85cc62321b84-kube-api-access-q27r9\") pod \"b9efeadd-d54b-4f78-a0ad-85cc62321b84\" (UID: \"b9efeadd-d54b-4f78-a0ad-85cc62321b84\") " Dec 10 23:09:00 crc kubenswrapper[4791]: I1210 23:09:00.838113 4791 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/b9efeadd-d54b-4f78-a0ad-85cc62321b84-config-data\") pod \"b9efeadd-d54b-4f78-a0ad-85cc62321b84\" (UID: \"b9efeadd-d54b-4f78-a0ad-85cc62321b84\") " Dec 10 23:09:00 crc kubenswrapper[4791]: I1210 23:09:00.838154 4791 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/b9efeadd-d54b-4f78-a0ad-85cc62321b84-combined-ca-bundle\") pod \"b9efeadd-d54b-4f78-a0ad-85cc62321b84\" (UID: \"b9efeadd-d54b-4f78-a0ad-85cc62321b84\") " Dec 10 23:09:00 crc kubenswrapper[4791]: I1210 23:09:00.844543 4791 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/b9efeadd-d54b-4f78-a0ad-85cc62321b84-logs" (OuterVolumeSpecName: "logs") pod "b9efeadd-d54b-4f78-a0ad-85cc62321b84" (UID: "b9efeadd-d54b-4f78-a0ad-85cc62321b84"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 10 23:09:00 crc kubenswrapper[4791]: I1210 23:09:00.857590 4791 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/b9efeadd-d54b-4f78-a0ad-85cc62321b84-kube-api-access-q27r9" (OuterVolumeSpecName: "kube-api-access-q27r9") pod "b9efeadd-d54b-4f78-a0ad-85cc62321b84" (UID: "b9efeadd-d54b-4f78-a0ad-85cc62321b84"). InnerVolumeSpecName "kube-api-access-q27r9". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 10 23:09:00 crc kubenswrapper[4791]: I1210 23:09:00.859724 4791 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/b9efeadd-d54b-4f78-a0ad-85cc62321b84-scripts" (OuterVolumeSpecName: "scripts") pod "b9efeadd-d54b-4f78-a0ad-85cc62321b84" (UID: "b9efeadd-d54b-4f78-a0ad-85cc62321b84"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 10 23:09:00 crc kubenswrapper[4791]: I1210 23:09:00.896501 4791 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/b9efeadd-d54b-4f78-a0ad-85cc62321b84-config-data" (OuterVolumeSpecName: "config-data") pod "b9efeadd-d54b-4f78-a0ad-85cc62321b84" (UID: "b9efeadd-d54b-4f78-a0ad-85cc62321b84"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 10 23:09:00 crc kubenswrapper[4791]: I1210 23:09:00.915310 4791 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/b9efeadd-d54b-4f78-a0ad-85cc62321b84-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "b9efeadd-d54b-4f78-a0ad-85cc62321b84" (UID: "b9efeadd-d54b-4f78-a0ad-85cc62321b84"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 10 23:09:00 crc kubenswrapper[4791]: I1210 23:09:00.940391 4791 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/b9efeadd-d54b-4f78-a0ad-85cc62321b84-scripts\") on node \"crc\" DevicePath \"\"" Dec 10 23:09:00 crc kubenswrapper[4791]: I1210 23:09:00.940418 4791 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-q27r9\" (UniqueName: \"kubernetes.io/projected/b9efeadd-d54b-4f78-a0ad-85cc62321b84-kube-api-access-q27r9\") on node \"crc\" DevicePath \"\"" Dec 10 23:09:00 crc kubenswrapper[4791]: I1210 23:09:00.940427 4791 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/b9efeadd-d54b-4f78-a0ad-85cc62321b84-config-data\") on node \"crc\" DevicePath \"\"" Dec 10 23:09:00 crc kubenswrapper[4791]: I1210 23:09:00.940436 4791 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/b9efeadd-d54b-4f78-a0ad-85cc62321b84-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 10 23:09:00 crc kubenswrapper[4791]: I1210 23:09:00.940445 4791 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/b9efeadd-d54b-4f78-a0ad-85cc62321b84-logs\") on node \"crc\" DevicePath \"\"" Dec 10 23:09:00 crc kubenswrapper[4791]: I1210 23:09:00.953682 4791 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-internal-api-0" event={"ID":"9460828b-2713-404b-8c14-ca25566937f4","Type":"ContainerStarted","Data":"521f2aafeb6650480862ba7b822c1a7b87190962da901d6b233d82ced08d46f6"} Dec 10 23:09:00 crc kubenswrapper[4791]: I1210 23:09:00.955607 4791 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"e51944e5-b91a-4310-9693-5333a680b9ff","Type":"ContainerStarted","Data":"410370d8e7a6e6b4006e81059ec37dab14f19eb22582c9d4426ad2345c56d96e"} Dec 10 23:09:00 crc kubenswrapper[4791]: I1210 23:09:00.966645 4791 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-db-sync-svgbr" event={"ID":"b9efeadd-d54b-4f78-a0ad-85cc62321b84","Type":"ContainerDied","Data":"539bc24935d590728b85d4263f0e96b5faf6637750a2619fa9490a40c7dbf5c6"} Dec 10 23:09:00 crc kubenswrapper[4791]: I1210 23:09:00.966684 4791 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="539bc24935d590728b85d4263f0e96b5faf6637750a2619fa9490a40c7dbf5c6" Dec 10 23:09:00 crc kubenswrapper[4791]: I1210 23:09:00.966747 4791 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/placement-db-sync-svgbr" Dec 10 23:09:01 crc kubenswrapper[4791]: I1210 23:09:01.047821 4791 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/placement-6487596ddd-btmd7"] Dec 10 23:09:01 crc kubenswrapper[4791]: E1210 23:09:01.048217 4791 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="b9efeadd-d54b-4f78-a0ad-85cc62321b84" containerName="placement-db-sync" Dec 10 23:09:01 crc kubenswrapper[4791]: I1210 23:09:01.048229 4791 state_mem.go:107] "Deleted CPUSet assignment" podUID="b9efeadd-d54b-4f78-a0ad-85cc62321b84" containerName="placement-db-sync" Dec 10 23:09:01 crc kubenswrapper[4791]: I1210 23:09:01.057458 4791 memory_manager.go:354] "RemoveStaleState removing state" podUID="b9efeadd-d54b-4f78-a0ad-85cc62321b84" containerName="placement-db-sync" Dec 10 23:09:01 crc kubenswrapper[4791]: I1210 23:09:01.120065 4791 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/horizon-5f54ccbf57-wwpfk" Dec 10 23:09:01 crc kubenswrapper[4791]: I1210 23:09:01.120182 4791 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/placement-6487596ddd-btmd7" Dec 10 23:09:01 crc kubenswrapper[4791]: I1210 23:09:01.127498 4791 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"placement-placement-dockercfg-wklr7" Dec 10 23:09:01 crc kubenswrapper[4791]: I1210 23:09:01.127979 4791 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"placement-scripts" Dec 10 23:09:01 crc kubenswrapper[4791]: I1210 23:09:01.128160 4791 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-placement-public-svc" Dec 10 23:09:01 crc kubenswrapper[4791]: I1210 23:09:01.128392 4791 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"placement-config-data" Dec 10 23:09:01 crc kubenswrapper[4791]: I1210 23:09:01.128385 4791 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-placement-internal-svc" Dec 10 23:09:01 crc kubenswrapper[4791]: I1210 23:09:01.142630 4791 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/placement-6487596ddd-btmd7"] Dec 10 23:09:01 crc kubenswrapper[4791]: I1210 23:09:01.187209 4791 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/keystone-54b7f6dc79-9klwj"] Dec 10 23:09:01 crc kubenswrapper[4791]: W1210 23:09:01.211672 4791 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod5a5460d8_5939_40d3_8453_a487b3c4b58f.slice/crio-634d7ab0fcca42fb43382524bc4473dfd54583baa466484f8bf4823a8355e99a WatchSource:0}: Error finding container 634d7ab0fcca42fb43382524bc4473dfd54583baa466484f8bf4823a8355e99a: Status 404 returned error can't find the container with id 634d7ab0fcca42fb43382524bc4473dfd54583baa466484f8bf4823a8355e99a Dec 10 23:09:01 crc kubenswrapper[4791]: I1210 23:09:01.268396 4791 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/0babf4ec-03b9-48de-875d-a27a8b7ed119-config-data\") pod \"placement-6487596ddd-btmd7\" (UID: \"0babf4ec-03b9-48de-875d-a27a8b7ed119\") " pod="openstack/placement-6487596ddd-btmd7" Dec 10 23:09:01 crc kubenswrapper[4791]: I1210 23:09:01.268481 4791 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/0babf4ec-03b9-48de-875d-a27a8b7ed119-internal-tls-certs\") pod \"placement-6487596ddd-btmd7\" (UID: \"0babf4ec-03b9-48de-875d-a27a8b7ed119\") " pod="openstack/placement-6487596ddd-btmd7" Dec 10 23:09:01 crc kubenswrapper[4791]: I1210 23:09:01.268527 4791 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/0babf4ec-03b9-48de-875d-a27a8b7ed119-scripts\") pod \"placement-6487596ddd-btmd7\" (UID: \"0babf4ec-03b9-48de-875d-a27a8b7ed119\") " pod="openstack/placement-6487596ddd-btmd7" Dec 10 23:09:01 crc kubenswrapper[4791]: I1210 23:09:01.268566 4791 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-bhwbf\" (UniqueName: \"kubernetes.io/projected/0babf4ec-03b9-48de-875d-a27a8b7ed119-kube-api-access-bhwbf\") pod \"placement-6487596ddd-btmd7\" (UID: \"0babf4ec-03b9-48de-875d-a27a8b7ed119\") " pod="openstack/placement-6487596ddd-btmd7" Dec 10 23:09:01 crc kubenswrapper[4791]: I1210 23:09:01.268654 4791 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/0babf4ec-03b9-48de-875d-a27a8b7ed119-public-tls-certs\") pod \"placement-6487596ddd-btmd7\" (UID: \"0babf4ec-03b9-48de-875d-a27a8b7ed119\") " pod="openstack/placement-6487596ddd-btmd7" Dec 10 23:09:01 crc kubenswrapper[4791]: I1210 23:09:01.268721 4791 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/0babf4ec-03b9-48de-875d-a27a8b7ed119-logs\") pod \"placement-6487596ddd-btmd7\" (UID: \"0babf4ec-03b9-48de-875d-a27a8b7ed119\") " pod="openstack/placement-6487596ddd-btmd7" Dec 10 23:09:01 crc kubenswrapper[4791]: I1210 23:09:01.268750 4791 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/0babf4ec-03b9-48de-875d-a27a8b7ed119-combined-ca-bundle\") pod \"placement-6487596ddd-btmd7\" (UID: \"0babf4ec-03b9-48de-875d-a27a8b7ed119\") " pod="openstack/placement-6487596ddd-btmd7" Dec 10 23:09:01 crc kubenswrapper[4791]: I1210 23:09:01.373552 4791 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/0babf4ec-03b9-48de-875d-a27a8b7ed119-public-tls-certs\") pod \"placement-6487596ddd-btmd7\" (UID: \"0babf4ec-03b9-48de-875d-a27a8b7ed119\") " pod="openstack/placement-6487596ddd-btmd7" Dec 10 23:09:01 crc kubenswrapper[4791]: I1210 23:09:01.373651 4791 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/0babf4ec-03b9-48de-875d-a27a8b7ed119-logs\") pod \"placement-6487596ddd-btmd7\" (UID: \"0babf4ec-03b9-48de-875d-a27a8b7ed119\") " pod="openstack/placement-6487596ddd-btmd7" Dec 10 23:09:01 crc kubenswrapper[4791]: I1210 23:09:01.373676 4791 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/0babf4ec-03b9-48de-875d-a27a8b7ed119-combined-ca-bundle\") pod \"placement-6487596ddd-btmd7\" (UID: \"0babf4ec-03b9-48de-875d-a27a8b7ed119\") " pod="openstack/placement-6487596ddd-btmd7" Dec 10 23:09:01 crc kubenswrapper[4791]: I1210 23:09:01.373743 4791 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/0babf4ec-03b9-48de-875d-a27a8b7ed119-config-data\") pod \"placement-6487596ddd-btmd7\" (UID: \"0babf4ec-03b9-48de-875d-a27a8b7ed119\") " pod="openstack/placement-6487596ddd-btmd7" Dec 10 23:09:01 crc kubenswrapper[4791]: I1210 23:09:01.373806 4791 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/0babf4ec-03b9-48de-875d-a27a8b7ed119-internal-tls-certs\") pod \"placement-6487596ddd-btmd7\" (UID: \"0babf4ec-03b9-48de-875d-a27a8b7ed119\") " pod="openstack/placement-6487596ddd-btmd7" Dec 10 23:09:01 crc kubenswrapper[4791]: I1210 23:09:01.373830 4791 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/0babf4ec-03b9-48de-875d-a27a8b7ed119-scripts\") pod \"placement-6487596ddd-btmd7\" (UID: \"0babf4ec-03b9-48de-875d-a27a8b7ed119\") " pod="openstack/placement-6487596ddd-btmd7" Dec 10 23:09:01 crc kubenswrapper[4791]: I1210 23:09:01.373867 4791 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-bhwbf\" (UniqueName: \"kubernetes.io/projected/0babf4ec-03b9-48de-875d-a27a8b7ed119-kube-api-access-bhwbf\") pod \"placement-6487596ddd-btmd7\" (UID: \"0babf4ec-03b9-48de-875d-a27a8b7ed119\") " pod="openstack/placement-6487596ddd-btmd7" Dec 10 23:09:01 crc kubenswrapper[4791]: I1210 23:09:01.378197 4791 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/0babf4ec-03b9-48de-875d-a27a8b7ed119-logs\") pod \"placement-6487596ddd-btmd7\" (UID: \"0babf4ec-03b9-48de-875d-a27a8b7ed119\") " pod="openstack/placement-6487596ddd-btmd7" Dec 10 23:09:01 crc kubenswrapper[4791]: I1210 23:09:01.381843 4791 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/0babf4ec-03b9-48de-875d-a27a8b7ed119-combined-ca-bundle\") pod \"placement-6487596ddd-btmd7\" (UID: \"0babf4ec-03b9-48de-875d-a27a8b7ed119\") " pod="openstack/placement-6487596ddd-btmd7" Dec 10 23:09:01 crc kubenswrapper[4791]: I1210 23:09:01.382882 4791 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/0babf4ec-03b9-48de-875d-a27a8b7ed119-config-data\") pod \"placement-6487596ddd-btmd7\" (UID: \"0babf4ec-03b9-48de-875d-a27a8b7ed119\") " pod="openstack/placement-6487596ddd-btmd7" Dec 10 23:09:01 crc kubenswrapper[4791]: I1210 23:09:01.384399 4791 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/0babf4ec-03b9-48de-875d-a27a8b7ed119-public-tls-certs\") pod \"placement-6487596ddd-btmd7\" (UID: \"0babf4ec-03b9-48de-875d-a27a8b7ed119\") " pod="openstack/placement-6487596ddd-btmd7" Dec 10 23:09:01 crc kubenswrapper[4791]: I1210 23:09:01.394143 4791 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/0babf4ec-03b9-48de-875d-a27a8b7ed119-scripts\") pod \"placement-6487596ddd-btmd7\" (UID: \"0babf4ec-03b9-48de-875d-a27a8b7ed119\") " pod="openstack/placement-6487596ddd-btmd7" Dec 10 23:09:01 crc kubenswrapper[4791]: I1210 23:09:01.409130 4791 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/0babf4ec-03b9-48de-875d-a27a8b7ed119-internal-tls-certs\") pod \"placement-6487596ddd-btmd7\" (UID: \"0babf4ec-03b9-48de-875d-a27a8b7ed119\") " pod="openstack/placement-6487596ddd-btmd7" Dec 10 23:09:01 crc kubenswrapper[4791]: I1210 23:09:01.441167 4791 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-bhwbf\" (UniqueName: \"kubernetes.io/projected/0babf4ec-03b9-48de-875d-a27a8b7ed119-kube-api-access-bhwbf\") pod \"placement-6487596ddd-btmd7\" (UID: \"0babf4ec-03b9-48de-875d-a27a8b7ed119\") " pod="openstack/placement-6487596ddd-btmd7" Dec 10 23:09:01 crc kubenswrapper[4791]: I1210 23:09:01.462738 4791 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/placement-6487596ddd-btmd7" Dec 10 23:09:01 crc kubenswrapper[4791]: I1210 23:09:01.914670 4791 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/dnsmasq-dns-84b966f6c9-tz266" Dec 10 23:09:01 crc kubenswrapper[4791]: I1210 23:09:01.986427 4791 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-76fcf4b695-r7wvb"] Dec 10 23:09:01 crc kubenswrapper[4791]: I1210 23:09:01.986667 4791 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/dnsmasq-dns-76fcf4b695-r7wvb" podUID="00320ef6-9362-427b-a2eb-12329d652a0e" containerName="dnsmasq-dns" containerID="cri-o://f6ba39c429e2e3771ef8f6a8d0b8099d2c828d474f01bd5a46f1fd73a72e07ab" gracePeriod=10 Dec 10 23:09:02 crc kubenswrapper[4791]: I1210 23:09:02.019135 4791 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-54b7f6dc79-9klwj" event={"ID":"5a5460d8-5939-40d3-8453-a487b3c4b58f","Type":"ContainerStarted","Data":"634d7ab0fcca42fb43382524bc4473dfd54583baa466484f8bf4823a8355e99a"} Dec 10 23:09:02 crc kubenswrapper[4791]: I1210 23:09:02.145756 4791 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/placement-6487596ddd-btmd7"] Dec 10 23:09:02 crc kubenswrapper[4791]: W1210 23:09:02.156118 4791 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod0babf4ec_03b9_48de_875d_a27a8b7ed119.slice/crio-f523666d39427e5b76961de0fdbee7b725c540c4daa5d65ffde13e6969a82811 WatchSource:0}: Error finding container f523666d39427e5b76961de0fdbee7b725c540c4daa5d65ffde13e6969a82811: Status 404 returned error can't find the container with id f523666d39427e5b76961de0fdbee7b725c540c4daa5d65ffde13e6969a82811 Dec 10 23:09:03 crc kubenswrapper[4791]: I1210 23:09:03.099308 4791 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"e51944e5-b91a-4310-9693-5333a680b9ff","Type":"ContainerStarted","Data":"192f172550e5b82eaddebad8f678283f7d534c4c29b13ec6c9004f66e467599e"} Dec 10 23:09:03 crc kubenswrapper[4791]: I1210 23:09:03.130565 4791 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-internal-api-0" event={"ID":"9460828b-2713-404b-8c14-ca25566937f4","Type":"ContainerStarted","Data":"b4f9dd4411c75b64e8ac6561f42e13b64709dd0bfd4f266b0d540210a7768a27"} Dec 10 23:09:03 crc kubenswrapper[4791]: I1210 23:09:03.181372 4791 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/glance-default-external-api-0" podStartSLOduration=7.18135261 podStartE2EDuration="7.18135261s" podCreationTimestamp="2025-12-10 23:08:56 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-10 23:09:03.163886146 +0000 UTC m=+1177.593503759" watchObservedRunningTime="2025-12-10 23:09:03.18135261 +0000 UTC m=+1177.610970223" Dec 10 23:09:03 crc kubenswrapper[4791]: I1210 23:09:03.194601 4791 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-54b7f6dc79-9klwj" event={"ID":"5a5460d8-5939-40d3-8453-a487b3c4b58f","Type":"ContainerStarted","Data":"7951607ab9b918b32ff46fb4c3fef669da5e15ec58d00ed0d872f49a4cef08bc"} Dec 10 23:09:03 crc kubenswrapper[4791]: I1210 23:09:03.195402 4791 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/keystone-54b7f6dc79-9klwj" Dec 10 23:09:03 crc kubenswrapper[4791]: I1210 23:09:03.226654 4791 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-6487596ddd-btmd7" event={"ID":"0babf4ec-03b9-48de-875d-a27a8b7ed119","Type":"ContainerStarted","Data":"3ca9a060853feb3e44b1625befe9b6238d3093de4bfa0d480777c90f89ce42c5"} Dec 10 23:09:03 crc kubenswrapper[4791]: I1210 23:09:03.226907 4791 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-6487596ddd-btmd7" event={"ID":"0babf4ec-03b9-48de-875d-a27a8b7ed119","Type":"ContainerStarted","Data":"f523666d39427e5b76961de0fdbee7b725c540c4daa5d65ffde13e6969a82811"} Dec 10 23:09:03 crc kubenswrapper[4791]: I1210 23:09:03.227729 4791 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/placement-6487596ddd-btmd7" Dec 10 23:09:03 crc kubenswrapper[4791]: I1210 23:09:03.227830 4791 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/placement-6487596ddd-btmd7" Dec 10 23:09:03 crc kubenswrapper[4791]: I1210 23:09:03.257750 4791 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/glance-default-internal-api-0" podStartSLOduration=7.257727085 podStartE2EDuration="7.257727085s" podCreationTimestamp="2025-12-10 23:08:56 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-10 23:09:03.232567629 +0000 UTC m=+1177.662185252" watchObservedRunningTime="2025-12-10 23:09:03.257727085 +0000 UTC m=+1177.687344698" Dec 10 23:09:03 crc kubenswrapper[4791]: I1210 23:09:03.258293 4791 generic.go:334] "Generic (PLEG): container finished" podID="00320ef6-9362-427b-a2eb-12329d652a0e" containerID="f6ba39c429e2e3771ef8f6a8d0b8099d2c828d474f01bd5a46f1fd73a72e07ab" exitCode=0 Dec 10 23:09:03 crc kubenswrapper[4791]: I1210 23:09:03.258465 4791 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-76fcf4b695-r7wvb" event={"ID":"00320ef6-9362-427b-a2eb-12329d652a0e","Type":"ContainerDied","Data":"f6ba39c429e2e3771ef8f6a8d0b8099d2c828d474f01bd5a46f1fd73a72e07ab"} Dec 10 23:09:03 crc kubenswrapper[4791]: I1210 23:09:03.282569 4791 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-db-sync-g4s5p" event={"ID":"a85df2ed-be84-41b3-8993-c3d963daef08","Type":"ContainerStarted","Data":"8fdda0940a6cb89ce941ba64df5df3bda9bee4bd20519bf14e86899f148c962d"} Dec 10 23:09:03 crc kubenswrapper[4791]: I1210 23:09:03.286847 4791 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/placement-6487596ddd-btmd7" podStartSLOduration=2.286827495 podStartE2EDuration="2.286827495s" podCreationTimestamp="2025-12-10 23:09:01 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-10 23:09:03.282383377 +0000 UTC m=+1177.712000990" watchObservedRunningTime="2025-12-10 23:09:03.286827495 +0000 UTC m=+1177.716445108" Dec 10 23:09:03 crc kubenswrapper[4791]: I1210 23:09:03.391017 4791 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/cinder-db-sync-g4s5p" podStartSLOduration=4.337122167 podStartE2EDuration="43.390993192s" podCreationTimestamp="2025-12-10 23:08:20 +0000 UTC" firstStartedPulling="2025-12-10 23:08:22.014774918 +0000 UTC m=+1136.444392531" lastFinishedPulling="2025-12-10 23:09:01.068645943 +0000 UTC m=+1175.498263556" observedRunningTime="2025-12-10 23:09:03.351129561 +0000 UTC m=+1177.780747174" watchObservedRunningTime="2025-12-10 23:09:03.390993192 +0000 UTC m=+1177.820610805" Dec 10 23:09:03 crc kubenswrapper[4791]: I1210 23:09:03.392833 4791 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/keystone-54b7f6dc79-9klwj" podStartSLOduration=4.392824225 podStartE2EDuration="4.392824225s" podCreationTimestamp="2025-12-10 23:08:59 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-10 23:09:03.324689148 +0000 UTC m=+1177.754306761" watchObservedRunningTime="2025-12-10 23:09:03.392824225 +0000 UTC m=+1177.822441838" Dec 10 23:09:04 crc kubenswrapper[4791]: I1210 23:09:04.302661 4791 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-6487596ddd-btmd7" event={"ID":"0babf4ec-03b9-48de-875d-a27a8b7ed119","Type":"ContainerStarted","Data":"a6e2b542bf913adb5d7074120583ff585ee4e83cf88f65d05939d5f6539ddb2f"} Dec 10 23:09:07 crc kubenswrapper[4791]: I1210 23:09:07.297699 4791 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/glance-default-external-api-0" Dec 10 23:09:07 crc kubenswrapper[4791]: I1210 23:09:07.297765 4791 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/glance-default-external-api-0" Dec 10 23:09:07 crc kubenswrapper[4791]: I1210 23:09:07.356649 4791 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/glance-default-external-api-0" Dec 10 23:09:07 crc kubenswrapper[4791]: I1210 23:09:07.358940 4791 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/glance-default-external-api-0" Dec 10 23:09:07 crc kubenswrapper[4791]: I1210 23:09:07.359104 4791 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/glance-default-external-api-0" Dec 10 23:09:07 crc kubenswrapper[4791]: I1210 23:09:07.449803 4791 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-76fcf4b695-r7wvb" Dec 10 23:09:07 crc kubenswrapper[4791]: I1210 23:09:07.548629 4791 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/00320ef6-9362-427b-a2eb-12329d652a0e-ovsdbserver-nb\") pod \"00320ef6-9362-427b-a2eb-12329d652a0e\" (UID: \"00320ef6-9362-427b-a2eb-12329d652a0e\") " Dec 10 23:09:07 crc kubenswrapper[4791]: I1210 23:09:07.548683 4791 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/00320ef6-9362-427b-a2eb-12329d652a0e-ovsdbserver-sb\") pod \"00320ef6-9362-427b-a2eb-12329d652a0e\" (UID: \"00320ef6-9362-427b-a2eb-12329d652a0e\") " Dec 10 23:09:07 crc kubenswrapper[4791]: I1210 23:09:07.548749 4791 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-x949f\" (UniqueName: \"kubernetes.io/projected/00320ef6-9362-427b-a2eb-12329d652a0e-kube-api-access-x949f\") pod \"00320ef6-9362-427b-a2eb-12329d652a0e\" (UID: \"00320ef6-9362-427b-a2eb-12329d652a0e\") " Dec 10 23:09:07 crc kubenswrapper[4791]: I1210 23:09:07.548789 4791 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/00320ef6-9362-427b-a2eb-12329d652a0e-dns-svc\") pod \"00320ef6-9362-427b-a2eb-12329d652a0e\" (UID: \"00320ef6-9362-427b-a2eb-12329d652a0e\") " Dec 10 23:09:07 crc kubenswrapper[4791]: I1210 23:09:07.548837 4791 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/00320ef6-9362-427b-a2eb-12329d652a0e-config\") pod \"00320ef6-9362-427b-a2eb-12329d652a0e\" (UID: \"00320ef6-9362-427b-a2eb-12329d652a0e\") " Dec 10 23:09:07 crc kubenswrapper[4791]: I1210 23:09:07.548889 4791 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/00320ef6-9362-427b-a2eb-12329d652a0e-dns-swift-storage-0\") pod \"00320ef6-9362-427b-a2eb-12329d652a0e\" (UID: \"00320ef6-9362-427b-a2eb-12329d652a0e\") " Dec 10 23:09:07 crc kubenswrapper[4791]: I1210 23:09:07.567539 4791 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/00320ef6-9362-427b-a2eb-12329d652a0e-kube-api-access-x949f" (OuterVolumeSpecName: "kube-api-access-x949f") pod "00320ef6-9362-427b-a2eb-12329d652a0e" (UID: "00320ef6-9362-427b-a2eb-12329d652a0e"). InnerVolumeSpecName "kube-api-access-x949f". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 10 23:09:07 crc kubenswrapper[4791]: I1210 23:09:07.615197 4791 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/00320ef6-9362-427b-a2eb-12329d652a0e-ovsdbserver-nb" (OuterVolumeSpecName: "ovsdbserver-nb") pod "00320ef6-9362-427b-a2eb-12329d652a0e" (UID: "00320ef6-9362-427b-a2eb-12329d652a0e"). InnerVolumeSpecName "ovsdbserver-nb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 10 23:09:07 crc kubenswrapper[4791]: I1210 23:09:07.633795 4791 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/00320ef6-9362-427b-a2eb-12329d652a0e-config" (OuterVolumeSpecName: "config") pod "00320ef6-9362-427b-a2eb-12329d652a0e" (UID: "00320ef6-9362-427b-a2eb-12329d652a0e"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 10 23:09:07 crc kubenswrapper[4791]: I1210 23:09:07.648115 4791 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/00320ef6-9362-427b-a2eb-12329d652a0e-dns-swift-storage-0" (OuterVolumeSpecName: "dns-swift-storage-0") pod "00320ef6-9362-427b-a2eb-12329d652a0e" (UID: "00320ef6-9362-427b-a2eb-12329d652a0e"). InnerVolumeSpecName "dns-swift-storage-0". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 10 23:09:07 crc kubenswrapper[4791]: I1210 23:09:07.650849 4791 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/00320ef6-9362-427b-a2eb-12329d652a0e-ovsdbserver-nb\") on node \"crc\" DevicePath \"\"" Dec 10 23:09:07 crc kubenswrapper[4791]: I1210 23:09:07.650871 4791 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-x949f\" (UniqueName: \"kubernetes.io/projected/00320ef6-9362-427b-a2eb-12329d652a0e-kube-api-access-x949f\") on node \"crc\" DevicePath \"\"" Dec 10 23:09:07 crc kubenswrapper[4791]: I1210 23:09:07.650884 4791 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/00320ef6-9362-427b-a2eb-12329d652a0e-config\") on node \"crc\" DevicePath \"\"" Dec 10 23:09:07 crc kubenswrapper[4791]: I1210 23:09:07.650894 4791 reconciler_common.go:293] "Volume detached for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/00320ef6-9362-427b-a2eb-12329d652a0e-dns-swift-storage-0\") on node \"crc\" DevicePath \"\"" Dec 10 23:09:07 crc kubenswrapper[4791]: I1210 23:09:07.651538 4791 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/00320ef6-9362-427b-a2eb-12329d652a0e-ovsdbserver-sb" (OuterVolumeSpecName: "ovsdbserver-sb") pod "00320ef6-9362-427b-a2eb-12329d652a0e" (UID: "00320ef6-9362-427b-a2eb-12329d652a0e"). InnerVolumeSpecName "ovsdbserver-sb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 10 23:09:07 crc kubenswrapper[4791]: I1210 23:09:07.660796 4791 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/glance-default-internal-api-0" Dec 10 23:09:07 crc kubenswrapper[4791]: I1210 23:09:07.660940 4791 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/glance-default-internal-api-0" Dec 10 23:09:07 crc kubenswrapper[4791]: I1210 23:09:07.683644 4791 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/00320ef6-9362-427b-a2eb-12329d652a0e-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "00320ef6-9362-427b-a2eb-12329d652a0e" (UID: "00320ef6-9362-427b-a2eb-12329d652a0e"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 10 23:09:07 crc kubenswrapper[4791]: I1210 23:09:07.699392 4791 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/glance-default-internal-api-0" Dec 10 23:09:07 crc kubenswrapper[4791]: I1210 23:09:07.725992 4791 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/glance-default-internal-api-0" Dec 10 23:09:07 crc kubenswrapper[4791]: I1210 23:09:07.753879 4791 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/00320ef6-9362-427b-a2eb-12329d652a0e-ovsdbserver-sb\") on node \"crc\" DevicePath \"\"" Dec 10 23:09:07 crc kubenswrapper[4791]: I1210 23:09:07.753915 4791 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/00320ef6-9362-427b-a2eb-12329d652a0e-dns-svc\") on node \"crc\" DevicePath \"\"" Dec 10 23:09:08 crc kubenswrapper[4791]: I1210 23:09:08.345081 4791 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-db-sync-zllvm" event={"ID":"851556d7-fa38-4eec-ae74-de989d66d27b","Type":"ContainerStarted","Data":"96fce98c2678503c67e20f58365f73a09244ac81b06aa07e100ddbf7bb1368d9"} Dec 10 23:09:08 crc kubenswrapper[4791]: I1210 23:09:08.356751 4791 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"8eaa1b1d-715f-4059-ad18-fe9632c036a7","Type":"ContainerStarted","Data":"29aea3c141a06d60106b8924f521c5d11243d56785dfbc6ff7d120f58e023588"} Dec 10 23:09:08 crc kubenswrapper[4791]: I1210 23:09:08.377566 4791 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-76fcf4b695-r7wvb" Dec 10 23:09:08 crc kubenswrapper[4791]: I1210 23:09:08.377573 4791 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-76fcf4b695-r7wvb" event={"ID":"00320ef6-9362-427b-a2eb-12329d652a0e","Type":"ContainerDied","Data":"89a0427f33bd75ec7105ab124933fbf02951e2bcaf48b56a70dfc6beac2d88c3"} Dec 10 23:09:08 crc kubenswrapper[4791]: I1210 23:09:08.378106 4791 scope.go:117] "RemoveContainer" containerID="f6ba39c429e2e3771ef8f6a8d0b8099d2c828d474f01bd5a46f1fd73a72e07ab" Dec 10 23:09:08 crc kubenswrapper[4791]: I1210 23:09:08.384982 4791 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/barbican-db-sync-zllvm" podStartSLOduration=2.918957197 podStartE2EDuration="48.384963484s" podCreationTimestamp="2025-12-10 23:08:20 +0000 UTC" firstStartedPulling="2025-12-10 23:08:21.986815909 +0000 UTC m=+1136.416433532" lastFinishedPulling="2025-12-10 23:09:07.452822166 +0000 UTC m=+1181.882439819" observedRunningTime="2025-12-10 23:09:08.367766658 +0000 UTC m=+1182.797384301" watchObservedRunningTime="2025-12-10 23:09:08.384963484 +0000 UTC m=+1182.814581117" Dec 10 23:09:08 crc kubenswrapper[4791]: I1210 23:09:08.386823 4791 generic.go:334] "Generic (PLEG): container finished" podID="a85df2ed-be84-41b3-8993-c3d963daef08" containerID="8fdda0940a6cb89ce941ba64df5df3bda9bee4bd20519bf14e86899f148c962d" exitCode=0 Dec 10 23:09:08 crc kubenswrapper[4791]: I1210 23:09:08.386972 4791 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-db-sync-g4s5p" event={"ID":"a85df2ed-be84-41b3-8993-c3d963daef08","Type":"ContainerDied","Data":"8fdda0940a6cb89ce941ba64df5df3bda9bee4bd20519bf14e86899f148c962d"} Dec 10 23:09:08 crc kubenswrapper[4791]: I1210 23:09:08.387767 4791 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/glance-default-internal-api-0" Dec 10 23:09:08 crc kubenswrapper[4791]: I1210 23:09:08.387969 4791 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/glance-default-external-api-0" Dec 10 23:09:08 crc kubenswrapper[4791]: I1210 23:09:08.388439 4791 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/glance-default-internal-api-0" Dec 10 23:09:08 crc kubenswrapper[4791]: I1210 23:09:08.412888 4791 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-76fcf4b695-r7wvb"] Dec 10 23:09:08 crc kubenswrapper[4791]: I1210 23:09:08.423370 4791 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-76fcf4b695-r7wvb"] Dec 10 23:09:08 crc kubenswrapper[4791]: I1210 23:09:08.444798 4791 scope.go:117] "RemoveContainer" containerID="b4b22d6561fd9eb3dbbf460fde86eaf3110f3e0633b12d60b315543205b7c7a1" Dec 10 23:09:09 crc kubenswrapper[4791]: I1210 23:09:09.399730 4791 prober_manager.go:312] "Failed to trigger a manual run" probe="Readiness" Dec 10 23:09:09 crc kubenswrapper[4791]: I1210 23:09:09.412407 4791 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/horizon-7cc66567bd-pwb2l" podUID="55139c7c-e9fa-4539-b7b2-b2be7ee43b44" containerName="horizon" probeResult="failure" output="Get \"https://10.217.0.144:8443/dashboard/auth/login/?next=/dashboard/\": dial tcp 10.217.0.144:8443: connect: connection refused" Dec 10 23:09:09 crc kubenswrapper[4791]: I1210 23:09:09.496456 4791 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/horizon-6999c5bf5b-7cr9t" podUID="0fab3b41-ed57-4536-ade0-f5f6ccde40e2" containerName="horizon" probeResult="failure" output="Get \"https://10.217.0.145:8443/dashboard/auth/login/?next=/dashboard/\": dial tcp 10.217.0.145:8443: connect: connection refused" Dec 10 23:09:09 crc kubenswrapper[4791]: I1210 23:09:09.546183 4791 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/glance-default-external-api-0" Dec 10 23:09:09 crc kubenswrapper[4791]: I1210 23:09:09.790523 4791 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-db-sync-g4s5p" Dec 10 23:09:09 crc kubenswrapper[4791]: I1210 23:09:09.897639 4791 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="00320ef6-9362-427b-a2eb-12329d652a0e" path="/var/lib/kubelet/pods/00320ef6-9362-427b-a2eb-12329d652a0e/volumes" Dec 10 23:09:09 crc kubenswrapper[4791]: I1210 23:09:09.906725 4791 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/a85df2ed-be84-41b3-8993-c3d963daef08-etc-machine-id\") pod \"a85df2ed-be84-41b3-8993-c3d963daef08\" (UID: \"a85df2ed-be84-41b3-8993-c3d963daef08\") " Dec 10 23:09:09 crc kubenswrapper[4791]: I1210 23:09:09.906789 4791 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/a85df2ed-be84-41b3-8993-c3d963daef08-config-data\") pod \"a85df2ed-be84-41b3-8993-c3d963daef08\" (UID: \"a85df2ed-be84-41b3-8993-c3d963daef08\") " Dec 10 23:09:09 crc kubenswrapper[4791]: I1210 23:09:09.906833 4791 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a85df2ed-be84-41b3-8993-c3d963daef08-combined-ca-bundle\") pod \"a85df2ed-be84-41b3-8993-c3d963daef08\" (UID: \"a85df2ed-be84-41b3-8993-c3d963daef08\") " Dec 10 23:09:09 crc kubenswrapper[4791]: I1210 23:09:09.906891 4791 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-5fw8x\" (UniqueName: \"kubernetes.io/projected/a85df2ed-be84-41b3-8993-c3d963daef08-kube-api-access-5fw8x\") pod \"a85df2ed-be84-41b3-8993-c3d963daef08\" (UID: \"a85df2ed-be84-41b3-8993-c3d963daef08\") " Dec 10 23:09:09 crc kubenswrapper[4791]: I1210 23:09:09.906980 4791 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/a85df2ed-be84-41b3-8993-c3d963daef08-scripts\") pod \"a85df2ed-be84-41b3-8993-c3d963daef08\" (UID: \"a85df2ed-be84-41b3-8993-c3d963daef08\") " Dec 10 23:09:09 crc kubenswrapper[4791]: I1210 23:09:09.907117 4791 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/a85df2ed-be84-41b3-8993-c3d963daef08-db-sync-config-data\") pod \"a85df2ed-be84-41b3-8993-c3d963daef08\" (UID: \"a85df2ed-be84-41b3-8993-c3d963daef08\") " Dec 10 23:09:09 crc kubenswrapper[4791]: I1210 23:09:09.907585 4791 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/a85df2ed-be84-41b3-8993-c3d963daef08-etc-machine-id" (OuterVolumeSpecName: "etc-machine-id") pod "a85df2ed-be84-41b3-8993-c3d963daef08" (UID: "a85df2ed-be84-41b3-8993-c3d963daef08"). InnerVolumeSpecName "etc-machine-id". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 10 23:09:09 crc kubenswrapper[4791]: I1210 23:09:09.908222 4791 reconciler_common.go:293] "Volume detached for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/a85df2ed-be84-41b3-8993-c3d963daef08-etc-machine-id\") on node \"crc\" DevicePath \"\"" Dec 10 23:09:09 crc kubenswrapper[4791]: I1210 23:09:09.918211 4791 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/a85df2ed-be84-41b3-8993-c3d963daef08-kube-api-access-5fw8x" (OuterVolumeSpecName: "kube-api-access-5fw8x") pod "a85df2ed-be84-41b3-8993-c3d963daef08" (UID: "a85df2ed-be84-41b3-8993-c3d963daef08"). InnerVolumeSpecName "kube-api-access-5fw8x". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 10 23:09:09 crc kubenswrapper[4791]: I1210 23:09:09.920474 4791 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/a85df2ed-be84-41b3-8993-c3d963daef08-db-sync-config-data" (OuterVolumeSpecName: "db-sync-config-data") pod "a85df2ed-be84-41b3-8993-c3d963daef08" (UID: "a85df2ed-be84-41b3-8993-c3d963daef08"). InnerVolumeSpecName "db-sync-config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 10 23:09:09 crc kubenswrapper[4791]: I1210 23:09:09.935786 4791 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/a85df2ed-be84-41b3-8993-c3d963daef08-scripts" (OuterVolumeSpecName: "scripts") pod "a85df2ed-be84-41b3-8993-c3d963daef08" (UID: "a85df2ed-be84-41b3-8993-c3d963daef08"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 10 23:09:09 crc kubenswrapper[4791]: I1210 23:09:09.959596 4791 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/a85df2ed-be84-41b3-8993-c3d963daef08-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "a85df2ed-be84-41b3-8993-c3d963daef08" (UID: "a85df2ed-be84-41b3-8993-c3d963daef08"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 10 23:09:09 crc kubenswrapper[4791]: I1210 23:09:09.982150 4791 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/a85df2ed-be84-41b3-8993-c3d963daef08-config-data" (OuterVolumeSpecName: "config-data") pod "a85df2ed-be84-41b3-8993-c3d963daef08" (UID: "a85df2ed-be84-41b3-8993-c3d963daef08"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 10 23:09:10 crc kubenswrapper[4791]: I1210 23:09:10.009580 4791 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/a85df2ed-be84-41b3-8993-c3d963daef08-config-data\") on node \"crc\" DevicePath \"\"" Dec 10 23:09:10 crc kubenswrapper[4791]: I1210 23:09:10.009612 4791 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a85df2ed-be84-41b3-8993-c3d963daef08-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 10 23:09:10 crc kubenswrapper[4791]: I1210 23:09:10.009624 4791 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-5fw8x\" (UniqueName: \"kubernetes.io/projected/a85df2ed-be84-41b3-8993-c3d963daef08-kube-api-access-5fw8x\") on node \"crc\" DevicePath \"\"" Dec 10 23:09:10 crc kubenswrapper[4791]: I1210 23:09:10.009632 4791 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/a85df2ed-be84-41b3-8993-c3d963daef08-scripts\") on node \"crc\" DevicePath \"\"" Dec 10 23:09:10 crc kubenswrapper[4791]: I1210 23:09:10.009640 4791 reconciler_common.go:293] "Volume detached for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/a85df2ed-be84-41b3-8993-c3d963daef08-db-sync-config-data\") on node \"crc\" DevicePath \"\"" Dec 10 23:09:10 crc kubenswrapper[4791]: I1210 23:09:10.458639 4791 prober_manager.go:312] "Failed to trigger a manual run" probe="Readiness" Dec 10 23:09:10 crc kubenswrapper[4791]: I1210 23:09:10.459820 4791 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-db-sync-g4s5p" Dec 10 23:09:10 crc kubenswrapper[4791]: I1210 23:09:10.463781 4791 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-db-sync-g4s5p" event={"ID":"a85df2ed-be84-41b3-8993-c3d963daef08","Type":"ContainerDied","Data":"31c7a8174d2b3adae98a2a223b58b992d2148dcb534d2f097625927e1a32586e"} Dec 10 23:09:10 crc kubenswrapper[4791]: I1210 23:09:10.463936 4791 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="31c7a8174d2b3adae98a2a223b58b992d2148dcb534d2f097625927e1a32586e" Dec 10 23:09:10 crc kubenswrapper[4791]: I1210 23:09:10.664865 4791 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/cinder-scheduler-0"] Dec 10 23:09:10 crc kubenswrapper[4791]: E1210 23:09:10.665218 4791 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="a85df2ed-be84-41b3-8993-c3d963daef08" containerName="cinder-db-sync" Dec 10 23:09:10 crc kubenswrapper[4791]: I1210 23:09:10.665235 4791 state_mem.go:107] "Deleted CPUSet assignment" podUID="a85df2ed-be84-41b3-8993-c3d963daef08" containerName="cinder-db-sync" Dec 10 23:09:10 crc kubenswrapper[4791]: E1210 23:09:10.665254 4791 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="00320ef6-9362-427b-a2eb-12329d652a0e" containerName="init" Dec 10 23:09:10 crc kubenswrapper[4791]: I1210 23:09:10.665260 4791 state_mem.go:107] "Deleted CPUSet assignment" podUID="00320ef6-9362-427b-a2eb-12329d652a0e" containerName="init" Dec 10 23:09:10 crc kubenswrapper[4791]: E1210 23:09:10.665287 4791 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="00320ef6-9362-427b-a2eb-12329d652a0e" containerName="dnsmasq-dns" Dec 10 23:09:10 crc kubenswrapper[4791]: I1210 23:09:10.665293 4791 state_mem.go:107] "Deleted CPUSet assignment" podUID="00320ef6-9362-427b-a2eb-12329d652a0e" containerName="dnsmasq-dns" Dec 10 23:09:10 crc kubenswrapper[4791]: I1210 23:09:10.665483 4791 memory_manager.go:354] "RemoveStaleState removing state" podUID="a85df2ed-be84-41b3-8993-c3d963daef08" containerName="cinder-db-sync" Dec 10 23:09:10 crc kubenswrapper[4791]: I1210 23:09:10.665499 4791 memory_manager.go:354] "RemoveStaleState removing state" podUID="00320ef6-9362-427b-a2eb-12329d652a0e" containerName="dnsmasq-dns" Dec 10 23:09:10 crc kubenswrapper[4791]: I1210 23:09:10.668960 4791 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-scheduler-0" Dec 10 23:09:10 crc kubenswrapper[4791]: I1210 23:09:10.675844 4791 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cinder-cinder-dockercfg-lxnf7" Dec 10 23:09:10 crc kubenswrapper[4791]: I1210 23:09:10.675941 4791 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cinder-scripts" Dec 10 23:09:10 crc kubenswrapper[4791]: I1210 23:09:10.676307 4791 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cinder-config-data" Dec 10 23:09:10 crc kubenswrapper[4791]: I1210 23:09:10.676319 4791 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cinder-scheduler-config-data" Dec 10 23:09:10 crc kubenswrapper[4791]: I1210 23:09:10.697996 4791 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-scheduler-0"] Dec 10 23:09:10 crc kubenswrapper[4791]: I1210 23:09:10.717554 4791 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/glance-default-internal-api-0" Dec 10 23:09:10 crc kubenswrapper[4791]: I1210 23:09:10.717652 4791 prober_manager.go:312] "Failed to trigger a manual run" probe="Readiness" Dec 10 23:09:10 crc kubenswrapper[4791]: I1210 23:09:10.725525 4791 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/5e5e327e-b8dc-4bd6-af75-8aebc417ada9-etc-machine-id\") pod \"cinder-scheduler-0\" (UID: \"5e5e327e-b8dc-4bd6-af75-8aebc417ada9\") " pod="openstack/cinder-scheduler-0" Dec 10 23:09:10 crc kubenswrapper[4791]: I1210 23:09:10.725609 4791 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/5e5e327e-b8dc-4bd6-af75-8aebc417ada9-scripts\") pod \"cinder-scheduler-0\" (UID: \"5e5e327e-b8dc-4bd6-af75-8aebc417ada9\") " pod="openstack/cinder-scheduler-0" Dec 10 23:09:10 crc kubenswrapper[4791]: I1210 23:09:10.725686 4791 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/5e5e327e-b8dc-4bd6-af75-8aebc417ada9-combined-ca-bundle\") pod \"cinder-scheduler-0\" (UID: \"5e5e327e-b8dc-4bd6-af75-8aebc417ada9\") " pod="openstack/cinder-scheduler-0" Dec 10 23:09:10 crc kubenswrapper[4791]: I1210 23:09:10.725704 4791 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/5e5e327e-b8dc-4bd6-af75-8aebc417ada9-config-data\") pod \"cinder-scheduler-0\" (UID: \"5e5e327e-b8dc-4bd6-af75-8aebc417ada9\") " pod="openstack/cinder-scheduler-0" Dec 10 23:09:10 crc kubenswrapper[4791]: I1210 23:09:10.725770 4791 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/5e5e327e-b8dc-4bd6-af75-8aebc417ada9-config-data-custom\") pod \"cinder-scheduler-0\" (UID: \"5e5e327e-b8dc-4bd6-af75-8aebc417ada9\") " pod="openstack/cinder-scheduler-0" Dec 10 23:09:10 crc kubenswrapper[4791]: I1210 23:09:10.725802 4791 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-qc87z\" (UniqueName: \"kubernetes.io/projected/5e5e327e-b8dc-4bd6-af75-8aebc417ada9-kube-api-access-qc87z\") pod \"cinder-scheduler-0\" (UID: \"5e5e327e-b8dc-4bd6-af75-8aebc417ada9\") " pod="openstack/cinder-scheduler-0" Dec 10 23:09:10 crc kubenswrapper[4791]: I1210 23:09:10.780455 4791 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-d68b9cb4c-xp29n"] Dec 10 23:09:10 crc kubenswrapper[4791]: I1210 23:09:10.782644 4791 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-d68b9cb4c-xp29n" Dec 10 23:09:10 crc kubenswrapper[4791]: I1210 23:09:10.812705 4791 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-d68b9cb4c-xp29n"] Dec 10 23:09:10 crc kubenswrapper[4791]: I1210 23:09:10.860812 4791 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/5e5e327e-b8dc-4bd6-af75-8aebc417ada9-combined-ca-bundle\") pod \"cinder-scheduler-0\" (UID: \"5e5e327e-b8dc-4bd6-af75-8aebc417ada9\") " pod="openstack/cinder-scheduler-0" Dec 10 23:09:10 crc kubenswrapper[4791]: I1210 23:09:10.861365 4791 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/5e5e327e-b8dc-4bd6-af75-8aebc417ada9-config-data\") pod \"cinder-scheduler-0\" (UID: \"5e5e327e-b8dc-4bd6-af75-8aebc417ada9\") " pod="openstack/cinder-scheduler-0" Dec 10 23:09:10 crc kubenswrapper[4791]: I1210 23:09:10.861583 4791 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/5e5e327e-b8dc-4bd6-af75-8aebc417ada9-config-data-custom\") pod \"cinder-scheduler-0\" (UID: \"5e5e327e-b8dc-4bd6-af75-8aebc417ada9\") " pod="openstack/cinder-scheduler-0" Dec 10 23:09:10 crc kubenswrapper[4791]: I1210 23:09:10.862164 4791 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-qc87z\" (UniqueName: \"kubernetes.io/projected/5e5e327e-b8dc-4bd6-af75-8aebc417ada9-kube-api-access-qc87z\") pod \"cinder-scheduler-0\" (UID: \"5e5e327e-b8dc-4bd6-af75-8aebc417ada9\") " pod="openstack/cinder-scheduler-0" Dec 10 23:09:10 crc kubenswrapper[4791]: I1210 23:09:10.862300 4791 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/5e5e327e-b8dc-4bd6-af75-8aebc417ada9-etc-machine-id\") pod \"cinder-scheduler-0\" (UID: \"5e5e327e-b8dc-4bd6-af75-8aebc417ada9\") " pod="openstack/cinder-scheduler-0" Dec 10 23:09:10 crc kubenswrapper[4791]: I1210 23:09:10.862446 4791 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/5e5e327e-b8dc-4bd6-af75-8aebc417ada9-scripts\") pod \"cinder-scheduler-0\" (UID: \"5e5e327e-b8dc-4bd6-af75-8aebc417ada9\") " pod="openstack/cinder-scheduler-0" Dec 10 23:09:10 crc kubenswrapper[4791]: I1210 23:09:10.864108 4791 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/5e5e327e-b8dc-4bd6-af75-8aebc417ada9-etc-machine-id\") pod \"cinder-scheduler-0\" (UID: \"5e5e327e-b8dc-4bd6-af75-8aebc417ada9\") " pod="openstack/cinder-scheduler-0" Dec 10 23:09:10 crc kubenswrapper[4791]: I1210 23:09:10.894513 4791 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/5e5e327e-b8dc-4bd6-af75-8aebc417ada9-scripts\") pod \"cinder-scheduler-0\" (UID: \"5e5e327e-b8dc-4bd6-af75-8aebc417ada9\") " pod="openstack/cinder-scheduler-0" Dec 10 23:09:10 crc kubenswrapper[4791]: I1210 23:09:10.901288 4791 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/5e5e327e-b8dc-4bd6-af75-8aebc417ada9-combined-ca-bundle\") pod \"cinder-scheduler-0\" (UID: \"5e5e327e-b8dc-4bd6-af75-8aebc417ada9\") " pod="openstack/cinder-scheduler-0" Dec 10 23:09:10 crc kubenswrapper[4791]: I1210 23:09:10.910132 4791 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/5e5e327e-b8dc-4bd6-af75-8aebc417ada9-config-data-custom\") pod \"cinder-scheduler-0\" (UID: \"5e5e327e-b8dc-4bd6-af75-8aebc417ada9\") " pod="openstack/cinder-scheduler-0" Dec 10 23:09:10 crc kubenswrapper[4791]: I1210 23:09:10.917472 4791 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/5e5e327e-b8dc-4bd6-af75-8aebc417ada9-config-data\") pod \"cinder-scheduler-0\" (UID: \"5e5e327e-b8dc-4bd6-af75-8aebc417ada9\") " pod="openstack/cinder-scheduler-0" Dec 10 23:09:10 crc kubenswrapper[4791]: I1210 23:09:10.929768 4791 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-qc87z\" (UniqueName: \"kubernetes.io/projected/5e5e327e-b8dc-4bd6-af75-8aebc417ada9-kube-api-access-qc87z\") pod \"cinder-scheduler-0\" (UID: \"5e5e327e-b8dc-4bd6-af75-8aebc417ada9\") " pod="openstack/cinder-scheduler-0" Dec 10 23:09:10 crc kubenswrapper[4791]: I1210 23:09:10.964331 4791 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/ae2623b4-7b39-4661-b9af-caf32415d8b1-ovsdbserver-sb\") pod \"dnsmasq-dns-d68b9cb4c-xp29n\" (UID: \"ae2623b4-7b39-4661-b9af-caf32415d8b1\") " pod="openstack/dnsmasq-dns-d68b9cb4c-xp29n" Dec 10 23:09:10 crc kubenswrapper[4791]: I1210 23:09:10.966693 4791 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/ae2623b4-7b39-4661-b9af-caf32415d8b1-ovsdbserver-nb\") pod \"dnsmasq-dns-d68b9cb4c-xp29n\" (UID: \"ae2623b4-7b39-4661-b9af-caf32415d8b1\") " pod="openstack/dnsmasq-dns-d68b9cb4c-xp29n" Dec 10 23:09:10 crc kubenswrapper[4791]: I1210 23:09:10.966858 4791 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/ae2623b4-7b39-4661-b9af-caf32415d8b1-config\") pod \"dnsmasq-dns-d68b9cb4c-xp29n\" (UID: \"ae2623b4-7b39-4661-b9af-caf32415d8b1\") " pod="openstack/dnsmasq-dns-d68b9cb4c-xp29n" Dec 10 23:09:10 crc kubenswrapper[4791]: I1210 23:09:10.966985 4791 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/ae2623b4-7b39-4661-b9af-caf32415d8b1-dns-swift-storage-0\") pod \"dnsmasq-dns-d68b9cb4c-xp29n\" (UID: \"ae2623b4-7b39-4661-b9af-caf32415d8b1\") " pod="openstack/dnsmasq-dns-d68b9cb4c-xp29n" Dec 10 23:09:10 crc kubenswrapper[4791]: I1210 23:09:10.967194 4791 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/ae2623b4-7b39-4661-b9af-caf32415d8b1-dns-svc\") pod \"dnsmasq-dns-d68b9cb4c-xp29n\" (UID: \"ae2623b4-7b39-4661-b9af-caf32415d8b1\") " pod="openstack/dnsmasq-dns-d68b9cb4c-xp29n" Dec 10 23:09:10 crc kubenswrapper[4791]: I1210 23:09:10.967296 4791 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-hv5bp\" (UniqueName: \"kubernetes.io/projected/ae2623b4-7b39-4661-b9af-caf32415d8b1-kube-api-access-hv5bp\") pod \"dnsmasq-dns-d68b9cb4c-xp29n\" (UID: \"ae2623b4-7b39-4661-b9af-caf32415d8b1\") " pod="openstack/dnsmasq-dns-d68b9cb4c-xp29n" Dec 10 23:09:11 crc kubenswrapper[4791]: I1210 23:09:11.016839 4791 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-scheduler-0" Dec 10 23:09:11 crc kubenswrapper[4791]: I1210 23:09:11.068544 4791 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/cinder-api-0"] Dec 10 23:09:11 crc kubenswrapper[4791]: I1210 23:09:11.069325 4791 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/ae2623b4-7b39-4661-b9af-caf32415d8b1-config\") pod \"dnsmasq-dns-d68b9cb4c-xp29n\" (UID: \"ae2623b4-7b39-4661-b9af-caf32415d8b1\") " pod="openstack/dnsmasq-dns-d68b9cb4c-xp29n" Dec 10 23:09:11 crc kubenswrapper[4791]: I1210 23:09:11.069412 4791 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/ae2623b4-7b39-4661-b9af-caf32415d8b1-dns-swift-storage-0\") pod \"dnsmasq-dns-d68b9cb4c-xp29n\" (UID: \"ae2623b4-7b39-4661-b9af-caf32415d8b1\") " pod="openstack/dnsmasq-dns-d68b9cb4c-xp29n" Dec 10 23:09:11 crc kubenswrapper[4791]: I1210 23:09:11.069472 4791 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/ae2623b4-7b39-4661-b9af-caf32415d8b1-dns-svc\") pod \"dnsmasq-dns-d68b9cb4c-xp29n\" (UID: \"ae2623b4-7b39-4661-b9af-caf32415d8b1\") " pod="openstack/dnsmasq-dns-d68b9cb4c-xp29n" Dec 10 23:09:11 crc kubenswrapper[4791]: I1210 23:09:11.069535 4791 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-hv5bp\" (UniqueName: \"kubernetes.io/projected/ae2623b4-7b39-4661-b9af-caf32415d8b1-kube-api-access-hv5bp\") pod \"dnsmasq-dns-d68b9cb4c-xp29n\" (UID: \"ae2623b4-7b39-4661-b9af-caf32415d8b1\") " pod="openstack/dnsmasq-dns-d68b9cb4c-xp29n" Dec 10 23:09:11 crc kubenswrapper[4791]: I1210 23:09:11.069594 4791 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/ae2623b4-7b39-4661-b9af-caf32415d8b1-ovsdbserver-sb\") pod \"dnsmasq-dns-d68b9cb4c-xp29n\" (UID: \"ae2623b4-7b39-4661-b9af-caf32415d8b1\") " pod="openstack/dnsmasq-dns-d68b9cb4c-xp29n" Dec 10 23:09:11 crc kubenswrapper[4791]: I1210 23:09:11.069647 4791 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/ae2623b4-7b39-4661-b9af-caf32415d8b1-ovsdbserver-nb\") pod \"dnsmasq-dns-d68b9cb4c-xp29n\" (UID: \"ae2623b4-7b39-4661-b9af-caf32415d8b1\") " pod="openstack/dnsmasq-dns-d68b9cb4c-xp29n" Dec 10 23:09:11 crc kubenswrapper[4791]: I1210 23:09:11.070067 4791 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-api-0" Dec 10 23:09:11 crc kubenswrapper[4791]: I1210 23:09:11.070393 4791 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/ae2623b4-7b39-4661-b9af-caf32415d8b1-config\") pod \"dnsmasq-dns-d68b9cb4c-xp29n\" (UID: \"ae2623b4-7b39-4661-b9af-caf32415d8b1\") " pod="openstack/dnsmasq-dns-d68b9cb4c-xp29n" Dec 10 23:09:11 crc kubenswrapper[4791]: I1210 23:09:11.071737 4791 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/dnsmasq-dns-76fcf4b695-r7wvb" podUID="00320ef6-9362-427b-a2eb-12329d652a0e" containerName="dnsmasq-dns" probeResult="failure" output="dial tcp 10.217.0.141:5353: i/o timeout" Dec 10 23:09:11 crc kubenswrapper[4791]: I1210 23:09:11.073195 4791 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/ae2623b4-7b39-4661-b9af-caf32415d8b1-ovsdbserver-nb\") pod \"dnsmasq-dns-d68b9cb4c-xp29n\" (UID: \"ae2623b4-7b39-4661-b9af-caf32415d8b1\") " pod="openstack/dnsmasq-dns-d68b9cb4c-xp29n" Dec 10 23:09:11 crc kubenswrapper[4791]: I1210 23:09:11.075916 4791 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/ae2623b4-7b39-4661-b9af-caf32415d8b1-dns-swift-storage-0\") pod \"dnsmasq-dns-d68b9cb4c-xp29n\" (UID: \"ae2623b4-7b39-4661-b9af-caf32415d8b1\") " pod="openstack/dnsmasq-dns-d68b9cb4c-xp29n" Dec 10 23:09:11 crc kubenswrapper[4791]: I1210 23:09:11.078288 4791 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/ae2623b4-7b39-4661-b9af-caf32415d8b1-dns-svc\") pod \"dnsmasq-dns-d68b9cb4c-xp29n\" (UID: \"ae2623b4-7b39-4661-b9af-caf32415d8b1\") " pod="openstack/dnsmasq-dns-d68b9cb4c-xp29n" Dec 10 23:09:11 crc kubenswrapper[4791]: I1210 23:09:11.079387 4791 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/ae2623b4-7b39-4661-b9af-caf32415d8b1-ovsdbserver-sb\") pod \"dnsmasq-dns-d68b9cb4c-xp29n\" (UID: \"ae2623b4-7b39-4661-b9af-caf32415d8b1\") " pod="openstack/dnsmasq-dns-d68b9cb4c-xp29n" Dec 10 23:09:11 crc kubenswrapper[4791]: I1210 23:09:11.088882 4791 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/glance-default-internal-api-0" Dec 10 23:09:11 crc kubenswrapper[4791]: I1210 23:09:11.089429 4791 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cinder-api-config-data" Dec 10 23:09:11 crc kubenswrapper[4791]: I1210 23:09:11.089677 4791 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-api-0"] Dec 10 23:09:11 crc kubenswrapper[4791]: I1210 23:09:11.128879 4791 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-hv5bp\" (UniqueName: \"kubernetes.io/projected/ae2623b4-7b39-4661-b9af-caf32415d8b1-kube-api-access-hv5bp\") pod \"dnsmasq-dns-d68b9cb4c-xp29n\" (UID: \"ae2623b4-7b39-4661-b9af-caf32415d8b1\") " pod="openstack/dnsmasq-dns-d68b9cb4c-xp29n" Dec 10 23:09:11 crc kubenswrapper[4791]: I1210 23:09:11.176607 4791 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/1a1ca37c-dab5-4748-b8c2-eb985e66a02f-config-data\") pod \"cinder-api-0\" (UID: \"1a1ca37c-dab5-4748-b8c2-eb985e66a02f\") " pod="openstack/cinder-api-0" Dec 10 23:09:11 crc kubenswrapper[4791]: I1210 23:09:11.176656 4791 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/1a1ca37c-dab5-4748-b8c2-eb985e66a02f-scripts\") pod \"cinder-api-0\" (UID: \"1a1ca37c-dab5-4748-b8c2-eb985e66a02f\") " pod="openstack/cinder-api-0" Dec 10 23:09:11 crc kubenswrapper[4791]: I1210 23:09:11.176694 4791 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/1a1ca37c-dab5-4748-b8c2-eb985e66a02f-logs\") pod \"cinder-api-0\" (UID: \"1a1ca37c-dab5-4748-b8c2-eb985e66a02f\") " pod="openstack/cinder-api-0" Dec 10 23:09:11 crc kubenswrapper[4791]: I1210 23:09:11.176753 4791 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/1a1ca37c-dab5-4748-b8c2-eb985e66a02f-combined-ca-bundle\") pod \"cinder-api-0\" (UID: \"1a1ca37c-dab5-4748-b8c2-eb985e66a02f\") " pod="openstack/cinder-api-0" Dec 10 23:09:11 crc kubenswrapper[4791]: I1210 23:09:11.176786 4791 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/1a1ca37c-dab5-4748-b8c2-eb985e66a02f-config-data-custom\") pod \"cinder-api-0\" (UID: \"1a1ca37c-dab5-4748-b8c2-eb985e66a02f\") " pod="openstack/cinder-api-0" Dec 10 23:09:11 crc kubenswrapper[4791]: I1210 23:09:11.176860 4791 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/1a1ca37c-dab5-4748-b8c2-eb985e66a02f-etc-machine-id\") pod \"cinder-api-0\" (UID: \"1a1ca37c-dab5-4748-b8c2-eb985e66a02f\") " pod="openstack/cinder-api-0" Dec 10 23:09:11 crc kubenswrapper[4791]: I1210 23:09:11.176899 4791 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-vvvdb\" (UniqueName: \"kubernetes.io/projected/1a1ca37c-dab5-4748-b8c2-eb985e66a02f-kube-api-access-vvvdb\") pod \"cinder-api-0\" (UID: \"1a1ca37c-dab5-4748-b8c2-eb985e66a02f\") " pod="openstack/cinder-api-0" Dec 10 23:09:11 crc kubenswrapper[4791]: I1210 23:09:11.282515 4791 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/1a1ca37c-dab5-4748-b8c2-eb985e66a02f-combined-ca-bundle\") pod \"cinder-api-0\" (UID: \"1a1ca37c-dab5-4748-b8c2-eb985e66a02f\") " pod="openstack/cinder-api-0" Dec 10 23:09:11 crc kubenswrapper[4791]: I1210 23:09:11.282902 4791 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/1a1ca37c-dab5-4748-b8c2-eb985e66a02f-config-data-custom\") pod \"cinder-api-0\" (UID: \"1a1ca37c-dab5-4748-b8c2-eb985e66a02f\") " pod="openstack/cinder-api-0" Dec 10 23:09:11 crc kubenswrapper[4791]: I1210 23:09:11.282984 4791 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/1a1ca37c-dab5-4748-b8c2-eb985e66a02f-etc-machine-id\") pod \"cinder-api-0\" (UID: \"1a1ca37c-dab5-4748-b8c2-eb985e66a02f\") " pod="openstack/cinder-api-0" Dec 10 23:09:11 crc kubenswrapper[4791]: I1210 23:09:11.283021 4791 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-vvvdb\" (UniqueName: \"kubernetes.io/projected/1a1ca37c-dab5-4748-b8c2-eb985e66a02f-kube-api-access-vvvdb\") pod \"cinder-api-0\" (UID: \"1a1ca37c-dab5-4748-b8c2-eb985e66a02f\") " pod="openstack/cinder-api-0" Dec 10 23:09:11 crc kubenswrapper[4791]: I1210 23:09:11.283074 4791 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/1a1ca37c-dab5-4748-b8c2-eb985e66a02f-config-data\") pod \"cinder-api-0\" (UID: \"1a1ca37c-dab5-4748-b8c2-eb985e66a02f\") " pod="openstack/cinder-api-0" Dec 10 23:09:11 crc kubenswrapper[4791]: I1210 23:09:11.283099 4791 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/1a1ca37c-dab5-4748-b8c2-eb985e66a02f-scripts\") pod \"cinder-api-0\" (UID: \"1a1ca37c-dab5-4748-b8c2-eb985e66a02f\") " pod="openstack/cinder-api-0" Dec 10 23:09:11 crc kubenswrapper[4791]: I1210 23:09:11.283143 4791 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/1a1ca37c-dab5-4748-b8c2-eb985e66a02f-logs\") pod \"cinder-api-0\" (UID: \"1a1ca37c-dab5-4748-b8c2-eb985e66a02f\") " pod="openstack/cinder-api-0" Dec 10 23:09:11 crc kubenswrapper[4791]: I1210 23:09:11.289794 4791 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/1a1ca37c-dab5-4748-b8c2-eb985e66a02f-etc-machine-id\") pod \"cinder-api-0\" (UID: \"1a1ca37c-dab5-4748-b8c2-eb985e66a02f\") " pod="openstack/cinder-api-0" Dec 10 23:09:11 crc kubenswrapper[4791]: I1210 23:09:11.290116 4791 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/1a1ca37c-dab5-4748-b8c2-eb985e66a02f-logs\") pod \"cinder-api-0\" (UID: \"1a1ca37c-dab5-4748-b8c2-eb985e66a02f\") " pod="openstack/cinder-api-0" Dec 10 23:09:11 crc kubenswrapper[4791]: I1210 23:09:11.291946 4791 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-d68b9cb4c-xp29n" Dec 10 23:09:11 crc kubenswrapper[4791]: I1210 23:09:11.306756 4791 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/1a1ca37c-dab5-4748-b8c2-eb985e66a02f-config-data-custom\") pod \"cinder-api-0\" (UID: \"1a1ca37c-dab5-4748-b8c2-eb985e66a02f\") " pod="openstack/cinder-api-0" Dec 10 23:09:11 crc kubenswrapper[4791]: I1210 23:09:11.306763 4791 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/1a1ca37c-dab5-4748-b8c2-eb985e66a02f-combined-ca-bundle\") pod \"cinder-api-0\" (UID: \"1a1ca37c-dab5-4748-b8c2-eb985e66a02f\") " pod="openstack/cinder-api-0" Dec 10 23:09:11 crc kubenswrapper[4791]: I1210 23:09:11.310353 4791 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/1a1ca37c-dab5-4748-b8c2-eb985e66a02f-scripts\") pod \"cinder-api-0\" (UID: \"1a1ca37c-dab5-4748-b8c2-eb985e66a02f\") " pod="openstack/cinder-api-0" Dec 10 23:09:11 crc kubenswrapper[4791]: I1210 23:09:11.311637 4791 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/1a1ca37c-dab5-4748-b8c2-eb985e66a02f-config-data\") pod \"cinder-api-0\" (UID: \"1a1ca37c-dab5-4748-b8c2-eb985e66a02f\") " pod="openstack/cinder-api-0" Dec 10 23:09:11 crc kubenswrapper[4791]: I1210 23:09:11.313817 4791 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-vvvdb\" (UniqueName: \"kubernetes.io/projected/1a1ca37c-dab5-4748-b8c2-eb985e66a02f-kube-api-access-vvvdb\") pod \"cinder-api-0\" (UID: \"1a1ca37c-dab5-4748-b8c2-eb985e66a02f\") " pod="openstack/cinder-api-0" Dec 10 23:09:11 crc kubenswrapper[4791]: I1210 23:09:11.324732 4791 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/glance-default-external-api-0" Dec 10 23:09:11 crc kubenswrapper[4791]: I1210 23:09:11.542835 4791 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-api-0" Dec 10 23:09:11 crc kubenswrapper[4791]: I1210 23:09:11.682138 4791 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-scheduler-0"] Dec 10 23:09:11 crc kubenswrapper[4791]: I1210 23:09:11.964631 4791 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-d68b9cb4c-xp29n"] Dec 10 23:09:12 crc kubenswrapper[4791]: W1210 23:09:12.002587 4791 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podae2623b4_7b39_4661_b9af_caf32415d8b1.slice/crio-d1ea78e5d6411134a13966ef4046829678d14b12c7138cb29784cac9fbadbd5a WatchSource:0}: Error finding container d1ea78e5d6411134a13966ef4046829678d14b12c7138cb29784cac9fbadbd5a: Status 404 returned error can't find the container with id d1ea78e5d6411134a13966ef4046829678d14b12c7138cb29784cac9fbadbd5a Dec 10 23:09:12 crc kubenswrapper[4791]: I1210 23:09:12.246402 4791 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-api-0"] Dec 10 23:09:12 crc kubenswrapper[4791]: I1210 23:09:12.540114 4791 generic.go:334] "Generic (PLEG): container finished" podID="ae2623b4-7b39-4661-b9af-caf32415d8b1" containerID="766b794c9708b4cbfe2821348b03e234d1dd7e3877bf577377616c1025ff7564" exitCode=0 Dec 10 23:09:12 crc kubenswrapper[4791]: I1210 23:09:12.540201 4791 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-d68b9cb4c-xp29n" event={"ID":"ae2623b4-7b39-4661-b9af-caf32415d8b1","Type":"ContainerDied","Data":"766b794c9708b4cbfe2821348b03e234d1dd7e3877bf577377616c1025ff7564"} Dec 10 23:09:12 crc kubenswrapper[4791]: I1210 23:09:12.540230 4791 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-d68b9cb4c-xp29n" event={"ID":"ae2623b4-7b39-4661-b9af-caf32415d8b1","Type":"ContainerStarted","Data":"d1ea78e5d6411134a13966ef4046829678d14b12c7138cb29784cac9fbadbd5a"} Dec 10 23:09:12 crc kubenswrapper[4791]: I1210 23:09:12.543471 4791 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-api-0" event={"ID":"1a1ca37c-dab5-4748-b8c2-eb985e66a02f","Type":"ContainerStarted","Data":"b5a57d174634113df118d0024e1c1c33aaa029ba9541d56b2728ce26d0d5fffa"} Dec 10 23:09:12 crc kubenswrapper[4791]: I1210 23:09:12.552416 4791 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-scheduler-0" event={"ID":"5e5e327e-b8dc-4bd6-af75-8aebc417ada9","Type":"ContainerStarted","Data":"a3778b6f3bbf0ca69bdffd7ffe50f1e49fcc48b28bed8e76f5376905253abefe"} Dec 10 23:09:12 crc kubenswrapper[4791]: I1210 23:09:12.555995 4791 generic.go:334] "Generic (PLEG): container finished" podID="851556d7-fa38-4eec-ae74-de989d66d27b" containerID="96fce98c2678503c67e20f58365f73a09244ac81b06aa07e100ddbf7bb1368d9" exitCode=0 Dec 10 23:09:12 crc kubenswrapper[4791]: I1210 23:09:12.556069 4791 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-db-sync-zllvm" event={"ID":"851556d7-fa38-4eec-ae74-de989d66d27b","Type":"ContainerDied","Data":"96fce98c2678503c67e20f58365f73a09244ac81b06aa07e100ddbf7bb1368d9"} Dec 10 23:09:13 crc kubenswrapper[4791]: I1210 23:09:13.579429 4791 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-d68b9cb4c-xp29n" event={"ID":"ae2623b4-7b39-4661-b9af-caf32415d8b1","Type":"ContainerStarted","Data":"8abb2f7b1e355fe0279461505e35f3aae94f2f3d4097ceb931b5ef3c52902873"} Dec 10 23:09:13 crc kubenswrapper[4791]: I1210 23:09:13.579795 4791 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/dnsmasq-dns-d68b9cb4c-xp29n" Dec 10 23:09:13 crc kubenswrapper[4791]: I1210 23:09:13.593273 4791 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-api-0" event={"ID":"1a1ca37c-dab5-4748-b8c2-eb985e66a02f","Type":"ContainerStarted","Data":"831488e7d1696daf20de78611d91cff3c3a4661569db23dac805484a70013576"} Dec 10 23:09:13 crc kubenswrapper[4791]: I1210 23:09:13.631207 4791 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/dnsmasq-dns-d68b9cb4c-xp29n" podStartSLOduration=3.631188958 podStartE2EDuration="3.631188958s" podCreationTimestamp="2025-12-10 23:09:10 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-10 23:09:13.628795909 +0000 UTC m=+1188.058413522" watchObservedRunningTime="2025-12-10 23:09:13.631188958 +0000 UTC m=+1188.060806571" Dec 10 23:09:13 crc kubenswrapper[4791]: I1210 23:09:13.728288 4791 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/cinder-api-0"] Dec 10 23:09:14 crc kubenswrapper[4791]: I1210 23:09:14.229406 4791 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-db-sync-zllvm" Dec 10 23:09:14 crc kubenswrapper[4791]: I1210 23:09:14.285395 4791 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-crjrd\" (UniqueName: \"kubernetes.io/projected/851556d7-fa38-4eec-ae74-de989d66d27b-kube-api-access-crjrd\") pod \"851556d7-fa38-4eec-ae74-de989d66d27b\" (UID: \"851556d7-fa38-4eec-ae74-de989d66d27b\") " Dec 10 23:09:14 crc kubenswrapper[4791]: I1210 23:09:14.285480 4791 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/851556d7-fa38-4eec-ae74-de989d66d27b-combined-ca-bundle\") pod \"851556d7-fa38-4eec-ae74-de989d66d27b\" (UID: \"851556d7-fa38-4eec-ae74-de989d66d27b\") " Dec 10 23:09:14 crc kubenswrapper[4791]: I1210 23:09:14.285548 4791 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/851556d7-fa38-4eec-ae74-de989d66d27b-db-sync-config-data\") pod \"851556d7-fa38-4eec-ae74-de989d66d27b\" (UID: \"851556d7-fa38-4eec-ae74-de989d66d27b\") " Dec 10 23:09:14 crc kubenswrapper[4791]: I1210 23:09:14.295244 4791 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/851556d7-fa38-4eec-ae74-de989d66d27b-kube-api-access-crjrd" (OuterVolumeSpecName: "kube-api-access-crjrd") pod "851556d7-fa38-4eec-ae74-de989d66d27b" (UID: "851556d7-fa38-4eec-ae74-de989d66d27b"). InnerVolumeSpecName "kube-api-access-crjrd". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 10 23:09:14 crc kubenswrapper[4791]: I1210 23:09:14.295592 4791 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/851556d7-fa38-4eec-ae74-de989d66d27b-db-sync-config-data" (OuterVolumeSpecName: "db-sync-config-data") pod "851556d7-fa38-4eec-ae74-de989d66d27b" (UID: "851556d7-fa38-4eec-ae74-de989d66d27b"). InnerVolumeSpecName "db-sync-config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 10 23:09:14 crc kubenswrapper[4791]: I1210 23:09:14.327044 4791 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/851556d7-fa38-4eec-ae74-de989d66d27b-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "851556d7-fa38-4eec-ae74-de989d66d27b" (UID: "851556d7-fa38-4eec-ae74-de989d66d27b"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 10 23:09:14 crc kubenswrapper[4791]: I1210 23:09:14.387937 4791 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-crjrd\" (UniqueName: \"kubernetes.io/projected/851556d7-fa38-4eec-ae74-de989d66d27b-kube-api-access-crjrd\") on node \"crc\" DevicePath \"\"" Dec 10 23:09:14 crc kubenswrapper[4791]: I1210 23:09:14.388231 4791 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/851556d7-fa38-4eec-ae74-de989d66d27b-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 10 23:09:14 crc kubenswrapper[4791]: I1210 23:09:14.388243 4791 reconciler_common.go:293] "Volume detached for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/851556d7-fa38-4eec-ae74-de989d66d27b-db-sync-config-data\") on node \"crc\" DevicePath \"\"" Dec 10 23:09:14 crc kubenswrapper[4791]: I1210 23:09:14.619842 4791 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-db-sync-zllvm" event={"ID":"851556d7-fa38-4eec-ae74-de989d66d27b","Type":"ContainerDied","Data":"85645cc296d872c3ce2efa9f113d3205f9f481a970d015419a478a6a5c20d362"} Dec 10 23:09:14 crc kubenswrapper[4791]: I1210 23:09:14.619890 4791 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="85645cc296d872c3ce2efa9f113d3205f9f481a970d015419a478a6a5c20d362" Dec 10 23:09:14 crc kubenswrapper[4791]: I1210 23:09:14.619961 4791 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-db-sync-zllvm" Dec 10 23:09:14 crc kubenswrapper[4791]: I1210 23:09:14.631653 4791 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/cinder-api-0" podUID="1a1ca37c-dab5-4748-b8c2-eb985e66a02f" containerName="cinder-api-log" containerID="cri-o://831488e7d1696daf20de78611d91cff3c3a4661569db23dac805484a70013576" gracePeriod=30 Dec 10 23:09:14 crc kubenswrapper[4791]: I1210 23:09:14.631958 4791 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-api-0" event={"ID":"1a1ca37c-dab5-4748-b8c2-eb985e66a02f","Type":"ContainerStarted","Data":"c9727fa5dc2cf41c8e8fb4fe2a685691a2061b0a9272331473fef258a5ed514f"} Dec 10 23:09:14 crc kubenswrapper[4791]: I1210 23:09:14.632010 4791 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/cinder-api-0" Dec 10 23:09:14 crc kubenswrapper[4791]: I1210 23:09:14.632296 4791 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/cinder-api-0" podUID="1a1ca37c-dab5-4748-b8c2-eb985e66a02f" containerName="cinder-api" containerID="cri-o://c9727fa5dc2cf41c8e8fb4fe2a685691a2061b0a9272331473fef258a5ed514f" gracePeriod=30 Dec 10 23:09:14 crc kubenswrapper[4791]: I1210 23:09:14.676730 4791 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/cinder-api-0" podStartSLOduration=3.676710129 podStartE2EDuration="3.676710129s" podCreationTimestamp="2025-12-10 23:09:11 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-10 23:09:14.648082763 +0000 UTC m=+1189.077700376" watchObservedRunningTime="2025-12-10 23:09:14.676710129 +0000 UTC m=+1189.106327742" Dec 10 23:09:15 crc kubenswrapper[4791]: I1210 23:09:15.028038 4791 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/barbican-worker-77fb99879-xzsrk"] Dec 10 23:09:15 crc kubenswrapper[4791]: E1210 23:09:15.028570 4791 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="851556d7-fa38-4eec-ae74-de989d66d27b" containerName="barbican-db-sync" Dec 10 23:09:15 crc kubenswrapper[4791]: I1210 23:09:15.028590 4791 state_mem.go:107] "Deleted CPUSet assignment" podUID="851556d7-fa38-4eec-ae74-de989d66d27b" containerName="barbican-db-sync" Dec 10 23:09:15 crc kubenswrapper[4791]: I1210 23:09:15.028850 4791 memory_manager.go:354] "RemoveStaleState removing state" podUID="851556d7-fa38-4eec-ae74-de989d66d27b" containerName="barbican-db-sync" Dec 10 23:09:15 crc kubenswrapper[4791]: I1210 23:09:15.029939 4791 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-worker-77fb99879-xzsrk" Dec 10 23:09:15 crc kubenswrapper[4791]: I1210 23:09:15.036859 4791 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"barbican-barbican-dockercfg-rkxgd" Dec 10 23:09:15 crc kubenswrapper[4791]: I1210 23:09:15.037758 4791 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"barbican-config-data" Dec 10 23:09:15 crc kubenswrapper[4791]: I1210 23:09:15.040394 4791 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"barbican-worker-config-data" Dec 10 23:09:15 crc kubenswrapper[4791]: I1210 23:09:15.063474 4791 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/barbican-worker-77fb99879-xzsrk"] Dec 10 23:09:15 crc kubenswrapper[4791]: I1210 23:09:15.106708 4791 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-rv449\" (UniqueName: \"kubernetes.io/projected/8a302b7c-c2df-4023-95da-4ef8e86dbc0b-kube-api-access-rv449\") pod \"barbican-worker-77fb99879-xzsrk\" (UID: \"8a302b7c-c2df-4023-95da-4ef8e86dbc0b\") " pod="openstack/barbican-worker-77fb99879-xzsrk" Dec 10 23:09:15 crc kubenswrapper[4791]: I1210 23:09:15.106760 4791 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/8a302b7c-c2df-4023-95da-4ef8e86dbc0b-config-data\") pod \"barbican-worker-77fb99879-xzsrk\" (UID: \"8a302b7c-c2df-4023-95da-4ef8e86dbc0b\") " pod="openstack/barbican-worker-77fb99879-xzsrk" Dec 10 23:09:15 crc kubenswrapper[4791]: I1210 23:09:15.106807 4791 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/8a302b7c-c2df-4023-95da-4ef8e86dbc0b-config-data-custom\") pod \"barbican-worker-77fb99879-xzsrk\" (UID: \"8a302b7c-c2df-4023-95da-4ef8e86dbc0b\") " pod="openstack/barbican-worker-77fb99879-xzsrk" Dec 10 23:09:15 crc kubenswrapper[4791]: I1210 23:09:15.106822 4791 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/8a302b7c-c2df-4023-95da-4ef8e86dbc0b-combined-ca-bundle\") pod \"barbican-worker-77fb99879-xzsrk\" (UID: \"8a302b7c-c2df-4023-95da-4ef8e86dbc0b\") " pod="openstack/barbican-worker-77fb99879-xzsrk" Dec 10 23:09:15 crc kubenswrapper[4791]: I1210 23:09:15.106888 4791 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/8a302b7c-c2df-4023-95da-4ef8e86dbc0b-logs\") pod \"barbican-worker-77fb99879-xzsrk\" (UID: \"8a302b7c-c2df-4023-95da-4ef8e86dbc0b\") " pod="openstack/barbican-worker-77fb99879-xzsrk" Dec 10 23:09:15 crc kubenswrapper[4791]: I1210 23:09:15.164479 4791 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/barbican-keystone-listener-67867746d6-mvxp4"] Dec 10 23:09:15 crc kubenswrapper[4791]: I1210 23:09:15.167502 4791 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-keystone-listener-67867746d6-mvxp4" Dec 10 23:09:15 crc kubenswrapper[4791]: I1210 23:09:15.169078 4791 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"barbican-keystone-listener-config-data" Dec 10 23:09:15 crc kubenswrapper[4791]: I1210 23:09:15.210823 4791 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/barbican-keystone-listener-67867746d6-mvxp4"] Dec 10 23:09:15 crc kubenswrapper[4791]: I1210 23:09:15.213622 4791 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/4d7d941e-8494-46ea-990c-fe83db67dd2a-config-data-custom\") pod \"barbican-keystone-listener-67867746d6-mvxp4\" (UID: \"4d7d941e-8494-46ea-990c-fe83db67dd2a\") " pod="openstack/barbican-keystone-listener-67867746d6-mvxp4" Dec 10 23:09:15 crc kubenswrapper[4791]: I1210 23:09:15.213680 4791 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/4d7d941e-8494-46ea-990c-fe83db67dd2a-logs\") pod \"barbican-keystone-listener-67867746d6-mvxp4\" (UID: \"4d7d941e-8494-46ea-990c-fe83db67dd2a\") " pod="openstack/barbican-keystone-listener-67867746d6-mvxp4" Dec 10 23:09:15 crc kubenswrapper[4791]: I1210 23:09:15.213811 4791 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-rv449\" (UniqueName: \"kubernetes.io/projected/8a302b7c-c2df-4023-95da-4ef8e86dbc0b-kube-api-access-rv449\") pod \"barbican-worker-77fb99879-xzsrk\" (UID: \"8a302b7c-c2df-4023-95da-4ef8e86dbc0b\") " pod="openstack/barbican-worker-77fb99879-xzsrk" Dec 10 23:09:15 crc kubenswrapper[4791]: I1210 23:09:15.213843 4791 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-bl8qt\" (UniqueName: \"kubernetes.io/projected/4d7d941e-8494-46ea-990c-fe83db67dd2a-kube-api-access-bl8qt\") pod \"barbican-keystone-listener-67867746d6-mvxp4\" (UID: \"4d7d941e-8494-46ea-990c-fe83db67dd2a\") " pod="openstack/barbican-keystone-listener-67867746d6-mvxp4" Dec 10 23:09:15 crc kubenswrapper[4791]: I1210 23:09:15.213876 4791 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/8a302b7c-c2df-4023-95da-4ef8e86dbc0b-config-data\") pod \"barbican-worker-77fb99879-xzsrk\" (UID: \"8a302b7c-c2df-4023-95da-4ef8e86dbc0b\") " pod="openstack/barbican-worker-77fb99879-xzsrk" Dec 10 23:09:15 crc kubenswrapper[4791]: I1210 23:09:15.213933 4791 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/4d7d941e-8494-46ea-990c-fe83db67dd2a-config-data\") pod \"barbican-keystone-listener-67867746d6-mvxp4\" (UID: \"4d7d941e-8494-46ea-990c-fe83db67dd2a\") " pod="openstack/barbican-keystone-listener-67867746d6-mvxp4" Dec 10 23:09:15 crc kubenswrapper[4791]: I1210 23:09:15.213972 4791 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/8a302b7c-c2df-4023-95da-4ef8e86dbc0b-config-data-custom\") pod \"barbican-worker-77fb99879-xzsrk\" (UID: \"8a302b7c-c2df-4023-95da-4ef8e86dbc0b\") " pod="openstack/barbican-worker-77fb99879-xzsrk" Dec 10 23:09:15 crc kubenswrapper[4791]: I1210 23:09:15.213991 4791 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/8a302b7c-c2df-4023-95da-4ef8e86dbc0b-combined-ca-bundle\") pod \"barbican-worker-77fb99879-xzsrk\" (UID: \"8a302b7c-c2df-4023-95da-4ef8e86dbc0b\") " pod="openstack/barbican-worker-77fb99879-xzsrk" Dec 10 23:09:15 crc kubenswrapper[4791]: I1210 23:09:15.214014 4791 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/4d7d941e-8494-46ea-990c-fe83db67dd2a-combined-ca-bundle\") pod \"barbican-keystone-listener-67867746d6-mvxp4\" (UID: \"4d7d941e-8494-46ea-990c-fe83db67dd2a\") " pod="openstack/barbican-keystone-listener-67867746d6-mvxp4" Dec 10 23:09:15 crc kubenswrapper[4791]: I1210 23:09:15.214086 4791 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/8a302b7c-c2df-4023-95da-4ef8e86dbc0b-logs\") pod \"barbican-worker-77fb99879-xzsrk\" (UID: \"8a302b7c-c2df-4023-95da-4ef8e86dbc0b\") " pod="openstack/barbican-worker-77fb99879-xzsrk" Dec 10 23:09:15 crc kubenswrapper[4791]: I1210 23:09:15.214682 4791 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/8a302b7c-c2df-4023-95da-4ef8e86dbc0b-logs\") pod \"barbican-worker-77fb99879-xzsrk\" (UID: \"8a302b7c-c2df-4023-95da-4ef8e86dbc0b\") " pod="openstack/barbican-worker-77fb99879-xzsrk" Dec 10 23:09:15 crc kubenswrapper[4791]: I1210 23:09:15.227846 4791 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/8a302b7c-c2df-4023-95da-4ef8e86dbc0b-combined-ca-bundle\") pod \"barbican-worker-77fb99879-xzsrk\" (UID: \"8a302b7c-c2df-4023-95da-4ef8e86dbc0b\") " pod="openstack/barbican-worker-77fb99879-xzsrk" Dec 10 23:09:15 crc kubenswrapper[4791]: I1210 23:09:15.228311 4791 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/8a302b7c-c2df-4023-95da-4ef8e86dbc0b-config-data-custom\") pod \"barbican-worker-77fb99879-xzsrk\" (UID: \"8a302b7c-c2df-4023-95da-4ef8e86dbc0b\") " pod="openstack/barbican-worker-77fb99879-xzsrk" Dec 10 23:09:15 crc kubenswrapper[4791]: I1210 23:09:15.252888 4791 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-rv449\" (UniqueName: \"kubernetes.io/projected/8a302b7c-c2df-4023-95da-4ef8e86dbc0b-kube-api-access-rv449\") pod \"barbican-worker-77fb99879-xzsrk\" (UID: \"8a302b7c-c2df-4023-95da-4ef8e86dbc0b\") " pod="openstack/barbican-worker-77fb99879-xzsrk" Dec 10 23:09:15 crc kubenswrapper[4791]: I1210 23:09:15.256020 4791 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/8a302b7c-c2df-4023-95da-4ef8e86dbc0b-config-data\") pod \"barbican-worker-77fb99879-xzsrk\" (UID: \"8a302b7c-c2df-4023-95da-4ef8e86dbc0b\") " pod="openstack/barbican-worker-77fb99879-xzsrk" Dec 10 23:09:15 crc kubenswrapper[4791]: I1210 23:09:15.304660 4791 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-d68b9cb4c-xp29n"] Dec 10 23:09:15 crc kubenswrapper[4791]: I1210 23:09:15.316146 4791 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/4d7d941e-8494-46ea-990c-fe83db67dd2a-config-data\") pod \"barbican-keystone-listener-67867746d6-mvxp4\" (UID: \"4d7d941e-8494-46ea-990c-fe83db67dd2a\") " pod="openstack/barbican-keystone-listener-67867746d6-mvxp4" Dec 10 23:09:15 crc kubenswrapper[4791]: I1210 23:09:15.316365 4791 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/4d7d941e-8494-46ea-990c-fe83db67dd2a-combined-ca-bundle\") pod \"barbican-keystone-listener-67867746d6-mvxp4\" (UID: \"4d7d941e-8494-46ea-990c-fe83db67dd2a\") " pod="openstack/barbican-keystone-listener-67867746d6-mvxp4" Dec 10 23:09:15 crc kubenswrapper[4791]: I1210 23:09:15.316644 4791 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/4d7d941e-8494-46ea-990c-fe83db67dd2a-config-data-custom\") pod \"barbican-keystone-listener-67867746d6-mvxp4\" (UID: \"4d7d941e-8494-46ea-990c-fe83db67dd2a\") " pod="openstack/barbican-keystone-listener-67867746d6-mvxp4" Dec 10 23:09:15 crc kubenswrapper[4791]: I1210 23:09:15.316771 4791 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/4d7d941e-8494-46ea-990c-fe83db67dd2a-logs\") pod \"barbican-keystone-listener-67867746d6-mvxp4\" (UID: \"4d7d941e-8494-46ea-990c-fe83db67dd2a\") " pod="openstack/barbican-keystone-listener-67867746d6-mvxp4" Dec 10 23:09:15 crc kubenswrapper[4791]: I1210 23:09:15.316933 4791 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-bl8qt\" (UniqueName: \"kubernetes.io/projected/4d7d941e-8494-46ea-990c-fe83db67dd2a-kube-api-access-bl8qt\") pod \"barbican-keystone-listener-67867746d6-mvxp4\" (UID: \"4d7d941e-8494-46ea-990c-fe83db67dd2a\") " pod="openstack/barbican-keystone-listener-67867746d6-mvxp4" Dec 10 23:09:15 crc kubenswrapper[4791]: I1210 23:09:15.319240 4791 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/4d7d941e-8494-46ea-990c-fe83db67dd2a-logs\") pod \"barbican-keystone-listener-67867746d6-mvxp4\" (UID: \"4d7d941e-8494-46ea-990c-fe83db67dd2a\") " pod="openstack/barbican-keystone-listener-67867746d6-mvxp4" Dec 10 23:09:15 crc kubenswrapper[4791]: I1210 23:09:15.337752 4791 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/4d7d941e-8494-46ea-990c-fe83db67dd2a-config-data-custom\") pod \"barbican-keystone-listener-67867746d6-mvxp4\" (UID: \"4d7d941e-8494-46ea-990c-fe83db67dd2a\") " pod="openstack/barbican-keystone-listener-67867746d6-mvxp4" Dec 10 23:09:15 crc kubenswrapper[4791]: I1210 23:09:15.342350 4791 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-worker-77fb99879-xzsrk" Dec 10 23:09:15 crc kubenswrapper[4791]: I1210 23:09:15.347650 4791 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/4d7d941e-8494-46ea-990c-fe83db67dd2a-config-data\") pod \"barbican-keystone-listener-67867746d6-mvxp4\" (UID: \"4d7d941e-8494-46ea-990c-fe83db67dd2a\") " pod="openstack/barbican-keystone-listener-67867746d6-mvxp4" Dec 10 23:09:15 crc kubenswrapper[4791]: I1210 23:09:15.351406 4791 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-5784cf869f-ngjqj"] Dec 10 23:09:15 crc kubenswrapper[4791]: I1210 23:09:15.353839 4791 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-5784cf869f-ngjqj" Dec 10 23:09:15 crc kubenswrapper[4791]: I1210 23:09:15.369450 4791 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-5784cf869f-ngjqj"] Dec 10 23:09:15 crc kubenswrapper[4791]: I1210 23:09:15.374864 4791 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/4d7d941e-8494-46ea-990c-fe83db67dd2a-combined-ca-bundle\") pod \"barbican-keystone-listener-67867746d6-mvxp4\" (UID: \"4d7d941e-8494-46ea-990c-fe83db67dd2a\") " pod="openstack/barbican-keystone-listener-67867746d6-mvxp4" Dec 10 23:09:15 crc kubenswrapper[4791]: I1210 23:09:15.380130 4791 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-bl8qt\" (UniqueName: \"kubernetes.io/projected/4d7d941e-8494-46ea-990c-fe83db67dd2a-kube-api-access-bl8qt\") pod \"barbican-keystone-listener-67867746d6-mvxp4\" (UID: \"4d7d941e-8494-46ea-990c-fe83db67dd2a\") " pod="openstack/barbican-keystone-listener-67867746d6-mvxp4" Dec 10 23:09:15 crc kubenswrapper[4791]: I1210 23:09:15.396781 4791 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/barbican-api-795f5df4db-xf57n"] Dec 10 23:09:15 crc kubenswrapper[4791]: I1210 23:09:15.399116 4791 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-api-795f5df4db-xf57n" Dec 10 23:09:15 crc kubenswrapper[4791]: I1210 23:09:15.404361 4791 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"barbican-api-config-data" Dec 10 23:09:15 crc kubenswrapper[4791]: I1210 23:09:15.421294 4791 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/224a8e50-4270-42e8-a1df-268bc144b1d5-ovsdbserver-nb\") pod \"dnsmasq-dns-5784cf869f-ngjqj\" (UID: \"224a8e50-4270-42e8-a1df-268bc144b1d5\") " pod="openstack/dnsmasq-dns-5784cf869f-ngjqj" Dec 10 23:09:15 crc kubenswrapper[4791]: I1210 23:09:15.421509 4791 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/224a8e50-4270-42e8-a1df-268bc144b1d5-dns-swift-storage-0\") pod \"dnsmasq-dns-5784cf869f-ngjqj\" (UID: \"224a8e50-4270-42e8-a1df-268bc144b1d5\") " pod="openstack/dnsmasq-dns-5784cf869f-ngjqj" Dec 10 23:09:15 crc kubenswrapper[4791]: I1210 23:09:15.421539 4791 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-xfqw8\" (UniqueName: \"kubernetes.io/projected/224a8e50-4270-42e8-a1df-268bc144b1d5-kube-api-access-xfqw8\") pod \"dnsmasq-dns-5784cf869f-ngjqj\" (UID: \"224a8e50-4270-42e8-a1df-268bc144b1d5\") " pod="openstack/dnsmasq-dns-5784cf869f-ngjqj" Dec 10 23:09:15 crc kubenswrapper[4791]: I1210 23:09:15.421611 4791 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/224a8e50-4270-42e8-a1df-268bc144b1d5-config\") pod \"dnsmasq-dns-5784cf869f-ngjqj\" (UID: \"224a8e50-4270-42e8-a1df-268bc144b1d5\") " pod="openstack/dnsmasq-dns-5784cf869f-ngjqj" Dec 10 23:09:15 crc kubenswrapper[4791]: I1210 23:09:15.421649 4791 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/224a8e50-4270-42e8-a1df-268bc144b1d5-dns-svc\") pod \"dnsmasq-dns-5784cf869f-ngjqj\" (UID: \"224a8e50-4270-42e8-a1df-268bc144b1d5\") " pod="openstack/dnsmasq-dns-5784cf869f-ngjqj" Dec 10 23:09:15 crc kubenswrapper[4791]: I1210 23:09:15.421669 4791 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/224a8e50-4270-42e8-a1df-268bc144b1d5-ovsdbserver-sb\") pod \"dnsmasq-dns-5784cf869f-ngjqj\" (UID: \"224a8e50-4270-42e8-a1df-268bc144b1d5\") " pod="openstack/dnsmasq-dns-5784cf869f-ngjqj" Dec 10 23:09:15 crc kubenswrapper[4791]: I1210 23:09:15.437483 4791 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/barbican-api-795f5df4db-xf57n"] Dec 10 23:09:15 crc kubenswrapper[4791]: I1210 23:09:15.469757 4791 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-keystone-listener-67867746d6-mvxp4" Dec 10 23:09:15 crc kubenswrapper[4791]: I1210 23:09:15.534108 4791 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/9b340efd-bf7e-4caf-8c76-ddc500aa2085-logs\") pod \"barbican-api-795f5df4db-xf57n\" (UID: \"9b340efd-bf7e-4caf-8c76-ddc500aa2085\") " pod="openstack/barbican-api-795f5df4db-xf57n" Dec 10 23:09:15 crc kubenswrapper[4791]: I1210 23:09:15.534660 4791 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/9b340efd-bf7e-4caf-8c76-ddc500aa2085-config-data\") pod \"barbican-api-795f5df4db-xf57n\" (UID: \"9b340efd-bf7e-4caf-8c76-ddc500aa2085\") " pod="openstack/barbican-api-795f5df4db-xf57n" Dec 10 23:09:15 crc kubenswrapper[4791]: I1210 23:09:15.534785 4791 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/224a8e50-4270-42e8-a1df-268bc144b1d5-ovsdbserver-nb\") pod \"dnsmasq-dns-5784cf869f-ngjqj\" (UID: \"224a8e50-4270-42e8-a1df-268bc144b1d5\") " pod="openstack/dnsmasq-dns-5784cf869f-ngjqj" Dec 10 23:09:15 crc kubenswrapper[4791]: I1210 23:09:15.534986 4791 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-vnldb\" (UniqueName: \"kubernetes.io/projected/9b340efd-bf7e-4caf-8c76-ddc500aa2085-kube-api-access-vnldb\") pod \"barbican-api-795f5df4db-xf57n\" (UID: \"9b340efd-bf7e-4caf-8c76-ddc500aa2085\") " pod="openstack/barbican-api-795f5df4db-xf57n" Dec 10 23:09:15 crc kubenswrapper[4791]: I1210 23:09:15.535028 4791 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/224a8e50-4270-42e8-a1df-268bc144b1d5-dns-swift-storage-0\") pod \"dnsmasq-dns-5784cf869f-ngjqj\" (UID: \"224a8e50-4270-42e8-a1df-268bc144b1d5\") " pod="openstack/dnsmasq-dns-5784cf869f-ngjqj" Dec 10 23:09:15 crc kubenswrapper[4791]: I1210 23:09:15.535078 4791 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-xfqw8\" (UniqueName: \"kubernetes.io/projected/224a8e50-4270-42e8-a1df-268bc144b1d5-kube-api-access-xfqw8\") pod \"dnsmasq-dns-5784cf869f-ngjqj\" (UID: \"224a8e50-4270-42e8-a1df-268bc144b1d5\") " pod="openstack/dnsmasq-dns-5784cf869f-ngjqj" Dec 10 23:09:15 crc kubenswrapper[4791]: I1210 23:09:15.535104 4791 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/9b340efd-bf7e-4caf-8c76-ddc500aa2085-combined-ca-bundle\") pod \"barbican-api-795f5df4db-xf57n\" (UID: \"9b340efd-bf7e-4caf-8c76-ddc500aa2085\") " pod="openstack/barbican-api-795f5df4db-xf57n" Dec 10 23:09:15 crc kubenswrapper[4791]: I1210 23:09:15.535194 4791 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/9b340efd-bf7e-4caf-8c76-ddc500aa2085-config-data-custom\") pod \"barbican-api-795f5df4db-xf57n\" (UID: \"9b340efd-bf7e-4caf-8c76-ddc500aa2085\") " pod="openstack/barbican-api-795f5df4db-xf57n" Dec 10 23:09:15 crc kubenswrapper[4791]: I1210 23:09:15.535254 4791 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/224a8e50-4270-42e8-a1df-268bc144b1d5-config\") pod \"dnsmasq-dns-5784cf869f-ngjqj\" (UID: \"224a8e50-4270-42e8-a1df-268bc144b1d5\") " pod="openstack/dnsmasq-dns-5784cf869f-ngjqj" Dec 10 23:09:15 crc kubenswrapper[4791]: I1210 23:09:15.535294 4791 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/224a8e50-4270-42e8-a1df-268bc144b1d5-dns-svc\") pod \"dnsmasq-dns-5784cf869f-ngjqj\" (UID: \"224a8e50-4270-42e8-a1df-268bc144b1d5\") " pod="openstack/dnsmasq-dns-5784cf869f-ngjqj" Dec 10 23:09:15 crc kubenswrapper[4791]: I1210 23:09:15.535317 4791 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/224a8e50-4270-42e8-a1df-268bc144b1d5-ovsdbserver-sb\") pod \"dnsmasq-dns-5784cf869f-ngjqj\" (UID: \"224a8e50-4270-42e8-a1df-268bc144b1d5\") " pod="openstack/dnsmasq-dns-5784cf869f-ngjqj" Dec 10 23:09:15 crc kubenswrapper[4791]: I1210 23:09:15.536511 4791 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/224a8e50-4270-42e8-a1df-268bc144b1d5-ovsdbserver-sb\") pod \"dnsmasq-dns-5784cf869f-ngjqj\" (UID: \"224a8e50-4270-42e8-a1df-268bc144b1d5\") " pod="openstack/dnsmasq-dns-5784cf869f-ngjqj" Dec 10 23:09:15 crc kubenswrapper[4791]: I1210 23:09:15.537226 4791 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/224a8e50-4270-42e8-a1df-268bc144b1d5-ovsdbserver-nb\") pod \"dnsmasq-dns-5784cf869f-ngjqj\" (UID: \"224a8e50-4270-42e8-a1df-268bc144b1d5\") " pod="openstack/dnsmasq-dns-5784cf869f-ngjqj" Dec 10 23:09:15 crc kubenswrapper[4791]: I1210 23:09:15.540736 4791 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/224a8e50-4270-42e8-a1df-268bc144b1d5-dns-swift-storage-0\") pod \"dnsmasq-dns-5784cf869f-ngjqj\" (UID: \"224a8e50-4270-42e8-a1df-268bc144b1d5\") " pod="openstack/dnsmasq-dns-5784cf869f-ngjqj" Dec 10 23:09:15 crc kubenswrapper[4791]: I1210 23:09:15.541108 4791 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/224a8e50-4270-42e8-a1df-268bc144b1d5-config\") pod \"dnsmasq-dns-5784cf869f-ngjqj\" (UID: \"224a8e50-4270-42e8-a1df-268bc144b1d5\") " pod="openstack/dnsmasq-dns-5784cf869f-ngjqj" Dec 10 23:09:15 crc kubenswrapper[4791]: I1210 23:09:15.541957 4791 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/224a8e50-4270-42e8-a1df-268bc144b1d5-dns-svc\") pod \"dnsmasq-dns-5784cf869f-ngjqj\" (UID: \"224a8e50-4270-42e8-a1df-268bc144b1d5\") " pod="openstack/dnsmasq-dns-5784cf869f-ngjqj" Dec 10 23:09:15 crc kubenswrapper[4791]: I1210 23:09:15.573951 4791 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-xfqw8\" (UniqueName: \"kubernetes.io/projected/224a8e50-4270-42e8-a1df-268bc144b1d5-kube-api-access-xfqw8\") pod \"dnsmasq-dns-5784cf869f-ngjqj\" (UID: \"224a8e50-4270-42e8-a1df-268bc144b1d5\") " pod="openstack/dnsmasq-dns-5784cf869f-ngjqj" Dec 10 23:09:15 crc kubenswrapper[4791]: I1210 23:09:15.637999 4791 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-vnldb\" (UniqueName: \"kubernetes.io/projected/9b340efd-bf7e-4caf-8c76-ddc500aa2085-kube-api-access-vnldb\") pod \"barbican-api-795f5df4db-xf57n\" (UID: \"9b340efd-bf7e-4caf-8c76-ddc500aa2085\") " pod="openstack/barbican-api-795f5df4db-xf57n" Dec 10 23:09:15 crc kubenswrapper[4791]: I1210 23:09:15.638068 4791 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/9b340efd-bf7e-4caf-8c76-ddc500aa2085-combined-ca-bundle\") pod \"barbican-api-795f5df4db-xf57n\" (UID: \"9b340efd-bf7e-4caf-8c76-ddc500aa2085\") " pod="openstack/barbican-api-795f5df4db-xf57n" Dec 10 23:09:15 crc kubenswrapper[4791]: I1210 23:09:15.638139 4791 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/9b340efd-bf7e-4caf-8c76-ddc500aa2085-config-data-custom\") pod \"barbican-api-795f5df4db-xf57n\" (UID: \"9b340efd-bf7e-4caf-8c76-ddc500aa2085\") " pod="openstack/barbican-api-795f5df4db-xf57n" Dec 10 23:09:15 crc kubenswrapper[4791]: I1210 23:09:15.638176 4791 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/9b340efd-bf7e-4caf-8c76-ddc500aa2085-logs\") pod \"barbican-api-795f5df4db-xf57n\" (UID: \"9b340efd-bf7e-4caf-8c76-ddc500aa2085\") " pod="openstack/barbican-api-795f5df4db-xf57n" Dec 10 23:09:15 crc kubenswrapper[4791]: I1210 23:09:15.638208 4791 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/9b340efd-bf7e-4caf-8c76-ddc500aa2085-config-data\") pod \"barbican-api-795f5df4db-xf57n\" (UID: \"9b340efd-bf7e-4caf-8c76-ddc500aa2085\") " pod="openstack/barbican-api-795f5df4db-xf57n" Dec 10 23:09:15 crc kubenswrapper[4791]: I1210 23:09:15.648610 4791 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/9b340efd-bf7e-4caf-8c76-ddc500aa2085-logs\") pod \"barbican-api-795f5df4db-xf57n\" (UID: \"9b340efd-bf7e-4caf-8c76-ddc500aa2085\") " pod="openstack/barbican-api-795f5df4db-xf57n" Dec 10 23:09:15 crc kubenswrapper[4791]: I1210 23:09:15.651894 4791 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/9b340efd-bf7e-4caf-8c76-ddc500aa2085-combined-ca-bundle\") pod \"barbican-api-795f5df4db-xf57n\" (UID: \"9b340efd-bf7e-4caf-8c76-ddc500aa2085\") " pod="openstack/barbican-api-795f5df4db-xf57n" Dec 10 23:09:15 crc kubenswrapper[4791]: I1210 23:09:15.654595 4791 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/9b340efd-bf7e-4caf-8c76-ddc500aa2085-config-data\") pod \"barbican-api-795f5df4db-xf57n\" (UID: \"9b340efd-bf7e-4caf-8c76-ddc500aa2085\") " pod="openstack/barbican-api-795f5df4db-xf57n" Dec 10 23:09:15 crc kubenswrapper[4791]: I1210 23:09:15.664831 4791 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/9b340efd-bf7e-4caf-8c76-ddc500aa2085-config-data-custom\") pod \"barbican-api-795f5df4db-xf57n\" (UID: \"9b340efd-bf7e-4caf-8c76-ddc500aa2085\") " pod="openstack/barbican-api-795f5df4db-xf57n" Dec 10 23:09:15 crc kubenswrapper[4791]: I1210 23:09:15.683489 4791 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-vnldb\" (UniqueName: \"kubernetes.io/projected/9b340efd-bf7e-4caf-8c76-ddc500aa2085-kube-api-access-vnldb\") pod \"barbican-api-795f5df4db-xf57n\" (UID: \"9b340efd-bf7e-4caf-8c76-ddc500aa2085\") " pod="openstack/barbican-api-795f5df4db-xf57n" Dec 10 23:09:15 crc kubenswrapper[4791]: I1210 23:09:15.685766 4791 generic.go:334] "Generic (PLEG): container finished" podID="1a1ca37c-dab5-4748-b8c2-eb985e66a02f" containerID="831488e7d1696daf20de78611d91cff3c3a4661569db23dac805484a70013576" exitCode=143 Dec 10 23:09:15 crc kubenswrapper[4791]: I1210 23:09:15.685838 4791 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-api-0" event={"ID":"1a1ca37c-dab5-4748-b8c2-eb985e66a02f","Type":"ContainerDied","Data":"831488e7d1696daf20de78611d91cff3c3a4661569db23dac805484a70013576"} Dec 10 23:09:15 crc kubenswrapper[4791]: I1210 23:09:15.699468 4791 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-scheduler-0" event={"ID":"5e5e327e-b8dc-4bd6-af75-8aebc417ada9","Type":"ContainerStarted","Data":"468b836fc75ba6c8ac143ddca9787ab03833f45994d432d9850ebe13fc2251ed"} Dec 10 23:09:15 crc kubenswrapper[4791]: I1210 23:09:15.769814 4791 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-5784cf869f-ngjqj" Dec 10 23:09:15 crc kubenswrapper[4791]: I1210 23:09:15.791480 4791 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-api-795f5df4db-xf57n" Dec 10 23:09:15 crc kubenswrapper[4791]: I1210 23:09:15.985556 4791 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/barbican-worker-77fb99879-xzsrk"] Dec 10 23:09:16 crc kubenswrapper[4791]: I1210 23:09:16.123860 4791 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/barbican-keystone-listener-67867746d6-mvxp4"] Dec 10 23:09:16 crc kubenswrapper[4791]: I1210 23:09:16.436292 4791 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/barbican-api-795f5df4db-xf57n"] Dec 10 23:09:16 crc kubenswrapper[4791]: I1210 23:09:16.452021 4791 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-5784cf869f-ngjqj"] Dec 10 23:09:16 crc kubenswrapper[4791]: I1210 23:09:16.708658 4791 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-scheduler-0" event={"ID":"5e5e327e-b8dc-4bd6-af75-8aebc417ada9","Type":"ContainerStarted","Data":"0491f982573b4cf5b921d4eb64aecef9ffb9a9bed5a7cd16d5e70ffa514274f4"} Dec 10 23:09:16 crc kubenswrapper[4791]: I1210 23:09:16.720046 4791 generic.go:334] "Generic (PLEG): container finished" podID="1a1ca37c-dab5-4748-b8c2-eb985e66a02f" containerID="c9727fa5dc2cf41c8e8fb4fe2a685691a2061b0a9272331473fef258a5ed514f" exitCode=0 Dec 10 23:09:16 crc kubenswrapper[4791]: I1210 23:09:16.720137 4791 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-api-0" event={"ID":"1a1ca37c-dab5-4748-b8c2-eb985e66a02f","Type":"ContainerDied","Data":"c9727fa5dc2cf41c8e8fb4fe2a685691a2061b0a9272331473fef258a5ed514f"} Dec 10 23:09:16 crc kubenswrapper[4791]: I1210 23:09:16.720603 4791 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/dnsmasq-dns-d68b9cb4c-xp29n" podUID="ae2623b4-7b39-4661-b9af-caf32415d8b1" containerName="dnsmasq-dns" containerID="cri-o://8abb2f7b1e355fe0279461505e35f3aae94f2f3d4097ceb931b5ef3c52902873" gracePeriod=10 Dec 10 23:09:16 crc kubenswrapper[4791]: I1210 23:09:16.735229 4791 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/cinder-scheduler-0" podStartSLOduration=4.134141726 podStartE2EDuration="6.735212792s" podCreationTimestamp="2025-12-10 23:09:10 +0000 UTC" firstStartedPulling="2025-12-10 23:09:11.693660427 +0000 UTC m=+1186.123278040" lastFinishedPulling="2025-12-10 23:09:14.294731493 +0000 UTC m=+1188.724349106" observedRunningTime="2025-12-10 23:09:16.729915949 +0000 UTC m=+1191.159533572" watchObservedRunningTime="2025-12-10 23:09:16.735212792 +0000 UTC m=+1191.164830405" Dec 10 23:09:17 crc kubenswrapper[4791]: I1210 23:09:17.760466 4791 generic.go:334] "Generic (PLEG): container finished" podID="ae2623b4-7b39-4661-b9af-caf32415d8b1" containerID="8abb2f7b1e355fe0279461505e35f3aae94f2f3d4097ceb931b5ef3c52902873" exitCode=0 Dec 10 23:09:17 crc kubenswrapper[4791]: I1210 23:09:17.761497 4791 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-d68b9cb4c-xp29n" event={"ID":"ae2623b4-7b39-4661-b9af-caf32415d8b1","Type":"ContainerDied","Data":"8abb2f7b1e355fe0279461505e35f3aae94f2f3d4097ceb931b5ef3c52902873"} Dec 10 23:09:17 crc kubenswrapper[4791]: I1210 23:09:17.871153 4791 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/barbican-api-68fb764d56-4mfz4"] Dec 10 23:09:17 crc kubenswrapper[4791]: I1210 23:09:17.875744 4791 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-api-68fb764d56-4mfz4" Dec 10 23:09:17 crc kubenswrapper[4791]: I1210 23:09:17.878377 4791 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-barbican-internal-svc" Dec 10 23:09:17 crc kubenswrapper[4791]: I1210 23:09:17.879882 4791 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-barbican-public-svc" Dec 10 23:09:17 crc kubenswrapper[4791]: I1210 23:09:17.893592 4791 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/eb482cb0-7ac1-4097-8799-4a5d78b6b5e5-public-tls-certs\") pod \"barbican-api-68fb764d56-4mfz4\" (UID: \"eb482cb0-7ac1-4097-8799-4a5d78b6b5e5\") " pod="openstack/barbican-api-68fb764d56-4mfz4" Dec 10 23:09:17 crc kubenswrapper[4791]: I1210 23:09:17.893689 4791 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/eb482cb0-7ac1-4097-8799-4a5d78b6b5e5-internal-tls-certs\") pod \"barbican-api-68fb764d56-4mfz4\" (UID: \"eb482cb0-7ac1-4097-8799-4a5d78b6b5e5\") " pod="openstack/barbican-api-68fb764d56-4mfz4" Dec 10 23:09:17 crc kubenswrapper[4791]: I1210 23:09:17.893721 4791 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-t8v84\" (UniqueName: \"kubernetes.io/projected/eb482cb0-7ac1-4097-8799-4a5d78b6b5e5-kube-api-access-t8v84\") pod \"barbican-api-68fb764d56-4mfz4\" (UID: \"eb482cb0-7ac1-4097-8799-4a5d78b6b5e5\") " pod="openstack/barbican-api-68fb764d56-4mfz4" Dec 10 23:09:17 crc kubenswrapper[4791]: I1210 23:09:17.893766 4791 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/eb482cb0-7ac1-4097-8799-4a5d78b6b5e5-config-data\") pod \"barbican-api-68fb764d56-4mfz4\" (UID: \"eb482cb0-7ac1-4097-8799-4a5d78b6b5e5\") " pod="openstack/barbican-api-68fb764d56-4mfz4" Dec 10 23:09:17 crc kubenswrapper[4791]: I1210 23:09:17.893813 4791 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/eb482cb0-7ac1-4097-8799-4a5d78b6b5e5-logs\") pod \"barbican-api-68fb764d56-4mfz4\" (UID: \"eb482cb0-7ac1-4097-8799-4a5d78b6b5e5\") " pod="openstack/barbican-api-68fb764d56-4mfz4" Dec 10 23:09:17 crc kubenswrapper[4791]: I1210 23:09:17.893840 4791 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/eb482cb0-7ac1-4097-8799-4a5d78b6b5e5-combined-ca-bundle\") pod \"barbican-api-68fb764d56-4mfz4\" (UID: \"eb482cb0-7ac1-4097-8799-4a5d78b6b5e5\") " pod="openstack/barbican-api-68fb764d56-4mfz4" Dec 10 23:09:17 crc kubenswrapper[4791]: I1210 23:09:17.893885 4791 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/eb482cb0-7ac1-4097-8799-4a5d78b6b5e5-config-data-custom\") pod \"barbican-api-68fb764d56-4mfz4\" (UID: \"eb482cb0-7ac1-4097-8799-4a5d78b6b5e5\") " pod="openstack/barbican-api-68fb764d56-4mfz4" Dec 10 23:09:17 crc kubenswrapper[4791]: I1210 23:09:17.925807 4791 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/barbican-api-68fb764d56-4mfz4"] Dec 10 23:09:17 crc kubenswrapper[4791]: I1210 23:09:17.996323 4791 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/eb482cb0-7ac1-4097-8799-4a5d78b6b5e5-config-data\") pod \"barbican-api-68fb764d56-4mfz4\" (UID: \"eb482cb0-7ac1-4097-8799-4a5d78b6b5e5\") " pod="openstack/barbican-api-68fb764d56-4mfz4" Dec 10 23:09:17 crc kubenswrapper[4791]: I1210 23:09:17.996471 4791 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/eb482cb0-7ac1-4097-8799-4a5d78b6b5e5-logs\") pod \"barbican-api-68fb764d56-4mfz4\" (UID: \"eb482cb0-7ac1-4097-8799-4a5d78b6b5e5\") " pod="openstack/barbican-api-68fb764d56-4mfz4" Dec 10 23:09:17 crc kubenswrapper[4791]: I1210 23:09:17.996533 4791 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/eb482cb0-7ac1-4097-8799-4a5d78b6b5e5-combined-ca-bundle\") pod \"barbican-api-68fb764d56-4mfz4\" (UID: \"eb482cb0-7ac1-4097-8799-4a5d78b6b5e5\") " pod="openstack/barbican-api-68fb764d56-4mfz4" Dec 10 23:09:17 crc kubenswrapper[4791]: I1210 23:09:17.996614 4791 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/eb482cb0-7ac1-4097-8799-4a5d78b6b5e5-config-data-custom\") pod \"barbican-api-68fb764d56-4mfz4\" (UID: \"eb482cb0-7ac1-4097-8799-4a5d78b6b5e5\") " pod="openstack/barbican-api-68fb764d56-4mfz4" Dec 10 23:09:17 crc kubenswrapper[4791]: I1210 23:09:17.996712 4791 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/eb482cb0-7ac1-4097-8799-4a5d78b6b5e5-public-tls-certs\") pod \"barbican-api-68fb764d56-4mfz4\" (UID: \"eb482cb0-7ac1-4097-8799-4a5d78b6b5e5\") " pod="openstack/barbican-api-68fb764d56-4mfz4" Dec 10 23:09:17 crc kubenswrapper[4791]: I1210 23:09:17.996794 4791 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/eb482cb0-7ac1-4097-8799-4a5d78b6b5e5-internal-tls-certs\") pod \"barbican-api-68fb764d56-4mfz4\" (UID: \"eb482cb0-7ac1-4097-8799-4a5d78b6b5e5\") " pod="openstack/barbican-api-68fb764d56-4mfz4" Dec 10 23:09:17 crc kubenswrapper[4791]: I1210 23:09:17.996828 4791 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-t8v84\" (UniqueName: \"kubernetes.io/projected/eb482cb0-7ac1-4097-8799-4a5d78b6b5e5-kube-api-access-t8v84\") pod \"barbican-api-68fb764d56-4mfz4\" (UID: \"eb482cb0-7ac1-4097-8799-4a5d78b6b5e5\") " pod="openstack/barbican-api-68fb764d56-4mfz4" Dec 10 23:09:17 crc kubenswrapper[4791]: I1210 23:09:17.998140 4791 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/eb482cb0-7ac1-4097-8799-4a5d78b6b5e5-logs\") pod \"barbican-api-68fb764d56-4mfz4\" (UID: \"eb482cb0-7ac1-4097-8799-4a5d78b6b5e5\") " pod="openstack/barbican-api-68fb764d56-4mfz4" Dec 10 23:09:18 crc kubenswrapper[4791]: I1210 23:09:18.005715 4791 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/eb482cb0-7ac1-4097-8799-4a5d78b6b5e5-internal-tls-certs\") pod \"barbican-api-68fb764d56-4mfz4\" (UID: \"eb482cb0-7ac1-4097-8799-4a5d78b6b5e5\") " pod="openstack/barbican-api-68fb764d56-4mfz4" Dec 10 23:09:18 crc kubenswrapper[4791]: I1210 23:09:18.008226 4791 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/eb482cb0-7ac1-4097-8799-4a5d78b6b5e5-public-tls-certs\") pod \"barbican-api-68fb764d56-4mfz4\" (UID: \"eb482cb0-7ac1-4097-8799-4a5d78b6b5e5\") " pod="openstack/barbican-api-68fb764d56-4mfz4" Dec 10 23:09:18 crc kubenswrapper[4791]: I1210 23:09:18.025557 4791 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-t8v84\" (UniqueName: \"kubernetes.io/projected/eb482cb0-7ac1-4097-8799-4a5d78b6b5e5-kube-api-access-t8v84\") pod \"barbican-api-68fb764d56-4mfz4\" (UID: \"eb482cb0-7ac1-4097-8799-4a5d78b6b5e5\") " pod="openstack/barbican-api-68fb764d56-4mfz4" Dec 10 23:09:18 crc kubenswrapper[4791]: I1210 23:09:18.025634 4791 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/eb482cb0-7ac1-4097-8799-4a5d78b6b5e5-combined-ca-bundle\") pod \"barbican-api-68fb764d56-4mfz4\" (UID: \"eb482cb0-7ac1-4097-8799-4a5d78b6b5e5\") " pod="openstack/barbican-api-68fb764d56-4mfz4" Dec 10 23:09:18 crc kubenswrapper[4791]: I1210 23:09:18.025811 4791 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/eb482cb0-7ac1-4097-8799-4a5d78b6b5e5-config-data\") pod \"barbican-api-68fb764d56-4mfz4\" (UID: \"eb482cb0-7ac1-4097-8799-4a5d78b6b5e5\") " pod="openstack/barbican-api-68fb764d56-4mfz4" Dec 10 23:09:18 crc kubenswrapper[4791]: I1210 23:09:18.026775 4791 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/eb482cb0-7ac1-4097-8799-4a5d78b6b5e5-config-data-custom\") pod \"barbican-api-68fb764d56-4mfz4\" (UID: \"eb482cb0-7ac1-4097-8799-4a5d78b6b5e5\") " pod="openstack/barbican-api-68fb764d56-4mfz4" Dec 10 23:09:18 crc kubenswrapper[4791]: I1210 23:09:18.225083 4791 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-api-68fb764d56-4mfz4" Dec 10 23:09:19 crc kubenswrapper[4791]: I1210 23:09:19.410711 4791 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/horizon-7cc66567bd-pwb2l" podUID="55139c7c-e9fa-4539-b7b2-b2be7ee43b44" containerName="horizon" probeResult="failure" output="Get \"https://10.217.0.144:8443/dashboard/auth/login/?next=/dashboard/\": dial tcp 10.217.0.144:8443: connect: connection refused" Dec 10 23:09:21 crc kubenswrapper[4791]: I1210 23:09:21.018200 4791 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/cinder-scheduler-0" Dec 10 23:09:21 crc kubenswrapper[4791]: I1210 23:09:21.237396 4791 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/cinder-scheduler-0" Dec 10 23:09:21 crc kubenswrapper[4791]: W1210 23:09:21.267173 4791 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod4d7d941e_8494_46ea_990c_fe83db67dd2a.slice/crio-7b926f7c68d5f5a51768536fe872dfe34745e7da1a9634091fbc6175f21976ae WatchSource:0}: Error finding container 7b926f7c68d5f5a51768536fe872dfe34745e7da1a9634091fbc6175f21976ae: Status 404 returned error can't find the container with id 7b926f7c68d5f5a51768536fe872dfe34745e7da1a9634091fbc6175f21976ae Dec 10 23:09:21 crc kubenswrapper[4791]: W1210 23:09:21.267378 4791 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod9b340efd_bf7e_4caf_8c76_ddc500aa2085.slice/crio-8f95cb70af18d0c8b0d71d55cc909fb5e15dbe337f92069cd49742ac1b44c594 WatchSource:0}: Error finding container 8f95cb70af18d0c8b0d71d55cc909fb5e15dbe337f92069cd49742ac1b44c594: Status 404 returned error can't find the container with id 8f95cb70af18d0c8b0d71d55cc909fb5e15dbe337f92069cd49742ac1b44c594 Dec 10 23:09:21 crc kubenswrapper[4791]: W1210 23:09:21.283558 4791 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod224a8e50_4270_42e8_a1df_268bc144b1d5.slice/crio-7bacd3841f62b678498a49df871b9e42dccba73bee49d48be648aeb6a7d1abbb WatchSource:0}: Error finding container 7bacd3841f62b678498a49df871b9e42dccba73bee49d48be648aeb6a7d1abbb: Status 404 returned error can't find the container with id 7bacd3841f62b678498a49df871b9e42dccba73bee49d48be648aeb6a7d1abbb Dec 10 23:09:21 crc kubenswrapper[4791]: I1210 23:09:21.295633 4791 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/dnsmasq-dns-d68b9cb4c-xp29n" podUID="ae2623b4-7b39-4661-b9af-caf32415d8b1" containerName="dnsmasq-dns" probeResult="failure" output="dial tcp 10.217.0.158:5353: connect: connection refused" Dec 10 23:09:21 crc kubenswrapper[4791]: I1210 23:09:21.336191 4791 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/horizon-6999c5bf5b-7cr9t" Dec 10 23:09:21 crc kubenswrapper[4791]: I1210 23:09:21.814467 4791 generic.go:334] "Generic (PLEG): container finished" podID="14f46cbc-afe2-40bb-a2a3-5673b5ac53e7" containerID="1bf24a4c4c7be94d64b9919fd150ec64d3f744be3bf1b6f17541d5c0b7a27a30" exitCode=137 Dec 10 23:09:21 crc kubenswrapper[4791]: I1210 23:09:21.814854 4791 generic.go:334] "Generic (PLEG): container finished" podID="14f46cbc-afe2-40bb-a2a3-5673b5ac53e7" containerID="d5bed1e64f4f7d6b0cbb4910d9037ebbea49eb4ac927509fe80353feed0ed97d" exitCode=137 Dec 10 23:09:21 crc kubenswrapper[4791]: I1210 23:09:21.814640 4791 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/horizon-78fc5d56dc-58cjt" event={"ID":"14f46cbc-afe2-40bb-a2a3-5673b5ac53e7","Type":"ContainerDied","Data":"1bf24a4c4c7be94d64b9919fd150ec64d3f744be3bf1b6f17541d5c0b7a27a30"} Dec 10 23:09:21 crc kubenswrapper[4791]: I1210 23:09:21.814923 4791 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/horizon-78fc5d56dc-58cjt" event={"ID":"14f46cbc-afe2-40bb-a2a3-5673b5ac53e7","Type":"ContainerDied","Data":"d5bed1e64f4f7d6b0cbb4910d9037ebbea49eb4ac927509fe80353feed0ed97d"} Dec 10 23:09:21 crc kubenswrapper[4791]: I1210 23:09:21.817003 4791 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-5784cf869f-ngjqj" event={"ID":"224a8e50-4270-42e8-a1df-268bc144b1d5","Type":"ContainerStarted","Data":"7bacd3841f62b678498a49df871b9e42dccba73bee49d48be648aeb6a7d1abbb"} Dec 10 23:09:21 crc kubenswrapper[4791]: I1210 23:09:21.818924 4791 generic.go:334] "Generic (PLEG): container finished" podID="9e9c1419-c6f3-4947-8788-4915f48677d7" containerID="e8a0fecaed859f1a8a83f5c36eabd8013be451668e6363d31a5ebdbc8e9a0863" exitCode=137 Dec 10 23:09:21 crc kubenswrapper[4791]: I1210 23:09:21.818949 4791 generic.go:334] "Generic (PLEG): container finished" podID="9e9c1419-c6f3-4947-8788-4915f48677d7" containerID="9369387266f0bf17a3578a2973371e39479e41aeaf91821d2bd23efe9fbbe09e" exitCode=137 Dec 10 23:09:21 crc kubenswrapper[4791]: I1210 23:09:21.819011 4791 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/horizon-5cc98b55bf-7xpnj" event={"ID":"9e9c1419-c6f3-4947-8788-4915f48677d7","Type":"ContainerDied","Data":"e8a0fecaed859f1a8a83f5c36eabd8013be451668e6363d31a5ebdbc8e9a0863"} Dec 10 23:09:21 crc kubenswrapper[4791]: I1210 23:09:21.819027 4791 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/horizon-5cc98b55bf-7xpnj" event={"ID":"9e9c1419-c6f3-4947-8788-4915f48677d7","Type":"ContainerDied","Data":"9369387266f0bf17a3578a2973371e39479e41aeaf91821d2bd23efe9fbbe09e"} Dec 10 23:09:21 crc kubenswrapper[4791]: I1210 23:09:21.821433 4791 generic.go:334] "Generic (PLEG): container finished" podID="aa3a63ab-6092-46ce-aaf6-40360aa060ac" containerID="51262b7b3d283cb5b5b3f2953648269ed55c1d6a470bbdf743fc7e6a314d9407" exitCode=137 Dec 10 23:09:21 crc kubenswrapper[4791]: I1210 23:09:21.821450 4791 generic.go:334] "Generic (PLEG): container finished" podID="aa3a63ab-6092-46ce-aaf6-40360aa060ac" containerID="35844ea7a50bd278dafc5819fb34072df86afbe70e63e900d65a4f0f56c5cfc8" exitCode=137 Dec 10 23:09:21 crc kubenswrapper[4791]: I1210 23:09:21.821490 4791 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/horizon-5f54ccbf57-wwpfk" event={"ID":"aa3a63ab-6092-46ce-aaf6-40360aa060ac","Type":"ContainerDied","Data":"51262b7b3d283cb5b5b3f2953648269ed55c1d6a470bbdf743fc7e6a314d9407"} Dec 10 23:09:21 crc kubenswrapper[4791]: I1210 23:09:21.821506 4791 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/horizon-5f54ccbf57-wwpfk" event={"ID":"aa3a63ab-6092-46ce-aaf6-40360aa060ac","Type":"ContainerDied","Data":"35844ea7a50bd278dafc5819fb34072df86afbe70e63e900d65a4f0f56c5cfc8"} Dec 10 23:09:21 crc kubenswrapper[4791]: I1210 23:09:21.822705 4791 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-keystone-listener-67867746d6-mvxp4" event={"ID":"4d7d941e-8494-46ea-990c-fe83db67dd2a","Type":"ContainerStarted","Data":"7b926f7c68d5f5a51768536fe872dfe34745e7da1a9634091fbc6175f21976ae"} Dec 10 23:09:21 crc kubenswrapper[4791]: I1210 23:09:21.823947 4791 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-api-795f5df4db-xf57n" event={"ID":"9b340efd-bf7e-4caf-8c76-ddc500aa2085","Type":"ContainerStarted","Data":"8f95cb70af18d0c8b0d71d55cc909fb5e15dbe337f92069cd49742ac1b44c594"} Dec 10 23:09:21 crc kubenswrapper[4791]: I1210 23:09:21.826079 4791 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-worker-77fb99879-xzsrk" event={"ID":"8a302b7c-c2df-4023-95da-4ef8e86dbc0b","Type":"ContainerStarted","Data":"c4ee13d0a0fa94d5cc11d3cd4ab5d86b60a6279cfb68f032eb8bd39e813daa2e"} Dec 10 23:09:21 crc kubenswrapper[4791]: I1210 23:09:21.904714 4791 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/neutron-85794c6bd6-ht8sz" Dec 10 23:09:21 crc kubenswrapper[4791]: I1210 23:09:21.906060 4791 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/cinder-scheduler-0"] Dec 10 23:09:22 crc kubenswrapper[4791]: I1210 23:09:22.530773 4791 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-api-0" Dec 10 23:09:22 crc kubenswrapper[4791]: I1210 23:09:22.538951 4791 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-d68b9cb4c-xp29n" Dec 10 23:09:22 crc kubenswrapper[4791]: I1210 23:09:22.632208 4791 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/1a1ca37c-dab5-4748-b8c2-eb985e66a02f-etc-machine-id\") pod \"1a1ca37c-dab5-4748-b8c2-eb985e66a02f\" (UID: \"1a1ca37c-dab5-4748-b8c2-eb985e66a02f\") " Dec 10 23:09:22 crc kubenswrapper[4791]: I1210 23:09:22.632300 4791 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/1a1ca37c-dab5-4748-b8c2-eb985e66a02f-config-data-custom\") pod \"1a1ca37c-dab5-4748-b8c2-eb985e66a02f\" (UID: \"1a1ca37c-dab5-4748-b8c2-eb985e66a02f\") " Dec 10 23:09:22 crc kubenswrapper[4791]: I1210 23:09:22.632405 4791 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/1a1ca37c-dab5-4748-b8c2-eb985e66a02f-etc-machine-id" (OuterVolumeSpecName: "etc-machine-id") pod "1a1ca37c-dab5-4748-b8c2-eb985e66a02f" (UID: "1a1ca37c-dab5-4748-b8c2-eb985e66a02f"). InnerVolumeSpecName "etc-machine-id". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 10 23:09:22 crc kubenswrapper[4791]: I1210 23:09:22.632441 4791 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/1a1ca37c-dab5-4748-b8c2-eb985e66a02f-config-data\") pod \"1a1ca37c-dab5-4748-b8c2-eb985e66a02f\" (UID: \"1a1ca37c-dab5-4748-b8c2-eb985e66a02f\") " Dec 10 23:09:22 crc kubenswrapper[4791]: I1210 23:09:22.632520 4791 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/ae2623b4-7b39-4661-b9af-caf32415d8b1-dns-svc\") pod \"ae2623b4-7b39-4661-b9af-caf32415d8b1\" (UID: \"ae2623b4-7b39-4661-b9af-caf32415d8b1\") " Dec 10 23:09:22 crc kubenswrapper[4791]: I1210 23:09:22.632602 4791 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/ae2623b4-7b39-4661-b9af-caf32415d8b1-dns-swift-storage-0\") pod \"ae2623b4-7b39-4661-b9af-caf32415d8b1\" (UID: \"ae2623b4-7b39-4661-b9af-caf32415d8b1\") " Dec 10 23:09:22 crc kubenswrapper[4791]: I1210 23:09:22.632634 4791 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-hv5bp\" (UniqueName: \"kubernetes.io/projected/ae2623b4-7b39-4661-b9af-caf32415d8b1-kube-api-access-hv5bp\") pod \"ae2623b4-7b39-4661-b9af-caf32415d8b1\" (UID: \"ae2623b4-7b39-4661-b9af-caf32415d8b1\") " Dec 10 23:09:22 crc kubenswrapper[4791]: I1210 23:09:22.632664 4791 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/1a1ca37c-dab5-4748-b8c2-eb985e66a02f-logs\") pod \"1a1ca37c-dab5-4748-b8c2-eb985e66a02f\" (UID: \"1a1ca37c-dab5-4748-b8c2-eb985e66a02f\") " Dec 10 23:09:22 crc kubenswrapper[4791]: I1210 23:09:22.632711 4791 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/1a1ca37c-dab5-4748-b8c2-eb985e66a02f-scripts\") pod \"1a1ca37c-dab5-4748-b8c2-eb985e66a02f\" (UID: \"1a1ca37c-dab5-4748-b8c2-eb985e66a02f\") " Dec 10 23:09:22 crc kubenswrapper[4791]: I1210 23:09:22.632772 4791 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-vvvdb\" (UniqueName: \"kubernetes.io/projected/1a1ca37c-dab5-4748-b8c2-eb985e66a02f-kube-api-access-vvvdb\") pod \"1a1ca37c-dab5-4748-b8c2-eb985e66a02f\" (UID: \"1a1ca37c-dab5-4748-b8c2-eb985e66a02f\") " Dec 10 23:09:22 crc kubenswrapper[4791]: I1210 23:09:22.632815 4791 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/ae2623b4-7b39-4661-b9af-caf32415d8b1-config\") pod \"ae2623b4-7b39-4661-b9af-caf32415d8b1\" (UID: \"ae2623b4-7b39-4661-b9af-caf32415d8b1\") " Dec 10 23:09:22 crc kubenswrapper[4791]: I1210 23:09:22.632838 4791 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/1a1ca37c-dab5-4748-b8c2-eb985e66a02f-combined-ca-bundle\") pod \"1a1ca37c-dab5-4748-b8c2-eb985e66a02f\" (UID: \"1a1ca37c-dab5-4748-b8c2-eb985e66a02f\") " Dec 10 23:09:22 crc kubenswrapper[4791]: I1210 23:09:22.632876 4791 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/ae2623b4-7b39-4661-b9af-caf32415d8b1-ovsdbserver-nb\") pod \"ae2623b4-7b39-4661-b9af-caf32415d8b1\" (UID: \"ae2623b4-7b39-4661-b9af-caf32415d8b1\") " Dec 10 23:09:22 crc kubenswrapper[4791]: I1210 23:09:22.632902 4791 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/ae2623b4-7b39-4661-b9af-caf32415d8b1-ovsdbserver-sb\") pod \"ae2623b4-7b39-4661-b9af-caf32415d8b1\" (UID: \"ae2623b4-7b39-4661-b9af-caf32415d8b1\") " Dec 10 23:09:22 crc kubenswrapper[4791]: I1210 23:09:22.636532 4791 reconciler_common.go:293] "Volume detached for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/1a1ca37c-dab5-4748-b8c2-eb985e66a02f-etc-machine-id\") on node \"crc\" DevicePath \"\"" Dec 10 23:09:22 crc kubenswrapper[4791]: I1210 23:09:22.637798 4791 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/1a1ca37c-dab5-4748-b8c2-eb985e66a02f-logs" (OuterVolumeSpecName: "logs") pod "1a1ca37c-dab5-4748-b8c2-eb985e66a02f" (UID: "1a1ca37c-dab5-4748-b8c2-eb985e66a02f"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 10 23:09:22 crc kubenswrapper[4791]: I1210 23:09:22.645978 4791 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/1a1ca37c-dab5-4748-b8c2-eb985e66a02f-kube-api-access-vvvdb" (OuterVolumeSpecName: "kube-api-access-vvvdb") pod "1a1ca37c-dab5-4748-b8c2-eb985e66a02f" (UID: "1a1ca37c-dab5-4748-b8c2-eb985e66a02f"). InnerVolumeSpecName "kube-api-access-vvvdb". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 10 23:09:22 crc kubenswrapper[4791]: I1210 23:09:22.660511 4791 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/ae2623b4-7b39-4661-b9af-caf32415d8b1-kube-api-access-hv5bp" (OuterVolumeSpecName: "kube-api-access-hv5bp") pod "ae2623b4-7b39-4661-b9af-caf32415d8b1" (UID: "ae2623b4-7b39-4661-b9af-caf32415d8b1"). InnerVolumeSpecName "kube-api-access-hv5bp". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 10 23:09:22 crc kubenswrapper[4791]: I1210 23:09:22.660788 4791 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/1a1ca37c-dab5-4748-b8c2-eb985e66a02f-config-data-custom" (OuterVolumeSpecName: "config-data-custom") pod "1a1ca37c-dab5-4748-b8c2-eb985e66a02f" (UID: "1a1ca37c-dab5-4748-b8c2-eb985e66a02f"). InnerVolumeSpecName "config-data-custom". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 10 23:09:22 crc kubenswrapper[4791]: I1210 23:09:22.692940 4791 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/1a1ca37c-dab5-4748-b8c2-eb985e66a02f-scripts" (OuterVolumeSpecName: "scripts") pod "1a1ca37c-dab5-4748-b8c2-eb985e66a02f" (UID: "1a1ca37c-dab5-4748-b8c2-eb985e66a02f"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 10 23:09:22 crc kubenswrapper[4791]: I1210 23:09:22.702141 4791 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/1a1ca37c-dab5-4748-b8c2-eb985e66a02f-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "1a1ca37c-dab5-4748-b8c2-eb985e66a02f" (UID: "1a1ca37c-dab5-4748-b8c2-eb985e66a02f"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 10 23:09:22 crc kubenswrapper[4791]: I1210 23:09:22.737932 4791 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-vvvdb\" (UniqueName: \"kubernetes.io/projected/1a1ca37c-dab5-4748-b8c2-eb985e66a02f-kube-api-access-vvvdb\") on node \"crc\" DevicePath \"\"" Dec 10 23:09:22 crc kubenswrapper[4791]: I1210 23:09:22.738440 4791 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/1a1ca37c-dab5-4748-b8c2-eb985e66a02f-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 10 23:09:22 crc kubenswrapper[4791]: I1210 23:09:22.738450 4791 reconciler_common.go:293] "Volume detached for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/1a1ca37c-dab5-4748-b8c2-eb985e66a02f-config-data-custom\") on node \"crc\" DevicePath \"\"" Dec 10 23:09:22 crc kubenswrapper[4791]: I1210 23:09:22.738460 4791 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-hv5bp\" (UniqueName: \"kubernetes.io/projected/ae2623b4-7b39-4661-b9af-caf32415d8b1-kube-api-access-hv5bp\") on node \"crc\" DevicePath \"\"" Dec 10 23:09:22 crc kubenswrapper[4791]: I1210 23:09:22.738469 4791 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/1a1ca37c-dab5-4748-b8c2-eb985e66a02f-logs\") on node \"crc\" DevicePath \"\"" Dec 10 23:09:22 crc kubenswrapper[4791]: I1210 23:09:22.738479 4791 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/1a1ca37c-dab5-4748-b8c2-eb985e66a02f-scripts\") on node \"crc\" DevicePath \"\"" Dec 10 23:09:22 crc kubenswrapper[4791]: I1210 23:09:22.842538 4791 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-api-0" event={"ID":"1a1ca37c-dab5-4748-b8c2-eb985e66a02f","Type":"ContainerDied","Data":"b5a57d174634113df118d0024e1c1c33aaa029ba9541d56b2728ce26d0d5fffa"} Dec 10 23:09:22 crc kubenswrapper[4791]: I1210 23:09:22.842592 4791 scope.go:117] "RemoveContainer" containerID="c9727fa5dc2cf41c8e8fb4fe2a685691a2061b0a9272331473fef258a5ed514f" Dec 10 23:09:22 crc kubenswrapper[4791]: I1210 23:09:22.842558 4791 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-api-0" Dec 10 23:09:22 crc kubenswrapper[4791]: I1210 23:09:22.848110 4791 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-d68b9cb4c-xp29n" Dec 10 23:09:22 crc kubenswrapper[4791]: I1210 23:09:22.848138 4791 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-d68b9cb4c-xp29n" event={"ID":"ae2623b4-7b39-4661-b9af-caf32415d8b1","Type":"ContainerDied","Data":"d1ea78e5d6411134a13966ef4046829678d14b12c7138cb29784cac9fbadbd5a"} Dec 10 23:09:22 crc kubenswrapper[4791]: I1210 23:09:22.848465 4791 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/cinder-scheduler-0" podUID="5e5e327e-b8dc-4bd6-af75-8aebc417ada9" containerName="cinder-scheduler" containerID="cri-o://468b836fc75ba6c8ac143ddca9787ab03833f45994d432d9850ebe13fc2251ed" gracePeriod=30 Dec 10 23:09:22 crc kubenswrapper[4791]: I1210 23:09:22.848615 4791 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/cinder-scheduler-0" podUID="5e5e327e-b8dc-4bd6-af75-8aebc417ada9" containerName="probe" containerID="cri-o://0491f982573b4cf5b921d4eb64aecef9ffb9a9bed5a7cd16d5e70ffa514274f4" gracePeriod=30 Dec 10 23:09:22 crc kubenswrapper[4791]: I1210 23:09:22.875432 4791 scope.go:117] "RemoveContainer" containerID="831488e7d1696daf20de78611d91cff3c3a4661569db23dac805484a70013576" Dec 10 23:09:22 crc kubenswrapper[4791]: I1210 23:09:22.901385 4791 scope.go:117] "RemoveContainer" containerID="8abb2f7b1e355fe0279461505e35f3aae94f2f3d4097ceb931b5ef3c52902873" Dec 10 23:09:22 crc kubenswrapper[4791]: I1210 23:09:22.974259 4791 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/horizon-78fc5d56dc-58cjt" Dec 10 23:09:23 crc kubenswrapper[4791]: I1210 23:09:23.023509 4791 scope.go:117] "RemoveContainer" containerID="766b794c9708b4cbfe2821348b03e234d1dd7e3877bf577377616c1025ff7564" Dec 10 23:09:23 crc kubenswrapper[4791]: I1210 23:09:23.040238 4791 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/ae2623b4-7b39-4661-b9af-caf32415d8b1-ovsdbserver-nb" (OuterVolumeSpecName: "ovsdbserver-nb") pod "ae2623b4-7b39-4661-b9af-caf32415d8b1" (UID: "ae2623b4-7b39-4661-b9af-caf32415d8b1"). InnerVolumeSpecName "ovsdbserver-nb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 10 23:09:23 crc kubenswrapper[4791]: I1210 23:09:23.046706 4791 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/14f46cbc-afe2-40bb-a2a3-5673b5ac53e7-scripts\") pod \"14f46cbc-afe2-40bb-a2a3-5673b5ac53e7\" (UID: \"14f46cbc-afe2-40bb-a2a3-5673b5ac53e7\") " Dec 10 23:09:23 crc kubenswrapper[4791]: I1210 23:09:23.047108 4791 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/14f46cbc-afe2-40bb-a2a3-5673b5ac53e7-logs\") pod \"14f46cbc-afe2-40bb-a2a3-5673b5ac53e7\" (UID: \"14f46cbc-afe2-40bb-a2a3-5673b5ac53e7\") " Dec 10 23:09:23 crc kubenswrapper[4791]: I1210 23:09:23.047236 4791 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/14f46cbc-afe2-40bb-a2a3-5673b5ac53e7-config-data\") pod \"14f46cbc-afe2-40bb-a2a3-5673b5ac53e7\" (UID: \"14f46cbc-afe2-40bb-a2a3-5673b5ac53e7\") " Dec 10 23:09:23 crc kubenswrapper[4791]: I1210 23:09:23.047481 4791 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"horizon-secret-key\" (UniqueName: \"kubernetes.io/secret/14f46cbc-afe2-40bb-a2a3-5673b5ac53e7-horizon-secret-key\") pod \"14f46cbc-afe2-40bb-a2a3-5673b5ac53e7\" (UID: \"14f46cbc-afe2-40bb-a2a3-5673b5ac53e7\") " Dec 10 23:09:23 crc kubenswrapper[4791]: I1210 23:09:23.047614 4791 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-dk8dr\" (UniqueName: \"kubernetes.io/projected/14f46cbc-afe2-40bb-a2a3-5673b5ac53e7-kube-api-access-dk8dr\") pod \"14f46cbc-afe2-40bb-a2a3-5673b5ac53e7\" (UID: \"14f46cbc-afe2-40bb-a2a3-5673b5ac53e7\") " Dec 10 23:09:23 crc kubenswrapper[4791]: I1210 23:09:23.048416 4791 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/ae2623b4-7b39-4661-b9af-caf32415d8b1-ovsdbserver-nb\") on node \"crc\" DevicePath \"\"" Dec 10 23:09:23 crc kubenswrapper[4791]: I1210 23:09:23.051360 4791 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/14f46cbc-afe2-40bb-a2a3-5673b5ac53e7-logs" (OuterVolumeSpecName: "logs") pod "14f46cbc-afe2-40bb-a2a3-5673b5ac53e7" (UID: "14f46cbc-afe2-40bb-a2a3-5673b5ac53e7"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 10 23:09:23 crc kubenswrapper[4791]: I1210 23:09:23.055255 4791 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/ae2623b4-7b39-4661-b9af-caf32415d8b1-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "ae2623b4-7b39-4661-b9af-caf32415d8b1" (UID: "ae2623b4-7b39-4661-b9af-caf32415d8b1"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 10 23:09:23 crc kubenswrapper[4791]: I1210 23:09:23.060645 4791 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/14f46cbc-afe2-40bb-a2a3-5673b5ac53e7-horizon-secret-key" (OuterVolumeSpecName: "horizon-secret-key") pod "14f46cbc-afe2-40bb-a2a3-5673b5ac53e7" (UID: "14f46cbc-afe2-40bb-a2a3-5673b5ac53e7"). InnerVolumeSpecName "horizon-secret-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 10 23:09:23 crc kubenswrapper[4791]: I1210 23:09:23.060893 4791 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/ae2623b4-7b39-4661-b9af-caf32415d8b1-config" (OuterVolumeSpecName: "config") pod "ae2623b4-7b39-4661-b9af-caf32415d8b1" (UID: "ae2623b4-7b39-4661-b9af-caf32415d8b1"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 10 23:09:23 crc kubenswrapper[4791]: I1210 23:09:23.061281 4791 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/14f46cbc-afe2-40bb-a2a3-5673b5ac53e7-kube-api-access-dk8dr" (OuterVolumeSpecName: "kube-api-access-dk8dr") pod "14f46cbc-afe2-40bb-a2a3-5673b5ac53e7" (UID: "14f46cbc-afe2-40bb-a2a3-5673b5ac53e7"). InnerVolumeSpecName "kube-api-access-dk8dr". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 10 23:09:23 crc kubenswrapper[4791]: I1210 23:09:23.066987 4791 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/ae2623b4-7b39-4661-b9af-caf32415d8b1-dns-swift-storage-0" (OuterVolumeSpecName: "dns-swift-storage-0") pod "ae2623b4-7b39-4661-b9af-caf32415d8b1" (UID: "ae2623b4-7b39-4661-b9af-caf32415d8b1"). InnerVolumeSpecName "dns-swift-storage-0". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 10 23:09:23 crc kubenswrapper[4791]: I1210 23:09:23.095161 4791 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/1a1ca37c-dab5-4748-b8c2-eb985e66a02f-config-data" (OuterVolumeSpecName: "config-data") pod "1a1ca37c-dab5-4748-b8c2-eb985e66a02f" (UID: "1a1ca37c-dab5-4748-b8c2-eb985e66a02f"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 10 23:09:23 crc kubenswrapper[4791]: I1210 23:09:23.107809 4791 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/horizon-6999c5bf5b-7cr9t" Dec 10 23:09:23 crc kubenswrapper[4791]: I1210 23:09:23.147985 4791 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/ae2623b4-7b39-4661-b9af-caf32415d8b1-ovsdbserver-sb" (OuterVolumeSpecName: "ovsdbserver-sb") pod "ae2623b4-7b39-4661-b9af-caf32415d8b1" (UID: "ae2623b4-7b39-4661-b9af-caf32415d8b1"). InnerVolumeSpecName "ovsdbserver-sb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 10 23:09:23 crc kubenswrapper[4791]: I1210 23:09:23.151792 4791 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/ae2623b4-7b39-4661-b9af-caf32415d8b1-config\") on node \"crc\" DevicePath \"\"" Dec 10 23:09:23 crc kubenswrapper[4791]: I1210 23:09:23.155042 4791 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/ae2623b4-7b39-4661-b9af-caf32415d8b1-ovsdbserver-sb\") on node \"crc\" DevicePath \"\"" Dec 10 23:09:23 crc kubenswrapper[4791]: I1210 23:09:23.155066 4791 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/14f46cbc-afe2-40bb-a2a3-5673b5ac53e7-logs\") on node \"crc\" DevicePath \"\"" Dec 10 23:09:23 crc kubenswrapper[4791]: I1210 23:09:23.155075 4791 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/1a1ca37c-dab5-4748-b8c2-eb985e66a02f-config-data\") on node \"crc\" DevicePath \"\"" Dec 10 23:09:23 crc kubenswrapper[4791]: I1210 23:09:23.155086 4791 reconciler_common.go:293] "Volume detached for volume \"horizon-secret-key\" (UniqueName: \"kubernetes.io/secret/14f46cbc-afe2-40bb-a2a3-5673b5ac53e7-horizon-secret-key\") on node \"crc\" DevicePath \"\"" Dec 10 23:09:23 crc kubenswrapper[4791]: I1210 23:09:23.155097 4791 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-dk8dr\" (UniqueName: \"kubernetes.io/projected/14f46cbc-afe2-40bb-a2a3-5673b5ac53e7-kube-api-access-dk8dr\") on node \"crc\" DevicePath \"\"" Dec 10 23:09:23 crc kubenswrapper[4791]: I1210 23:09:23.155106 4791 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/ae2623b4-7b39-4661-b9af-caf32415d8b1-dns-svc\") on node \"crc\" DevicePath \"\"" Dec 10 23:09:23 crc kubenswrapper[4791]: I1210 23:09:23.155115 4791 reconciler_common.go:293] "Volume detached for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/ae2623b4-7b39-4661-b9af-caf32415d8b1-dns-swift-storage-0\") on node \"crc\" DevicePath \"\"" Dec 10 23:09:23 crc kubenswrapper[4791]: I1210 23:09:23.173445 4791 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/horizon-7cc66567bd-pwb2l"] Dec 10 23:09:23 crc kubenswrapper[4791]: I1210 23:09:23.173675 4791 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/horizon-7cc66567bd-pwb2l" podUID="55139c7c-e9fa-4539-b7b2-b2be7ee43b44" containerName="horizon-log" containerID="cri-o://7b1ada7642cab048e60fe308d453e086b82053b893d130082d56f6aaf472608a" gracePeriod=30 Dec 10 23:09:23 crc kubenswrapper[4791]: I1210 23:09:23.173782 4791 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/horizon-7cc66567bd-pwb2l" podUID="55139c7c-e9fa-4539-b7b2-b2be7ee43b44" containerName="horizon" containerID="cri-o://a441e9dd03ea293d73ed83e0ce82cd1f391a27dfc4d4c0f86b23bba57c39c2c5" gracePeriod=30 Dec 10 23:09:23 crc kubenswrapper[4791]: I1210 23:09:23.176200 4791 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/14f46cbc-afe2-40bb-a2a3-5673b5ac53e7-scripts" (OuterVolumeSpecName: "scripts") pod "14f46cbc-afe2-40bb-a2a3-5673b5ac53e7" (UID: "14f46cbc-afe2-40bb-a2a3-5673b5ac53e7"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 10 23:09:23 crc kubenswrapper[4791]: I1210 23:09:23.176585 4791 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/14f46cbc-afe2-40bb-a2a3-5673b5ac53e7-config-data" (OuterVolumeSpecName: "config-data") pod "14f46cbc-afe2-40bb-a2a3-5673b5ac53e7" (UID: "14f46cbc-afe2-40bb-a2a3-5673b5ac53e7"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 10 23:09:23 crc kubenswrapper[4791]: I1210 23:09:23.247410 4791 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-d68b9cb4c-xp29n"] Dec 10 23:09:23 crc kubenswrapper[4791]: I1210 23:09:23.262628 4791 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/14f46cbc-afe2-40bb-a2a3-5673b5ac53e7-config-data\") on node \"crc\" DevicePath \"\"" Dec 10 23:09:23 crc kubenswrapper[4791]: I1210 23:09:23.262665 4791 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/14f46cbc-afe2-40bb-a2a3-5673b5ac53e7-scripts\") on node \"crc\" DevicePath \"\"" Dec 10 23:09:23 crc kubenswrapper[4791]: I1210 23:09:23.262837 4791 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/horizon-5f54ccbf57-wwpfk" Dec 10 23:09:23 crc kubenswrapper[4791]: I1210 23:09:23.263226 4791 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/horizon-5cc98b55bf-7xpnj" Dec 10 23:09:23 crc kubenswrapper[4791]: I1210 23:09:23.290397 4791 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-d68b9cb4c-xp29n"] Dec 10 23:09:23 crc kubenswrapper[4791]: I1210 23:09:23.324215 4791 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/barbican-api-68fb764d56-4mfz4"] Dec 10 23:09:23 crc kubenswrapper[4791]: I1210 23:09:23.369668 4791 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/cinder-api-0"] Dec 10 23:09:23 crc kubenswrapper[4791]: I1210 23:09:23.372910 4791 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/aa3a63ab-6092-46ce-aaf6-40360aa060ac-logs\") pod \"aa3a63ab-6092-46ce-aaf6-40360aa060ac\" (UID: \"aa3a63ab-6092-46ce-aaf6-40360aa060ac\") " Dec 10 23:09:23 crc kubenswrapper[4791]: I1210 23:09:23.372985 4791 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/aa3a63ab-6092-46ce-aaf6-40360aa060ac-config-data\") pod \"aa3a63ab-6092-46ce-aaf6-40360aa060ac\" (UID: \"aa3a63ab-6092-46ce-aaf6-40360aa060ac\") " Dec 10 23:09:23 crc kubenswrapper[4791]: I1210 23:09:23.373013 4791 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-v97s8\" (UniqueName: \"kubernetes.io/projected/9e9c1419-c6f3-4947-8788-4915f48677d7-kube-api-access-v97s8\") pod \"9e9c1419-c6f3-4947-8788-4915f48677d7\" (UID: \"9e9c1419-c6f3-4947-8788-4915f48677d7\") " Dec 10 23:09:23 crc kubenswrapper[4791]: I1210 23:09:23.373031 4791 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/aa3a63ab-6092-46ce-aaf6-40360aa060ac-scripts\") pod \"aa3a63ab-6092-46ce-aaf6-40360aa060ac\" (UID: \"aa3a63ab-6092-46ce-aaf6-40360aa060ac\") " Dec 10 23:09:23 crc kubenswrapper[4791]: I1210 23:09:23.373070 4791 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/9e9c1419-c6f3-4947-8788-4915f48677d7-logs\") pod \"9e9c1419-c6f3-4947-8788-4915f48677d7\" (UID: \"9e9c1419-c6f3-4947-8788-4915f48677d7\") " Dec 10 23:09:23 crc kubenswrapper[4791]: I1210 23:09:23.373177 4791 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"horizon-secret-key\" (UniqueName: \"kubernetes.io/secret/9e9c1419-c6f3-4947-8788-4915f48677d7-horizon-secret-key\") pod \"9e9c1419-c6f3-4947-8788-4915f48677d7\" (UID: \"9e9c1419-c6f3-4947-8788-4915f48677d7\") " Dec 10 23:09:23 crc kubenswrapper[4791]: I1210 23:09:23.373213 4791 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/9e9c1419-c6f3-4947-8788-4915f48677d7-config-data\") pod \"9e9c1419-c6f3-4947-8788-4915f48677d7\" (UID: \"9e9c1419-c6f3-4947-8788-4915f48677d7\") " Dec 10 23:09:23 crc kubenswrapper[4791]: I1210 23:09:23.373234 4791 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-85n5v\" (UniqueName: \"kubernetes.io/projected/aa3a63ab-6092-46ce-aaf6-40360aa060ac-kube-api-access-85n5v\") pod \"aa3a63ab-6092-46ce-aaf6-40360aa060ac\" (UID: \"aa3a63ab-6092-46ce-aaf6-40360aa060ac\") " Dec 10 23:09:23 crc kubenswrapper[4791]: I1210 23:09:23.373271 4791 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"horizon-secret-key\" (UniqueName: \"kubernetes.io/secret/aa3a63ab-6092-46ce-aaf6-40360aa060ac-horizon-secret-key\") pod \"aa3a63ab-6092-46ce-aaf6-40360aa060ac\" (UID: \"aa3a63ab-6092-46ce-aaf6-40360aa060ac\") " Dec 10 23:09:23 crc kubenswrapper[4791]: I1210 23:09:23.373308 4791 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/9e9c1419-c6f3-4947-8788-4915f48677d7-scripts\") pod \"9e9c1419-c6f3-4947-8788-4915f48677d7\" (UID: \"9e9c1419-c6f3-4947-8788-4915f48677d7\") " Dec 10 23:09:23 crc kubenswrapper[4791]: I1210 23:09:23.373806 4791 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/9e9c1419-c6f3-4947-8788-4915f48677d7-logs" (OuterVolumeSpecName: "logs") pod "9e9c1419-c6f3-4947-8788-4915f48677d7" (UID: "9e9c1419-c6f3-4947-8788-4915f48677d7"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 10 23:09:23 crc kubenswrapper[4791]: I1210 23:09:23.374148 4791 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/aa3a63ab-6092-46ce-aaf6-40360aa060ac-logs" (OuterVolumeSpecName: "logs") pod "aa3a63ab-6092-46ce-aaf6-40360aa060ac" (UID: "aa3a63ab-6092-46ce-aaf6-40360aa060ac"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 10 23:09:23 crc kubenswrapper[4791]: I1210 23:09:23.385939 4791 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/cinder-api-0"] Dec 10 23:09:23 crc kubenswrapper[4791]: I1210 23:09:23.468106 4791 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/aa3a63ab-6092-46ce-aaf6-40360aa060ac-kube-api-access-85n5v" (OuterVolumeSpecName: "kube-api-access-85n5v") pod "aa3a63ab-6092-46ce-aaf6-40360aa060ac" (UID: "aa3a63ab-6092-46ce-aaf6-40360aa060ac"). InnerVolumeSpecName "kube-api-access-85n5v". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 10 23:09:23 crc kubenswrapper[4791]: I1210 23:09:23.469318 4791 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/9e9c1419-c6f3-4947-8788-4915f48677d7-scripts" (OuterVolumeSpecName: "scripts") pod "9e9c1419-c6f3-4947-8788-4915f48677d7" (UID: "9e9c1419-c6f3-4947-8788-4915f48677d7"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 10 23:09:23 crc kubenswrapper[4791]: I1210 23:09:23.470145 4791 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/9e9c1419-c6f3-4947-8788-4915f48677d7-horizon-secret-key" (OuterVolumeSpecName: "horizon-secret-key") pod "9e9c1419-c6f3-4947-8788-4915f48677d7" (UID: "9e9c1419-c6f3-4947-8788-4915f48677d7"). InnerVolumeSpecName "horizon-secret-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 10 23:09:23 crc kubenswrapper[4791]: I1210 23:09:23.506413 4791 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/9e9c1419-c6f3-4947-8788-4915f48677d7-kube-api-access-v97s8" (OuterVolumeSpecName: "kube-api-access-v97s8") pod "9e9c1419-c6f3-4947-8788-4915f48677d7" (UID: "9e9c1419-c6f3-4947-8788-4915f48677d7"). InnerVolumeSpecName "kube-api-access-v97s8". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 10 23:09:23 crc kubenswrapper[4791]: I1210 23:09:23.507782 4791 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/9e9c1419-c6f3-4947-8788-4915f48677d7-config-data" (OuterVolumeSpecName: "config-data") pod "9e9c1419-c6f3-4947-8788-4915f48677d7" (UID: "9e9c1419-c6f3-4947-8788-4915f48677d7"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 10 23:09:23 crc kubenswrapper[4791]: I1210 23:09:23.507868 4791 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/aa3a63ab-6092-46ce-aaf6-40360aa060ac-horizon-secret-key" (OuterVolumeSpecName: "horizon-secret-key") pod "aa3a63ab-6092-46ce-aaf6-40360aa060ac" (UID: "aa3a63ab-6092-46ce-aaf6-40360aa060ac"). InnerVolumeSpecName "horizon-secret-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 10 23:09:23 crc kubenswrapper[4791]: I1210 23:09:23.511761 4791 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/aa3a63ab-6092-46ce-aaf6-40360aa060ac-scripts" (OuterVolumeSpecName: "scripts") pod "aa3a63ab-6092-46ce-aaf6-40360aa060ac" (UID: "aa3a63ab-6092-46ce-aaf6-40360aa060ac"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 10 23:09:23 crc kubenswrapper[4791]: I1210 23:09:23.511162 4791 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/9e9c1419-c6f3-4947-8788-4915f48677d7-logs\") on node \"crc\" DevicePath \"\"" Dec 10 23:09:23 crc kubenswrapper[4791]: I1210 23:09:23.512512 4791 reconciler_common.go:293] "Volume detached for volume \"horizon-secret-key\" (UniqueName: \"kubernetes.io/secret/9e9c1419-c6f3-4947-8788-4915f48677d7-horizon-secret-key\") on node \"crc\" DevicePath \"\"" Dec 10 23:09:23 crc kubenswrapper[4791]: I1210 23:09:23.512667 4791 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-85n5v\" (UniqueName: \"kubernetes.io/projected/aa3a63ab-6092-46ce-aaf6-40360aa060ac-kube-api-access-85n5v\") on node \"crc\" DevicePath \"\"" Dec 10 23:09:23 crc kubenswrapper[4791]: I1210 23:09:23.512754 4791 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/9e9c1419-c6f3-4947-8788-4915f48677d7-scripts\") on node \"crc\" DevicePath \"\"" Dec 10 23:09:23 crc kubenswrapper[4791]: I1210 23:09:23.512852 4791 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/aa3a63ab-6092-46ce-aaf6-40360aa060ac-logs\") on node \"crc\" DevicePath \"\"" Dec 10 23:09:23 crc kubenswrapper[4791]: I1210 23:09:23.516625 4791 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/aa3a63ab-6092-46ce-aaf6-40360aa060ac-config-data" (OuterVolumeSpecName: "config-data") pod "aa3a63ab-6092-46ce-aaf6-40360aa060ac" (UID: "aa3a63ab-6092-46ce-aaf6-40360aa060ac"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 10 23:09:23 crc kubenswrapper[4791]: I1210 23:09:23.535272 4791 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/cinder-api-0"] Dec 10 23:09:23 crc kubenswrapper[4791]: E1210 23:09:23.536248 4791 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="aa3a63ab-6092-46ce-aaf6-40360aa060ac" containerName="horizon" Dec 10 23:09:23 crc kubenswrapper[4791]: I1210 23:09:23.536266 4791 state_mem.go:107] "Deleted CPUSet assignment" podUID="aa3a63ab-6092-46ce-aaf6-40360aa060ac" containerName="horizon" Dec 10 23:09:23 crc kubenswrapper[4791]: E1210 23:09:23.536278 4791 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="ae2623b4-7b39-4661-b9af-caf32415d8b1" containerName="dnsmasq-dns" Dec 10 23:09:23 crc kubenswrapper[4791]: I1210 23:09:23.536290 4791 state_mem.go:107] "Deleted CPUSet assignment" podUID="ae2623b4-7b39-4661-b9af-caf32415d8b1" containerName="dnsmasq-dns" Dec 10 23:09:23 crc kubenswrapper[4791]: E1210 23:09:23.536306 4791 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="aa3a63ab-6092-46ce-aaf6-40360aa060ac" containerName="horizon-log" Dec 10 23:09:23 crc kubenswrapper[4791]: I1210 23:09:23.536313 4791 state_mem.go:107] "Deleted CPUSet assignment" podUID="aa3a63ab-6092-46ce-aaf6-40360aa060ac" containerName="horizon-log" Dec 10 23:09:23 crc kubenswrapper[4791]: E1210 23:09:23.536355 4791 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="ae2623b4-7b39-4661-b9af-caf32415d8b1" containerName="init" Dec 10 23:09:23 crc kubenswrapper[4791]: I1210 23:09:23.536362 4791 state_mem.go:107] "Deleted CPUSet assignment" podUID="ae2623b4-7b39-4661-b9af-caf32415d8b1" containerName="init" Dec 10 23:09:23 crc kubenswrapper[4791]: E1210 23:09:23.536379 4791 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="1a1ca37c-dab5-4748-b8c2-eb985e66a02f" containerName="cinder-api-log" Dec 10 23:09:23 crc kubenswrapper[4791]: I1210 23:09:23.536385 4791 state_mem.go:107] "Deleted CPUSet assignment" podUID="1a1ca37c-dab5-4748-b8c2-eb985e66a02f" containerName="cinder-api-log" Dec 10 23:09:23 crc kubenswrapper[4791]: E1210 23:09:23.536394 4791 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="1a1ca37c-dab5-4748-b8c2-eb985e66a02f" containerName="cinder-api" Dec 10 23:09:23 crc kubenswrapper[4791]: I1210 23:09:23.536399 4791 state_mem.go:107] "Deleted CPUSet assignment" podUID="1a1ca37c-dab5-4748-b8c2-eb985e66a02f" containerName="cinder-api" Dec 10 23:09:23 crc kubenswrapper[4791]: E1210 23:09:23.536414 4791 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="14f46cbc-afe2-40bb-a2a3-5673b5ac53e7" containerName="horizon-log" Dec 10 23:09:23 crc kubenswrapper[4791]: I1210 23:09:23.536422 4791 state_mem.go:107] "Deleted CPUSet assignment" podUID="14f46cbc-afe2-40bb-a2a3-5673b5ac53e7" containerName="horizon-log" Dec 10 23:09:23 crc kubenswrapper[4791]: E1210 23:09:23.536437 4791 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="9e9c1419-c6f3-4947-8788-4915f48677d7" containerName="horizon-log" Dec 10 23:09:23 crc kubenswrapper[4791]: I1210 23:09:23.536444 4791 state_mem.go:107] "Deleted CPUSet assignment" podUID="9e9c1419-c6f3-4947-8788-4915f48677d7" containerName="horizon-log" Dec 10 23:09:23 crc kubenswrapper[4791]: E1210 23:09:23.536464 4791 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="14f46cbc-afe2-40bb-a2a3-5673b5ac53e7" containerName="horizon" Dec 10 23:09:23 crc kubenswrapper[4791]: I1210 23:09:23.536471 4791 state_mem.go:107] "Deleted CPUSet assignment" podUID="14f46cbc-afe2-40bb-a2a3-5673b5ac53e7" containerName="horizon" Dec 10 23:09:23 crc kubenswrapper[4791]: E1210 23:09:23.536490 4791 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="9e9c1419-c6f3-4947-8788-4915f48677d7" containerName="horizon" Dec 10 23:09:23 crc kubenswrapper[4791]: I1210 23:09:23.536496 4791 state_mem.go:107] "Deleted CPUSet assignment" podUID="9e9c1419-c6f3-4947-8788-4915f48677d7" containerName="horizon" Dec 10 23:09:23 crc kubenswrapper[4791]: I1210 23:09:23.536837 4791 memory_manager.go:354] "RemoveStaleState removing state" podUID="aa3a63ab-6092-46ce-aaf6-40360aa060ac" containerName="horizon-log" Dec 10 23:09:23 crc kubenswrapper[4791]: I1210 23:09:23.536856 4791 memory_manager.go:354] "RemoveStaleState removing state" podUID="9e9c1419-c6f3-4947-8788-4915f48677d7" containerName="horizon" Dec 10 23:09:23 crc kubenswrapper[4791]: I1210 23:09:23.536881 4791 memory_manager.go:354] "RemoveStaleState removing state" podUID="ae2623b4-7b39-4661-b9af-caf32415d8b1" containerName="dnsmasq-dns" Dec 10 23:09:23 crc kubenswrapper[4791]: I1210 23:09:23.536901 4791 memory_manager.go:354] "RemoveStaleState removing state" podUID="1a1ca37c-dab5-4748-b8c2-eb985e66a02f" containerName="cinder-api" Dec 10 23:09:23 crc kubenswrapper[4791]: I1210 23:09:23.536926 4791 memory_manager.go:354] "RemoveStaleState removing state" podUID="14f46cbc-afe2-40bb-a2a3-5673b5ac53e7" containerName="horizon" Dec 10 23:09:23 crc kubenswrapper[4791]: I1210 23:09:23.536939 4791 memory_manager.go:354] "RemoveStaleState removing state" podUID="14f46cbc-afe2-40bb-a2a3-5673b5ac53e7" containerName="horizon-log" Dec 10 23:09:23 crc kubenswrapper[4791]: I1210 23:09:23.536966 4791 memory_manager.go:354] "RemoveStaleState removing state" podUID="1a1ca37c-dab5-4748-b8c2-eb985e66a02f" containerName="cinder-api-log" Dec 10 23:09:23 crc kubenswrapper[4791]: I1210 23:09:23.536990 4791 memory_manager.go:354] "RemoveStaleState removing state" podUID="9e9c1419-c6f3-4947-8788-4915f48677d7" containerName="horizon-log" Dec 10 23:09:23 crc kubenswrapper[4791]: I1210 23:09:23.537011 4791 memory_manager.go:354] "RemoveStaleState removing state" podUID="aa3a63ab-6092-46ce-aaf6-40360aa060ac" containerName="horizon" Dec 10 23:09:23 crc kubenswrapper[4791]: I1210 23:09:23.538625 4791 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-api-0" Dec 10 23:09:23 crc kubenswrapper[4791]: I1210 23:09:23.552210 4791 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-cinder-public-svc" Dec 10 23:09:23 crc kubenswrapper[4791]: I1210 23:09:23.553326 4791 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-cinder-internal-svc" Dec 10 23:09:23 crc kubenswrapper[4791]: I1210 23:09:23.553805 4791 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cinder-api-config-data" Dec 10 23:09:23 crc kubenswrapper[4791]: I1210 23:09:23.560777 4791 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-api-0"] Dec 10 23:09:23 crc kubenswrapper[4791]: I1210 23:09:23.615411 4791 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-7fhmf\" (UniqueName: \"kubernetes.io/projected/f807e752-6cf5-4ad6-bc21-f22777b62c83-kube-api-access-7fhmf\") pod \"cinder-api-0\" (UID: \"f807e752-6cf5-4ad6-bc21-f22777b62c83\") " pod="openstack/cinder-api-0" Dec 10 23:09:23 crc kubenswrapper[4791]: I1210 23:09:23.615465 4791 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/f807e752-6cf5-4ad6-bc21-f22777b62c83-public-tls-certs\") pod \"cinder-api-0\" (UID: \"f807e752-6cf5-4ad6-bc21-f22777b62c83\") " pod="openstack/cinder-api-0" Dec 10 23:09:23 crc kubenswrapper[4791]: I1210 23:09:23.615491 4791 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/f807e752-6cf5-4ad6-bc21-f22777b62c83-scripts\") pod \"cinder-api-0\" (UID: \"f807e752-6cf5-4ad6-bc21-f22777b62c83\") " pod="openstack/cinder-api-0" Dec 10 23:09:23 crc kubenswrapper[4791]: I1210 23:09:23.615530 4791 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/f807e752-6cf5-4ad6-bc21-f22777b62c83-internal-tls-certs\") pod \"cinder-api-0\" (UID: \"f807e752-6cf5-4ad6-bc21-f22777b62c83\") " pod="openstack/cinder-api-0" Dec 10 23:09:23 crc kubenswrapper[4791]: I1210 23:09:23.615547 4791 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/f807e752-6cf5-4ad6-bc21-f22777b62c83-etc-machine-id\") pod \"cinder-api-0\" (UID: \"f807e752-6cf5-4ad6-bc21-f22777b62c83\") " pod="openstack/cinder-api-0" Dec 10 23:09:23 crc kubenswrapper[4791]: I1210 23:09:23.615564 4791 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/f807e752-6cf5-4ad6-bc21-f22777b62c83-config-data\") pod \"cinder-api-0\" (UID: \"f807e752-6cf5-4ad6-bc21-f22777b62c83\") " pod="openstack/cinder-api-0" Dec 10 23:09:23 crc kubenswrapper[4791]: I1210 23:09:23.615646 4791 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/f807e752-6cf5-4ad6-bc21-f22777b62c83-config-data-custom\") pod \"cinder-api-0\" (UID: \"f807e752-6cf5-4ad6-bc21-f22777b62c83\") " pod="openstack/cinder-api-0" Dec 10 23:09:23 crc kubenswrapper[4791]: I1210 23:09:23.615679 4791 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f807e752-6cf5-4ad6-bc21-f22777b62c83-combined-ca-bundle\") pod \"cinder-api-0\" (UID: \"f807e752-6cf5-4ad6-bc21-f22777b62c83\") " pod="openstack/cinder-api-0" Dec 10 23:09:23 crc kubenswrapper[4791]: I1210 23:09:23.615705 4791 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/f807e752-6cf5-4ad6-bc21-f22777b62c83-logs\") pod \"cinder-api-0\" (UID: \"f807e752-6cf5-4ad6-bc21-f22777b62c83\") " pod="openstack/cinder-api-0" Dec 10 23:09:23 crc kubenswrapper[4791]: I1210 23:09:23.615783 4791 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/aa3a63ab-6092-46ce-aaf6-40360aa060ac-config-data\") on node \"crc\" DevicePath \"\"" Dec 10 23:09:23 crc kubenswrapper[4791]: I1210 23:09:23.615795 4791 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-v97s8\" (UniqueName: \"kubernetes.io/projected/9e9c1419-c6f3-4947-8788-4915f48677d7-kube-api-access-v97s8\") on node \"crc\" DevicePath \"\"" Dec 10 23:09:23 crc kubenswrapper[4791]: I1210 23:09:23.615807 4791 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/aa3a63ab-6092-46ce-aaf6-40360aa060ac-scripts\") on node \"crc\" DevicePath \"\"" Dec 10 23:09:23 crc kubenswrapper[4791]: I1210 23:09:23.615831 4791 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/9e9c1419-c6f3-4947-8788-4915f48677d7-config-data\") on node \"crc\" DevicePath \"\"" Dec 10 23:09:23 crc kubenswrapper[4791]: I1210 23:09:23.615839 4791 reconciler_common.go:293] "Volume detached for volume \"horizon-secret-key\" (UniqueName: \"kubernetes.io/secret/aa3a63ab-6092-46ce-aaf6-40360aa060ac-horizon-secret-key\") on node \"crc\" DevicePath \"\"" Dec 10 23:09:23 crc kubenswrapper[4791]: E1210 23:09:23.680052 4791 cadvisor_stats_provider.go:516] "Partial failure issuing cadvisor.ContainerInfoV2" err="partial failures: [\"/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod55139c7c_e9fa_4539_b7b2_b2be7ee43b44.slice/crio-conmon-a441e9dd03ea293d73ed83e0ce82cd1f391a27dfc4d4c0f86b23bba57c39c2c5.scope\": RecentStats: unable to find data in memory cache]" Dec 10 23:09:23 crc kubenswrapper[4791]: I1210 23:09:23.717913 4791 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/f807e752-6cf5-4ad6-bc21-f22777b62c83-config-data-custom\") pod \"cinder-api-0\" (UID: \"f807e752-6cf5-4ad6-bc21-f22777b62c83\") " pod="openstack/cinder-api-0" Dec 10 23:09:23 crc kubenswrapper[4791]: I1210 23:09:23.717953 4791 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f807e752-6cf5-4ad6-bc21-f22777b62c83-combined-ca-bundle\") pod \"cinder-api-0\" (UID: \"f807e752-6cf5-4ad6-bc21-f22777b62c83\") " pod="openstack/cinder-api-0" Dec 10 23:09:23 crc kubenswrapper[4791]: I1210 23:09:23.717982 4791 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/f807e752-6cf5-4ad6-bc21-f22777b62c83-logs\") pod \"cinder-api-0\" (UID: \"f807e752-6cf5-4ad6-bc21-f22777b62c83\") " pod="openstack/cinder-api-0" Dec 10 23:09:23 crc kubenswrapper[4791]: I1210 23:09:23.718046 4791 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-7fhmf\" (UniqueName: \"kubernetes.io/projected/f807e752-6cf5-4ad6-bc21-f22777b62c83-kube-api-access-7fhmf\") pod \"cinder-api-0\" (UID: \"f807e752-6cf5-4ad6-bc21-f22777b62c83\") " pod="openstack/cinder-api-0" Dec 10 23:09:23 crc kubenswrapper[4791]: I1210 23:09:23.718063 4791 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/f807e752-6cf5-4ad6-bc21-f22777b62c83-public-tls-certs\") pod \"cinder-api-0\" (UID: \"f807e752-6cf5-4ad6-bc21-f22777b62c83\") " pod="openstack/cinder-api-0" Dec 10 23:09:23 crc kubenswrapper[4791]: I1210 23:09:23.718090 4791 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/f807e752-6cf5-4ad6-bc21-f22777b62c83-scripts\") pod \"cinder-api-0\" (UID: \"f807e752-6cf5-4ad6-bc21-f22777b62c83\") " pod="openstack/cinder-api-0" Dec 10 23:09:23 crc kubenswrapper[4791]: I1210 23:09:23.718116 4791 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/f807e752-6cf5-4ad6-bc21-f22777b62c83-internal-tls-certs\") pod \"cinder-api-0\" (UID: \"f807e752-6cf5-4ad6-bc21-f22777b62c83\") " pod="openstack/cinder-api-0" Dec 10 23:09:23 crc kubenswrapper[4791]: I1210 23:09:23.718139 4791 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/f807e752-6cf5-4ad6-bc21-f22777b62c83-etc-machine-id\") pod \"cinder-api-0\" (UID: \"f807e752-6cf5-4ad6-bc21-f22777b62c83\") " pod="openstack/cinder-api-0" Dec 10 23:09:23 crc kubenswrapper[4791]: I1210 23:09:23.718160 4791 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/f807e752-6cf5-4ad6-bc21-f22777b62c83-config-data\") pod \"cinder-api-0\" (UID: \"f807e752-6cf5-4ad6-bc21-f22777b62c83\") " pod="openstack/cinder-api-0" Dec 10 23:09:23 crc kubenswrapper[4791]: I1210 23:09:23.720060 4791 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/f807e752-6cf5-4ad6-bc21-f22777b62c83-etc-machine-id\") pod \"cinder-api-0\" (UID: \"f807e752-6cf5-4ad6-bc21-f22777b62c83\") " pod="openstack/cinder-api-0" Dec 10 23:09:23 crc kubenswrapper[4791]: I1210 23:09:23.721206 4791 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/f807e752-6cf5-4ad6-bc21-f22777b62c83-logs\") pod \"cinder-api-0\" (UID: \"f807e752-6cf5-4ad6-bc21-f22777b62c83\") " pod="openstack/cinder-api-0" Dec 10 23:09:23 crc kubenswrapper[4791]: I1210 23:09:23.723687 4791 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/f807e752-6cf5-4ad6-bc21-f22777b62c83-scripts\") pod \"cinder-api-0\" (UID: \"f807e752-6cf5-4ad6-bc21-f22777b62c83\") " pod="openstack/cinder-api-0" Dec 10 23:09:23 crc kubenswrapper[4791]: I1210 23:09:23.724791 4791 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/f807e752-6cf5-4ad6-bc21-f22777b62c83-config-data-custom\") pod \"cinder-api-0\" (UID: \"f807e752-6cf5-4ad6-bc21-f22777b62c83\") " pod="openstack/cinder-api-0" Dec 10 23:09:23 crc kubenswrapper[4791]: I1210 23:09:23.724817 4791 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/f807e752-6cf5-4ad6-bc21-f22777b62c83-internal-tls-certs\") pod \"cinder-api-0\" (UID: \"f807e752-6cf5-4ad6-bc21-f22777b62c83\") " pod="openstack/cinder-api-0" Dec 10 23:09:23 crc kubenswrapper[4791]: I1210 23:09:23.725138 4791 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f807e752-6cf5-4ad6-bc21-f22777b62c83-combined-ca-bundle\") pod \"cinder-api-0\" (UID: \"f807e752-6cf5-4ad6-bc21-f22777b62c83\") " pod="openstack/cinder-api-0" Dec 10 23:09:23 crc kubenswrapper[4791]: I1210 23:09:23.725621 4791 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/f807e752-6cf5-4ad6-bc21-f22777b62c83-public-tls-certs\") pod \"cinder-api-0\" (UID: \"f807e752-6cf5-4ad6-bc21-f22777b62c83\") " pod="openstack/cinder-api-0" Dec 10 23:09:23 crc kubenswrapper[4791]: I1210 23:09:23.725784 4791 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/f807e752-6cf5-4ad6-bc21-f22777b62c83-config-data\") pod \"cinder-api-0\" (UID: \"f807e752-6cf5-4ad6-bc21-f22777b62c83\") " pod="openstack/cinder-api-0" Dec 10 23:09:23 crc kubenswrapper[4791]: I1210 23:09:23.737534 4791 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-7fhmf\" (UniqueName: \"kubernetes.io/projected/f807e752-6cf5-4ad6-bc21-f22777b62c83-kube-api-access-7fhmf\") pod \"cinder-api-0\" (UID: \"f807e752-6cf5-4ad6-bc21-f22777b62c83\") " pod="openstack/cinder-api-0" Dec 10 23:09:23 crc kubenswrapper[4791]: I1210 23:09:23.862761 4791 generic.go:334] "Generic (PLEG): container finished" podID="224a8e50-4270-42e8-a1df-268bc144b1d5" containerID="930aefa334c3a39fdcf6c92de149a6110c9b76856c19a726d39b0c5feff817ad" exitCode=0 Dec 10 23:09:23 crc kubenswrapper[4791]: I1210 23:09:23.862862 4791 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-5784cf869f-ngjqj" event={"ID":"224a8e50-4270-42e8-a1df-268bc144b1d5","Type":"ContainerDied","Data":"930aefa334c3a39fdcf6c92de149a6110c9b76856c19a726d39b0c5feff817ad"} Dec 10 23:09:23 crc kubenswrapper[4791]: I1210 23:09:23.868587 4791 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-api-68fb764d56-4mfz4" event={"ID":"eb482cb0-7ac1-4097-8799-4a5d78b6b5e5","Type":"ContainerStarted","Data":"a2bba91961a1ed46f84b823ae193b87e5ccd7758ddeaaf6603b3037ca69400b6"} Dec 10 23:09:23 crc kubenswrapper[4791]: I1210 23:09:23.918280 4791 generic.go:334] "Generic (PLEG): container finished" podID="55139c7c-e9fa-4539-b7b2-b2be7ee43b44" containerID="a441e9dd03ea293d73ed83e0ce82cd1f391a27dfc4d4c0f86b23bba57c39c2c5" exitCode=0 Dec 10 23:09:23 crc kubenswrapper[4791]: I1210 23:09:23.918929 4791 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-api-0" Dec 10 23:09:23 crc kubenswrapper[4791]: I1210 23:09:23.923955 4791 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/horizon-78fc5d56dc-58cjt" Dec 10 23:09:23 crc kubenswrapper[4791]: I1210 23:09:23.927121 4791 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/horizon-5cc98b55bf-7xpnj" Dec 10 23:09:23 crc kubenswrapper[4791]: I1210 23:09:23.932613 4791 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="8eaa1b1d-715f-4059-ad18-fe9632c036a7" containerName="sg-core" containerID="cri-o://29aea3c141a06d60106b8924f521c5d11243d56785dfbc6ff7d120f58e023588" gracePeriod=30 Dec 10 23:09:23 crc kubenswrapper[4791]: I1210 23:09:23.932614 4791 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="8eaa1b1d-715f-4059-ad18-fe9632c036a7" containerName="proxy-httpd" containerID="cri-o://72e65227cdfd1c319b35d8f974f3175e29919dfd3b2cce0325ebe6425722d9b2" gracePeriod=30 Dec 10 23:09:23 crc kubenswrapper[4791]: I1210 23:09:23.933001 4791 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="8eaa1b1d-715f-4059-ad18-fe9632c036a7" containerName="ceilometer-notification-agent" containerID="cri-o://a6fb5e073d90b22471fee522d523321fa4497ebdf36b79f38fd1ee68b1a79815" gracePeriod=30 Dec 10 23:09:23 crc kubenswrapper[4791]: I1210 23:09:23.932927 4791 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="8eaa1b1d-715f-4059-ad18-fe9632c036a7" containerName="ceilometer-central-agent" containerID="cri-o://5ae0fb7182370bec008f1c24c5d245fd6cb390d075a883adda305056ffab03b5" gracePeriod=30 Dec 10 23:09:23 crc kubenswrapper[4791]: I1210 23:09:23.938651 4791 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/horizon-5f54ccbf57-wwpfk" Dec 10 23:09:23 crc kubenswrapper[4791]: I1210 23:09:23.948193 4791 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/barbican-api-795f5df4db-xf57n" podStartSLOduration=8.94817501 podStartE2EDuration="8.94817501s" podCreationTimestamp="2025-12-10 23:09:15 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-10 23:09:23.946673037 +0000 UTC m=+1198.376290650" watchObservedRunningTime="2025-12-10 23:09:23.94817501 +0000 UTC m=+1198.377792623" Dec 10 23:09:23 crc kubenswrapper[4791]: I1210 23:09:23.965067 4791 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="1a1ca37c-dab5-4748-b8c2-eb985e66a02f" path="/var/lib/kubelet/pods/1a1ca37c-dab5-4748-b8c2-eb985e66a02f/volumes" Dec 10 23:09:23 crc kubenswrapper[4791]: I1210 23:09:23.972427 4791 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ceilometer-0" podStartSLOduration=3.165734911 podStartE2EDuration="1m3.97240801s" podCreationTimestamp="2025-12-10 23:08:20 +0000 UTC" firstStartedPulling="2025-12-10 23:08:21.719979466 +0000 UTC m=+1136.149597069" lastFinishedPulling="2025-12-10 23:09:22.526652535 +0000 UTC m=+1196.956270168" observedRunningTime="2025-12-10 23:09:23.969030282 +0000 UTC m=+1198.398647895" watchObservedRunningTime="2025-12-10 23:09:23.97240801 +0000 UTC m=+1198.402025623" Dec 10 23:09:23 crc kubenswrapper[4791]: I1210 23:09:23.978901 4791 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="ae2623b4-7b39-4661-b9af-caf32415d8b1" path="/var/lib/kubelet/pods/ae2623b4-7b39-4661-b9af-caf32415d8b1/volumes" Dec 10 23:09:23 crc kubenswrapper[4791]: I1210 23:09:23.991889 4791 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/barbican-api-795f5df4db-xf57n" Dec 10 23:09:23 crc kubenswrapper[4791]: I1210 23:09:23.992610 4791 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/barbican-api-795f5df4db-xf57n" Dec 10 23:09:23 crc kubenswrapper[4791]: I1210 23:09:23.992639 4791 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/ceilometer-0" Dec 10 23:09:23 crc kubenswrapper[4791]: I1210 23:09:23.992659 4791 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/horizon-7cc66567bd-pwb2l" event={"ID":"55139c7c-e9fa-4539-b7b2-b2be7ee43b44","Type":"ContainerDied","Data":"a441e9dd03ea293d73ed83e0ce82cd1f391a27dfc4d4c0f86b23bba57c39c2c5"} Dec 10 23:09:23 crc kubenswrapper[4791]: I1210 23:09:23.992686 4791 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-api-795f5df4db-xf57n" event={"ID":"9b340efd-bf7e-4caf-8c76-ddc500aa2085","Type":"ContainerStarted","Data":"31193dfd2f9e85871df3d79e4aef1d62fa1b1a1b269896d2e71dbdb4f4784a21"} Dec 10 23:09:23 crc kubenswrapper[4791]: I1210 23:09:23.992700 4791 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-api-795f5df4db-xf57n" event={"ID":"9b340efd-bf7e-4caf-8c76-ddc500aa2085","Type":"ContainerStarted","Data":"f05d3eef09f65c594818648ae666e092eb1a3e05925b8b910f123fdb1170cde9"} Dec 10 23:09:23 crc kubenswrapper[4791]: I1210 23:09:23.992713 4791 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/horizon-78fc5d56dc-58cjt" event={"ID":"14f46cbc-afe2-40bb-a2a3-5673b5ac53e7","Type":"ContainerDied","Data":"24570b6eae27ba7121cf1d2b73b3c0e94c1598dfd6a06e134fc597203116ee98"} Dec 10 23:09:23 crc kubenswrapper[4791]: I1210 23:09:23.992729 4791 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/horizon-5cc98b55bf-7xpnj" event={"ID":"9e9c1419-c6f3-4947-8788-4915f48677d7","Type":"ContainerDied","Data":"4a159644255a6732d7d58d79d7ad6c6e1295b53c7457c9ca2759ec3d5712484b"} Dec 10 23:09:23 crc kubenswrapper[4791]: I1210 23:09:23.992745 4791 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"8eaa1b1d-715f-4059-ad18-fe9632c036a7","Type":"ContainerStarted","Data":"72e65227cdfd1c319b35d8f974f3175e29919dfd3b2cce0325ebe6425722d9b2"} Dec 10 23:09:23 crc kubenswrapper[4791]: I1210 23:09:23.992756 4791 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/horizon-5f54ccbf57-wwpfk" event={"ID":"aa3a63ab-6092-46ce-aaf6-40360aa060ac","Type":"ContainerDied","Data":"cd3a5fe7eb6d3ca058066e281d4e2debdd7d84c3d08f93c34d63063dbabf1a7d"} Dec 10 23:09:23 crc kubenswrapper[4791]: I1210 23:09:23.992781 4791 scope.go:117] "RemoveContainer" containerID="1bf24a4c4c7be94d64b9919fd150ec64d3f744be3bf1b6f17541d5c0b7a27a30" Dec 10 23:09:24 crc kubenswrapper[4791]: I1210 23:09:24.007841 4791 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/horizon-5cc98b55bf-7xpnj"] Dec 10 23:09:24 crc kubenswrapper[4791]: I1210 23:09:24.027202 4791 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/horizon-5cc98b55bf-7xpnj"] Dec 10 23:09:24 crc kubenswrapper[4791]: I1210 23:09:24.036831 4791 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/horizon-78fc5d56dc-58cjt"] Dec 10 23:09:24 crc kubenswrapper[4791]: I1210 23:09:24.044094 4791 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/horizon-78fc5d56dc-58cjt"] Dec 10 23:09:24 crc kubenswrapper[4791]: I1210 23:09:24.055400 4791 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/horizon-5f54ccbf57-wwpfk"] Dec 10 23:09:24 crc kubenswrapper[4791]: I1210 23:09:24.061442 4791 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/horizon-5f54ccbf57-wwpfk"] Dec 10 23:09:24 crc kubenswrapper[4791]: I1210 23:09:24.553120 4791 scope.go:117] "RemoveContainer" containerID="d5bed1e64f4f7d6b0cbb4910d9037ebbea49eb4ac927509fe80353feed0ed97d" Dec 10 23:09:24 crc kubenswrapper[4791]: I1210 23:09:24.663441 4791 scope.go:117] "RemoveContainer" containerID="e8a0fecaed859f1a8a83f5c36eabd8013be451668e6363d31a5ebdbc8e9a0863" Dec 10 23:09:24 crc kubenswrapper[4791]: I1210 23:09:24.895133 4791 scope.go:117] "RemoveContainer" containerID="9369387266f0bf17a3578a2973371e39479e41aeaf91821d2bd23efe9fbbe09e" Dec 10 23:09:24 crc kubenswrapper[4791]: I1210 23:09:24.920179 4791 scope.go:117] "RemoveContainer" containerID="51262b7b3d283cb5b5b3f2953648269ed55c1d6a470bbdf743fc7e6a314d9407" Dec 10 23:09:24 crc kubenswrapper[4791]: I1210 23:09:24.950654 4791 generic.go:334] "Generic (PLEG): container finished" podID="5e5e327e-b8dc-4bd6-af75-8aebc417ada9" containerID="0491f982573b4cf5b921d4eb64aecef9ffb9a9bed5a7cd16d5e70ffa514274f4" exitCode=0 Dec 10 23:09:24 crc kubenswrapper[4791]: I1210 23:09:24.950708 4791 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-scheduler-0" event={"ID":"5e5e327e-b8dc-4bd6-af75-8aebc417ada9","Type":"ContainerDied","Data":"0491f982573b4cf5b921d4eb64aecef9ffb9a9bed5a7cd16d5e70ffa514274f4"} Dec 10 23:09:24 crc kubenswrapper[4791]: I1210 23:09:24.952535 4791 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-api-68fb764d56-4mfz4" event={"ID":"eb482cb0-7ac1-4097-8799-4a5d78b6b5e5","Type":"ContainerStarted","Data":"7c0a77b49c0868ac45952fdb5c9f904149d7e67ad62f453d5cfa68d082c4c685"} Dec 10 23:09:24 crc kubenswrapper[4791]: I1210 23:09:24.952644 4791 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-api-68fb764d56-4mfz4" event={"ID":"eb482cb0-7ac1-4097-8799-4a5d78b6b5e5","Type":"ContainerStarted","Data":"0e7c75a93a8584bd5df8575dd1a6ea44ffa8571a5892b4012804f62b3d85876a"} Dec 10 23:09:24 crc kubenswrapper[4791]: I1210 23:09:24.952659 4791 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/barbican-api-68fb764d56-4mfz4" Dec 10 23:09:24 crc kubenswrapper[4791]: I1210 23:09:24.952693 4791 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/barbican-api-68fb764d56-4mfz4" Dec 10 23:09:24 crc kubenswrapper[4791]: I1210 23:09:24.955531 4791 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-worker-77fb99879-xzsrk" event={"ID":"8a302b7c-c2df-4023-95da-4ef8e86dbc0b","Type":"ContainerStarted","Data":"28161fcb6a70f0e0e11c0a0fc2a5a7bd8d4131cd417e16511a0e8298309e5de2"} Dec 10 23:09:24 crc kubenswrapper[4791]: I1210 23:09:24.958834 4791 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-5784cf869f-ngjqj" event={"ID":"224a8e50-4270-42e8-a1df-268bc144b1d5","Type":"ContainerStarted","Data":"d0b7aa50e7609ff3f59cefa955b1e1e05f6d96ae4a3e4143c8e51f4065c45be1"} Dec 10 23:09:24 crc kubenswrapper[4791]: I1210 23:09:24.959365 4791 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/dnsmasq-dns-5784cf869f-ngjqj" Dec 10 23:09:24 crc kubenswrapper[4791]: I1210 23:09:24.963376 4791 generic.go:334] "Generic (PLEG): container finished" podID="8eaa1b1d-715f-4059-ad18-fe9632c036a7" containerID="72e65227cdfd1c319b35d8f974f3175e29919dfd3b2cce0325ebe6425722d9b2" exitCode=0 Dec 10 23:09:24 crc kubenswrapper[4791]: I1210 23:09:24.963398 4791 generic.go:334] "Generic (PLEG): container finished" podID="8eaa1b1d-715f-4059-ad18-fe9632c036a7" containerID="29aea3c141a06d60106b8924f521c5d11243d56785dfbc6ff7d120f58e023588" exitCode=2 Dec 10 23:09:24 crc kubenswrapper[4791]: I1210 23:09:24.963406 4791 generic.go:334] "Generic (PLEG): container finished" podID="8eaa1b1d-715f-4059-ad18-fe9632c036a7" containerID="5ae0fb7182370bec008f1c24c5d245fd6cb390d075a883adda305056ffab03b5" exitCode=0 Dec 10 23:09:24 crc kubenswrapper[4791]: I1210 23:09:24.963443 4791 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"8eaa1b1d-715f-4059-ad18-fe9632c036a7","Type":"ContainerDied","Data":"72e65227cdfd1c319b35d8f974f3175e29919dfd3b2cce0325ebe6425722d9b2"} Dec 10 23:09:24 crc kubenswrapper[4791]: I1210 23:09:24.963470 4791 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"8eaa1b1d-715f-4059-ad18-fe9632c036a7","Type":"ContainerDied","Data":"29aea3c141a06d60106b8924f521c5d11243d56785dfbc6ff7d120f58e023588"} Dec 10 23:09:24 crc kubenswrapper[4791]: I1210 23:09:24.963484 4791 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"8eaa1b1d-715f-4059-ad18-fe9632c036a7","Type":"ContainerDied","Data":"5ae0fb7182370bec008f1c24c5d245fd6cb390d075a883adda305056ffab03b5"} Dec 10 23:09:24 crc kubenswrapper[4791]: I1210 23:09:24.980269 4791 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/barbican-api-68fb764d56-4mfz4" podStartSLOduration=7.980253104 podStartE2EDuration="7.980253104s" podCreationTimestamp="2025-12-10 23:09:17 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-10 23:09:24.968058692 +0000 UTC m=+1199.397676315" watchObservedRunningTime="2025-12-10 23:09:24.980253104 +0000 UTC m=+1199.409870717" Dec 10 23:09:24 crc kubenswrapper[4791]: I1210 23:09:24.992926 4791 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/dnsmasq-dns-5784cf869f-ngjqj" podStartSLOduration=9.992910989 podStartE2EDuration="9.992910989s" podCreationTimestamp="2025-12-10 23:09:15 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-10 23:09:24.986544555 +0000 UTC m=+1199.416162168" watchObservedRunningTime="2025-12-10 23:09:24.992910989 +0000 UTC m=+1199.422528602" Dec 10 23:09:25 crc kubenswrapper[4791]: I1210 23:09:25.040548 4791 patch_prober.go:28] interesting pod/machine-config-daemon-5rb5l container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 10 23:09:25 crc kubenswrapper[4791]: I1210 23:09:25.040623 4791 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-5rb5l" podUID="ba35653c-6e06-4cee-a4d6-137764090d18" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 10 23:09:25 crc kubenswrapper[4791]: I1210 23:09:25.040672 4791 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-5rb5l" Dec 10 23:09:25 crc kubenswrapper[4791]: I1210 23:09:25.041427 4791 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"2b6bbc8a626f5244441ef712b0db42022f2a15236c1846adb0b1f8c097905a73"} pod="openshift-machine-config-operator/machine-config-daemon-5rb5l" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Dec 10 23:09:25 crc kubenswrapper[4791]: I1210 23:09:25.041488 4791 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-5rb5l" podUID="ba35653c-6e06-4cee-a4d6-137764090d18" containerName="machine-config-daemon" containerID="cri-o://2b6bbc8a626f5244441ef712b0db42022f2a15236c1846adb0b1f8c097905a73" gracePeriod=600 Dec 10 23:09:25 crc kubenswrapper[4791]: I1210 23:09:25.069086 4791 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/neutron-856d8bb659-x5kwg" Dec 10 23:09:25 crc kubenswrapper[4791]: I1210 23:09:25.128770 4791 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-api-0"] Dec 10 23:09:25 crc kubenswrapper[4791]: I1210 23:09:25.176011 4791 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/neutron-85794c6bd6-ht8sz"] Dec 10 23:09:25 crc kubenswrapper[4791]: I1210 23:09:25.176199 4791 scope.go:117] "RemoveContainer" containerID="35844ea7a50bd278dafc5819fb34072df86afbe70e63e900d65a4f0f56c5cfc8" Dec 10 23:09:25 crc kubenswrapper[4791]: I1210 23:09:25.176238 4791 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/neutron-85794c6bd6-ht8sz" podUID="75eb971a-04fc-40f5-9890-d5a0b4daccd1" containerName="neutron-api" containerID="cri-o://dd2fb7dddb229ce25296aefb965623011367c1640da7f68c5f11bd17f39afbdc" gracePeriod=30 Dec 10 23:09:25 crc kubenswrapper[4791]: I1210 23:09:25.176475 4791 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/neutron-85794c6bd6-ht8sz" podUID="75eb971a-04fc-40f5-9890-d5a0b4daccd1" containerName="neutron-httpd" containerID="cri-o://a1c20ac32bb42bda941e83eca0be34d9e4905b4b8eff8dab079f24798af60c6b" gracePeriod=30 Dec 10 23:09:25 crc kubenswrapper[4791]: I1210 23:09:25.909019 4791 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="14f46cbc-afe2-40bb-a2a3-5673b5ac53e7" path="/var/lib/kubelet/pods/14f46cbc-afe2-40bb-a2a3-5673b5ac53e7/volumes" Dec 10 23:09:25 crc kubenswrapper[4791]: I1210 23:09:25.910686 4791 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="9e9c1419-c6f3-4947-8788-4915f48677d7" path="/var/lib/kubelet/pods/9e9c1419-c6f3-4947-8788-4915f48677d7/volumes" Dec 10 23:09:25 crc kubenswrapper[4791]: I1210 23:09:25.915936 4791 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="aa3a63ab-6092-46ce-aaf6-40360aa060ac" path="/var/lib/kubelet/pods/aa3a63ab-6092-46ce-aaf6-40360aa060ac/volumes" Dec 10 23:09:25 crc kubenswrapper[4791]: I1210 23:09:25.977449 4791 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Dec 10 23:09:25 crc kubenswrapper[4791]: I1210 23:09:25.988568 4791 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-keystone-listener-67867746d6-mvxp4" event={"ID":"4d7d941e-8494-46ea-990c-fe83db67dd2a","Type":"ContainerStarted","Data":"ecda741f9ae5bd0089d3cd51e1ec65a73479cf035f0bfbae37a16a3dd7630bf9"} Dec 10 23:09:25 crc kubenswrapper[4791]: I1210 23:09:25.988607 4791 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-keystone-listener-67867746d6-mvxp4" event={"ID":"4d7d941e-8494-46ea-990c-fe83db67dd2a","Type":"ContainerStarted","Data":"0e25db1d89edafb7a655c2d17e34975f762599ae084921ab785cee000fe699d8"} Dec 10 23:09:25 crc kubenswrapper[4791]: I1210 23:09:25.994388 4791 generic.go:334] "Generic (PLEG): container finished" podID="ba35653c-6e06-4cee-a4d6-137764090d18" containerID="2b6bbc8a626f5244441ef712b0db42022f2a15236c1846adb0b1f8c097905a73" exitCode=0 Dec 10 23:09:25 crc kubenswrapper[4791]: I1210 23:09:25.994439 4791 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-5rb5l" event={"ID":"ba35653c-6e06-4cee-a4d6-137764090d18","Type":"ContainerDied","Data":"2b6bbc8a626f5244441ef712b0db42022f2a15236c1846adb0b1f8c097905a73"} Dec 10 23:09:25 crc kubenswrapper[4791]: I1210 23:09:25.994460 4791 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-5rb5l" event={"ID":"ba35653c-6e06-4cee-a4d6-137764090d18","Type":"ContainerStarted","Data":"27ea6396cad6699de1e6134e18b2c18010b7367f4d673a0bdef71d4d1c354c04"} Dec 10 23:09:25 crc kubenswrapper[4791]: I1210 23:09:25.994477 4791 scope.go:117] "RemoveContainer" containerID="6bffb04499032491e4580b048fb3dc3c71ca5a93b7a98e3a62cb913d0da4b5c3" Dec 10 23:09:25 crc kubenswrapper[4791]: I1210 23:09:25.998930 4791 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-worker-77fb99879-xzsrk" event={"ID":"8a302b7c-c2df-4023-95da-4ef8e86dbc0b","Type":"ContainerStarted","Data":"35eed59344d6120aad4ed666324fe54050390a300efa146c4a451364ec3f740d"} Dec 10 23:09:26 crc kubenswrapper[4791]: I1210 23:09:26.021140 4791 generic.go:334] "Generic (PLEG): container finished" podID="8eaa1b1d-715f-4059-ad18-fe9632c036a7" containerID="a6fb5e073d90b22471fee522d523321fa4497ebdf36b79f38fd1ee68b1a79815" exitCode=0 Dec 10 23:09:26 crc kubenswrapper[4791]: I1210 23:09:26.021213 4791 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"8eaa1b1d-715f-4059-ad18-fe9632c036a7","Type":"ContainerDied","Data":"a6fb5e073d90b22471fee522d523321fa4497ebdf36b79f38fd1ee68b1a79815"} Dec 10 23:09:26 crc kubenswrapper[4791]: I1210 23:09:26.021257 4791 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"8eaa1b1d-715f-4059-ad18-fe9632c036a7","Type":"ContainerDied","Data":"759f432e76faae97192a0bd18f244848d3975025239dc1ed6f67864d6c26fd4a"} Dec 10 23:09:26 crc kubenswrapper[4791]: I1210 23:09:26.021363 4791 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Dec 10 23:09:26 crc kubenswrapper[4791]: I1210 23:09:26.032036 4791 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-api-0" event={"ID":"f807e752-6cf5-4ad6-bc21-f22777b62c83","Type":"ContainerStarted","Data":"b71444cbde07780c98f8344c34363ea0b2dbd1ba69d5eea4c1ff268d0e679199"} Dec 10 23:09:26 crc kubenswrapper[4791]: I1210 23:09:26.048642 4791 generic.go:334] "Generic (PLEG): container finished" podID="75eb971a-04fc-40f5-9890-d5a0b4daccd1" containerID="a1c20ac32bb42bda941e83eca0be34d9e4905b4b8eff8dab079f24798af60c6b" exitCode=0 Dec 10 23:09:26 crc kubenswrapper[4791]: I1210 23:09:26.048707 4791 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-85794c6bd6-ht8sz" event={"ID":"75eb971a-04fc-40f5-9890-d5a0b4daccd1","Type":"ContainerDied","Data":"a1c20ac32bb42bda941e83eca0be34d9e4905b4b8eff8dab079f24798af60c6b"} Dec 10 23:09:26 crc kubenswrapper[4791]: I1210 23:09:26.073026 4791 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/8eaa1b1d-715f-4059-ad18-fe9632c036a7-log-httpd\") pod \"8eaa1b1d-715f-4059-ad18-fe9632c036a7\" (UID: \"8eaa1b1d-715f-4059-ad18-fe9632c036a7\") " Dec 10 23:09:26 crc kubenswrapper[4791]: I1210 23:09:26.073107 4791 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/8eaa1b1d-715f-4059-ad18-fe9632c036a7-combined-ca-bundle\") pod \"8eaa1b1d-715f-4059-ad18-fe9632c036a7\" (UID: \"8eaa1b1d-715f-4059-ad18-fe9632c036a7\") " Dec 10 23:09:26 crc kubenswrapper[4791]: I1210 23:09:26.073158 4791 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-7zjdl\" (UniqueName: \"kubernetes.io/projected/8eaa1b1d-715f-4059-ad18-fe9632c036a7-kube-api-access-7zjdl\") pod \"8eaa1b1d-715f-4059-ad18-fe9632c036a7\" (UID: \"8eaa1b1d-715f-4059-ad18-fe9632c036a7\") " Dec 10 23:09:26 crc kubenswrapper[4791]: I1210 23:09:26.073232 4791 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/8eaa1b1d-715f-4059-ad18-fe9632c036a7-run-httpd\") pod \"8eaa1b1d-715f-4059-ad18-fe9632c036a7\" (UID: \"8eaa1b1d-715f-4059-ad18-fe9632c036a7\") " Dec 10 23:09:26 crc kubenswrapper[4791]: I1210 23:09:26.073273 4791 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/8eaa1b1d-715f-4059-ad18-fe9632c036a7-scripts\") pod \"8eaa1b1d-715f-4059-ad18-fe9632c036a7\" (UID: \"8eaa1b1d-715f-4059-ad18-fe9632c036a7\") " Dec 10 23:09:26 crc kubenswrapper[4791]: I1210 23:09:26.073317 4791 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/8eaa1b1d-715f-4059-ad18-fe9632c036a7-config-data\") pod \"8eaa1b1d-715f-4059-ad18-fe9632c036a7\" (UID: \"8eaa1b1d-715f-4059-ad18-fe9632c036a7\") " Dec 10 23:09:26 crc kubenswrapper[4791]: I1210 23:09:26.073429 4791 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/8eaa1b1d-715f-4059-ad18-fe9632c036a7-sg-core-conf-yaml\") pod \"8eaa1b1d-715f-4059-ad18-fe9632c036a7\" (UID: \"8eaa1b1d-715f-4059-ad18-fe9632c036a7\") " Dec 10 23:09:26 crc kubenswrapper[4791]: I1210 23:09:26.073717 4791 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/barbican-keystone-listener-67867746d6-mvxp4" podStartSLOduration=7.7198858139999995 podStartE2EDuration="11.073692929s" podCreationTimestamp="2025-12-10 23:09:15 +0000 UTC" firstStartedPulling="2025-12-10 23:09:21.271217014 +0000 UTC m=+1195.700834627" lastFinishedPulling="2025-12-10 23:09:24.625024129 +0000 UTC m=+1199.054641742" observedRunningTime="2025-12-10 23:09:26.070917328 +0000 UTC m=+1200.500534941" watchObservedRunningTime="2025-12-10 23:09:26.073692929 +0000 UTC m=+1200.503310542" Dec 10 23:09:26 crc kubenswrapper[4791]: I1210 23:09:26.075146 4791 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/8eaa1b1d-715f-4059-ad18-fe9632c036a7-log-httpd" (OuterVolumeSpecName: "log-httpd") pod "8eaa1b1d-715f-4059-ad18-fe9632c036a7" (UID: "8eaa1b1d-715f-4059-ad18-fe9632c036a7"). InnerVolumeSpecName "log-httpd". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 10 23:09:26 crc kubenswrapper[4791]: I1210 23:09:26.079421 4791 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/8eaa1b1d-715f-4059-ad18-fe9632c036a7-run-httpd" (OuterVolumeSpecName: "run-httpd") pod "8eaa1b1d-715f-4059-ad18-fe9632c036a7" (UID: "8eaa1b1d-715f-4059-ad18-fe9632c036a7"). InnerVolumeSpecName "run-httpd". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 10 23:09:26 crc kubenswrapper[4791]: I1210 23:09:26.086027 4791 scope.go:117] "RemoveContainer" containerID="72e65227cdfd1c319b35d8f974f3175e29919dfd3b2cce0325ebe6425722d9b2" Dec 10 23:09:26 crc kubenswrapper[4791]: I1210 23:09:26.095729 4791 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/8eaa1b1d-715f-4059-ad18-fe9632c036a7-kube-api-access-7zjdl" (OuterVolumeSpecName: "kube-api-access-7zjdl") pod "8eaa1b1d-715f-4059-ad18-fe9632c036a7" (UID: "8eaa1b1d-715f-4059-ad18-fe9632c036a7"). InnerVolumeSpecName "kube-api-access-7zjdl". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 10 23:09:26 crc kubenswrapper[4791]: I1210 23:09:26.103088 4791 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/barbican-worker-77fb99879-xzsrk" podStartSLOduration=8.777327252 podStartE2EDuration="12.103064586s" podCreationTimestamp="2025-12-10 23:09:14 +0000 UTC" firstStartedPulling="2025-12-10 23:09:21.278819784 +0000 UTC m=+1195.708437397" lastFinishedPulling="2025-12-10 23:09:24.604557118 +0000 UTC m=+1199.034174731" observedRunningTime="2025-12-10 23:09:26.095967252 +0000 UTC m=+1200.525584875" watchObservedRunningTime="2025-12-10 23:09:26.103064586 +0000 UTC m=+1200.532682199" Dec 10 23:09:26 crc kubenswrapper[4791]: I1210 23:09:26.125595 4791 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/8eaa1b1d-715f-4059-ad18-fe9632c036a7-scripts" (OuterVolumeSpecName: "scripts") pod "8eaa1b1d-715f-4059-ad18-fe9632c036a7" (UID: "8eaa1b1d-715f-4059-ad18-fe9632c036a7"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 10 23:09:26 crc kubenswrapper[4791]: I1210 23:09:26.154488 4791 scope.go:117] "RemoveContainer" containerID="29aea3c141a06d60106b8924f521c5d11243d56785dfbc6ff7d120f58e023588" Dec 10 23:09:26 crc kubenswrapper[4791]: I1210 23:09:26.176661 4791 reconciler_common.go:293] "Volume detached for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/8eaa1b1d-715f-4059-ad18-fe9632c036a7-run-httpd\") on node \"crc\" DevicePath \"\"" Dec 10 23:09:26 crc kubenswrapper[4791]: I1210 23:09:26.176726 4791 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/8eaa1b1d-715f-4059-ad18-fe9632c036a7-scripts\") on node \"crc\" DevicePath \"\"" Dec 10 23:09:26 crc kubenswrapper[4791]: I1210 23:09:26.176741 4791 reconciler_common.go:293] "Volume detached for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/8eaa1b1d-715f-4059-ad18-fe9632c036a7-log-httpd\") on node \"crc\" DevicePath \"\"" Dec 10 23:09:26 crc kubenswrapper[4791]: I1210 23:09:26.176752 4791 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-7zjdl\" (UniqueName: \"kubernetes.io/projected/8eaa1b1d-715f-4059-ad18-fe9632c036a7-kube-api-access-7zjdl\") on node \"crc\" DevicePath \"\"" Dec 10 23:09:26 crc kubenswrapper[4791]: I1210 23:09:26.186968 4791 scope.go:117] "RemoveContainer" containerID="a6fb5e073d90b22471fee522d523321fa4497ebdf36b79f38fd1ee68b1a79815" Dec 10 23:09:26 crc kubenswrapper[4791]: I1210 23:09:26.199334 4791 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/8eaa1b1d-715f-4059-ad18-fe9632c036a7-sg-core-conf-yaml" (OuterVolumeSpecName: "sg-core-conf-yaml") pod "8eaa1b1d-715f-4059-ad18-fe9632c036a7" (UID: "8eaa1b1d-715f-4059-ad18-fe9632c036a7"). InnerVolumeSpecName "sg-core-conf-yaml". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 10 23:09:26 crc kubenswrapper[4791]: I1210 23:09:26.226019 4791 scope.go:117] "RemoveContainer" containerID="5ae0fb7182370bec008f1c24c5d245fd6cb390d075a883adda305056ffab03b5" Dec 10 23:09:26 crc kubenswrapper[4791]: I1210 23:09:26.250515 4791 scope.go:117] "RemoveContainer" containerID="72e65227cdfd1c319b35d8f974f3175e29919dfd3b2cce0325ebe6425722d9b2" Dec 10 23:09:26 crc kubenswrapper[4791]: E1210 23:09:26.250909 4791 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"72e65227cdfd1c319b35d8f974f3175e29919dfd3b2cce0325ebe6425722d9b2\": container with ID starting with 72e65227cdfd1c319b35d8f974f3175e29919dfd3b2cce0325ebe6425722d9b2 not found: ID does not exist" containerID="72e65227cdfd1c319b35d8f974f3175e29919dfd3b2cce0325ebe6425722d9b2" Dec 10 23:09:26 crc kubenswrapper[4791]: I1210 23:09:26.250964 4791 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"72e65227cdfd1c319b35d8f974f3175e29919dfd3b2cce0325ebe6425722d9b2"} err="failed to get container status \"72e65227cdfd1c319b35d8f974f3175e29919dfd3b2cce0325ebe6425722d9b2\": rpc error: code = NotFound desc = could not find container \"72e65227cdfd1c319b35d8f974f3175e29919dfd3b2cce0325ebe6425722d9b2\": container with ID starting with 72e65227cdfd1c319b35d8f974f3175e29919dfd3b2cce0325ebe6425722d9b2 not found: ID does not exist" Dec 10 23:09:26 crc kubenswrapper[4791]: I1210 23:09:26.250997 4791 scope.go:117] "RemoveContainer" containerID="29aea3c141a06d60106b8924f521c5d11243d56785dfbc6ff7d120f58e023588" Dec 10 23:09:26 crc kubenswrapper[4791]: E1210 23:09:26.251340 4791 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"29aea3c141a06d60106b8924f521c5d11243d56785dfbc6ff7d120f58e023588\": container with ID starting with 29aea3c141a06d60106b8924f521c5d11243d56785dfbc6ff7d120f58e023588 not found: ID does not exist" containerID="29aea3c141a06d60106b8924f521c5d11243d56785dfbc6ff7d120f58e023588" Dec 10 23:09:26 crc kubenswrapper[4791]: I1210 23:09:26.251386 4791 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"29aea3c141a06d60106b8924f521c5d11243d56785dfbc6ff7d120f58e023588"} err="failed to get container status \"29aea3c141a06d60106b8924f521c5d11243d56785dfbc6ff7d120f58e023588\": rpc error: code = NotFound desc = could not find container \"29aea3c141a06d60106b8924f521c5d11243d56785dfbc6ff7d120f58e023588\": container with ID starting with 29aea3c141a06d60106b8924f521c5d11243d56785dfbc6ff7d120f58e023588 not found: ID does not exist" Dec 10 23:09:26 crc kubenswrapper[4791]: I1210 23:09:26.251407 4791 scope.go:117] "RemoveContainer" containerID="a6fb5e073d90b22471fee522d523321fa4497ebdf36b79f38fd1ee68b1a79815" Dec 10 23:09:26 crc kubenswrapper[4791]: E1210 23:09:26.252054 4791 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"a6fb5e073d90b22471fee522d523321fa4497ebdf36b79f38fd1ee68b1a79815\": container with ID starting with a6fb5e073d90b22471fee522d523321fa4497ebdf36b79f38fd1ee68b1a79815 not found: ID does not exist" containerID="a6fb5e073d90b22471fee522d523321fa4497ebdf36b79f38fd1ee68b1a79815" Dec 10 23:09:26 crc kubenswrapper[4791]: I1210 23:09:26.252087 4791 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"a6fb5e073d90b22471fee522d523321fa4497ebdf36b79f38fd1ee68b1a79815"} err="failed to get container status \"a6fb5e073d90b22471fee522d523321fa4497ebdf36b79f38fd1ee68b1a79815\": rpc error: code = NotFound desc = could not find container \"a6fb5e073d90b22471fee522d523321fa4497ebdf36b79f38fd1ee68b1a79815\": container with ID starting with a6fb5e073d90b22471fee522d523321fa4497ebdf36b79f38fd1ee68b1a79815 not found: ID does not exist" Dec 10 23:09:26 crc kubenswrapper[4791]: I1210 23:09:26.252110 4791 scope.go:117] "RemoveContainer" containerID="5ae0fb7182370bec008f1c24c5d245fd6cb390d075a883adda305056ffab03b5" Dec 10 23:09:26 crc kubenswrapper[4791]: E1210 23:09:26.252669 4791 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"5ae0fb7182370bec008f1c24c5d245fd6cb390d075a883adda305056ffab03b5\": container with ID starting with 5ae0fb7182370bec008f1c24c5d245fd6cb390d075a883adda305056ffab03b5 not found: ID does not exist" containerID="5ae0fb7182370bec008f1c24c5d245fd6cb390d075a883adda305056ffab03b5" Dec 10 23:09:26 crc kubenswrapper[4791]: I1210 23:09:26.252700 4791 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"5ae0fb7182370bec008f1c24c5d245fd6cb390d075a883adda305056ffab03b5"} err="failed to get container status \"5ae0fb7182370bec008f1c24c5d245fd6cb390d075a883adda305056ffab03b5\": rpc error: code = NotFound desc = could not find container \"5ae0fb7182370bec008f1c24c5d245fd6cb390d075a883adda305056ffab03b5\": container with ID starting with 5ae0fb7182370bec008f1c24c5d245fd6cb390d075a883adda305056ffab03b5 not found: ID does not exist" Dec 10 23:09:26 crc kubenswrapper[4791]: I1210 23:09:26.276430 4791 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/8eaa1b1d-715f-4059-ad18-fe9632c036a7-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "8eaa1b1d-715f-4059-ad18-fe9632c036a7" (UID: "8eaa1b1d-715f-4059-ad18-fe9632c036a7"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 10 23:09:26 crc kubenswrapper[4791]: I1210 23:09:26.279536 4791 reconciler_common.go:293] "Volume detached for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/8eaa1b1d-715f-4059-ad18-fe9632c036a7-sg-core-conf-yaml\") on node \"crc\" DevicePath \"\"" Dec 10 23:09:26 crc kubenswrapper[4791]: I1210 23:09:26.279608 4791 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/8eaa1b1d-715f-4059-ad18-fe9632c036a7-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 10 23:09:26 crc kubenswrapper[4791]: I1210 23:09:26.299582 4791 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/8eaa1b1d-715f-4059-ad18-fe9632c036a7-config-data" (OuterVolumeSpecName: "config-data") pod "8eaa1b1d-715f-4059-ad18-fe9632c036a7" (UID: "8eaa1b1d-715f-4059-ad18-fe9632c036a7"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 10 23:09:26 crc kubenswrapper[4791]: I1210 23:09:26.388840 4791 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/8eaa1b1d-715f-4059-ad18-fe9632c036a7-config-data\") on node \"crc\" DevicePath \"\"" Dec 10 23:09:26 crc kubenswrapper[4791]: I1210 23:09:26.485142 4791 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ceilometer-0"] Dec 10 23:09:26 crc kubenswrapper[4791]: I1210 23:09:26.502257 4791 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/ceilometer-0"] Dec 10 23:09:26 crc kubenswrapper[4791]: I1210 23:09:26.513236 4791 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ceilometer-0"] Dec 10 23:09:26 crc kubenswrapper[4791]: E1210 23:09:26.513656 4791 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="8eaa1b1d-715f-4059-ad18-fe9632c036a7" containerName="ceilometer-central-agent" Dec 10 23:09:26 crc kubenswrapper[4791]: I1210 23:09:26.513679 4791 state_mem.go:107] "Deleted CPUSet assignment" podUID="8eaa1b1d-715f-4059-ad18-fe9632c036a7" containerName="ceilometer-central-agent" Dec 10 23:09:26 crc kubenswrapper[4791]: E1210 23:09:26.513704 4791 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="8eaa1b1d-715f-4059-ad18-fe9632c036a7" containerName="proxy-httpd" Dec 10 23:09:26 crc kubenswrapper[4791]: I1210 23:09:26.513711 4791 state_mem.go:107] "Deleted CPUSet assignment" podUID="8eaa1b1d-715f-4059-ad18-fe9632c036a7" containerName="proxy-httpd" Dec 10 23:09:26 crc kubenswrapper[4791]: E1210 23:09:26.513722 4791 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="8eaa1b1d-715f-4059-ad18-fe9632c036a7" containerName="ceilometer-notification-agent" Dec 10 23:09:26 crc kubenswrapper[4791]: I1210 23:09:26.513727 4791 state_mem.go:107] "Deleted CPUSet assignment" podUID="8eaa1b1d-715f-4059-ad18-fe9632c036a7" containerName="ceilometer-notification-agent" Dec 10 23:09:26 crc kubenswrapper[4791]: E1210 23:09:26.513736 4791 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="8eaa1b1d-715f-4059-ad18-fe9632c036a7" containerName="sg-core" Dec 10 23:09:26 crc kubenswrapper[4791]: I1210 23:09:26.513742 4791 state_mem.go:107] "Deleted CPUSet assignment" podUID="8eaa1b1d-715f-4059-ad18-fe9632c036a7" containerName="sg-core" Dec 10 23:09:26 crc kubenswrapper[4791]: I1210 23:09:26.513934 4791 memory_manager.go:354] "RemoveStaleState removing state" podUID="8eaa1b1d-715f-4059-ad18-fe9632c036a7" containerName="sg-core" Dec 10 23:09:26 crc kubenswrapper[4791]: I1210 23:09:26.513951 4791 memory_manager.go:354] "RemoveStaleState removing state" podUID="8eaa1b1d-715f-4059-ad18-fe9632c036a7" containerName="proxy-httpd" Dec 10 23:09:26 crc kubenswrapper[4791]: I1210 23:09:26.513963 4791 memory_manager.go:354] "RemoveStaleState removing state" podUID="8eaa1b1d-715f-4059-ad18-fe9632c036a7" containerName="ceilometer-central-agent" Dec 10 23:09:26 crc kubenswrapper[4791]: I1210 23:09:26.513972 4791 memory_manager.go:354] "RemoveStaleState removing state" podUID="8eaa1b1d-715f-4059-ad18-fe9632c036a7" containerName="ceilometer-notification-agent" Dec 10 23:09:26 crc kubenswrapper[4791]: I1210 23:09:26.515627 4791 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Dec 10 23:09:26 crc kubenswrapper[4791]: I1210 23:09:26.518677 4791 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-config-data" Dec 10 23:09:26 crc kubenswrapper[4791]: I1210 23:09:26.519005 4791 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-scripts" Dec 10 23:09:26 crc kubenswrapper[4791]: I1210 23:09:26.527509 4791 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceilometer-0"] Dec 10 23:09:26 crc kubenswrapper[4791]: I1210 23:09:26.546013 4791 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/cinder-api-0" podUID="1a1ca37c-dab5-4748-b8c2-eb985e66a02f" containerName="cinder-api" probeResult="failure" output="Get \"http://10.217.0.159:8776/healthcheck\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" Dec 10 23:09:26 crc kubenswrapper[4791]: I1210 23:09:26.594718 4791 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/58167832-d8da-48de-8b8b-cf51fc54d92c-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"58167832-d8da-48de-8b8b-cf51fc54d92c\") " pod="openstack/ceilometer-0" Dec 10 23:09:26 crc kubenswrapper[4791]: I1210 23:09:26.594818 4791 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/58167832-d8da-48de-8b8b-cf51fc54d92c-config-data\") pod \"ceilometer-0\" (UID: \"58167832-d8da-48de-8b8b-cf51fc54d92c\") " pod="openstack/ceilometer-0" Dec 10 23:09:26 crc kubenswrapper[4791]: I1210 23:09:26.594905 4791 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/58167832-d8da-48de-8b8b-cf51fc54d92c-log-httpd\") pod \"ceilometer-0\" (UID: \"58167832-d8da-48de-8b8b-cf51fc54d92c\") " pod="openstack/ceilometer-0" Dec 10 23:09:26 crc kubenswrapper[4791]: I1210 23:09:26.594962 4791 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/58167832-d8da-48de-8b8b-cf51fc54d92c-scripts\") pod \"ceilometer-0\" (UID: \"58167832-d8da-48de-8b8b-cf51fc54d92c\") " pod="openstack/ceilometer-0" Dec 10 23:09:26 crc kubenswrapper[4791]: I1210 23:09:26.594988 4791 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-mxm4d\" (UniqueName: \"kubernetes.io/projected/58167832-d8da-48de-8b8b-cf51fc54d92c-kube-api-access-mxm4d\") pod \"ceilometer-0\" (UID: \"58167832-d8da-48de-8b8b-cf51fc54d92c\") " pod="openstack/ceilometer-0" Dec 10 23:09:26 crc kubenswrapper[4791]: I1210 23:09:26.595041 4791 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/58167832-d8da-48de-8b8b-cf51fc54d92c-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"58167832-d8da-48de-8b8b-cf51fc54d92c\") " pod="openstack/ceilometer-0" Dec 10 23:09:26 crc kubenswrapper[4791]: I1210 23:09:26.595147 4791 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/58167832-d8da-48de-8b8b-cf51fc54d92c-run-httpd\") pod \"ceilometer-0\" (UID: \"58167832-d8da-48de-8b8b-cf51fc54d92c\") " pod="openstack/ceilometer-0" Dec 10 23:09:26 crc kubenswrapper[4791]: I1210 23:09:26.696702 4791 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/58167832-d8da-48de-8b8b-cf51fc54d92c-run-httpd\") pod \"ceilometer-0\" (UID: \"58167832-d8da-48de-8b8b-cf51fc54d92c\") " pod="openstack/ceilometer-0" Dec 10 23:09:26 crc kubenswrapper[4791]: I1210 23:09:26.696814 4791 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/58167832-d8da-48de-8b8b-cf51fc54d92c-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"58167832-d8da-48de-8b8b-cf51fc54d92c\") " pod="openstack/ceilometer-0" Dec 10 23:09:26 crc kubenswrapper[4791]: I1210 23:09:26.696870 4791 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/58167832-d8da-48de-8b8b-cf51fc54d92c-config-data\") pod \"ceilometer-0\" (UID: \"58167832-d8da-48de-8b8b-cf51fc54d92c\") " pod="openstack/ceilometer-0" Dec 10 23:09:26 crc kubenswrapper[4791]: I1210 23:09:26.696900 4791 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/58167832-d8da-48de-8b8b-cf51fc54d92c-log-httpd\") pod \"ceilometer-0\" (UID: \"58167832-d8da-48de-8b8b-cf51fc54d92c\") " pod="openstack/ceilometer-0" Dec 10 23:09:26 crc kubenswrapper[4791]: I1210 23:09:26.696933 4791 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/58167832-d8da-48de-8b8b-cf51fc54d92c-scripts\") pod \"ceilometer-0\" (UID: \"58167832-d8da-48de-8b8b-cf51fc54d92c\") " pod="openstack/ceilometer-0" Dec 10 23:09:26 crc kubenswrapper[4791]: I1210 23:09:26.696958 4791 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-mxm4d\" (UniqueName: \"kubernetes.io/projected/58167832-d8da-48de-8b8b-cf51fc54d92c-kube-api-access-mxm4d\") pod \"ceilometer-0\" (UID: \"58167832-d8da-48de-8b8b-cf51fc54d92c\") " pod="openstack/ceilometer-0" Dec 10 23:09:26 crc kubenswrapper[4791]: I1210 23:09:26.696989 4791 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/58167832-d8da-48de-8b8b-cf51fc54d92c-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"58167832-d8da-48de-8b8b-cf51fc54d92c\") " pod="openstack/ceilometer-0" Dec 10 23:09:26 crc kubenswrapper[4791]: I1210 23:09:26.698737 4791 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/58167832-d8da-48de-8b8b-cf51fc54d92c-run-httpd\") pod \"ceilometer-0\" (UID: \"58167832-d8da-48de-8b8b-cf51fc54d92c\") " pod="openstack/ceilometer-0" Dec 10 23:09:26 crc kubenswrapper[4791]: I1210 23:09:26.698877 4791 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/58167832-d8da-48de-8b8b-cf51fc54d92c-log-httpd\") pod \"ceilometer-0\" (UID: \"58167832-d8da-48de-8b8b-cf51fc54d92c\") " pod="openstack/ceilometer-0" Dec 10 23:09:26 crc kubenswrapper[4791]: I1210 23:09:26.703121 4791 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/58167832-d8da-48de-8b8b-cf51fc54d92c-scripts\") pod \"ceilometer-0\" (UID: \"58167832-d8da-48de-8b8b-cf51fc54d92c\") " pod="openstack/ceilometer-0" Dec 10 23:09:26 crc kubenswrapper[4791]: I1210 23:09:26.704020 4791 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/58167832-d8da-48de-8b8b-cf51fc54d92c-config-data\") pod \"ceilometer-0\" (UID: \"58167832-d8da-48de-8b8b-cf51fc54d92c\") " pod="openstack/ceilometer-0" Dec 10 23:09:26 crc kubenswrapper[4791]: I1210 23:09:26.705373 4791 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/58167832-d8da-48de-8b8b-cf51fc54d92c-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"58167832-d8da-48de-8b8b-cf51fc54d92c\") " pod="openstack/ceilometer-0" Dec 10 23:09:26 crc kubenswrapper[4791]: I1210 23:09:26.718783 4791 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/58167832-d8da-48de-8b8b-cf51fc54d92c-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"58167832-d8da-48de-8b8b-cf51fc54d92c\") " pod="openstack/ceilometer-0" Dec 10 23:09:26 crc kubenswrapper[4791]: I1210 23:09:26.719366 4791 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-mxm4d\" (UniqueName: \"kubernetes.io/projected/58167832-d8da-48de-8b8b-cf51fc54d92c-kube-api-access-mxm4d\") pod \"ceilometer-0\" (UID: \"58167832-d8da-48de-8b8b-cf51fc54d92c\") " pod="openstack/ceilometer-0" Dec 10 23:09:26 crc kubenswrapper[4791]: I1210 23:09:26.853112 4791 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Dec 10 23:09:27 crc kubenswrapper[4791]: I1210 23:09:27.068121 4791 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-api-0" event={"ID":"f807e752-6cf5-4ad6-bc21-f22777b62c83","Type":"ContainerStarted","Data":"3016fc63d85b13db30fde16bf46b74a7506d9adedd039cd4c5e045a41f1d5078"} Dec 10 23:09:27 crc kubenswrapper[4791]: I1210 23:09:27.349084 4791 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceilometer-0"] Dec 10 23:09:27 crc kubenswrapper[4791]: I1210 23:09:27.753568 4791 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-scheduler-0" Dec 10 23:09:27 crc kubenswrapper[4791]: I1210 23:09:27.830724 4791 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-qc87z\" (UniqueName: \"kubernetes.io/projected/5e5e327e-b8dc-4bd6-af75-8aebc417ada9-kube-api-access-qc87z\") pod \"5e5e327e-b8dc-4bd6-af75-8aebc417ada9\" (UID: \"5e5e327e-b8dc-4bd6-af75-8aebc417ada9\") " Dec 10 23:09:27 crc kubenswrapper[4791]: I1210 23:09:27.831211 4791 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/5e5e327e-b8dc-4bd6-af75-8aebc417ada9-config-data\") pod \"5e5e327e-b8dc-4bd6-af75-8aebc417ada9\" (UID: \"5e5e327e-b8dc-4bd6-af75-8aebc417ada9\") " Dec 10 23:09:27 crc kubenswrapper[4791]: I1210 23:09:27.831272 4791 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/5e5e327e-b8dc-4bd6-af75-8aebc417ada9-combined-ca-bundle\") pod \"5e5e327e-b8dc-4bd6-af75-8aebc417ada9\" (UID: \"5e5e327e-b8dc-4bd6-af75-8aebc417ada9\") " Dec 10 23:09:27 crc kubenswrapper[4791]: I1210 23:09:27.831314 4791 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/5e5e327e-b8dc-4bd6-af75-8aebc417ada9-config-data-custom\") pod \"5e5e327e-b8dc-4bd6-af75-8aebc417ada9\" (UID: \"5e5e327e-b8dc-4bd6-af75-8aebc417ada9\") " Dec 10 23:09:27 crc kubenswrapper[4791]: I1210 23:09:27.831422 4791 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/5e5e327e-b8dc-4bd6-af75-8aebc417ada9-scripts\") pod \"5e5e327e-b8dc-4bd6-af75-8aebc417ada9\" (UID: \"5e5e327e-b8dc-4bd6-af75-8aebc417ada9\") " Dec 10 23:09:27 crc kubenswrapper[4791]: I1210 23:09:27.831525 4791 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/5e5e327e-b8dc-4bd6-af75-8aebc417ada9-etc-machine-id\") pod \"5e5e327e-b8dc-4bd6-af75-8aebc417ada9\" (UID: \"5e5e327e-b8dc-4bd6-af75-8aebc417ada9\") " Dec 10 23:09:27 crc kubenswrapper[4791]: I1210 23:09:27.832020 4791 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/5e5e327e-b8dc-4bd6-af75-8aebc417ada9-etc-machine-id" (OuterVolumeSpecName: "etc-machine-id") pod "5e5e327e-b8dc-4bd6-af75-8aebc417ada9" (UID: "5e5e327e-b8dc-4bd6-af75-8aebc417ada9"). InnerVolumeSpecName "etc-machine-id". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 10 23:09:27 crc kubenswrapper[4791]: I1210 23:09:27.848986 4791 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/5e5e327e-b8dc-4bd6-af75-8aebc417ada9-config-data-custom" (OuterVolumeSpecName: "config-data-custom") pod "5e5e327e-b8dc-4bd6-af75-8aebc417ada9" (UID: "5e5e327e-b8dc-4bd6-af75-8aebc417ada9"). InnerVolumeSpecName "config-data-custom". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 10 23:09:27 crc kubenswrapper[4791]: I1210 23:09:27.849111 4791 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/5e5e327e-b8dc-4bd6-af75-8aebc417ada9-kube-api-access-qc87z" (OuterVolumeSpecName: "kube-api-access-qc87z") pod "5e5e327e-b8dc-4bd6-af75-8aebc417ada9" (UID: "5e5e327e-b8dc-4bd6-af75-8aebc417ada9"). InnerVolumeSpecName "kube-api-access-qc87z". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 10 23:09:27 crc kubenswrapper[4791]: I1210 23:09:27.853078 4791 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/5e5e327e-b8dc-4bd6-af75-8aebc417ada9-scripts" (OuterVolumeSpecName: "scripts") pod "5e5e327e-b8dc-4bd6-af75-8aebc417ada9" (UID: "5e5e327e-b8dc-4bd6-af75-8aebc417ada9"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 10 23:09:27 crc kubenswrapper[4791]: I1210 23:09:27.900026 4791 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="8eaa1b1d-715f-4059-ad18-fe9632c036a7" path="/var/lib/kubelet/pods/8eaa1b1d-715f-4059-ad18-fe9632c036a7/volumes" Dec 10 23:09:27 crc kubenswrapper[4791]: I1210 23:09:27.921702 4791 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/5e5e327e-b8dc-4bd6-af75-8aebc417ada9-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "5e5e327e-b8dc-4bd6-af75-8aebc417ada9" (UID: "5e5e327e-b8dc-4bd6-af75-8aebc417ada9"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 10 23:09:27 crc kubenswrapper[4791]: I1210 23:09:27.933901 4791 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-qc87z\" (UniqueName: \"kubernetes.io/projected/5e5e327e-b8dc-4bd6-af75-8aebc417ada9-kube-api-access-qc87z\") on node \"crc\" DevicePath \"\"" Dec 10 23:09:27 crc kubenswrapper[4791]: I1210 23:09:27.933931 4791 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/5e5e327e-b8dc-4bd6-af75-8aebc417ada9-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 10 23:09:27 crc kubenswrapper[4791]: I1210 23:09:27.933945 4791 reconciler_common.go:293] "Volume detached for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/5e5e327e-b8dc-4bd6-af75-8aebc417ada9-config-data-custom\") on node \"crc\" DevicePath \"\"" Dec 10 23:09:27 crc kubenswrapper[4791]: I1210 23:09:27.933953 4791 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/5e5e327e-b8dc-4bd6-af75-8aebc417ada9-scripts\") on node \"crc\" DevicePath \"\"" Dec 10 23:09:27 crc kubenswrapper[4791]: I1210 23:09:27.934101 4791 reconciler_common.go:293] "Volume detached for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/5e5e327e-b8dc-4bd6-af75-8aebc417ada9-etc-machine-id\") on node \"crc\" DevicePath \"\"" Dec 10 23:09:27 crc kubenswrapper[4791]: I1210 23:09:27.990167 4791 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/5e5e327e-b8dc-4bd6-af75-8aebc417ada9-config-data" (OuterVolumeSpecName: "config-data") pod "5e5e327e-b8dc-4bd6-af75-8aebc417ada9" (UID: "5e5e327e-b8dc-4bd6-af75-8aebc417ada9"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 10 23:09:28 crc kubenswrapper[4791]: I1210 23:09:28.037331 4791 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/5e5e327e-b8dc-4bd6-af75-8aebc417ada9-config-data\") on node \"crc\" DevicePath \"\"" Dec 10 23:09:28 crc kubenswrapper[4791]: I1210 23:09:28.098576 4791 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-api-0" event={"ID":"f807e752-6cf5-4ad6-bc21-f22777b62c83","Type":"ContainerStarted","Data":"70d0063bec1e5c33965680fdb7d6445cc3220a4d2d0194dce41bd2b87ae35dea"} Dec 10 23:09:28 crc kubenswrapper[4791]: I1210 23:09:28.106531 4791 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"58167832-d8da-48de-8b8b-cf51fc54d92c","Type":"ContainerStarted","Data":"ca930728dc321c5022ce2263f6f361be1de64fd0569818e4c3e7dca6b76ec3df"} Dec 10 23:09:28 crc kubenswrapper[4791]: I1210 23:09:28.110258 4791 generic.go:334] "Generic (PLEG): container finished" podID="5e5e327e-b8dc-4bd6-af75-8aebc417ada9" containerID="468b836fc75ba6c8ac143ddca9787ab03833f45994d432d9850ebe13fc2251ed" exitCode=0 Dec 10 23:09:28 crc kubenswrapper[4791]: I1210 23:09:28.110312 4791 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-scheduler-0" event={"ID":"5e5e327e-b8dc-4bd6-af75-8aebc417ada9","Type":"ContainerDied","Data":"468b836fc75ba6c8ac143ddca9787ab03833f45994d432d9850ebe13fc2251ed"} Dec 10 23:09:28 crc kubenswrapper[4791]: I1210 23:09:28.110291 4791 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-scheduler-0" Dec 10 23:09:28 crc kubenswrapper[4791]: I1210 23:09:28.110385 4791 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-scheduler-0" event={"ID":"5e5e327e-b8dc-4bd6-af75-8aebc417ada9","Type":"ContainerDied","Data":"a3778b6f3bbf0ca69bdffd7ffe50f1e49fcc48b28bed8e76f5376905253abefe"} Dec 10 23:09:28 crc kubenswrapper[4791]: I1210 23:09:28.110409 4791 scope.go:117] "RemoveContainer" containerID="0491f982573b4cf5b921d4eb64aecef9ffb9a9bed5a7cd16d5e70ffa514274f4" Dec 10 23:09:28 crc kubenswrapper[4791]: I1210 23:09:28.140759 4791 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/cinder-api-0" podStartSLOduration=5.140738988 podStartE2EDuration="5.140738988s" podCreationTimestamp="2025-12-10 23:09:23 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-10 23:09:28.12072385 +0000 UTC m=+1202.550341453" watchObservedRunningTime="2025-12-10 23:09:28.140738988 +0000 UTC m=+1202.570356601" Dec 10 23:09:28 crc kubenswrapper[4791]: I1210 23:09:28.146036 4791 scope.go:117] "RemoveContainer" containerID="468b836fc75ba6c8ac143ddca9787ab03833f45994d432d9850ebe13fc2251ed" Dec 10 23:09:28 crc kubenswrapper[4791]: I1210 23:09:28.167570 4791 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/cinder-scheduler-0"] Dec 10 23:09:28 crc kubenswrapper[4791]: I1210 23:09:28.175688 4791 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/cinder-scheduler-0"] Dec 10 23:09:28 crc kubenswrapper[4791]: I1210 23:09:28.180916 4791 scope.go:117] "RemoveContainer" containerID="0491f982573b4cf5b921d4eb64aecef9ffb9a9bed5a7cd16d5e70ffa514274f4" Dec 10 23:09:28 crc kubenswrapper[4791]: E1210 23:09:28.181468 4791 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"0491f982573b4cf5b921d4eb64aecef9ffb9a9bed5a7cd16d5e70ffa514274f4\": container with ID starting with 0491f982573b4cf5b921d4eb64aecef9ffb9a9bed5a7cd16d5e70ffa514274f4 not found: ID does not exist" containerID="0491f982573b4cf5b921d4eb64aecef9ffb9a9bed5a7cd16d5e70ffa514274f4" Dec 10 23:09:28 crc kubenswrapper[4791]: I1210 23:09:28.181498 4791 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"0491f982573b4cf5b921d4eb64aecef9ffb9a9bed5a7cd16d5e70ffa514274f4"} err="failed to get container status \"0491f982573b4cf5b921d4eb64aecef9ffb9a9bed5a7cd16d5e70ffa514274f4\": rpc error: code = NotFound desc = could not find container \"0491f982573b4cf5b921d4eb64aecef9ffb9a9bed5a7cd16d5e70ffa514274f4\": container with ID starting with 0491f982573b4cf5b921d4eb64aecef9ffb9a9bed5a7cd16d5e70ffa514274f4 not found: ID does not exist" Dec 10 23:09:28 crc kubenswrapper[4791]: I1210 23:09:28.181528 4791 scope.go:117] "RemoveContainer" containerID="468b836fc75ba6c8ac143ddca9787ab03833f45994d432d9850ebe13fc2251ed" Dec 10 23:09:28 crc kubenswrapper[4791]: E1210 23:09:28.181884 4791 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"468b836fc75ba6c8ac143ddca9787ab03833f45994d432d9850ebe13fc2251ed\": container with ID starting with 468b836fc75ba6c8ac143ddca9787ab03833f45994d432d9850ebe13fc2251ed not found: ID does not exist" containerID="468b836fc75ba6c8ac143ddca9787ab03833f45994d432d9850ebe13fc2251ed" Dec 10 23:09:28 crc kubenswrapper[4791]: I1210 23:09:28.181907 4791 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"468b836fc75ba6c8ac143ddca9787ab03833f45994d432d9850ebe13fc2251ed"} err="failed to get container status \"468b836fc75ba6c8ac143ddca9787ab03833f45994d432d9850ebe13fc2251ed\": rpc error: code = NotFound desc = could not find container \"468b836fc75ba6c8ac143ddca9787ab03833f45994d432d9850ebe13fc2251ed\": container with ID starting with 468b836fc75ba6c8ac143ddca9787ab03833f45994d432d9850ebe13fc2251ed not found: ID does not exist" Dec 10 23:09:28 crc kubenswrapper[4791]: I1210 23:09:28.200399 4791 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/cinder-scheduler-0"] Dec 10 23:09:28 crc kubenswrapper[4791]: E1210 23:09:28.200832 4791 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="5e5e327e-b8dc-4bd6-af75-8aebc417ada9" containerName="probe" Dec 10 23:09:28 crc kubenswrapper[4791]: I1210 23:09:28.200851 4791 state_mem.go:107] "Deleted CPUSet assignment" podUID="5e5e327e-b8dc-4bd6-af75-8aebc417ada9" containerName="probe" Dec 10 23:09:28 crc kubenswrapper[4791]: E1210 23:09:28.200876 4791 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="5e5e327e-b8dc-4bd6-af75-8aebc417ada9" containerName="cinder-scheduler" Dec 10 23:09:28 crc kubenswrapper[4791]: I1210 23:09:28.200882 4791 state_mem.go:107] "Deleted CPUSet assignment" podUID="5e5e327e-b8dc-4bd6-af75-8aebc417ada9" containerName="cinder-scheduler" Dec 10 23:09:28 crc kubenswrapper[4791]: I1210 23:09:28.201058 4791 memory_manager.go:354] "RemoveStaleState removing state" podUID="5e5e327e-b8dc-4bd6-af75-8aebc417ada9" containerName="probe" Dec 10 23:09:28 crc kubenswrapper[4791]: I1210 23:09:28.201080 4791 memory_manager.go:354] "RemoveStaleState removing state" podUID="5e5e327e-b8dc-4bd6-af75-8aebc417ada9" containerName="cinder-scheduler" Dec 10 23:09:28 crc kubenswrapper[4791]: I1210 23:09:28.202079 4791 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-scheduler-0" Dec 10 23:09:28 crc kubenswrapper[4791]: I1210 23:09:28.208585 4791 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cinder-scheduler-config-data" Dec 10 23:09:28 crc kubenswrapper[4791]: I1210 23:09:28.208790 4791 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-scheduler-0"] Dec 10 23:09:28 crc kubenswrapper[4791]: I1210 23:09:28.342533 4791 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-78992\" (UniqueName: \"kubernetes.io/projected/e703a39e-cce5-4b36-b017-611836d0fa18-kube-api-access-78992\") pod \"cinder-scheduler-0\" (UID: \"e703a39e-cce5-4b36-b017-611836d0fa18\") " pod="openstack/cinder-scheduler-0" Dec 10 23:09:28 crc kubenswrapper[4791]: I1210 23:09:28.342601 4791 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/e703a39e-cce5-4b36-b017-611836d0fa18-config-data-custom\") pod \"cinder-scheduler-0\" (UID: \"e703a39e-cce5-4b36-b017-611836d0fa18\") " pod="openstack/cinder-scheduler-0" Dec 10 23:09:28 crc kubenswrapper[4791]: I1210 23:09:28.342916 4791 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/e703a39e-cce5-4b36-b017-611836d0fa18-etc-machine-id\") pod \"cinder-scheduler-0\" (UID: \"e703a39e-cce5-4b36-b017-611836d0fa18\") " pod="openstack/cinder-scheduler-0" Dec 10 23:09:28 crc kubenswrapper[4791]: I1210 23:09:28.342998 4791 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/e703a39e-cce5-4b36-b017-611836d0fa18-scripts\") pod \"cinder-scheduler-0\" (UID: \"e703a39e-cce5-4b36-b017-611836d0fa18\") " pod="openstack/cinder-scheduler-0" Dec 10 23:09:28 crc kubenswrapper[4791]: I1210 23:09:28.343041 4791 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/e703a39e-cce5-4b36-b017-611836d0fa18-combined-ca-bundle\") pod \"cinder-scheduler-0\" (UID: \"e703a39e-cce5-4b36-b017-611836d0fa18\") " pod="openstack/cinder-scheduler-0" Dec 10 23:09:28 crc kubenswrapper[4791]: I1210 23:09:28.343274 4791 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/e703a39e-cce5-4b36-b017-611836d0fa18-config-data\") pod \"cinder-scheduler-0\" (UID: \"e703a39e-cce5-4b36-b017-611836d0fa18\") " pod="openstack/cinder-scheduler-0" Dec 10 23:09:28 crc kubenswrapper[4791]: I1210 23:09:28.446519 4791 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/e703a39e-cce5-4b36-b017-611836d0fa18-config-data\") pod \"cinder-scheduler-0\" (UID: \"e703a39e-cce5-4b36-b017-611836d0fa18\") " pod="openstack/cinder-scheduler-0" Dec 10 23:09:28 crc kubenswrapper[4791]: I1210 23:09:28.446613 4791 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-78992\" (UniqueName: \"kubernetes.io/projected/e703a39e-cce5-4b36-b017-611836d0fa18-kube-api-access-78992\") pod \"cinder-scheduler-0\" (UID: \"e703a39e-cce5-4b36-b017-611836d0fa18\") " pod="openstack/cinder-scheduler-0" Dec 10 23:09:28 crc kubenswrapper[4791]: I1210 23:09:28.446646 4791 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/e703a39e-cce5-4b36-b017-611836d0fa18-config-data-custom\") pod \"cinder-scheduler-0\" (UID: \"e703a39e-cce5-4b36-b017-611836d0fa18\") " pod="openstack/cinder-scheduler-0" Dec 10 23:09:28 crc kubenswrapper[4791]: I1210 23:09:28.446802 4791 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/e703a39e-cce5-4b36-b017-611836d0fa18-etc-machine-id\") pod \"cinder-scheduler-0\" (UID: \"e703a39e-cce5-4b36-b017-611836d0fa18\") " pod="openstack/cinder-scheduler-0" Dec 10 23:09:28 crc kubenswrapper[4791]: I1210 23:09:28.446875 4791 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/e703a39e-cce5-4b36-b017-611836d0fa18-scripts\") pod \"cinder-scheduler-0\" (UID: \"e703a39e-cce5-4b36-b017-611836d0fa18\") " pod="openstack/cinder-scheduler-0" Dec 10 23:09:28 crc kubenswrapper[4791]: I1210 23:09:28.446921 4791 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/e703a39e-cce5-4b36-b017-611836d0fa18-combined-ca-bundle\") pod \"cinder-scheduler-0\" (UID: \"e703a39e-cce5-4b36-b017-611836d0fa18\") " pod="openstack/cinder-scheduler-0" Dec 10 23:09:28 crc kubenswrapper[4791]: I1210 23:09:28.447043 4791 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/e703a39e-cce5-4b36-b017-611836d0fa18-etc-machine-id\") pod \"cinder-scheduler-0\" (UID: \"e703a39e-cce5-4b36-b017-611836d0fa18\") " pod="openstack/cinder-scheduler-0" Dec 10 23:09:28 crc kubenswrapper[4791]: I1210 23:09:28.453879 4791 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/e703a39e-cce5-4b36-b017-611836d0fa18-scripts\") pod \"cinder-scheduler-0\" (UID: \"e703a39e-cce5-4b36-b017-611836d0fa18\") " pod="openstack/cinder-scheduler-0" Dec 10 23:09:28 crc kubenswrapper[4791]: I1210 23:09:28.454702 4791 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/e703a39e-cce5-4b36-b017-611836d0fa18-combined-ca-bundle\") pod \"cinder-scheduler-0\" (UID: \"e703a39e-cce5-4b36-b017-611836d0fa18\") " pod="openstack/cinder-scheduler-0" Dec 10 23:09:28 crc kubenswrapper[4791]: I1210 23:09:28.455734 4791 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/e703a39e-cce5-4b36-b017-611836d0fa18-config-data\") pod \"cinder-scheduler-0\" (UID: \"e703a39e-cce5-4b36-b017-611836d0fa18\") " pod="openstack/cinder-scheduler-0" Dec 10 23:09:28 crc kubenswrapper[4791]: I1210 23:09:28.468866 4791 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-78992\" (UniqueName: \"kubernetes.io/projected/e703a39e-cce5-4b36-b017-611836d0fa18-kube-api-access-78992\") pod \"cinder-scheduler-0\" (UID: \"e703a39e-cce5-4b36-b017-611836d0fa18\") " pod="openstack/cinder-scheduler-0" Dec 10 23:09:28 crc kubenswrapper[4791]: I1210 23:09:28.475688 4791 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/e703a39e-cce5-4b36-b017-611836d0fa18-config-data-custom\") pod \"cinder-scheduler-0\" (UID: \"e703a39e-cce5-4b36-b017-611836d0fa18\") " pod="openstack/cinder-scheduler-0" Dec 10 23:09:28 crc kubenswrapper[4791]: I1210 23:09:28.543668 4791 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-scheduler-0" Dec 10 23:09:28 crc kubenswrapper[4791]: I1210 23:09:28.920216 4791 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/cinder-api-0" Dec 10 23:09:29 crc kubenswrapper[4791]: I1210 23:09:29.046634 4791 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-scheduler-0"] Dec 10 23:09:29 crc kubenswrapper[4791]: I1210 23:09:29.129749 4791 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"58167832-d8da-48de-8b8b-cf51fc54d92c","Type":"ContainerStarted","Data":"b32ac48681950cbc60cb9f131ea5fd50ebfc4c397b7e8f40d8325d077bbfcbcf"} Dec 10 23:09:29 crc kubenswrapper[4791]: I1210 23:09:29.135597 4791 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-scheduler-0" event={"ID":"e703a39e-cce5-4b36-b017-611836d0fa18","Type":"ContainerStarted","Data":"96f9334239ffd7942c6c55aaec43032eb33df5e3641ef94e2ccf15f27d2651ce"} Dec 10 23:09:29 crc kubenswrapper[4791]: I1210 23:09:29.897073 4791 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="5e5e327e-b8dc-4bd6-af75-8aebc417ada9" path="/var/lib/kubelet/pods/5e5e327e-b8dc-4bd6-af75-8aebc417ada9/volumes" Dec 10 23:09:29 crc kubenswrapper[4791]: I1210 23:09:29.968699 4791 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-85794c6bd6-ht8sz" Dec 10 23:09:30 crc kubenswrapper[4791]: I1210 23:09:30.090195 4791 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"httpd-config\" (UniqueName: \"kubernetes.io/secret/75eb971a-04fc-40f5-9890-d5a0b4daccd1-httpd-config\") pod \"75eb971a-04fc-40f5-9890-d5a0b4daccd1\" (UID: \"75eb971a-04fc-40f5-9890-d5a0b4daccd1\") " Dec 10 23:09:30 crc kubenswrapper[4791]: I1210 23:09:30.090291 4791 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/secret/75eb971a-04fc-40f5-9890-d5a0b4daccd1-config\") pod \"75eb971a-04fc-40f5-9890-d5a0b4daccd1\" (UID: \"75eb971a-04fc-40f5-9890-d5a0b4daccd1\") " Dec 10 23:09:30 crc kubenswrapper[4791]: I1210 23:09:30.090397 4791 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/75eb971a-04fc-40f5-9890-d5a0b4daccd1-combined-ca-bundle\") pod \"75eb971a-04fc-40f5-9890-d5a0b4daccd1\" (UID: \"75eb971a-04fc-40f5-9890-d5a0b4daccd1\") " Dec 10 23:09:30 crc kubenswrapper[4791]: I1210 23:09:30.090482 4791 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovndb-tls-certs\" (UniqueName: \"kubernetes.io/secret/75eb971a-04fc-40f5-9890-d5a0b4daccd1-ovndb-tls-certs\") pod \"75eb971a-04fc-40f5-9890-d5a0b4daccd1\" (UID: \"75eb971a-04fc-40f5-9890-d5a0b4daccd1\") " Dec 10 23:09:30 crc kubenswrapper[4791]: I1210 23:09:30.090538 4791 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-9cgk8\" (UniqueName: \"kubernetes.io/projected/75eb971a-04fc-40f5-9890-d5a0b4daccd1-kube-api-access-9cgk8\") pod \"75eb971a-04fc-40f5-9890-d5a0b4daccd1\" (UID: \"75eb971a-04fc-40f5-9890-d5a0b4daccd1\") " Dec 10 23:09:30 crc kubenswrapper[4791]: I1210 23:09:30.095876 4791 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/75eb971a-04fc-40f5-9890-d5a0b4daccd1-kube-api-access-9cgk8" (OuterVolumeSpecName: "kube-api-access-9cgk8") pod "75eb971a-04fc-40f5-9890-d5a0b4daccd1" (UID: "75eb971a-04fc-40f5-9890-d5a0b4daccd1"). InnerVolumeSpecName "kube-api-access-9cgk8". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 10 23:09:30 crc kubenswrapper[4791]: I1210 23:09:30.106937 4791 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/75eb971a-04fc-40f5-9890-d5a0b4daccd1-httpd-config" (OuterVolumeSpecName: "httpd-config") pod "75eb971a-04fc-40f5-9890-d5a0b4daccd1" (UID: "75eb971a-04fc-40f5-9890-d5a0b4daccd1"). InnerVolumeSpecName "httpd-config". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 10 23:09:30 crc kubenswrapper[4791]: I1210 23:09:30.156763 4791 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/75eb971a-04fc-40f5-9890-d5a0b4daccd1-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "75eb971a-04fc-40f5-9890-d5a0b4daccd1" (UID: "75eb971a-04fc-40f5-9890-d5a0b4daccd1"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 10 23:09:30 crc kubenswrapper[4791]: I1210 23:09:30.159350 4791 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/75eb971a-04fc-40f5-9890-d5a0b4daccd1-config" (OuterVolumeSpecName: "config") pod "75eb971a-04fc-40f5-9890-d5a0b4daccd1" (UID: "75eb971a-04fc-40f5-9890-d5a0b4daccd1"). InnerVolumeSpecName "config". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 10 23:09:30 crc kubenswrapper[4791]: I1210 23:09:30.163562 4791 generic.go:334] "Generic (PLEG): container finished" podID="75eb971a-04fc-40f5-9890-d5a0b4daccd1" containerID="dd2fb7dddb229ce25296aefb965623011367c1640da7f68c5f11bd17f39afbdc" exitCode=0 Dec 10 23:09:30 crc kubenswrapper[4791]: I1210 23:09:30.163793 4791 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-85794c6bd6-ht8sz" Dec 10 23:09:30 crc kubenswrapper[4791]: I1210 23:09:30.164186 4791 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-85794c6bd6-ht8sz" event={"ID":"75eb971a-04fc-40f5-9890-d5a0b4daccd1","Type":"ContainerDied","Data":"dd2fb7dddb229ce25296aefb965623011367c1640da7f68c5f11bd17f39afbdc"} Dec 10 23:09:30 crc kubenswrapper[4791]: I1210 23:09:30.164248 4791 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-85794c6bd6-ht8sz" event={"ID":"75eb971a-04fc-40f5-9890-d5a0b4daccd1","Type":"ContainerDied","Data":"dd2f20735a4b4cd7f75dfe9937d9cb51e2dc61038563f55ba640cae578951ee2"} Dec 10 23:09:30 crc kubenswrapper[4791]: I1210 23:09:30.164274 4791 scope.go:117] "RemoveContainer" containerID="a1c20ac32bb42bda941e83eca0be34d9e4905b4b8eff8dab079f24798af60c6b" Dec 10 23:09:30 crc kubenswrapper[4791]: I1210 23:09:30.169636 4791 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-scheduler-0" event={"ID":"e703a39e-cce5-4b36-b017-611836d0fa18","Type":"ContainerStarted","Data":"fd16ab1ba11ab626d213f4f354e10b3f457c2de87cb627d8f5019978fd73f160"} Dec 10 23:09:30 crc kubenswrapper[4791]: I1210 23:09:30.174412 4791 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/75eb971a-04fc-40f5-9890-d5a0b4daccd1-ovndb-tls-certs" (OuterVolumeSpecName: "ovndb-tls-certs") pod "75eb971a-04fc-40f5-9890-d5a0b4daccd1" (UID: "75eb971a-04fc-40f5-9890-d5a0b4daccd1"). InnerVolumeSpecName "ovndb-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 10 23:09:30 crc kubenswrapper[4791]: I1210 23:09:30.194255 4791 reconciler_common.go:293] "Volume detached for volume \"httpd-config\" (UniqueName: \"kubernetes.io/secret/75eb971a-04fc-40f5-9890-d5a0b4daccd1-httpd-config\") on node \"crc\" DevicePath \"\"" Dec 10 23:09:30 crc kubenswrapper[4791]: I1210 23:09:30.194434 4791 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/secret/75eb971a-04fc-40f5-9890-d5a0b4daccd1-config\") on node \"crc\" DevicePath \"\"" Dec 10 23:09:30 crc kubenswrapper[4791]: I1210 23:09:30.194455 4791 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/75eb971a-04fc-40f5-9890-d5a0b4daccd1-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 10 23:09:30 crc kubenswrapper[4791]: I1210 23:09:30.194528 4791 reconciler_common.go:293] "Volume detached for volume \"ovndb-tls-certs\" (UniqueName: \"kubernetes.io/secret/75eb971a-04fc-40f5-9890-d5a0b4daccd1-ovndb-tls-certs\") on node \"crc\" DevicePath \"\"" Dec 10 23:09:30 crc kubenswrapper[4791]: I1210 23:09:30.194542 4791 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-9cgk8\" (UniqueName: \"kubernetes.io/projected/75eb971a-04fc-40f5-9890-d5a0b4daccd1-kube-api-access-9cgk8\") on node \"crc\" DevicePath \"\"" Dec 10 23:09:30 crc kubenswrapper[4791]: I1210 23:09:30.194636 4791 scope.go:117] "RemoveContainer" containerID="dd2fb7dddb229ce25296aefb965623011367c1640da7f68c5f11bd17f39afbdc" Dec 10 23:09:30 crc kubenswrapper[4791]: I1210 23:09:30.219557 4791 scope.go:117] "RemoveContainer" containerID="a1c20ac32bb42bda941e83eca0be34d9e4905b4b8eff8dab079f24798af60c6b" Dec 10 23:09:30 crc kubenswrapper[4791]: E1210 23:09:30.220218 4791 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"a1c20ac32bb42bda941e83eca0be34d9e4905b4b8eff8dab079f24798af60c6b\": container with ID starting with a1c20ac32bb42bda941e83eca0be34d9e4905b4b8eff8dab079f24798af60c6b not found: ID does not exist" containerID="a1c20ac32bb42bda941e83eca0be34d9e4905b4b8eff8dab079f24798af60c6b" Dec 10 23:09:30 crc kubenswrapper[4791]: I1210 23:09:30.220284 4791 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"a1c20ac32bb42bda941e83eca0be34d9e4905b4b8eff8dab079f24798af60c6b"} err="failed to get container status \"a1c20ac32bb42bda941e83eca0be34d9e4905b4b8eff8dab079f24798af60c6b\": rpc error: code = NotFound desc = could not find container \"a1c20ac32bb42bda941e83eca0be34d9e4905b4b8eff8dab079f24798af60c6b\": container with ID starting with a1c20ac32bb42bda941e83eca0be34d9e4905b4b8eff8dab079f24798af60c6b not found: ID does not exist" Dec 10 23:09:30 crc kubenswrapper[4791]: I1210 23:09:30.220325 4791 scope.go:117] "RemoveContainer" containerID="dd2fb7dddb229ce25296aefb965623011367c1640da7f68c5f11bd17f39afbdc" Dec 10 23:09:30 crc kubenswrapper[4791]: E1210 23:09:30.220820 4791 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"dd2fb7dddb229ce25296aefb965623011367c1640da7f68c5f11bd17f39afbdc\": container with ID starting with dd2fb7dddb229ce25296aefb965623011367c1640da7f68c5f11bd17f39afbdc not found: ID does not exist" containerID="dd2fb7dddb229ce25296aefb965623011367c1640da7f68c5f11bd17f39afbdc" Dec 10 23:09:30 crc kubenswrapper[4791]: I1210 23:09:30.220863 4791 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"dd2fb7dddb229ce25296aefb965623011367c1640da7f68c5f11bd17f39afbdc"} err="failed to get container status \"dd2fb7dddb229ce25296aefb965623011367c1640da7f68c5f11bd17f39afbdc\": rpc error: code = NotFound desc = could not find container \"dd2fb7dddb229ce25296aefb965623011367c1640da7f68c5f11bd17f39afbdc\": container with ID starting with dd2fb7dddb229ce25296aefb965623011367c1640da7f68c5f11bd17f39afbdc not found: ID does not exist" Dec 10 23:09:30 crc kubenswrapper[4791]: I1210 23:09:30.509921 4791 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/neutron-85794c6bd6-ht8sz"] Dec 10 23:09:30 crc kubenswrapper[4791]: I1210 23:09:30.517062 4791 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/neutron-85794c6bd6-ht8sz"] Dec 10 23:09:30 crc kubenswrapper[4791]: I1210 23:09:30.774509 4791 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/dnsmasq-dns-5784cf869f-ngjqj" Dec 10 23:09:30 crc kubenswrapper[4791]: I1210 23:09:30.873959 4791 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-84b966f6c9-tz266"] Dec 10 23:09:30 crc kubenswrapper[4791]: I1210 23:09:30.874233 4791 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/dnsmasq-dns-84b966f6c9-tz266" podUID="3a4fcb67-0b78-43d6-a6a1-8d319bf19184" containerName="dnsmasq-dns" containerID="cri-o://4294aa9ba3a7a9440f41cfc9d38f97e9bcc1fa80423767e0df5c6d4016638bf9" gracePeriod=10 Dec 10 23:09:31 crc kubenswrapper[4791]: I1210 23:09:31.206659 4791 generic.go:334] "Generic (PLEG): container finished" podID="3a4fcb67-0b78-43d6-a6a1-8d319bf19184" containerID="4294aa9ba3a7a9440f41cfc9d38f97e9bcc1fa80423767e0df5c6d4016638bf9" exitCode=0 Dec 10 23:09:31 crc kubenswrapper[4791]: I1210 23:09:31.207099 4791 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-84b966f6c9-tz266" event={"ID":"3a4fcb67-0b78-43d6-a6a1-8d319bf19184","Type":"ContainerDied","Data":"4294aa9ba3a7a9440f41cfc9d38f97e9bcc1fa80423767e0df5c6d4016638bf9"} Dec 10 23:09:31 crc kubenswrapper[4791]: I1210 23:09:31.221902 4791 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"58167832-d8da-48de-8b8b-cf51fc54d92c","Type":"ContainerStarted","Data":"b8f078f30aed9ddb6d5815433c894f5ac4a00e3c7489055d52e0cb4879bb16cf"} Dec 10 23:09:31 crc kubenswrapper[4791]: I1210 23:09:31.224473 4791 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-scheduler-0" event={"ID":"e703a39e-cce5-4b36-b017-611836d0fa18","Type":"ContainerStarted","Data":"327fded2077d4cecfd014d59cb0a9300de0b8de61b37515d2c8a2e65dc877c36"} Dec 10 23:09:31 crc kubenswrapper[4791]: I1210 23:09:31.253548 4791 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/cinder-scheduler-0" podStartSLOduration=3.253528526 podStartE2EDuration="3.253528526s" podCreationTimestamp="2025-12-10 23:09:28 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-10 23:09:31.249714196 +0000 UTC m=+1205.679331829" watchObservedRunningTime="2025-12-10 23:09:31.253528526 +0000 UTC m=+1205.683146139" Dec 10 23:09:31 crc kubenswrapper[4791]: I1210 23:09:31.571448 4791 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-84b966f6c9-tz266" Dec 10 23:09:31 crc kubenswrapper[4791]: I1210 23:09:31.732231 4791 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/3a4fcb67-0b78-43d6-a6a1-8d319bf19184-ovsdbserver-sb\") pod \"3a4fcb67-0b78-43d6-a6a1-8d319bf19184\" (UID: \"3a4fcb67-0b78-43d6-a6a1-8d319bf19184\") " Dec 10 23:09:31 crc kubenswrapper[4791]: I1210 23:09:31.732658 4791 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/3a4fcb67-0b78-43d6-a6a1-8d319bf19184-ovsdbserver-nb\") pod \"3a4fcb67-0b78-43d6-a6a1-8d319bf19184\" (UID: \"3a4fcb67-0b78-43d6-a6a1-8d319bf19184\") " Dec 10 23:09:31 crc kubenswrapper[4791]: I1210 23:09:31.732717 4791 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/3a4fcb67-0b78-43d6-a6a1-8d319bf19184-dns-svc\") pod \"3a4fcb67-0b78-43d6-a6a1-8d319bf19184\" (UID: \"3a4fcb67-0b78-43d6-a6a1-8d319bf19184\") " Dec 10 23:09:31 crc kubenswrapper[4791]: I1210 23:09:31.732758 4791 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/3a4fcb67-0b78-43d6-a6a1-8d319bf19184-dns-swift-storage-0\") pod \"3a4fcb67-0b78-43d6-a6a1-8d319bf19184\" (UID: \"3a4fcb67-0b78-43d6-a6a1-8d319bf19184\") " Dec 10 23:09:31 crc kubenswrapper[4791]: I1210 23:09:31.732811 4791 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-pg94c\" (UniqueName: \"kubernetes.io/projected/3a4fcb67-0b78-43d6-a6a1-8d319bf19184-kube-api-access-pg94c\") pod \"3a4fcb67-0b78-43d6-a6a1-8d319bf19184\" (UID: \"3a4fcb67-0b78-43d6-a6a1-8d319bf19184\") " Dec 10 23:09:31 crc kubenswrapper[4791]: I1210 23:09:31.732885 4791 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/3a4fcb67-0b78-43d6-a6a1-8d319bf19184-config\") pod \"3a4fcb67-0b78-43d6-a6a1-8d319bf19184\" (UID: \"3a4fcb67-0b78-43d6-a6a1-8d319bf19184\") " Dec 10 23:09:31 crc kubenswrapper[4791]: I1210 23:09:31.751564 4791 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/3a4fcb67-0b78-43d6-a6a1-8d319bf19184-kube-api-access-pg94c" (OuterVolumeSpecName: "kube-api-access-pg94c") pod "3a4fcb67-0b78-43d6-a6a1-8d319bf19184" (UID: "3a4fcb67-0b78-43d6-a6a1-8d319bf19184"). InnerVolumeSpecName "kube-api-access-pg94c". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 10 23:09:31 crc kubenswrapper[4791]: I1210 23:09:31.836205 4791 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-pg94c\" (UniqueName: \"kubernetes.io/projected/3a4fcb67-0b78-43d6-a6a1-8d319bf19184-kube-api-access-pg94c\") on node \"crc\" DevicePath \"\"" Dec 10 23:09:31 crc kubenswrapper[4791]: I1210 23:09:31.880037 4791 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/3a4fcb67-0b78-43d6-a6a1-8d319bf19184-ovsdbserver-nb" (OuterVolumeSpecName: "ovsdbserver-nb") pod "3a4fcb67-0b78-43d6-a6a1-8d319bf19184" (UID: "3a4fcb67-0b78-43d6-a6a1-8d319bf19184"). InnerVolumeSpecName "ovsdbserver-nb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 10 23:09:31 crc kubenswrapper[4791]: I1210 23:09:31.880372 4791 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/3a4fcb67-0b78-43d6-a6a1-8d319bf19184-dns-swift-storage-0" (OuterVolumeSpecName: "dns-swift-storage-0") pod "3a4fcb67-0b78-43d6-a6a1-8d319bf19184" (UID: "3a4fcb67-0b78-43d6-a6a1-8d319bf19184"). InnerVolumeSpecName "dns-swift-storage-0". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 10 23:09:31 crc kubenswrapper[4791]: I1210 23:09:31.881026 4791 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/3a4fcb67-0b78-43d6-a6a1-8d319bf19184-config" (OuterVolumeSpecName: "config") pod "3a4fcb67-0b78-43d6-a6a1-8d319bf19184" (UID: "3a4fcb67-0b78-43d6-a6a1-8d319bf19184"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 10 23:09:31 crc kubenswrapper[4791]: I1210 23:09:31.881237 4791 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/3a4fcb67-0b78-43d6-a6a1-8d319bf19184-ovsdbserver-sb" (OuterVolumeSpecName: "ovsdbserver-sb") pod "3a4fcb67-0b78-43d6-a6a1-8d319bf19184" (UID: "3a4fcb67-0b78-43d6-a6a1-8d319bf19184"). InnerVolumeSpecName "ovsdbserver-sb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 10 23:09:31 crc kubenswrapper[4791]: I1210 23:09:31.885697 4791 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/3a4fcb67-0b78-43d6-a6a1-8d319bf19184-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "3a4fcb67-0b78-43d6-a6a1-8d319bf19184" (UID: "3a4fcb67-0b78-43d6-a6a1-8d319bf19184"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 10 23:09:31 crc kubenswrapper[4791]: I1210 23:09:31.919368 4791 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="75eb971a-04fc-40f5-9890-d5a0b4daccd1" path="/var/lib/kubelet/pods/75eb971a-04fc-40f5-9890-d5a0b4daccd1/volumes" Dec 10 23:09:31 crc kubenswrapper[4791]: I1210 23:09:31.938397 4791 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/3a4fcb67-0b78-43d6-a6a1-8d319bf19184-ovsdbserver-sb\") on node \"crc\" DevicePath \"\"" Dec 10 23:09:31 crc kubenswrapper[4791]: I1210 23:09:31.938427 4791 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/3a4fcb67-0b78-43d6-a6a1-8d319bf19184-ovsdbserver-nb\") on node \"crc\" DevicePath \"\"" Dec 10 23:09:31 crc kubenswrapper[4791]: I1210 23:09:31.938436 4791 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/3a4fcb67-0b78-43d6-a6a1-8d319bf19184-dns-svc\") on node \"crc\" DevicePath \"\"" Dec 10 23:09:31 crc kubenswrapper[4791]: I1210 23:09:31.938446 4791 reconciler_common.go:293] "Volume detached for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/3a4fcb67-0b78-43d6-a6a1-8d319bf19184-dns-swift-storage-0\") on node \"crc\" DevicePath \"\"" Dec 10 23:09:31 crc kubenswrapper[4791]: I1210 23:09:31.938455 4791 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/3a4fcb67-0b78-43d6-a6a1-8d319bf19184-config\") on node \"crc\" DevicePath \"\"" Dec 10 23:09:32 crc kubenswrapper[4791]: I1210 23:09:32.235971 4791 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"58167832-d8da-48de-8b8b-cf51fc54d92c","Type":"ContainerStarted","Data":"d7021f57af9ab5eef5206bcbbcf0c53acba5e7dfcef13ecaa64e06f7d4fe6fc2"} Dec 10 23:09:32 crc kubenswrapper[4791]: I1210 23:09:32.238238 4791 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-84b966f6c9-tz266" Dec 10 23:09:32 crc kubenswrapper[4791]: I1210 23:09:32.238656 4791 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-84b966f6c9-tz266" event={"ID":"3a4fcb67-0b78-43d6-a6a1-8d319bf19184","Type":"ContainerDied","Data":"47471e5879b32ac699bed8b99af0a4c42ce82327f2054dd5c2ccb69ce6de9494"} Dec 10 23:09:32 crc kubenswrapper[4791]: I1210 23:09:32.238684 4791 scope.go:117] "RemoveContainer" containerID="4294aa9ba3a7a9440f41cfc9d38f97e9bcc1fa80423767e0df5c6d4016638bf9" Dec 10 23:09:32 crc kubenswrapper[4791]: I1210 23:09:32.264932 4791 scope.go:117] "RemoveContainer" containerID="e3d0d1d6e6602ea537ea9b3c0b7772d80cba882bb829676e2912ef4ae01fb09f" Dec 10 23:09:32 crc kubenswrapper[4791]: I1210 23:09:32.268812 4791 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-84b966f6c9-tz266"] Dec 10 23:09:32 crc kubenswrapper[4791]: I1210 23:09:32.284256 4791 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-84b966f6c9-tz266"] Dec 10 23:09:32 crc kubenswrapper[4791]: I1210 23:09:32.434533 4791 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/keystone-54b7f6dc79-9klwj" Dec 10 23:09:33 crc kubenswrapper[4791]: I1210 23:09:33.115424 4791 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/openstackclient"] Dec 10 23:09:33 crc kubenswrapper[4791]: E1210 23:09:33.116137 4791 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="75eb971a-04fc-40f5-9890-d5a0b4daccd1" containerName="neutron-api" Dec 10 23:09:33 crc kubenswrapper[4791]: I1210 23:09:33.116161 4791 state_mem.go:107] "Deleted CPUSet assignment" podUID="75eb971a-04fc-40f5-9890-d5a0b4daccd1" containerName="neutron-api" Dec 10 23:09:33 crc kubenswrapper[4791]: E1210 23:09:33.116182 4791 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="3a4fcb67-0b78-43d6-a6a1-8d319bf19184" containerName="dnsmasq-dns" Dec 10 23:09:33 crc kubenswrapper[4791]: I1210 23:09:33.116190 4791 state_mem.go:107] "Deleted CPUSet assignment" podUID="3a4fcb67-0b78-43d6-a6a1-8d319bf19184" containerName="dnsmasq-dns" Dec 10 23:09:33 crc kubenswrapper[4791]: E1210 23:09:33.116210 4791 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="75eb971a-04fc-40f5-9890-d5a0b4daccd1" containerName="neutron-httpd" Dec 10 23:09:33 crc kubenswrapper[4791]: I1210 23:09:33.116216 4791 state_mem.go:107] "Deleted CPUSet assignment" podUID="75eb971a-04fc-40f5-9890-d5a0b4daccd1" containerName="neutron-httpd" Dec 10 23:09:33 crc kubenswrapper[4791]: E1210 23:09:33.116236 4791 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="3a4fcb67-0b78-43d6-a6a1-8d319bf19184" containerName="init" Dec 10 23:09:33 crc kubenswrapper[4791]: I1210 23:09:33.116243 4791 state_mem.go:107] "Deleted CPUSet assignment" podUID="3a4fcb67-0b78-43d6-a6a1-8d319bf19184" containerName="init" Dec 10 23:09:33 crc kubenswrapper[4791]: I1210 23:09:33.116441 4791 memory_manager.go:354] "RemoveStaleState removing state" podUID="75eb971a-04fc-40f5-9890-d5a0b4daccd1" containerName="neutron-httpd" Dec 10 23:09:33 crc kubenswrapper[4791]: I1210 23:09:33.116468 4791 memory_manager.go:354] "RemoveStaleState removing state" podUID="75eb971a-04fc-40f5-9890-d5a0b4daccd1" containerName="neutron-api" Dec 10 23:09:33 crc kubenswrapper[4791]: I1210 23:09:33.116485 4791 memory_manager.go:354] "RemoveStaleState removing state" podUID="3a4fcb67-0b78-43d6-a6a1-8d319bf19184" containerName="dnsmasq-dns" Dec 10 23:09:33 crc kubenswrapper[4791]: I1210 23:09:33.117078 4791 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/openstackclient" Dec 10 23:09:33 crc kubenswrapper[4791]: I1210 23:09:33.123263 4791 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-config" Dec 10 23:09:33 crc kubenswrapper[4791]: I1210 23:09:33.123553 4791 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-config-secret" Dec 10 23:09:33 crc kubenswrapper[4791]: I1210 23:09:33.123741 4791 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstackclient-openstackclient-dockercfg-s94s7" Dec 10 23:09:33 crc kubenswrapper[4791]: I1210 23:09:33.155607 4791 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/openstackclient"] Dec 10 23:09:33 crc kubenswrapper[4791]: I1210 23:09:33.269545 4791 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"openstack-config\" (UniqueName: \"kubernetes.io/configmap/2d25921f-5a9e-4e3b-9c51-f07f95aa02a1-openstack-config\") pod \"openstackclient\" (UID: \"2d25921f-5a9e-4e3b-9c51-f07f95aa02a1\") " pod="openstack/openstackclient" Dec 10 23:09:33 crc kubenswrapper[4791]: I1210 23:09:33.269927 4791 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/2d25921f-5a9e-4e3b-9c51-f07f95aa02a1-combined-ca-bundle\") pod \"openstackclient\" (UID: \"2d25921f-5a9e-4e3b-9c51-f07f95aa02a1\") " pod="openstack/openstackclient" Dec 10 23:09:33 crc kubenswrapper[4791]: I1210 23:09:33.269977 4791 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-z7gxq\" (UniqueName: \"kubernetes.io/projected/2d25921f-5a9e-4e3b-9c51-f07f95aa02a1-kube-api-access-z7gxq\") pod \"openstackclient\" (UID: \"2d25921f-5a9e-4e3b-9c51-f07f95aa02a1\") " pod="openstack/openstackclient" Dec 10 23:09:33 crc kubenswrapper[4791]: I1210 23:09:33.270010 4791 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"openstack-config-secret\" (UniqueName: \"kubernetes.io/secret/2d25921f-5a9e-4e3b-9c51-f07f95aa02a1-openstack-config-secret\") pod \"openstackclient\" (UID: \"2d25921f-5a9e-4e3b-9c51-f07f95aa02a1\") " pod="openstack/openstackclient" Dec 10 23:09:33 crc kubenswrapper[4791]: I1210 23:09:33.387269 4791 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"openstack-config\" (UniqueName: \"kubernetes.io/configmap/2d25921f-5a9e-4e3b-9c51-f07f95aa02a1-openstack-config\") pod \"openstackclient\" (UID: \"2d25921f-5a9e-4e3b-9c51-f07f95aa02a1\") " pod="openstack/openstackclient" Dec 10 23:09:33 crc kubenswrapper[4791]: I1210 23:09:33.387323 4791 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/2d25921f-5a9e-4e3b-9c51-f07f95aa02a1-combined-ca-bundle\") pod \"openstackclient\" (UID: \"2d25921f-5a9e-4e3b-9c51-f07f95aa02a1\") " pod="openstack/openstackclient" Dec 10 23:09:33 crc kubenswrapper[4791]: I1210 23:09:33.387372 4791 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-z7gxq\" (UniqueName: \"kubernetes.io/projected/2d25921f-5a9e-4e3b-9c51-f07f95aa02a1-kube-api-access-z7gxq\") pod \"openstackclient\" (UID: \"2d25921f-5a9e-4e3b-9c51-f07f95aa02a1\") " pod="openstack/openstackclient" Dec 10 23:09:33 crc kubenswrapper[4791]: I1210 23:09:33.387393 4791 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"openstack-config-secret\" (UniqueName: \"kubernetes.io/secret/2d25921f-5a9e-4e3b-9c51-f07f95aa02a1-openstack-config-secret\") pod \"openstackclient\" (UID: \"2d25921f-5a9e-4e3b-9c51-f07f95aa02a1\") " pod="openstack/openstackclient" Dec 10 23:09:33 crc kubenswrapper[4791]: I1210 23:09:33.388814 4791 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"openstack-config\" (UniqueName: \"kubernetes.io/configmap/2d25921f-5a9e-4e3b-9c51-f07f95aa02a1-openstack-config\") pod \"openstackclient\" (UID: \"2d25921f-5a9e-4e3b-9c51-f07f95aa02a1\") " pod="openstack/openstackclient" Dec 10 23:09:33 crc kubenswrapper[4791]: I1210 23:09:33.407413 4791 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/2d25921f-5a9e-4e3b-9c51-f07f95aa02a1-combined-ca-bundle\") pod \"openstackclient\" (UID: \"2d25921f-5a9e-4e3b-9c51-f07f95aa02a1\") " pod="openstack/openstackclient" Dec 10 23:09:33 crc kubenswrapper[4791]: I1210 23:09:33.432890 4791 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"openstack-config-secret\" (UniqueName: \"kubernetes.io/secret/2d25921f-5a9e-4e3b-9c51-f07f95aa02a1-openstack-config-secret\") pod \"openstackclient\" (UID: \"2d25921f-5a9e-4e3b-9c51-f07f95aa02a1\") " pod="openstack/openstackclient" Dec 10 23:09:33 crc kubenswrapper[4791]: I1210 23:09:33.432898 4791 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-z7gxq\" (UniqueName: \"kubernetes.io/projected/2d25921f-5a9e-4e3b-9c51-f07f95aa02a1-kube-api-access-z7gxq\") pod \"openstackclient\" (UID: \"2d25921f-5a9e-4e3b-9c51-f07f95aa02a1\") " pod="openstack/openstackclient" Dec 10 23:09:33 crc kubenswrapper[4791]: I1210 23:09:33.457445 4791 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/openstackclient" Dec 10 23:09:33 crc kubenswrapper[4791]: I1210 23:09:33.548546 4791 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/cinder-scheduler-0" Dec 10 23:09:33 crc kubenswrapper[4791]: I1210 23:09:33.872805 4791 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/placement-6487596ddd-btmd7" Dec 10 23:09:33 crc kubenswrapper[4791]: I1210 23:09:33.908685 4791 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="3a4fcb67-0b78-43d6-a6a1-8d319bf19184" path="/var/lib/kubelet/pods/3a4fcb67-0b78-43d6-a6a1-8d319bf19184/volumes" Dec 10 23:09:33 crc kubenswrapper[4791]: I1210 23:09:33.985000 4791 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/placement-6487596ddd-btmd7" Dec 10 23:09:34 crc kubenswrapper[4791]: I1210 23:09:34.097432 4791 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/openstackclient"] Dec 10 23:09:34 crc kubenswrapper[4791]: I1210 23:09:34.135097 4791 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/barbican-api-795f5df4db-xf57n" Dec 10 23:09:34 crc kubenswrapper[4791]: I1210 23:09:34.268218 4791 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/openstackclient" event={"ID":"2d25921f-5a9e-4e3b-9c51-f07f95aa02a1","Type":"ContainerStarted","Data":"05b680ce5868779eb00af73b1047499e7fa0cb1145c81a5a8c5fed55396613cf"} Dec 10 23:09:34 crc kubenswrapper[4791]: I1210 23:09:34.281490 4791 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"58167832-d8da-48de-8b8b-cf51fc54d92c","Type":"ContainerStarted","Data":"d3db4aa0578d2e480256777270f40b9150d19368ae770c08c87f98f42a115b57"} Dec 10 23:09:34 crc kubenswrapper[4791]: I1210 23:09:34.281600 4791 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/ceilometer-0" Dec 10 23:09:34 crc kubenswrapper[4791]: I1210 23:09:34.314048 4791 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ceilometer-0" podStartSLOduration=2.431517441 podStartE2EDuration="8.314025864s" podCreationTimestamp="2025-12-10 23:09:26 +0000 UTC" firstStartedPulling="2025-12-10 23:09:27.372626224 +0000 UTC m=+1201.802243837" lastFinishedPulling="2025-12-10 23:09:33.255134647 +0000 UTC m=+1207.684752260" observedRunningTime="2025-12-10 23:09:34.303364986 +0000 UTC m=+1208.732982599" watchObservedRunningTime="2025-12-10 23:09:34.314025864 +0000 UTC m=+1208.743643477" Dec 10 23:09:34 crc kubenswrapper[4791]: I1210 23:09:34.973728 4791 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/barbican-api-795f5df4db-xf57n" Dec 10 23:09:35 crc kubenswrapper[4791]: I1210 23:09:35.411066 4791 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/barbican-api-68fb764d56-4mfz4" Dec 10 23:09:36 crc kubenswrapper[4791]: I1210 23:09:36.241681 4791 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/barbican-api-68fb764d56-4mfz4" Dec 10 23:09:36 crc kubenswrapper[4791]: I1210 23:09:36.322707 4791 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/barbican-api-795f5df4db-xf57n"] Dec 10 23:09:36 crc kubenswrapper[4791]: I1210 23:09:36.323158 4791 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/barbican-api-795f5df4db-xf57n" podUID="9b340efd-bf7e-4caf-8c76-ddc500aa2085" containerName="barbican-api-log" containerID="cri-o://f05d3eef09f65c594818648ae666e092eb1a3e05925b8b910f123fdb1170cde9" gracePeriod=30 Dec 10 23:09:36 crc kubenswrapper[4791]: I1210 23:09:36.323605 4791 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/barbican-api-795f5df4db-xf57n" podUID="9b340efd-bf7e-4caf-8c76-ddc500aa2085" containerName="barbican-api" containerID="cri-o://31193dfd2f9e85871df3d79e4aef1d62fa1b1a1b269896d2e71dbdb4f4784a21" gracePeriod=30 Dec 10 23:09:36 crc kubenswrapper[4791]: I1210 23:09:36.330275 4791 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/barbican-api-795f5df4db-xf57n" podUID="9b340efd-bf7e-4caf-8c76-ddc500aa2085" containerName="barbican-api-log" probeResult="failure" output="Get \"http://10.217.0.163:9311/healthcheck\": EOF" Dec 10 23:09:36 crc kubenswrapper[4791]: I1210 23:09:36.332408 4791 prober.go:107] "Probe failed" probeType="Liveness" pod="openstack/barbican-api-795f5df4db-xf57n" podUID="9b340efd-bf7e-4caf-8c76-ddc500aa2085" containerName="barbican-api-log" probeResult="failure" output="Get \"http://10.217.0.163:9311/healthcheck\": EOF" Dec 10 23:09:36 crc kubenswrapper[4791]: I1210 23:09:36.332580 4791 prober.go:107] "Probe failed" probeType="Liveness" pod="openstack/barbican-api-795f5df4db-xf57n" podUID="9b340efd-bf7e-4caf-8c76-ddc500aa2085" containerName="barbican-api" probeResult="failure" output="Get \"http://10.217.0.163:9311/healthcheck\": EOF" Dec 10 23:09:37 crc kubenswrapper[4791]: I1210 23:09:37.099890 4791 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/cinder-api-0" Dec 10 23:09:37 crc kubenswrapper[4791]: I1210 23:09:37.323251 4791 generic.go:334] "Generic (PLEG): container finished" podID="9b340efd-bf7e-4caf-8c76-ddc500aa2085" containerID="f05d3eef09f65c594818648ae666e092eb1a3e05925b8b910f123fdb1170cde9" exitCode=143 Dec 10 23:09:37 crc kubenswrapper[4791]: I1210 23:09:37.323294 4791 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-api-795f5df4db-xf57n" event={"ID":"9b340efd-bf7e-4caf-8c76-ddc500aa2085","Type":"ContainerDied","Data":"f05d3eef09f65c594818648ae666e092eb1a3e05925b8b910f123fdb1170cde9"} Dec 10 23:09:38 crc kubenswrapper[4791]: I1210 23:09:38.745586 4791 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/cinder-scheduler-0" Dec 10 23:09:40 crc kubenswrapper[4791]: I1210 23:09:40.174719 4791 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-api-db-create-cvjnf"] Dec 10 23:09:40 crc kubenswrapper[4791]: I1210 23:09:40.179677 4791 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-db-create-cvjnf" Dec 10 23:09:40 crc kubenswrapper[4791]: I1210 23:09:40.188260 4791 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-api-db-create-cvjnf"] Dec 10 23:09:40 crc kubenswrapper[4791]: I1210 23:09:40.275069 4791 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-pqv5l\" (UniqueName: \"kubernetes.io/projected/18c84e26-d0bb-4cd5-ba6a-49bf21155889-kube-api-access-pqv5l\") pod \"nova-api-db-create-cvjnf\" (UID: \"18c84e26-d0bb-4cd5-ba6a-49bf21155889\") " pod="openstack/nova-api-db-create-cvjnf" Dec 10 23:09:40 crc kubenswrapper[4791]: I1210 23:09:40.275614 4791 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/18c84e26-d0bb-4cd5-ba6a-49bf21155889-operator-scripts\") pod \"nova-api-db-create-cvjnf\" (UID: \"18c84e26-d0bb-4cd5-ba6a-49bf21155889\") " pod="openstack/nova-api-db-create-cvjnf" Dec 10 23:09:40 crc kubenswrapper[4791]: I1210 23:09:40.288101 4791 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-api-8832-account-create-update-wp7p5"] Dec 10 23:09:40 crc kubenswrapper[4791]: I1210 23:09:40.289631 4791 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-8832-account-create-update-wp7p5" Dec 10 23:09:40 crc kubenswrapper[4791]: I1210 23:09:40.293943 4791 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-api-db-secret" Dec 10 23:09:40 crc kubenswrapper[4791]: I1210 23:09:40.301489 4791 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-cell0-db-create-z2v58"] Dec 10 23:09:40 crc kubenswrapper[4791]: I1210 23:09:40.302993 4791 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-db-create-z2v58" Dec 10 23:09:40 crc kubenswrapper[4791]: I1210 23:09:40.306929 4791 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-api-8832-account-create-update-wp7p5"] Dec 10 23:09:40 crc kubenswrapper[4791]: I1210 23:09:40.323614 4791 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell0-db-create-z2v58"] Dec 10 23:09:40 crc kubenswrapper[4791]: I1210 23:09:40.377335 4791 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-6fnsl\" (UniqueName: \"kubernetes.io/projected/0a94efe1-86a8-4768-9199-f47051935bcf-kube-api-access-6fnsl\") pod \"nova-api-8832-account-create-update-wp7p5\" (UID: \"0a94efe1-86a8-4768-9199-f47051935bcf\") " pod="openstack/nova-api-8832-account-create-update-wp7p5" Dec 10 23:09:40 crc kubenswrapper[4791]: I1210 23:09:40.377422 4791 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-pqv5l\" (UniqueName: \"kubernetes.io/projected/18c84e26-d0bb-4cd5-ba6a-49bf21155889-kube-api-access-pqv5l\") pod \"nova-api-db-create-cvjnf\" (UID: \"18c84e26-d0bb-4cd5-ba6a-49bf21155889\") " pod="openstack/nova-api-db-create-cvjnf" Dec 10 23:09:40 crc kubenswrapper[4791]: I1210 23:09:40.377467 4791 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/18c84e26-d0bb-4cd5-ba6a-49bf21155889-operator-scripts\") pod \"nova-api-db-create-cvjnf\" (UID: \"18c84e26-d0bb-4cd5-ba6a-49bf21155889\") " pod="openstack/nova-api-db-create-cvjnf" Dec 10 23:09:40 crc kubenswrapper[4791]: I1210 23:09:40.377496 4791 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/0a94efe1-86a8-4768-9199-f47051935bcf-operator-scripts\") pod \"nova-api-8832-account-create-update-wp7p5\" (UID: \"0a94efe1-86a8-4768-9199-f47051935bcf\") " pod="openstack/nova-api-8832-account-create-update-wp7p5" Dec 10 23:09:40 crc kubenswrapper[4791]: I1210 23:09:40.378478 4791 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/18c84e26-d0bb-4cd5-ba6a-49bf21155889-operator-scripts\") pod \"nova-api-db-create-cvjnf\" (UID: \"18c84e26-d0bb-4cd5-ba6a-49bf21155889\") " pod="openstack/nova-api-db-create-cvjnf" Dec 10 23:09:40 crc kubenswrapper[4791]: I1210 23:09:40.401197 4791 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-cell1-db-create-tz8dr"] Dec 10 23:09:40 crc kubenswrapper[4791]: I1210 23:09:40.403452 4791 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-db-create-tz8dr" Dec 10 23:09:40 crc kubenswrapper[4791]: I1210 23:09:40.412278 4791 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-db-create-tz8dr"] Dec 10 23:09:40 crc kubenswrapper[4791]: I1210 23:09:40.425969 4791 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-pqv5l\" (UniqueName: \"kubernetes.io/projected/18c84e26-d0bb-4cd5-ba6a-49bf21155889-kube-api-access-pqv5l\") pod \"nova-api-db-create-cvjnf\" (UID: \"18c84e26-d0bb-4cd5-ba6a-49bf21155889\") " pod="openstack/nova-api-db-create-cvjnf" Dec 10 23:09:40 crc kubenswrapper[4791]: I1210 23:09:40.485097 4791 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/b8b6cd29-edfe-4177-bf14-14b8580d86c2-operator-scripts\") pod \"nova-cell0-db-create-z2v58\" (UID: \"b8b6cd29-edfe-4177-bf14-14b8580d86c2\") " pod="openstack/nova-cell0-db-create-z2v58" Dec 10 23:09:40 crc kubenswrapper[4791]: I1210 23:09:40.485178 4791 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-6fnsl\" (UniqueName: \"kubernetes.io/projected/0a94efe1-86a8-4768-9199-f47051935bcf-kube-api-access-6fnsl\") pod \"nova-api-8832-account-create-update-wp7p5\" (UID: \"0a94efe1-86a8-4768-9199-f47051935bcf\") " pod="openstack/nova-api-8832-account-create-update-wp7p5" Dec 10 23:09:40 crc kubenswrapper[4791]: I1210 23:09:40.485249 4791 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-2cg8h\" (UniqueName: \"kubernetes.io/projected/b8b6cd29-edfe-4177-bf14-14b8580d86c2-kube-api-access-2cg8h\") pod \"nova-cell0-db-create-z2v58\" (UID: \"b8b6cd29-edfe-4177-bf14-14b8580d86c2\") " pod="openstack/nova-cell0-db-create-z2v58" Dec 10 23:09:40 crc kubenswrapper[4791]: I1210 23:09:40.485304 4791 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/0a94efe1-86a8-4768-9199-f47051935bcf-operator-scripts\") pod \"nova-api-8832-account-create-update-wp7p5\" (UID: \"0a94efe1-86a8-4768-9199-f47051935bcf\") " pod="openstack/nova-api-8832-account-create-update-wp7p5" Dec 10 23:09:40 crc kubenswrapper[4791]: I1210 23:09:40.486299 4791 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/0a94efe1-86a8-4768-9199-f47051935bcf-operator-scripts\") pod \"nova-api-8832-account-create-update-wp7p5\" (UID: \"0a94efe1-86a8-4768-9199-f47051935bcf\") " pod="openstack/nova-api-8832-account-create-update-wp7p5" Dec 10 23:09:40 crc kubenswrapper[4791]: I1210 23:09:40.497559 4791 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-cell0-e207-account-create-update-hlmck"] Dec 10 23:09:40 crc kubenswrapper[4791]: I1210 23:09:40.498692 4791 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-e207-account-create-update-hlmck" Dec 10 23:09:40 crc kubenswrapper[4791]: I1210 23:09:40.501022 4791 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell0-db-secret" Dec 10 23:09:40 crc kubenswrapper[4791]: I1210 23:09:40.501915 4791 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-db-create-cvjnf" Dec 10 23:09:40 crc kubenswrapper[4791]: I1210 23:09:40.512629 4791 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell0-e207-account-create-update-hlmck"] Dec 10 23:09:40 crc kubenswrapper[4791]: I1210 23:09:40.523030 4791 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-6fnsl\" (UniqueName: \"kubernetes.io/projected/0a94efe1-86a8-4768-9199-f47051935bcf-kube-api-access-6fnsl\") pod \"nova-api-8832-account-create-update-wp7p5\" (UID: \"0a94efe1-86a8-4768-9199-f47051935bcf\") " pod="openstack/nova-api-8832-account-create-update-wp7p5" Dec 10 23:09:40 crc kubenswrapper[4791]: I1210 23:09:40.586735 4791 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/b8b6cd29-edfe-4177-bf14-14b8580d86c2-operator-scripts\") pod \"nova-cell0-db-create-z2v58\" (UID: \"b8b6cd29-edfe-4177-bf14-14b8580d86c2\") " pod="openstack/nova-cell0-db-create-z2v58" Dec 10 23:09:40 crc kubenswrapper[4791]: I1210 23:09:40.586848 4791 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/100aa755-c120-4c5a-9135-23bb98661f3d-operator-scripts\") pod \"nova-cell1-db-create-tz8dr\" (UID: \"100aa755-c120-4c5a-9135-23bb98661f3d\") " pod="openstack/nova-cell1-db-create-tz8dr" Dec 10 23:09:40 crc kubenswrapper[4791]: I1210 23:09:40.586888 4791 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-kg2hz\" (UniqueName: \"kubernetes.io/projected/cf0595b8-0690-448d-9ef7-a64227bad653-kube-api-access-kg2hz\") pod \"nova-cell0-e207-account-create-update-hlmck\" (UID: \"cf0595b8-0690-448d-9ef7-a64227bad653\") " pod="openstack/nova-cell0-e207-account-create-update-hlmck" Dec 10 23:09:40 crc kubenswrapper[4791]: I1210 23:09:40.586918 4791 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-2cg8h\" (UniqueName: \"kubernetes.io/projected/b8b6cd29-edfe-4177-bf14-14b8580d86c2-kube-api-access-2cg8h\") pod \"nova-cell0-db-create-z2v58\" (UID: \"b8b6cd29-edfe-4177-bf14-14b8580d86c2\") " pod="openstack/nova-cell0-db-create-z2v58" Dec 10 23:09:40 crc kubenswrapper[4791]: I1210 23:09:40.586956 4791 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/cf0595b8-0690-448d-9ef7-a64227bad653-operator-scripts\") pod \"nova-cell0-e207-account-create-update-hlmck\" (UID: \"cf0595b8-0690-448d-9ef7-a64227bad653\") " pod="openstack/nova-cell0-e207-account-create-update-hlmck" Dec 10 23:09:40 crc kubenswrapper[4791]: I1210 23:09:40.587031 4791 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-xrlpr\" (UniqueName: \"kubernetes.io/projected/100aa755-c120-4c5a-9135-23bb98661f3d-kube-api-access-xrlpr\") pod \"nova-cell1-db-create-tz8dr\" (UID: \"100aa755-c120-4c5a-9135-23bb98661f3d\") " pod="openstack/nova-cell1-db-create-tz8dr" Dec 10 23:09:40 crc kubenswrapper[4791]: I1210 23:09:40.587939 4791 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/b8b6cd29-edfe-4177-bf14-14b8580d86c2-operator-scripts\") pod \"nova-cell0-db-create-z2v58\" (UID: \"b8b6cd29-edfe-4177-bf14-14b8580d86c2\") " pod="openstack/nova-cell0-db-create-z2v58" Dec 10 23:09:40 crc kubenswrapper[4791]: I1210 23:09:40.604298 4791 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-8832-account-create-update-wp7p5" Dec 10 23:09:40 crc kubenswrapper[4791]: I1210 23:09:40.607884 4791 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-2cg8h\" (UniqueName: \"kubernetes.io/projected/b8b6cd29-edfe-4177-bf14-14b8580d86c2-kube-api-access-2cg8h\") pod \"nova-cell0-db-create-z2v58\" (UID: \"b8b6cd29-edfe-4177-bf14-14b8580d86c2\") " pod="openstack/nova-cell0-db-create-z2v58" Dec 10 23:09:40 crc kubenswrapper[4791]: I1210 23:09:40.622794 4791 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-db-create-z2v58" Dec 10 23:09:40 crc kubenswrapper[4791]: I1210 23:09:40.689685 4791 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/100aa755-c120-4c5a-9135-23bb98661f3d-operator-scripts\") pod \"nova-cell1-db-create-tz8dr\" (UID: \"100aa755-c120-4c5a-9135-23bb98661f3d\") " pod="openstack/nova-cell1-db-create-tz8dr" Dec 10 23:09:40 crc kubenswrapper[4791]: I1210 23:09:40.689768 4791 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-kg2hz\" (UniqueName: \"kubernetes.io/projected/cf0595b8-0690-448d-9ef7-a64227bad653-kube-api-access-kg2hz\") pod \"nova-cell0-e207-account-create-update-hlmck\" (UID: \"cf0595b8-0690-448d-9ef7-a64227bad653\") " pod="openstack/nova-cell0-e207-account-create-update-hlmck" Dec 10 23:09:40 crc kubenswrapper[4791]: I1210 23:09:40.689821 4791 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/cf0595b8-0690-448d-9ef7-a64227bad653-operator-scripts\") pod \"nova-cell0-e207-account-create-update-hlmck\" (UID: \"cf0595b8-0690-448d-9ef7-a64227bad653\") " pod="openstack/nova-cell0-e207-account-create-update-hlmck" Dec 10 23:09:40 crc kubenswrapper[4791]: I1210 23:09:40.689902 4791 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-xrlpr\" (UniqueName: \"kubernetes.io/projected/100aa755-c120-4c5a-9135-23bb98661f3d-kube-api-access-xrlpr\") pod \"nova-cell1-db-create-tz8dr\" (UID: \"100aa755-c120-4c5a-9135-23bb98661f3d\") " pod="openstack/nova-cell1-db-create-tz8dr" Dec 10 23:09:40 crc kubenswrapper[4791]: I1210 23:09:40.691298 4791 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/cf0595b8-0690-448d-9ef7-a64227bad653-operator-scripts\") pod \"nova-cell0-e207-account-create-update-hlmck\" (UID: \"cf0595b8-0690-448d-9ef7-a64227bad653\") " pod="openstack/nova-cell0-e207-account-create-update-hlmck" Dec 10 23:09:40 crc kubenswrapper[4791]: I1210 23:09:40.692559 4791 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/100aa755-c120-4c5a-9135-23bb98661f3d-operator-scripts\") pod \"nova-cell1-db-create-tz8dr\" (UID: \"100aa755-c120-4c5a-9135-23bb98661f3d\") " pod="openstack/nova-cell1-db-create-tz8dr" Dec 10 23:09:40 crc kubenswrapper[4791]: I1210 23:09:40.704406 4791 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-cell1-b007-account-create-update-fkwxb"] Dec 10 23:09:40 crc kubenswrapper[4791]: I1210 23:09:40.706001 4791 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-b007-account-create-update-fkwxb" Dec 10 23:09:40 crc kubenswrapper[4791]: I1210 23:09:40.714505 4791 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-b007-account-create-update-fkwxb"] Dec 10 23:09:40 crc kubenswrapper[4791]: I1210 23:09:40.716672 4791 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell1-db-secret" Dec 10 23:09:40 crc kubenswrapper[4791]: I1210 23:09:40.720616 4791 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-kg2hz\" (UniqueName: \"kubernetes.io/projected/cf0595b8-0690-448d-9ef7-a64227bad653-kube-api-access-kg2hz\") pod \"nova-cell0-e207-account-create-update-hlmck\" (UID: \"cf0595b8-0690-448d-9ef7-a64227bad653\") " pod="openstack/nova-cell0-e207-account-create-update-hlmck" Dec 10 23:09:40 crc kubenswrapper[4791]: I1210 23:09:40.723743 4791 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-xrlpr\" (UniqueName: \"kubernetes.io/projected/100aa755-c120-4c5a-9135-23bb98661f3d-kube-api-access-xrlpr\") pod \"nova-cell1-db-create-tz8dr\" (UID: \"100aa755-c120-4c5a-9135-23bb98661f3d\") " pod="openstack/nova-cell1-db-create-tz8dr" Dec 10 23:09:40 crc kubenswrapper[4791]: I1210 23:09:40.791248 4791 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/83ddd1be-5119-45a6-a080-1b63f0c8f7a1-operator-scripts\") pod \"nova-cell1-b007-account-create-update-fkwxb\" (UID: \"83ddd1be-5119-45a6-a080-1b63f0c8f7a1\") " pod="openstack/nova-cell1-b007-account-create-update-fkwxb" Dec 10 23:09:40 crc kubenswrapper[4791]: I1210 23:09:40.791300 4791 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-kdmwm\" (UniqueName: \"kubernetes.io/projected/83ddd1be-5119-45a6-a080-1b63f0c8f7a1-kube-api-access-kdmwm\") pod \"nova-cell1-b007-account-create-update-fkwxb\" (UID: \"83ddd1be-5119-45a6-a080-1b63f0c8f7a1\") " pod="openstack/nova-cell1-b007-account-create-update-fkwxb" Dec 10 23:09:40 crc kubenswrapper[4791]: I1210 23:09:40.791706 4791 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-db-create-tz8dr" Dec 10 23:09:40 crc kubenswrapper[4791]: I1210 23:09:40.871320 4791 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-e207-account-create-update-hlmck" Dec 10 23:09:40 crc kubenswrapper[4791]: I1210 23:09:40.894207 4791 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/83ddd1be-5119-45a6-a080-1b63f0c8f7a1-operator-scripts\") pod \"nova-cell1-b007-account-create-update-fkwxb\" (UID: \"83ddd1be-5119-45a6-a080-1b63f0c8f7a1\") " pod="openstack/nova-cell1-b007-account-create-update-fkwxb" Dec 10 23:09:40 crc kubenswrapper[4791]: I1210 23:09:40.894557 4791 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-kdmwm\" (UniqueName: \"kubernetes.io/projected/83ddd1be-5119-45a6-a080-1b63f0c8f7a1-kube-api-access-kdmwm\") pod \"nova-cell1-b007-account-create-update-fkwxb\" (UID: \"83ddd1be-5119-45a6-a080-1b63f0c8f7a1\") " pod="openstack/nova-cell1-b007-account-create-update-fkwxb" Dec 10 23:09:40 crc kubenswrapper[4791]: I1210 23:09:40.895087 4791 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/83ddd1be-5119-45a6-a080-1b63f0c8f7a1-operator-scripts\") pod \"nova-cell1-b007-account-create-update-fkwxb\" (UID: \"83ddd1be-5119-45a6-a080-1b63f0c8f7a1\") " pod="openstack/nova-cell1-b007-account-create-update-fkwxb" Dec 10 23:09:40 crc kubenswrapper[4791]: I1210 23:09:40.913620 4791 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-kdmwm\" (UniqueName: \"kubernetes.io/projected/83ddd1be-5119-45a6-a080-1b63f0c8f7a1-kube-api-access-kdmwm\") pod \"nova-cell1-b007-account-create-update-fkwxb\" (UID: \"83ddd1be-5119-45a6-a080-1b63f0c8f7a1\") " pod="openstack/nova-cell1-b007-account-create-update-fkwxb" Dec 10 23:09:41 crc kubenswrapper[4791]: I1210 23:09:41.112891 4791 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-b007-account-create-update-fkwxb" Dec 10 23:09:41 crc kubenswrapper[4791]: I1210 23:09:41.377017 4791 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/barbican-api-795f5df4db-xf57n" podUID="9b340efd-bf7e-4caf-8c76-ddc500aa2085" containerName="barbican-api-log" probeResult="failure" output="Get \"http://10.217.0.163:9311/healthcheck\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" Dec 10 23:09:41 crc kubenswrapper[4791]: I1210 23:09:41.430772 4791 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/swift-proxy-c7bbc5ff-rlprn"] Dec 10 23:09:41 crc kubenswrapper[4791]: I1210 23:09:41.497464 4791 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/swift-proxy-c7bbc5ff-rlprn" Dec 10 23:09:41 crc kubenswrapper[4791]: I1210 23:09:41.509433 4791 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-swift-internal-svc" Dec 10 23:09:41 crc kubenswrapper[4791]: I1210 23:09:41.509640 4791 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"swift-proxy-config-data" Dec 10 23:09:41 crc kubenswrapper[4791]: I1210 23:09:41.510086 4791 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/swift-proxy-c7bbc5ff-rlprn"] Dec 10 23:09:41 crc kubenswrapper[4791]: I1210 23:09:41.548435 4791 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-swift-public-svc" Dec 10 23:09:41 crc kubenswrapper[4791]: I1210 23:09:41.615357 4791 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/644890b3-8cae-45bc-b3eb-e0e6bfa8d957-config-data\") pod \"swift-proxy-c7bbc5ff-rlprn\" (UID: \"644890b3-8cae-45bc-b3eb-e0e6bfa8d957\") " pod="openstack/swift-proxy-c7bbc5ff-rlprn" Dec 10 23:09:41 crc kubenswrapper[4791]: I1210 23:09:41.615690 4791 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/projected/644890b3-8cae-45bc-b3eb-e0e6bfa8d957-etc-swift\") pod \"swift-proxy-c7bbc5ff-rlprn\" (UID: \"644890b3-8cae-45bc-b3eb-e0e6bfa8d957\") " pod="openstack/swift-proxy-c7bbc5ff-rlprn" Dec 10 23:09:41 crc kubenswrapper[4791]: I1210 23:09:41.615762 4791 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/644890b3-8cae-45bc-b3eb-e0e6bfa8d957-run-httpd\") pod \"swift-proxy-c7bbc5ff-rlprn\" (UID: \"644890b3-8cae-45bc-b3eb-e0e6bfa8d957\") " pod="openstack/swift-proxy-c7bbc5ff-rlprn" Dec 10 23:09:41 crc kubenswrapper[4791]: I1210 23:09:41.615795 4791 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/644890b3-8cae-45bc-b3eb-e0e6bfa8d957-combined-ca-bundle\") pod \"swift-proxy-c7bbc5ff-rlprn\" (UID: \"644890b3-8cae-45bc-b3eb-e0e6bfa8d957\") " pod="openstack/swift-proxy-c7bbc5ff-rlprn" Dec 10 23:09:41 crc kubenswrapper[4791]: I1210 23:09:41.615877 4791 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-vmctf\" (UniqueName: \"kubernetes.io/projected/644890b3-8cae-45bc-b3eb-e0e6bfa8d957-kube-api-access-vmctf\") pod \"swift-proxy-c7bbc5ff-rlprn\" (UID: \"644890b3-8cae-45bc-b3eb-e0e6bfa8d957\") " pod="openstack/swift-proxy-c7bbc5ff-rlprn" Dec 10 23:09:41 crc kubenswrapper[4791]: I1210 23:09:41.615904 4791 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/644890b3-8cae-45bc-b3eb-e0e6bfa8d957-internal-tls-certs\") pod \"swift-proxy-c7bbc5ff-rlprn\" (UID: \"644890b3-8cae-45bc-b3eb-e0e6bfa8d957\") " pod="openstack/swift-proxy-c7bbc5ff-rlprn" Dec 10 23:09:41 crc kubenswrapper[4791]: I1210 23:09:41.615942 4791 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/644890b3-8cae-45bc-b3eb-e0e6bfa8d957-public-tls-certs\") pod \"swift-proxy-c7bbc5ff-rlprn\" (UID: \"644890b3-8cae-45bc-b3eb-e0e6bfa8d957\") " pod="openstack/swift-proxy-c7bbc5ff-rlprn" Dec 10 23:09:41 crc kubenswrapper[4791]: I1210 23:09:41.616022 4791 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/644890b3-8cae-45bc-b3eb-e0e6bfa8d957-log-httpd\") pod \"swift-proxy-c7bbc5ff-rlprn\" (UID: \"644890b3-8cae-45bc-b3eb-e0e6bfa8d957\") " pod="openstack/swift-proxy-c7bbc5ff-rlprn" Dec 10 23:09:41 crc kubenswrapper[4791]: I1210 23:09:41.718221 4791 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/projected/644890b3-8cae-45bc-b3eb-e0e6bfa8d957-etc-swift\") pod \"swift-proxy-c7bbc5ff-rlprn\" (UID: \"644890b3-8cae-45bc-b3eb-e0e6bfa8d957\") " pod="openstack/swift-proxy-c7bbc5ff-rlprn" Dec 10 23:09:41 crc kubenswrapper[4791]: I1210 23:09:41.718282 4791 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/644890b3-8cae-45bc-b3eb-e0e6bfa8d957-run-httpd\") pod \"swift-proxy-c7bbc5ff-rlprn\" (UID: \"644890b3-8cae-45bc-b3eb-e0e6bfa8d957\") " pod="openstack/swift-proxy-c7bbc5ff-rlprn" Dec 10 23:09:41 crc kubenswrapper[4791]: I1210 23:09:41.718307 4791 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/644890b3-8cae-45bc-b3eb-e0e6bfa8d957-combined-ca-bundle\") pod \"swift-proxy-c7bbc5ff-rlprn\" (UID: \"644890b3-8cae-45bc-b3eb-e0e6bfa8d957\") " pod="openstack/swift-proxy-c7bbc5ff-rlprn" Dec 10 23:09:41 crc kubenswrapper[4791]: I1210 23:09:41.718369 4791 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-vmctf\" (UniqueName: \"kubernetes.io/projected/644890b3-8cae-45bc-b3eb-e0e6bfa8d957-kube-api-access-vmctf\") pod \"swift-proxy-c7bbc5ff-rlprn\" (UID: \"644890b3-8cae-45bc-b3eb-e0e6bfa8d957\") " pod="openstack/swift-proxy-c7bbc5ff-rlprn" Dec 10 23:09:41 crc kubenswrapper[4791]: I1210 23:09:41.718393 4791 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/644890b3-8cae-45bc-b3eb-e0e6bfa8d957-internal-tls-certs\") pod \"swift-proxy-c7bbc5ff-rlprn\" (UID: \"644890b3-8cae-45bc-b3eb-e0e6bfa8d957\") " pod="openstack/swift-proxy-c7bbc5ff-rlprn" Dec 10 23:09:41 crc kubenswrapper[4791]: I1210 23:09:41.718425 4791 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/644890b3-8cae-45bc-b3eb-e0e6bfa8d957-public-tls-certs\") pod \"swift-proxy-c7bbc5ff-rlprn\" (UID: \"644890b3-8cae-45bc-b3eb-e0e6bfa8d957\") " pod="openstack/swift-proxy-c7bbc5ff-rlprn" Dec 10 23:09:41 crc kubenswrapper[4791]: I1210 23:09:41.718468 4791 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/644890b3-8cae-45bc-b3eb-e0e6bfa8d957-log-httpd\") pod \"swift-proxy-c7bbc5ff-rlprn\" (UID: \"644890b3-8cae-45bc-b3eb-e0e6bfa8d957\") " pod="openstack/swift-proxy-c7bbc5ff-rlprn" Dec 10 23:09:41 crc kubenswrapper[4791]: I1210 23:09:41.718538 4791 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/644890b3-8cae-45bc-b3eb-e0e6bfa8d957-config-data\") pod \"swift-proxy-c7bbc5ff-rlprn\" (UID: \"644890b3-8cae-45bc-b3eb-e0e6bfa8d957\") " pod="openstack/swift-proxy-c7bbc5ff-rlprn" Dec 10 23:09:41 crc kubenswrapper[4791]: I1210 23:09:41.719776 4791 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/644890b3-8cae-45bc-b3eb-e0e6bfa8d957-run-httpd\") pod \"swift-proxy-c7bbc5ff-rlprn\" (UID: \"644890b3-8cae-45bc-b3eb-e0e6bfa8d957\") " pod="openstack/swift-proxy-c7bbc5ff-rlprn" Dec 10 23:09:41 crc kubenswrapper[4791]: I1210 23:09:41.720207 4791 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/644890b3-8cae-45bc-b3eb-e0e6bfa8d957-log-httpd\") pod \"swift-proxy-c7bbc5ff-rlprn\" (UID: \"644890b3-8cae-45bc-b3eb-e0e6bfa8d957\") " pod="openstack/swift-proxy-c7bbc5ff-rlprn" Dec 10 23:09:41 crc kubenswrapper[4791]: I1210 23:09:41.726391 4791 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/644890b3-8cae-45bc-b3eb-e0e6bfa8d957-config-data\") pod \"swift-proxy-c7bbc5ff-rlprn\" (UID: \"644890b3-8cae-45bc-b3eb-e0e6bfa8d957\") " pod="openstack/swift-proxy-c7bbc5ff-rlprn" Dec 10 23:09:41 crc kubenswrapper[4791]: I1210 23:09:41.728445 4791 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-swift\" (UniqueName: \"kubernetes.io/projected/644890b3-8cae-45bc-b3eb-e0e6bfa8d957-etc-swift\") pod \"swift-proxy-c7bbc5ff-rlprn\" (UID: \"644890b3-8cae-45bc-b3eb-e0e6bfa8d957\") " pod="openstack/swift-proxy-c7bbc5ff-rlprn" Dec 10 23:09:41 crc kubenswrapper[4791]: I1210 23:09:41.729666 4791 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/644890b3-8cae-45bc-b3eb-e0e6bfa8d957-public-tls-certs\") pod \"swift-proxy-c7bbc5ff-rlprn\" (UID: \"644890b3-8cae-45bc-b3eb-e0e6bfa8d957\") " pod="openstack/swift-proxy-c7bbc5ff-rlprn" Dec 10 23:09:41 crc kubenswrapper[4791]: I1210 23:09:41.738835 4791 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/644890b3-8cae-45bc-b3eb-e0e6bfa8d957-internal-tls-certs\") pod \"swift-proxy-c7bbc5ff-rlprn\" (UID: \"644890b3-8cae-45bc-b3eb-e0e6bfa8d957\") " pod="openstack/swift-proxy-c7bbc5ff-rlprn" Dec 10 23:09:41 crc kubenswrapper[4791]: I1210 23:09:41.742038 4791 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/644890b3-8cae-45bc-b3eb-e0e6bfa8d957-combined-ca-bundle\") pod \"swift-proxy-c7bbc5ff-rlprn\" (UID: \"644890b3-8cae-45bc-b3eb-e0e6bfa8d957\") " pod="openstack/swift-proxy-c7bbc5ff-rlprn" Dec 10 23:09:41 crc kubenswrapper[4791]: I1210 23:09:41.755314 4791 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/barbican-api-795f5df4db-xf57n" podUID="9b340efd-bf7e-4caf-8c76-ddc500aa2085" containerName="barbican-api" probeResult="failure" output="Get \"http://10.217.0.163:9311/healthcheck\": read tcp 10.217.0.2:38312->10.217.0.163:9311: read: connection reset by peer" Dec 10 23:09:41 crc kubenswrapper[4791]: I1210 23:09:41.755368 4791 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/barbican-api-795f5df4db-xf57n" podUID="9b340efd-bf7e-4caf-8c76-ddc500aa2085" containerName="barbican-api-log" probeResult="failure" output="Get \"http://10.217.0.163:9311/healthcheck\": read tcp 10.217.0.2:38326->10.217.0.163:9311: read: connection reset by peer" Dec 10 23:09:41 crc kubenswrapper[4791]: I1210 23:09:41.755451 4791 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/barbican-api-795f5df4db-xf57n" Dec 10 23:09:41 crc kubenswrapper[4791]: I1210 23:09:41.757560 4791 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-vmctf\" (UniqueName: \"kubernetes.io/projected/644890b3-8cae-45bc-b3eb-e0e6bfa8d957-kube-api-access-vmctf\") pod \"swift-proxy-c7bbc5ff-rlprn\" (UID: \"644890b3-8cae-45bc-b3eb-e0e6bfa8d957\") " pod="openstack/swift-proxy-c7bbc5ff-rlprn" Dec 10 23:09:41 crc kubenswrapper[4791]: I1210 23:09:41.859808 4791 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/swift-proxy-c7bbc5ff-rlprn" Dec 10 23:09:42 crc kubenswrapper[4791]: I1210 23:09:42.391441 4791 generic.go:334] "Generic (PLEG): container finished" podID="9b340efd-bf7e-4caf-8c76-ddc500aa2085" containerID="31193dfd2f9e85871df3d79e4aef1d62fa1b1a1b269896d2e71dbdb4f4784a21" exitCode=0 Dec 10 23:09:42 crc kubenswrapper[4791]: I1210 23:09:42.391591 4791 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-api-795f5df4db-xf57n" event={"ID":"9b340efd-bf7e-4caf-8c76-ddc500aa2085","Type":"ContainerDied","Data":"31193dfd2f9e85871df3d79e4aef1d62fa1b1a1b269896d2e71dbdb4f4784a21"} Dec 10 23:09:42 crc kubenswrapper[4791]: I1210 23:09:42.881719 4791 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ceilometer-0"] Dec 10 23:09:42 crc kubenswrapper[4791]: I1210 23:09:42.881969 4791 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="58167832-d8da-48de-8b8b-cf51fc54d92c" containerName="ceilometer-central-agent" containerID="cri-o://b32ac48681950cbc60cb9f131ea5fd50ebfc4c397b7e8f40d8325d077bbfcbcf" gracePeriod=30 Dec 10 23:09:42 crc kubenswrapper[4791]: I1210 23:09:42.882035 4791 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="58167832-d8da-48de-8b8b-cf51fc54d92c" containerName="proxy-httpd" containerID="cri-o://d3db4aa0578d2e480256777270f40b9150d19368ae770c08c87f98f42a115b57" gracePeriod=30 Dec 10 23:09:42 crc kubenswrapper[4791]: I1210 23:09:42.882071 4791 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="58167832-d8da-48de-8b8b-cf51fc54d92c" containerName="sg-core" containerID="cri-o://d7021f57af9ab5eef5206bcbbcf0c53acba5e7dfcef13ecaa64e06f7d4fe6fc2" gracePeriod=30 Dec 10 23:09:42 crc kubenswrapper[4791]: I1210 23:09:42.882105 4791 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="58167832-d8da-48de-8b8b-cf51fc54d92c" containerName="ceilometer-notification-agent" containerID="cri-o://b8f078f30aed9ddb6d5815433c894f5ac4a00e3c7489055d52e0cb4879bb16cf" gracePeriod=30 Dec 10 23:09:43 crc kubenswrapper[4791]: I1210 23:09:43.406257 4791 generic.go:334] "Generic (PLEG): container finished" podID="58167832-d8da-48de-8b8b-cf51fc54d92c" containerID="d3db4aa0578d2e480256777270f40b9150d19368ae770c08c87f98f42a115b57" exitCode=0 Dec 10 23:09:43 crc kubenswrapper[4791]: I1210 23:09:43.406292 4791 generic.go:334] "Generic (PLEG): container finished" podID="58167832-d8da-48de-8b8b-cf51fc54d92c" containerID="d7021f57af9ab5eef5206bcbbcf0c53acba5e7dfcef13ecaa64e06f7d4fe6fc2" exitCode=2 Dec 10 23:09:43 crc kubenswrapper[4791]: I1210 23:09:43.406314 4791 generic.go:334] "Generic (PLEG): container finished" podID="58167832-d8da-48de-8b8b-cf51fc54d92c" containerID="b8f078f30aed9ddb6d5815433c894f5ac4a00e3c7489055d52e0cb4879bb16cf" exitCode=0 Dec 10 23:09:43 crc kubenswrapper[4791]: I1210 23:09:43.406321 4791 generic.go:334] "Generic (PLEG): container finished" podID="58167832-d8da-48de-8b8b-cf51fc54d92c" containerID="b32ac48681950cbc60cb9f131ea5fd50ebfc4c397b7e8f40d8325d077bbfcbcf" exitCode=0 Dec 10 23:09:43 crc kubenswrapper[4791]: I1210 23:09:43.406356 4791 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"58167832-d8da-48de-8b8b-cf51fc54d92c","Type":"ContainerDied","Data":"d3db4aa0578d2e480256777270f40b9150d19368ae770c08c87f98f42a115b57"} Dec 10 23:09:43 crc kubenswrapper[4791]: I1210 23:09:43.406413 4791 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"58167832-d8da-48de-8b8b-cf51fc54d92c","Type":"ContainerDied","Data":"d7021f57af9ab5eef5206bcbbcf0c53acba5e7dfcef13ecaa64e06f7d4fe6fc2"} Dec 10 23:09:43 crc kubenswrapper[4791]: I1210 23:09:43.406427 4791 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"58167832-d8da-48de-8b8b-cf51fc54d92c","Type":"ContainerDied","Data":"b8f078f30aed9ddb6d5815433c894f5ac4a00e3c7489055d52e0cb4879bb16cf"} Dec 10 23:09:43 crc kubenswrapper[4791]: I1210 23:09:43.406440 4791 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"58167832-d8da-48de-8b8b-cf51fc54d92c","Type":"ContainerDied","Data":"b32ac48681950cbc60cb9f131ea5fd50ebfc4c397b7e8f40d8325d077bbfcbcf"} Dec 10 23:09:45 crc kubenswrapper[4791]: I1210 23:09:45.792668 4791 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/barbican-api-795f5df4db-xf57n" podUID="9b340efd-bf7e-4caf-8c76-ddc500aa2085" containerName="barbican-api" probeResult="failure" output="Get \"http://10.217.0.163:9311/healthcheck\": dial tcp 10.217.0.163:9311: connect: connection refused" Dec 10 23:09:45 crc kubenswrapper[4791]: I1210 23:09:45.792687 4791 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/barbican-api-795f5df4db-xf57n" podUID="9b340efd-bf7e-4caf-8c76-ddc500aa2085" containerName="barbican-api-log" probeResult="failure" output="Get \"http://10.217.0.163:9311/healthcheck\": dial tcp 10.217.0.163:9311: connect: connection refused" Dec 10 23:09:46 crc kubenswrapper[4791]: I1210 23:09:46.177886 4791 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-api-795f5df4db-xf57n" Dec 10 23:09:46 crc kubenswrapper[4791]: I1210 23:09:46.307868 4791 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/9b340efd-bf7e-4caf-8c76-ddc500aa2085-config-data-custom\") pod \"9b340efd-bf7e-4caf-8c76-ddc500aa2085\" (UID: \"9b340efd-bf7e-4caf-8c76-ddc500aa2085\") " Dec 10 23:09:46 crc kubenswrapper[4791]: I1210 23:09:46.307935 4791 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-vnldb\" (UniqueName: \"kubernetes.io/projected/9b340efd-bf7e-4caf-8c76-ddc500aa2085-kube-api-access-vnldb\") pod \"9b340efd-bf7e-4caf-8c76-ddc500aa2085\" (UID: \"9b340efd-bf7e-4caf-8c76-ddc500aa2085\") " Dec 10 23:09:46 crc kubenswrapper[4791]: I1210 23:09:46.308031 4791 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/9b340efd-bf7e-4caf-8c76-ddc500aa2085-config-data\") pod \"9b340efd-bf7e-4caf-8c76-ddc500aa2085\" (UID: \"9b340efd-bf7e-4caf-8c76-ddc500aa2085\") " Dec 10 23:09:46 crc kubenswrapper[4791]: I1210 23:09:46.308079 4791 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/9b340efd-bf7e-4caf-8c76-ddc500aa2085-logs\") pod \"9b340efd-bf7e-4caf-8c76-ddc500aa2085\" (UID: \"9b340efd-bf7e-4caf-8c76-ddc500aa2085\") " Dec 10 23:09:46 crc kubenswrapper[4791]: I1210 23:09:46.308116 4791 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/9b340efd-bf7e-4caf-8c76-ddc500aa2085-combined-ca-bundle\") pod \"9b340efd-bf7e-4caf-8c76-ddc500aa2085\" (UID: \"9b340efd-bf7e-4caf-8c76-ddc500aa2085\") " Dec 10 23:09:46 crc kubenswrapper[4791]: I1210 23:09:46.315297 4791 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/9b340efd-bf7e-4caf-8c76-ddc500aa2085-logs" (OuterVolumeSpecName: "logs") pod "9b340efd-bf7e-4caf-8c76-ddc500aa2085" (UID: "9b340efd-bf7e-4caf-8c76-ddc500aa2085"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 10 23:09:46 crc kubenswrapper[4791]: I1210 23:09:46.319295 4791 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/9b340efd-bf7e-4caf-8c76-ddc500aa2085-kube-api-access-vnldb" (OuterVolumeSpecName: "kube-api-access-vnldb") pod "9b340efd-bf7e-4caf-8c76-ddc500aa2085" (UID: "9b340efd-bf7e-4caf-8c76-ddc500aa2085"). InnerVolumeSpecName "kube-api-access-vnldb". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 10 23:09:46 crc kubenswrapper[4791]: I1210 23:09:46.355285 4791 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/9b340efd-bf7e-4caf-8c76-ddc500aa2085-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "9b340efd-bf7e-4caf-8c76-ddc500aa2085" (UID: "9b340efd-bf7e-4caf-8c76-ddc500aa2085"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 10 23:09:46 crc kubenswrapper[4791]: I1210 23:09:46.355283 4791 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/9b340efd-bf7e-4caf-8c76-ddc500aa2085-config-data-custom" (OuterVolumeSpecName: "config-data-custom") pod "9b340efd-bf7e-4caf-8c76-ddc500aa2085" (UID: "9b340efd-bf7e-4caf-8c76-ddc500aa2085"). InnerVolumeSpecName "config-data-custom". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 10 23:09:46 crc kubenswrapper[4791]: I1210 23:09:46.388144 4791 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Dec 10 23:09:46 crc kubenswrapper[4791]: I1210 23:09:46.389577 4791 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/9b340efd-bf7e-4caf-8c76-ddc500aa2085-config-data" (OuterVolumeSpecName: "config-data") pod "9b340efd-bf7e-4caf-8c76-ddc500aa2085" (UID: "9b340efd-bf7e-4caf-8c76-ddc500aa2085"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 10 23:09:46 crc kubenswrapper[4791]: I1210 23:09:46.409782 4791 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-vnldb\" (UniqueName: \"kubernetes.io/projected/9b340efd-bf7e-4caf-8c76-ddc500aa2085-kube-api-access-vnldb\") on node \"crc\" DevicePath \"\"" Dec 10 23:09:46 crc kubenswrapper[4791]: I1210 23:09:46.409815 4791 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/9b340efd-bf7e-4caf-8c76-ddc500aa2085-config-data\") on node \"crc\" DevicePath \"\"" Dec 10 23:09:46 crc kubenswrapper[4791]: I1210 23:09:46.409826 4791 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/9b340efd-bf7e-4caf-8c76-ddc500aa2085-logs\") on node \"crc\" DevicePath \"\"" Dec 10 23:09:46 crc kubenswrapper[4791]: I1210 23:09:46.409834 4791 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/9b340efd-bf7e-4caf-8c76-ddc500aa2085-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 10 23:09:46 crc kubenswrapper[4791]: I1210 23:09:46.409843 4791 reconciler_common.go:293] "Volume detached for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/9b340efd-bf7e-4caf-8c76-ddc500aa2085-config-data-custom\") on node \"crc\" DevicePath \"\"" Dec 10 23:09:46 crc kubenswrapper[4791]: I1210 23:09:46.447060 4791 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"58167832-d8da-48de-8b8b-cf51fc54d92c","Type":"ContainerDied","Data":"ca930728dc321c5022ce2263f6f361be1de64fd0569818e4c3e7dca6b76ec3df"} Dec 10 23:09:46 crc kubenswrapper[4791]: I1210 23:09:46.447288 4791 scope.go:117] "RemoveContainer" containerID="d3db4aa0578d2e480256777270f40b9150d19368ae770c08c87f98f42a115b57" Dec 10 23:09:46 crc kubenswrapper[4791]: I1210 23:09:46.447531 4791 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Dec 10 23:09:46 crc kubenswrapper[4791]: I1210 23:09:46.451811 4791 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-api-795f5df4db-xf57n" event={"ID":"9b340efd-bf7e-4caf-8c76-ddc500aa2085","Type":"ContainerDied","Data":"8f95cb70af18d0c8b0d71d55cc909fb5e15dbe337f92069cd49742ac1b44c594"} Dec 10 23:09:46 crc kubenswrapper[4791]: I1210 23:09:46.451973 4791 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-api-795f5df4db-xf57n" Dec 10 23:09:46 crc kubenswrapper[4791]: I1210 23:09:46.465397 4791 scope.go:117] "RemoveContainer" containerID="d7021f57af9ab5eef5206bcbbcf0c53acba5e7dfcef13ecaa64e06f7d4fe6fc2" Dec 10 23:09:46 crc kubenswrapper[4791]: I1210 23:09:46.496632 4791 scope.go:117] "RemoveContainer" containerID="b8f078f30aed9ddb6d5815433c894f5ac4a00e3c7489055d52e0cb4879bb16cf" Dec 10 23:09:46 crc kubenswrapper[4791]: I1210 23:09:46.510598 4791 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/58167832-d8da-48de-8b8b-cf51fc54d92c-scripts\") pod \"58167832-d8da-48de-8b8b-cf51fc54d92c\" (UID: \"58167832-d8da-48de-8b8b-cf51fc54d92c\") " Dec 10 23:09:46 crc kubenswrapper[4791]: I1210 23:09:46.510666 4791 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-mxm4d\" (UniqueName: \"kubernetes.io/projected/58167832-d8da-48de-8b8b-cf51fc54d92c-kube-api-access-mxm4d\") pod \"58167832-d8da-48de-8b8b-cf51fc54d92c\" (UID: \"58167832-d8da-48de-8b8b-cf51fc54d92c\") " Dec 10 23:09:46 crc kubenswrapper[4791]: I1210 23:09:46.510708 4791 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/58167832-d8da-48de-8b8b-cf51fc54d92c-log-httpd\") pod \"58167832-d8da-48de-8b8b-cf51fc54d92c\" (UID: \"58167832-d8da-48de-8b8b-cf51fc54d92c\") " Dec 10 23:09:46 crc kubenswrapper[4791]: I1210 23:09:46.510788 4791 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/58167832-d8da-48de-8b8b-cf51fc54d92c-sg-core-conf-yaml\") pod \"58167832-d8da-48de-8b8b-cf51fc54d92c\" (UID: \"58167832-d8da-48de-8b8b-cf51fc54d92c\") " Dec 10 23:09:46 crc kubenswrapper[4791]: I1210 23:09:46.510819 4791 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/58167832-d8da-48de-8b8b-cf51fc54d92c-config-data\") pod \"58167832-d8da-48de-8b8b-cf51fc54d92c\" (UID: \"58167832-d8da-48de-8b8b-cf51fc54d92c\") " Dec 10 23:09:46 crc kubenswrapper[4791]: I1210 23:09:46.510868 4791 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/58167832-d8da-48de-8b8b-cf51fc54d92c-combined-ca-bundle\") pod \"58167832-d8da-48de-8b8b-cf51fc54d92c\" (UID: \"58167832-d8da-48de-8b8b-cf51fc54d92c\") " Dec 10 23:09:46 crc kubenswrapper[4791]: I1210 23:09:46.510953 4791 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/58167832-d8da-48de-8b8b-cf51fc54d92c-run-httpd\") pod \"58167832-d8da-48de-8b8b-cf51fc54d92c\" (UID: \"58167832-d8da-48de-8b8b-cf51fc54d92c\") " Dec 10 23:09:46 crc kubenswrapper[4791]: I1210 23:09:46.511881 4791 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/58167832-d8da-48de-8b8b-cf51fc54d92c-run-httpd" (OuterVolumeSpecName: "run-httpd") pod "58167832-d8da-48de-8b8b-cf51fc54d92c" (UID: "58167832-d8da-48de-8b8b-cf51fc54d92c"). InnerVolumeSpecName "run-httpd". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 10 23:09:46 crc kubenswrapper[4791]: I1210 23:09:46.520966 4791 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/58167832-d8da-48de-8b8b-cf51fc54d92c-log-httpd" (OuterVolumeSpecName: "log-httpd") pod "58167832-d8da-48de-8b8b-cf51fc54d92c" (UID: "58167832-d8da-48de-8b8b-cf51fc54d92c"). InnerVolumeSpecName "log-httpd". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 10 23:09:46 crc kubenswrapper[4791]: I1210 23:09:46.521654 4791 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/58167832-d8da-48de-8b8b-cf51fc54d92c-scripts" (OuterVolumeSpecName: "scripts") pod "58167832-d8da-48de-8b8b-cf51fc54d92c" (UID: "58167832-d8da-48de-8b8b-cf51fc54d92c"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 10 23:09:46 crc kubenswrapper[4791]: I1210 23:09:46.524974 4791 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/barbican-api-795f5df4db-xf57n"] Dec 10 23:09:46 crc kubenswrapper[4791]: I1210 23:09:46.529055 4791 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/58167832-d8da-48de-8b8b-cf51fc54d92c-kube-api-access-mxm4d" (OuterVolumeSpecName: "kube-api-access-mxm4d") pod "58167832-d8da-48de-8b8b-cf51fc54d92c" (UID: "58167832-d8da-48de-8b8b-cf51fc54d92c"). InnerVolumeSpecName "kube-api-access-mxm4d". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 10 23:09:46 crc kubenswrapper[4791]: I1210 23:09:46.545103 4791 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/barbican-api-795f5df4db-xf57n"] Dec 10 23:09:46 crc kubenswrapper[4791]: I1210 23:09:46.554849 4791 scope.go:117] "RemoveContainer" containerID="b32ac48681950cbc60cb9f131ea5fd50ebfc4c397b7e8f40d8325d077bbfcbcf" Dec 10 23:09:46 crc kubenswrapper[4791]: I1210 23:09:46.562654 4791 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/58167832-d8da-48de-8b8b-cf51fc54d92c-sg-core-conf-yaml" (OuterVolumeSpecName: "sg-core-conf-yaml") pod "58167832-d8da-48de-8b8b-cf51fc54d92c" (UID: "58167832-d8da-48de-8b8b-cf51fc54d92c"). InnerVolumeSpecName "sg-core-conf-yaml". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 10 23:09:46 crc kubenswrapper[4791]: I1210 23:09:46.572234 4791 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-api-db-create-cvjnf"] Dec 10 23:09:46 crc kubenswrapper[4791]: I1210 23:09:46.584767 4791 scope.go:117] "RemoveContainer" containerID="31193dfd2f9e85871df3d79e4aef1d62fa1b1a1b269896d2e71dbdb4f4784a21" Dec 10 23:09:46 crc kubenswrapper[4791]: I1210 23:09:46.613403 4791 reconciler_common.go:293] "Volume detached for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/58167832-d8da-48de-8b8b-cf51fc54d92c-run-httpd\") on node \"crc\" DevicePath \"\"" Dec 10 23:09:46 crc kubenswrapper[4791]: I1210 23:09:46.613627 4791 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/58167832-d8da-48de-8b8b-cf51fc54d92c-scripts\") on node \"crc\" DevicePath \"\"" Dec 10 23:09:46 crc kubenswrapper[4791]: I1210 23:09:46.613716 4791 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-mxm4d\" (UniqueName: \"kubernetes.io/projected/58167832-d8da-48de-8b8b-cf51fc54d92c-kube-api-access-mxm4d\") on node \"crc\" DevicePath \"\"" Dec 10 23:09:46 crc kubenswrapper[4791]: I1210 23:09:46.613818 4791 reconciler_common.go:293] "Volume detached for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/58167832-d8da-48de-8b8b-cf51fc54d92c-log-httpd\") on node \"crc\" DevicePath \"\"" Dec 10 23:09:46 crc kubenswrapper[4791]: I1210 23:09:46.613896 4791 reconciler_common.go:293] "Volume detached for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/58167832-d8da-48de-8b8b-cf51fc54d92c-sg-core-conf-yaml\") on node \"crc\" DevicePath \"\"" Dec 10 23:09:46 crc kubenswrapper[4791]: I1210 23:09:46.621895 4791 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/58167832-d8da-48de-8b8b-cf51fc54d92c-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "58167832-d8da-48de-8b8b-cf51fc54d92c" (UID: "58167832-d8da-48de-8b8b-cf51fc54d92c"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 10 23:09:46 crc kubenswrapper[4791]: I1210 23:09:46.626191 4791 scope.go:117] "RemoveContainer" containerID="f05d3eef09f65c594818648ae666e092eb1a3e05925b8b910f123fdb1170cde9" Dec 10 23:09:46 crc kubenswrapper[4791]: I1210 23:09:46.652509 4791 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/58167832-d8da-48de-8b8b-cf51fc54d92c-config-data" (OuterVolumeSpecName: "config-data") pod "58167832-d8da-48de-8b8b-cf51fc54d92c" (UID: "58167832-d8da-48de-8b8b-cf51fc54d92c"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 10 23:09:46 crc kubenswrapper[4791]: I1210 23:09:46.716705 4791 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/58167832-d8da-48de-8b8b-cf51fc54d92c-config-data\") on node \"crc\" DevicePath \"\"" Dec 10 23:09:46 crc kubenswrapper[4791]: I1210 23:09:46.716740 4791 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/58167832-d8da-48de-8b8b-cf51fc54d92c-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 10 23:09:46 crc kubenswrapper[4791]: I1210 23:09:46.787575 4791 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ceilometer-0"] Dec 10 23:09:46 crc kubenswrapper[4791]: I1210 23:09:46.806899 4791 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/ceilometer-0"] Dec 10 23:09:46 crc kubenswrapper[4791]: I1210 23:09:46.823252 4791 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ceilometer-0"] Dec 10 23:09:46 crc kubenswrapper[4791]: E1210 23:09:46.823677 4791 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="9b340efd-bf7e-4caf-8c76-ddc500aa2085" containerName="barbican-api" Dec 10 23:09:46 crc kubenswrapper[4791]: I1210 23:09:46.823704 4791 state_mem.go:107] "Deleted CPUSet assignment" podUID="9b340efd-bf7e-4caf-8c76-ddc500aa2085" containerName="barbican-api" Dec 10 23:09:46 crc kubenswrapper[4791]: E1210 23:09:46.823727 4791 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="9b340efd-bf7e-4caf-8c76-ddc500aa2085" containerName="barbican-api-log" Dec 10 23:09:46 crc kubenswrapper[4791]: I1210 23:09:46.823737 4791 state_mem.go:107] "Deleted CPUSet assignment" podUID="9b340efd-bf7e-4caf-8c76-ddc500aa2085" containerName="barbican-api-log" Dec 10 23:09:46 crc kubenswrapper[4791]: E1210 23:09:46.823747 4791 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="58167832-d8da-48de-8b8b-cf51fc54d92c" containerName="ceilometer-central-agent" Dec 10 23:09:46 crc kubenswrapper[4791]: I1210 23:09:46.823755 4791 state_mem.go:107] "Deleted CPUSet assignment" podUID="58167832-d8da-48de-8b8b-cf51fc54d92c" containerName="ceilometer-central-agent" Dec 10 23:09:46 crc kubenswrapper[4791]: E1210 23:09:46.823776 4791 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="58167832-d8da-48de-8b8b-cf51fc54d92c" containerName="sg-core" Dec 10 23:09:46 crc kubenswrapper[4791]: I1210 23:09:46.823783 4791 state_mem.go:107] "Deleted CPUSet assignment" podUID="58167832-d8da-48de-8b8b-cf51fc54d92c" containerName="sg-core" Dec 10 23:09:46 crc kubenswrapper[4791]: E1210 23:09:46.823796 4791 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="58167832-d8da-48de-8b8b-cf51fc54d92c" containerName="ceilometer-notification-agent" Dec 10 23:09:46 crc kubenswrapper[4791]: I1210 23:09:46.823804 4791 state_mem.go:107] "Deleted CPUSet assignment" podUID="58167832-d8da-48de-8b8b-cf51fc54d92c" containerName="ceilometer-notification-agent" Dec 10 23:09:46 crc kubenswrapper[4791]: E1210 23:09:46.823825 4791 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="58167832-d8da-48de-8b8b-cf51fc54d92c" containerName="proxy-httpd" Dec 10 23:09:46 crc kubenswrapper[4791]: I1210 23:09:46.823848 4791 state_mem.go:107] "Deleted CPUSet assignment" podUID="58167832-d8da-48de-8b8b-cf51fc54d92c" containerName="proxy-httpd" Dec 10 23:09:46 crc kubenswrapper[4791]: I1210 23:09:46.824039 4791 memory_manager.go:354] "RemoveStaleState removing state" podUID="9b340efd-bf7e-4caf-8c76-ddc500aa2085" containerName="barbican-api" Dec 10 23:09:46 crc kubenswrapper[4791]: I1210 23:09:46.824057 4791 memory_manager.go:354] "RemoveStaleState removing state" podUID="58167832-d8da-48de-8b8b-cf51fc54d92c" containerName="proxy-httpd" Dec 10 23:09:46 crc kubenswrapper[4791]: I1210 23:09:46.824068 4791 memory_manager.go:354] "RemoveStaleState removing state" podUID="9b340efd-bf7e-4caf-8c76-ddc500aa2085" containerName="barbican-api-log" Dec 10 23:09:46 crc kubenswrapper[4791]: I1210 23:09:46.824085 4791 memory_manager.go:354] "RemoveStaleState removing state" podUID="58167832-d8da-48de-8b8b-cf51fc54d92c" containerName="ceilometer-central-agent" Dec 10 23:09:46 crc kubenswrapper[4791]: I1210 23:09:46.824096 4791 memory_manager.go:354] "RemoveStaleState removing state" podUID="58167832-d8da-48de-8b8b-cf51fc54d92c" containerName="sg-core" Dec 10 23:09:46 crc kubenswrapper[4791]: I1210 23:09:46.824106 4791 memory_manager.go:354] "RemoveStaleState removing state" podUID="58167832-d8da-48de-8b8b-cf51fc54d92c" containerName="ceilometer-notification-agent" Dec 10 23:09:46 crc kubenswrapper[4791]: I1210 23:09:46.826574 4791 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Dec 10 23:09:46 crc kubenswrapper[4791]: I1210 23:09:46.830892 4791 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-scripts" Dec 10 23:09:46 crc kubenswrapper[4791]: I1210 23:09:46.831154 4791 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-config-data" Dec 10 23:09:46 crc kubenswrapper[4791]: I1210 23:09:46.834997 4791 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceilometer-0"] Dec 10 23:09:46 crc kubenswrapper[4791]: I1210 23:09:46.852261 4791 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-b007-account-create-update-fkwxb"] Dec 10 23:09:46 crc kubenswrapper[4791]: I1210 23:09:46.878262 4791 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell0-e207-account-create-update-hlmck"] Dec 10 23:09:46 crc kubenswrapper[4791]: I1210 23:09:46.921539 4791 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-wrsg2\" (UniqueName: \"kubernetes.io/projected/053f7590-725f-4504-a204-be83d801bad2-kube-api-access-wrsg2\") pod \"ceilometer-0\" (UID: \"053f7590-725f-4504-a204-be83d801bad2\") " pod="openstack/ceilometer-0" Dec 10 23:09:46 crc kubenswrapper[4791]: I1210 23:09:46.921643 4791 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/053f7590-725f-4504-a204-be83d801bad2-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"053f7590-725f-4504-a204-be83d801bad2\") " pod="openstack/ceilometer-0" Dec 10 23:09:46 crc kubenswrapper[4791]: I1210 23:09:46.921681 4791 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/053f7590-725f-4504-a204-be83d801bad2-scripts\") pod \"ceilometer-0\" (UID: \"053f7590-725f-4504-a204-be83d801bad2\") " pod="openstack/ceilometer-0" Dec 10 23:09:46 crc kubenswrapper[4791]: I1210 23:09:46.921726 4791 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/053f7590-725f-4504-a204-be83d801bad2-run-httpd\") pod \"ceilometer-0\" (UID: \"053f7590-725f-4504-a204-be83d801bad2\") " pod="openstack/ceilometer-0" Dec 10 23:09:46 crc kubenswrapper[4791]: I1210 23:09:46.921752 4791 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/053f7590-725f-4504-a204-be83d801bad2-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"053f7590-725f-4504-a204-be83d801bad2\") " pod="openstack/ceilometer-0" Dec 10 23:09:46 crc kubenswrapper[4791]: I1210 23:09:46.921800 4791 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/053f7590-725f-4504-a204-be83d801bad2-config-data\") pod \"ceilometer-0\" (UID: \"053f7590-725f-4504-a204-be83d801bad2\") " pod="openstack/ceilometer-0" Dec 10 23:09:46 crc kubenswrapper[4791]: I1210 23:09:46.921832 4791 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/053f7590-725f-4504-a204-be83d801bad2-log-httpd\") pod \"ceilometer-0\" (UID: \"053f7590-725f-4504-a204-be83d801bad2\") " pod="openstack/ceilometer-0" Dec 10 23:09:47 crc kubenswrapper[4791]: I1210 23:09:47.005950 4791 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell0-db-create-z2v58"] Dec 10 23:09:47 crc kubenswrapper[4791]: I1210 23:09:47.025707 4791 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-wrsg2\" (UniqueName: \"kubernetes.io/projected/053f7590-725f-4504-a204-be83d801bad2-kube-api-access-wrsg2\") pod \"ceilometer-0\" (UID: \"053f7590-725f-4504-a204-be83d801bad2\") " pod="openstack/ceilometer-0" Dec 10 23:09:47 crc kubenswrapper[4791]: I1210 23:09:47.025823 4791 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/053f7590-725f-4504-a204-be83d801bad2-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"053f7590-725f-4504-a204-be83d801bad2\") " pod="openstack/ceilometer-0" Dec 10 23:09:47 crc kubenswrapper[4791]: I1210 23:09:47.025865 4791 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/053f7590-725f-4504-a204-be83d801bad2-scripts\") pod \"ceilometer-0\" (UID: \"053f7590-725f-4504-a204-be83d801bad2\") " pod="openstack/ceilometer-0" Dec 10 23:09:47 crc kubenswrapper[4791]: I1210 23:09:47.025915 4791 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/053f7590-725f-4504-a204-be83d801bad2-run-httpd\") pod \"ceilometer-0\" (UID: \"053f7590-725f-4504-a204-be83d801bad2\") " pod="openstack/ceilometer-0" Dec 10 23:09:47 crc kubenswrapper[4791]: I1210 23:09:47.026727 4791 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/053f7590-725f-4504-a204-be83d801bad2-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"053f7590-725f-4504-a204-be83d801bad2\") " pod="openstack/ceilometer-0" Dec 10 23:09:47 crc kubenswrapper[4791]: I1210 23:09:47.026786 4791 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-api-8832-account-create-update-wp7p5"] Dec 10 23:09:47 crc kubenswrapper[4791]: I1210 23:09:47.026836 4791 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/053f7590-725f-4504-a204-be83d801bad2-config-data\") pod \"ceilometer-0\" (UID: \"053f7590-725f-4504-a204-be83d801bad2\") " pod="openstack/ceilometer-0" Dec 10 23:09:47 crc kubenswrapper[4791]: I1210 23:09:47.026877 4791 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/053f7590-725f-4504-a204-be83d801bad2-log-httpd\") pod \"ceilometer-0\" (UID: \"053f7590-725f-4504-a204-be83d801bad2\") " pod="openstack/ceilometer-0" Dec 10 23:09:47 crc kubenswrapper[4791]: I1210 23:09:47.027457 4791 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/053f7590-725f-4504-a204-be83d801bad2-log-httpd\") pod \"ceilometer-0\" (UID: \"053f7590-725f-4504-a204-be83d801bad2\") " pod="openstack/ceilometer-0" Dec 10 23:09:47 crc kubenswrapper[4791]: I1210 23:09:47.027729 4791 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/053f7590-725f-4504-a204-be83d801bad2-run-httpd\") pod \"ceilometer-0\" (UID: \"053f7590-725f-4504-a204-be83d801bad2\") " pod="openstack/ceilometer-0" Dec 10 23:09:47 crc kubenswrapper[4791]: I1210 23:09:47.038996 4791 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/053f7590-725f-4504-a204-be83d801bad2-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"053f7590-725f-4504-a204-be83d801bad2\") " pod="openstack/ceilometer-0" Dec 10 23:09:47 crc kubenswrapper[4791]: I1210 23:09:47.040432 4791 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/053f7590-725f-4504-a204-be83d801bad2-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"053f7590-725f-4504-a204-be83d801bad2\") " pod="openstack/ceilometer-0" Dec 10 23:09:47 crc kubenswrapper[4791]: I1210 23:09:47.042128 4791 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/053f7590-725f-4504-a204-be83d801bad2-scripts\") pod \"ceilometer-0\" (UID: \"053f7590-725f-4504-a204-be83d801bad2\") " pod="openstack/ceilometer-0" Dec 10 23:09:47 crc kubenswrapper[4791]: I1210 23:09:47.043006 4791 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/053f7590-725f-4504-a204-be83d801bad2-config-data\") pod \"ceilometer-0\" (UID: \"053f7590-725f-4504-a204-be83d801bad2\") " pod="openstack/ceilometer-0" Dec 10 23:09:47 crc kubenswrapper[4791]: I1210 23:09:47.046148 4791 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/swift-proxy-c7bbc5ff-rlprn"] Dec 10 23:09:47 crc kubenswrapper[4791]: I1210 23:09:47.054710 4791 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-wrsg2\" (UniqueName: \"kubernetes.io/projected/053f7590-725f-4504-a204-be83d801bad2-kube-api-access-wrsg2\") pod \"ceilometer-0\" (UID: \"053f7590-725f-4504-a204-be83d801bad2\") " pod="openstack/ceilometer-0" Dec 10 23:09:47 crc kubenswrapper[4791]: I1210 23:09:47.061973 4791 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-db-create-tz8dr"] Dec 10 23:09:47 crc kubenswrapper[4791]: I1210 23:09:47.295639 4791 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Dec 10 23:09:47 crc kubenswrapper[4791]: I1210 23:09:47.470510 4791 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-8832-account-create-update-wp7p5" event={"ID":"0a94efe1-86a8-4768-9199-f47051935bcf","Type":"ContainerStarted","Data":"473c10a544d131d442414c7e9d016b5e15a352a6c22d2a9114caba72711977f5"} Dec 10 23:09:47 crc kubenswrapper[4791]: I1210 23:09:47.470923 4791 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-8832-account-create-update-wp7p5" event={"ID":"0a94efe1-86a8-4768-9199-f47051935bcf","Type":"ContainerStarted","Data":"4586f22cc6cca6c7efa8850ef98273336200096b846dee143e924162d165de69"} Dec 10 23:09:47 crc kubenswrapper[4791]: I1210 23:09:47.478567 4791 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-b007-account-create-update-fkwxb" event={"ID":"83ddd1be-5119-45a6-a080-1b63f0c8f7a1","Type":"ContainerStarted","Data":"25ee7ca4ad03fcd0c892fe328f36b3e3b779fd605dd652f8b53cf1cf97a7e30e"} Dec 10 23:09:47 crc kubenswrapper[4791]: I1210 23:09:47.478606 4791 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-b007-account-create-update-fkwxb" event={"ID":"83ddd1be-5119-45a6-a080-1b63f0c8f7a1","Type":"ContainerStarted","Data":"9b4ec8ea53c365d50c632d7add917b0aab3ece5693497f267402b895a87f4e84"} Dec 10 23:09:47 crc kubenswrapper[4791]: I1210 23:09:47.502798 4791 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/openstackclient" event={"ID":"2d25921f-5a9e-4e3b-9c51-f07f95aa02a1","Type":"ContainerStarted","Data":"1a15b26de6961ad3710b1b466d6c96ac29e6c67e9c3aec2b3781f233fd24aa44"} Dec 10 23:09:47 crc kubenswrapper[4791]: I1210 23:09:47.516131 4791 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-db-create-z2v58" event={"ID":"b8b6cd29-edfe-4177-bf14-14b8580d86c2","Type":"ContainerStarted","Data":"fe91f12d8f47a3a6a0ed46cd6dbd4e49f33e7fde61170871d208c67fa35d5b77"} Dec 10 23:09:47 crc kubenswrapper[4791]: I1210 23:09:47.517066 4791 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-api-8832-account-create-update-wp7p5" podStartSLOduration=7.517042549 podStartE2EDuration="7.517042549s" podCreationTimestamp="2025-12-10 23:09:40 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-10 23:09:47.493857239 +0000 UTC m=+1221.923474852" watchObservedRunningTime="2025-12-10 23:09:47.517042549 +0000 UTC m=+1221.946660162" Dec 10 23:09:47 crc kubenswrapper[4791]: I1210 23:09:47.539431 4791 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-cell1-b007-account-create-update-fkwxb" podStartSLOduration=7.539413125 podStartE2EDuration="7.539413125s" podCreationTimestamp="2025-12-10 23:09:40 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-10 23:09:47.509269584 +0000 UTC m=+1221.938887197" watchObservedRunningTime="2025-12-10 23:09:47.539413125 +0000 UTC m=+1221.969030738" Dec 10 23:09:47 crc kubenswrapper[4791]: I1210 23:09:47.558039 4791 generic.go:334] "Generic (PLEG): container finished" podID="18c84e26-d0bb-4cd5-ba6a-49bf21155889" containerID="d9f21c9e92337da1466d284b85b264870b7e25dac547e118e080479b8e75affa" exitCode=0 Dec 10 23:09:47 crc kubenswrapper[4791]: I1210 23:09:47.558146 4791 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-db-create-cvjnf" event={"ID":"18c84e26-d0bb-4cd5-ba6a-49bf21155889","Type":"ContainerDied","Data":"d9f21c9e92337da1466d284b85b264870b7e25dac547e118e080479b8e75affa"} Dec 10 23:09:47 crc kubenswrapper[4791]: I1210 23:09:47.558171 4791 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-db-create-cvjnf" event={"ID":"18c84e26-d0bb-4cd5-ba6a-49bf21155889","Type":"ContainerStarted","Data":"db806c62ccdcf7f92f63d063ef3d251ac98177541edfea6b3e53e454f5d15e96"} Dec 10 23:09:47 crc kubenswrapper[4791]: I1210 23:09:47.559164 4791 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/openstackclient" podStartSLOduration=2.594655603 podStartE2EDuration="14.559119243s" podCreationTimestamp="2025-12-10 23:09:33 +0000 UTC" firstStartedPulling="2025-12-10 23:09:34.108205873 +0000 UTC m=+1208.537823486" lastFinishedPulling="2025-12-10 23:09:46.072669513 +0000 UTC m=+1220.502287126" observedRunningTime="2025-12-10 23:09:47.532759122 +0000 UTC m=+1221.962376745" watchObservedRunningTime="2025-12-10 23:09:47.559119243 +0000 UTC m=+1221.988736856" Dec 10 23:09:47 crc kubenswrapper[4791]: I1210 23:09:47.559801 4791 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-proxy-c7bbc5ff-rlprn" event={"ID":"644890b3-8cae-45bc-b3eb-e0e6bfa8d957","Type":"ContainerStarted","Data":"99f2e35a579fb20957d64063ef1c18924255eb4fb61b23cc99e658dc20c96fce"} Dec 10 23:09:47 crc kubenswrapper[4791]: I1210 23:09:47.559820 4791 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-proxy-c7bbc5ff-rlprn" event={"ID":"644890b3-8cae-45bc-b3eb-e0e6bfa8d957","Type":"ContainerStarted","Data":"49aacfc20393e45fa7adf117d10b8a7b2eece78a63319e4ed8912557491fdd70"} Dec 10 23:09:47 crc kubenswrapper[4791]: I1210 23:09:47.560668 4791 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-db-create-tz8dr" event={"ID":"100aa755-c120-4c5a-9135-23bb98661f3d","Type":"ContainerStarted","Data":"e63790203ba9fd9cd2790cabecfb42628c365964dca974e26ff03d91ff75d79d"} Dec 10 23:09:47 crc kubenswrapper[4791]: I1210 23:09:47.560691 4791 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-db-create-tz8dr" event={"ID":"100aa755-c120-4c5a-9135-23bb98661f3d","Type":"ContainerStarted","Data":"0f38ac94433531ed03f4993d66c9066b954d3e61b35dd4324309de81da6fb0ee"} Dec 10 23:09:47 crc kubenswrapper[4791]: I1210 23:09:47.563278 4791 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-e207-account-create-update-hlmck" event={"ID":"cf0595b8-0690-448d-9ef7-a64227bad653","Type":"ContainerStarted","Data":"6b1df2861daf4c05537a80c73e098b117f3695e83c011f0b5ae42d4c9ac0ddc6"} Dec 10 23:09:47 crc kubenswrapper[4791]: I1210 23:09:47.563303 4791 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-e207-account-create-update-hlmck" event={"ID":"cf0595b8-0690-448d-9ef7-a64227bad653","Type":"ContainerStarted","Data":"0adc35a87f508325498682c7c8f1a921d9cb89d48aa878fde128269d62043983"} Dec 10 23:09:47 crc kubenswrapper[4791]: I1210 23:09:47.585706 4791 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-cell0-db-create-z2v58" podStartSLOduration=7.58568169 podStartE2EDuration="7.58568169s" podCreationTimestamp="2025-12-10 23:09:40 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-10 23:09:47.557057704 +0000 UTC m=+1221.986675317" watchObservedRunningTime="2025-12-10 23:09:47.58568169 +0000 UTC m=+1222.015299303" Dec 10 23:09:47 crc kubenswrapper[4791]: I1210 23:09:47.660185 4791 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-cell0-e207-account-create-update-hlmck" podStartSLOduration=7.66016227 podStartE2EDuration="7.66016227s" podCreationTimestamp="2025-12-10 23:09:40 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-10 23:09:47.606898923 +0000 UTC m=+1222.036516546" watchObservedRunningTime="2025-12-10 23:09:47.66016227 +0000 UTC m=+1222.089779883" Dec 10 23:09:47 crc kubenswrapper[4791]: I1210 23:09:47.675527 4791 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-cell1-db-create-tz8dr" podStartSLOduration=7.675505463 podStartE2EDuration="7.675505463s" podCreationTimestamp="2025-12-10 23:09:40 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-10 23:09:47.62586781 +0000 UTC m=+1222.055485423" watchObservedRunningTime="2025-12-10 23:09:47.675505463 +0000 UTC m=+1222.105123076" Dec 10 23:09:47 crc kubenswrapper[4791]: I1210 23:09:47.850811 4791 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceilometer-0"] Dec 10 23:09:47 crc kubenswrapper[4791]: I1210 23:09:47.910145 4791 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="58167832-d8da-48de-8b8b-cf51fc54d92c" path="/var/lib/kubelet/pods/58167832-d8da-48de-8b8b-cf51fc54d92c/volumes" Dec 10 23:09:47 crc kubenswrapper[4791]: I1210 23:09:47.911060 4791 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="9b340efd-bf7e-4caf-8c76-ddc500aa2085" path="/var/lib/kubelet/pods/9b340efd-bf7e-4caf-8c76-ddc500aa2085/volumes" Dec 10 23:09:48 crc kubenswrapper[4791]: I1210 23:09:48.573329 4791 generic.go:334] "Generic (PLEG): container finished" podID="b8b6cd29-edfe-4177-bf14-14b8580d86c2" containerID="c18bd4b8e12d3f0f752d9f47afa76b3cc91ab1912ba9618ed62c16cf23683866" exitCode=0 Dec 10 23:09:48 crc kubenswrapper[4791]: I1210 23:09:48.573440 4791 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-db-create-z2v58" event={"ID":"b8b6cd29-edfe-4177-bf14-14b8580d86c2","Type":"ContainerDied","Data":"c18bd4b8e12d3f0f752d9f47afa76b3cc91ab1912ba9618ed62c16cf23683866"} Dec 10 23:09:48 crc kubenswrapper[4791]: I1210 23:09:48.575702 4791 generic.go:334] "Generic (PLEG): container finished" podID="0a94efe1-86a8-4768-9199-f47051935bcf" containerID="473c10a544d131d442414c7e9d016b5e15a352a6c22d2a9114caba72711977f5" exitCode=0 Dec 10 23:09:48 crc kubenswrapper[4791]: I1210 23:09:48.575776 4791 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-8832-account-create-update-wp7p5" event={"ID":"0a94efe1-86a8-4768-9199-f47051935bcf","Type":"ContainerDied","Data":"473c10a544d131d442414c7e9d016b5e15a352a6c22d2a9114caba72711977f5"} Dec 10 23:09:48 crc kubenswrapper[4791]: I1210 23:09:48.577710 4791 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"053f7590-725f-4504-a204-be83d801bad2","Type":"ContainerStarted","Data":"b599db3cab1becab15d15a46369973f0d8c5999cd6e208610c4736b0264bab56"} Dec 10 23:09:48 crc kubenswrapper[4791]: I1210 23:09:48.579501 4791 generic.go:334] "Generic (PLEG): container finished" podID="83ddd1be-5119-45a6-a080-1b63f0c8f7a1" containerID="25ee7ca4ad03fcd0c892fe328f36b3e3b779fd605dd652f8b53cf1cf97a7e30e" exitCode=0 Dec 10 23:09:48 crc kubenswrapper[4791]: I1210 23:09:48.579573 4791 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-b007-account-create-update-fkwxb" event={"ID":"83ddd1be-5119-45a6-a080-1b63f0c8f7a1","Type":"ContainerDied","Data":"25ee7ca4ad03fcd0c892fe328f36b3e3b779fd605dd652f8b53cf1cf97a7e30e"} Dec 10 23:09:48 crc kubenswrapper[4791]: I1210 23:09:48.581911 4791 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-proxy-c7bbc5ff-rlprn" event={"ID":"644890b3-8cae-45bc-b3eb-e0e6bfa8d957","Type":"ContainerStarted","Data":"f8a8130f6b8007b724b30a9fbdccf1ec54270b9d896a91dbfb528c69917181af"} Dec 10 23:09:48 crc kubenswrapper[4791]: I1210 23:09:48.582051 4791 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/swift-proxy-c7bbc5ff-rlprn" Dec 10 23:09:48 crc kubenswrapper[4791]: I1210 23:09:48.584799 4791 generic.go:334] "Generic (PLEG): container finished" podID="100aa755-c120-4c5a-9135-23bb98661f3d" containerID="e63790203ba9fd9cd2790cabecfb42628c365964dca974e26ff03d91ff75d79d" exitCode=0 Dec 10 23:09:48 crc kubenswrapper[4791]: I1210 23:09:48.584864 4791 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-db-create-tz8dr" event={"ID":"100aa755-c120-4c5a-9135-23bb98661f3d","Type":"ContainerDied","Data":"e63790203ba9fd9cd2790cabecfb42628c365964dca974e26ff03d91ff75d79d"} Dec 10 23:09:48 crc kubenswrapper[4791]: I1210 23:09:48.586246 4791 generic.go:334] "Generic (PLEG): container finished" podID="cf0595b8-0690-448d-9ef7-a64227bad653" containerID="6b1df2861daf4c05537a80c73e098b117f3695e83c011f0b5ae42d4c9ac0ddc6" exitCode=0 Dec 10 23:09:48 crc kubenswrapper[4791]: I1210 23:09:48.586313 4791 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-e207-account-create-update-hlmck" event={"ID":"cf0595b8-0690-448d-9ef7-a64227bad653","Type":"ContainerDied","Data":"6b1df2861daf4c05537a80c73e098b117f3695e83c011f0b5ae42d4c9ac0ddc6"} Dec 10 23:09:48 crc kubenswrapper[4791]: I1210 23:09:48.708577 4791 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/swift-proxy-c7bbc5ff-rlprn" podStartSLOduration=7.708555594 podStartE2EDuration="7.708555594s" podCreationTimestamp="2025-12-10 23:09:41 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-10 23:09:48.701228762 +0000 UTC m=+1223.130846375" watchObservedRunningTime="2025-12-10 23:09:48.708555594 +0000 UTC m=+1223.138173207" Dec 10 23:09:49 crc kubenswrapper[4791]: I1210 23:09:49.034103 4791 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-db-create-cvjnf" Dec 10 23:09:49 crc kubenswrapper[4791]: I1210 23:09:49.054088 4791 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ceilometer-0"] Dec 10 23:09:49 crc kubenswrapper[4791]: I1210 23:09:49.075939 4791 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/18c84e26-d0bb-4cd5-ba6a-49bf21155889-operator-scripts\") pod \"18c84e26-d0bb-4cd5-ba6a-49bf21155889\" (UID: \"18c84e26-d0bb-4cd5-ba6a-49bf21155889\") " Dec 10 23:09:49 crc kubenswrapper[4791]: I1210 23:09:49.084423 4791 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-pqv5l\" (UniqueName: \"kubernetes.io/projected/18c84e26-d0bb-4cd5-ba6a-49bf21155889-kube-api-access-pqv5l\") pod \"18c84e26-d0bb-4cd5-ba6a-49bf21155889\" (UID: \"18c84e26-d0bb-4cd5-ba6a-49bf21155889\") " Dec 10 23:09:49 crc kubenswrapper[4791]: I1210 23:09:49.078856 4791 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/18c84e26-d0bb-4cd5-ba6a-49bf21155889-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "18c84e26-d0bb-4cd5-ba6a-49bf21155889" (UID: "18c84e26-d0bb-4cd5-ba6a-49bf21155889"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 10 23:09:49 crc kubenswrapper[4791]: I1210 23:09:49.085499 4791 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/18c84e26-d0bb-4cd5-ba6a-49bf21155889-operator-scripts\") on node \"crc\" DevicePath \"\"" Dec 10 23:09:49 crc kubenswrapper[4791]: I1210 23:09:49.092365 4791 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/18c84e26-d0bb-4cd5-ba6a-49bf21155889-kube-api-access-pqv5l" (OuterVolumeSpecName: "kube-api-access-pqv5l") pod "18c84e26-d0bb-4cd5-ba6a-49bf21155889" (UID: "18c84e26-d0bb-4cd5-ba6a-49bf21155889"). InnerVolumeSpecName "kube-api-access-pqv5l". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 10 23:09:49 crc kubenswrapper[4791]: I1210 23:09:49.186926 4791 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-pqv5l\" (UniqueName: \"kubernetes.io/projected/18c84e26-d0bb-4cd5-ba6a-49bf21155889-kube-api-access-pqv5l\") on node \"crc\" DevicePath \"\"" Dec 10 23:09:49 crc kubenswrapper[4791]: I1210 23:09:49.596042 4791 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-db-create-cvjnf" event={"ID":"18c84e26-d0bb-4cd5-ba6a-49bf21155889","Type":"ContainerDied","Data":"db806c62ccdcf7f92f63d063ef3d251ac98177541edfea6b3e53e454f5d15e96"} Dec 10 23:09:49 crc kubenswrapper[4791]: I1210 23:09:49.596296 4791 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="db806c62ccdcf7f92f63d063ef3d251ac98177541edfea6b3e53e454f5d15e96" Dec 10 23:09:49 crc kubenswrapper[4791]: I1210 23:09:49.596122 4791 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-db-create-cvjnf" Dec 10 23:09:49 crc kubenswrapper[4791]: I1210 23:09:49.597877 4791 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"053f7590-725f-4504-a204-be83d801bad2","Type":"ContainerStarted","Data":"1cd4657d399538b1cc65c711e3e11791d84c93aef6bed6e7009f412fb56b3453"} Dec 10 23:09:49 crc kubenswrapper[4791]: I1210 23:09:49.597938 4791 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"053f7590-725f-4504-a204-be83d801bad2","Type":"ContainerStarted","Data":"3a843f3dd0a7f3fa2fe6719f2d7fcb8bb9ae41ef430830e5a912a792e2fe3c4a"} Dec 10 23:09:49 crc kubenswrapper[4791]: I1210 23:09:49.598295 4791 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/swift-proxy-c7bbc5ff-rlprn" Dec 10 23:09:49 crc kubenswrapper[4791]: I1210 23:09:49.914096 4791 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-b007-account-create-update-fkwxb" Dec 10 23:09:50 crc kubenswrapper[4791]: I1210 23:09:50.000779 4791 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-kdmwm\" (UniqueName: \"kubernetes.io/projected/83ddd1be-5119-45a6-a080-1b63f0c8f7a1-kube-api-access-kdmwm\") pod \"83ddd1be-5119-45a6-a080-1b63f0c8f7a1\" (UID: \"83ddd1be-5119-45a6-a080-1b63f0c8f7a1\") " Dec 10 23:09:50 crc kubenswrapper[4791]: I1210 23:09:50.000939 4791 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/83ddd1be-5119-45a6-a080-1b63f0c8f7a1-operator-scripts\") pod \"83ddd1be-5119-45a6-a080-1b63f0c8f7a1\" (UID: \"83ddd1be-5119-45a6-a080-1b63f0c8f7a1\") " Dec 10 23:09:50 crc kubenswrapper[4791]: I1210 23:09:50.001888 4791 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/83ddd1be-5119-45a6-a080-1b63f0c8f7a1-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "83ddd1be-5119-45a6-a080-1b63f0c8f7a1" (UID: "83ddd1be-5119-45a6-a080-1b63f0c8f7a1"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 10 23:09:50 crc kubenswrapper[4791]: I1210 23:09:50.004818 4791 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/83ddd1be-5119-45a6-a080-1b63f0c8f7a1-kube-api-access-kdmwm" (OuterVolumeSpecName: "kube-api-access-kdmwm") pod "83ddd1be-5119-45a6-a080-1b63f0c8f7a1" (UID: "83ddd1be-5119-45a6-a080-1b63f0c8f7a1"). InnerVolumeSpecName "kube-api-access-kdmwm". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 10 23:09:50 crc kubenswrapper[4791]: I1210 23:09:50.103455 4791 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-kdmwm\" (UniqueName: \"kubernetes.io/projected/83ddd1be-5119-45a6-a080-1b63f0c8f7a1-kube-api-access-kdmwm\") on node \"crc\" DevicePath \"\"" Dec 10 23:09:50 crc kubenswrapper[4791]: I1210 23:09:50.103485 4791 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/83ddd1be-5119-45a6-a080-1b63f0c8f7a1-operator-scripts\") on node \"crc\" DevicePath \"\"" Dec 10 23:09:50 crc kubenswrapper[4791]: I1210 23:09:50.225732 4791 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-db-create-tz8dr" Dec 10 23:09:50 crc kubenswrapper[4791]: I1210 23:09:50.235166 4791 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-e207-account-create-update-hlmck" Dec 10 23:09:50 crc kubenswrapper[4791]: I1210 23:09:50.284544 4791 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-8832-account-create-update-wp7p5" Dec 10 23:09:50 crc kubenswrapper[4791]: I1210 23:09:50.291795 4791 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-db-create-z2v58" Dec 10 23:09:50 crc kubenswrapper[4791]: I1210 23:09:50.305940 4791 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-kg2hz\" (UniqueName: \"kubernetes.io/projected/cf0595b8-0690-448d-9ef7-a64227bad653-kube-api-access-kg2hz\") pod \"cf0595b8-0690-448d-9ef7-a64227bad653\" (UID: \"cf0595b8-0690-448d-9ef7-a64227bad653\") " Dec 10 23:09:50 crc kubenswrapper[4791]: I1210 23:09:50.306071 4791 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-xrlpr\" (UniqueName: \"kubernetes.io/projected/100aa755-c120-4c5a-9135-23bb98661f3d-kube-api-access-xrlpr\") pod \"100aa755-c120-4c5a-9135-23bb98661f3d\" (UID: \"100aa755-c120-4c5a-9135-23bb98661f3d\") " Dec 10 23:09:50 crc kubenswrapper[4791]: I1210 23:09:50.306115 4791 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/100aa755-c120-4c5a-9135-23bb98661f3d-operator-scripts\") pod \"100aa755-c120-4c5a-9135-23bb98661f3d\" (UID: \"100aa755-c120-4c5a-9135-23bb98661f3d\") " Dec 10 23:09:50 crc kubenswrapper[4791]: I1210 23:09:50.306137 4791 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/cf0595b8-0690-448d-9ef7-a64227bad653-operator-scripts\") pod \"cf0595b8-0690-448d-9ef7-a64227bad653\" (UID: \"cf0595b8-0690-448d-9ef7-a64227bad653\") " Dec 10 23:09:50 crc kubenswrapper[4791]: I1210 23:09:50.307060 4791 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/100aa755-c120-4c5a-9135-23bb98661f3d-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "100aa755-c120-4c5a-9135-23bb98661f3d" (UID: "100aa755-c120-4c5a-9135-23bb98661f3d"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 10 23:09:50 crc kubenswrapper[4791]: I1210 23:09:50.307593 4791 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/cf0595b8-0690-448d-9ef7-a64227bad653-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "cf0595b8-0690-448d-9ef7-a64227bad653" (UID: "cf0595b8-0690-448d-9ef7-a64227bad653"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 10 23:09:50 crc kubenswrapper[4791]: I1210 23:09:50.312618 4791 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/100aa755-c120-4c5a-9135-23bb98661f3d-kube-api-access-xrlpr" (OuterVolumeSpecName: "kube-api-access-xrlpr") pod "100aa755-c120-4c5a-9135-23bb98661f3d" (UID: "100aa755-c120-4c5a-9135-23bb98661f3d"). InnerVolumeSpecName "kube-api-access-xrlpr". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 10 23:09:50 crc kubenswrapper[4791]: I1210 23:09:50.329600 4791 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/cf0595b8-0690-448d-9ef7-a64227bad653-kube-api-access-kg2hz" (OuterVolumeSpecName: "kube-api-access-kg2hz") pod "cf0595b8-0690-448d-9ef7-a64227bad653" (UID: "cf0595b8-0690-448d-9ef7-a64227bad653"). InnerVolumeSpecName "kube-api-access-kg2hz". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 10 23:09:50 crc kubenswrapper[4791]: I1210 23:09:50.408027 4791 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-2cg8h\" (UniqueName: \"kubernetes.io/projected/b8b6cd29-edfe-4177-bf14-14b8580d86c2-kube-api-access-2cg8h\") pod \"b8b6cd29-edfe-4177-bf14-14b8580d86c2\" (UID: \"b8b6cd29-edfe-4177-bf14-14b8580d86c2\") " Dec 10 23:09:50 crc kubenswrapper[4791]: I1210 23:09:50.408168 4791 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/0a94efe1-86a8-4768-9199-f47051935bcf-operator-scripts\") pod \"0a94efe1-86a8-4768-9199-f47051935bcf\" (UID: \"0a94efe1-86a8-4768-9199-f47051935bcf\") " Dec 10 23:09:50 crc kubenswrapper[4791]: I1210 23:09:50.408245 4791 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/b8b6cd29-edfe-4177-bf14-14b8580d86c2-operator-scripts\") pod \"b8b6cd29-edfe-4177-bf14-14b8580d86c2\" (UID: \"b8b6cd29-edfe-4177-bf14-14b8580d86c2\") " Dec 10 23:09:50 crc kubenswrapper[4791]: I1210 23:09:50.408354 4791 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-6fnsl\" (UniqueName: \"kubernetes.io/projected/0a94efe1-86a8-4768-9199-f47051935bcf-kube-api-access-6fnsl\") pod \"0a94efe1-86a8-4768-9199-f47051935bcf\" (UID: \"0a94efe1-86a8-4768-9199-f47051935bcf\") " Dec 10 23:09:50 crc kubenswrapper[4791]: I1210 23:09:50.408752 4791 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/0a94efe1-86a8-4768-9199-f47051935bcf-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "0a94efe1-86a8-4768-9199-f47051935bcf" (UID: "0a94efe1-86a8-4768-9199-f47051935bcf"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 10 23:09:50 crc kubenswrapper[4791]: I1210 23:09:50.408917 4791 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/b8b6cd29-edfe-4177-bf14-14b8580d86c2-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "b8b6cd29-edfe-4177-bf14-14b8580d86c2" (UID: "b8b6cd29-edfe-4177-bf14-14b8580d86c2"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 10 23:09:50 crc kubenswrapper[4791]: I1210 23:09:50.409016 4791 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-xrlpr\" (UniqueName: \"kubernetes.io/projected/100aa755-c120-4c5a-9135-23bb98661f3d-kube-api-access-xrlpr\") on node \"crc\" DevicePath \"\"" Dec 10 23:09:50 crc kubenswrapper[4791]: I1210 23:09:50.409030 4791 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/100aa755-c120-4c5a-9135-23bb98661f3d-operator-scripts\") on node \"crc\" DevicePath \"\"" Dec 10 23:09:50 crc kubenswrapper[4791]: I1210 23:09:50.409040 4791 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/cf0595b8-0690-448d-9ef7-a64227bad653-operator-scripts\") on node \"crc\" DevicePath \"\"" Dec 10 23:09:50 crc kubenswrapper[4791]: I1210 23:09:50.409048 4791 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/0a94efe1-86a8-4768-9199-f47051935bcf-operator-scripts\") on node \"crc\" DevicePath \"\"" Dec 10 23:09:50 crc kubenswrapper[4791]: I1210 23:09:50.409060 4791 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-kg2hz\" (UniqueName: \"kubernetes.io/projected/cf0595b8-0690-448d-9ef7-a64227bad653-kube-api-access-kg2hz\") on node \"crc\" DevicePath \"\"" Dec 10 23:09:50 crc kubenswrapper[4791]: I1210 23:09:50.412435 4791 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/0a94efe1-86a8-4768-9199-f47051935bcf-kube-api-access-6fnsl" (OuterVolumeSpecName: "kube-api-access-6fnsl") pod "0a94efe1-86a8-4768-9199-f47051935bcf" (UID: "0a94efe1-86a8-4768-9199-f47051935bcf"). InnerVolumeSpecName "kube-api-access-6fnsl". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 10 23:09:50 crc kubenswrapper[4791]: I1210 23:09:50.412503 4791 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/b8b6cd29-edfe-4177-bf14-14b8580d86c2-kube-api-access-2cg8h" (OuterVolumeSpecName: "kube-api-access-2cg8h") pod "b8b6cd29-edfe-4177-bf14-14b8580d86c2" (UID: "b8b6cd29-edfe-4177-bf14-14b8580d86c2"). InnerVolumeSpecName "kube-api-access-2cg8h". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 10 23:09:50 crc kubenswrapper[4791]: I1210 23:09:50.511112 4791 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/b8b6cd29-edfe-4177-bf14-14b8580d86c2-operator-scripts\") on node \"crc\" DevicePath \"\"" Dec 10 23:09:50 crc kubenswrapper[4791]: I1210 23:09:50.511152 4791 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-6fnsl\" (UniqueName: \"kubernetes.io/projected/0a94efe1-86a8-4768-9199-f47051935bcf-kube-api-access-6fnsl\") on node \"crc\" DevicePath \"\"" Dec 10 23:09:50 crc kubenswrapper[4791]: I1210 23:09:50.511172 4791 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-2cg8h\" (UniqueName: \"kubernetes.io/projected/b8b6cd29-edfe-4177-bf14-14b8580d86c2-kube-api-access-2cg8h\") on node \"crc\" DevicePath \"\"" Dec 10 23:09:50 crc kubenswrapper[4791]: I1210 23:09:50.618207 4791 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-db-create-tz8dr" Dec 10 23:09:50 crc kubenswrapper[4791]: I1210 23:09:50.618179 4791 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-db-create-tz8dr" event={"ID":"100aa755-c120-4c5a-9135-23bb98661f3d","Type":"ContainerDied","Data":"0f38ac94433531ed03f4993d66c9066b954d3e61b35dd4324309de81da6fb0ee"} Dec 10 23:09:50 crc kubenswrapper[4791]: I1210 23:09:50.618328 4791 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="0f38ac94433531ed03f4993d66c9066b954d3e61b35dd4324309de81da6fb0ee" Dec 10 23:09:50 crc kubenswrapper[4791]: I1210 23:09:50.630083 4791 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-db-create-z2v58" event={"ID":"b8b6cd29-edfe-4177-bf14-14b8580d86c2","Type":"ContainerDied","Data":"fe91f12d8f47a3a6a0ed46cd6dbd4e49f33e7fde61170871d208c67fa35d5b77"} Dec 10 23:09:50 crc kubenswrapper[4791]: I1210 23:09:50.630235 4791 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="fe91f12d8f47a3a6a0ed46cd6dbd4e49f33e7fde61170871d208c67fa35d5b77" Dec 10 23:09:50 crc kubenswrapper[4791]: I1210 23:09:50.630409 4791 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-db-create-z2v58" Dec 10 23:09:50 crc kubenswrapper[4791]: I1210 23:09:50.635397 4791 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-b007-account-create-update-fkwxb" event={"ID":"83ddd1be-5119-45a6-a080-1b63f0c8f7a1","Type":"ContainerDied","Data":"9b4ec8ea53c365d50c632d7add917b0aab3ece5693497f267402b895a87f4e84"} Dec 10 23:09:50 crc kubenswrapper[4791]: I1210 23:09:50.635557 4791 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="9b4ec8ea53c365d50c632d7add917b0aab3ece5693497f267402b895a87f4e84" Dec 10 23:09:50 crc kubenswrapper[4791]: I1210 23:09:50.635694 4791 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-b007-account-create-update-fkwxb" Dec 10 23:09:50 crc kubenswrapper[4791]: I1210 23:09:50.640235 4791 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"053f7590-725f-4504-a204-be83d801bad2","Type":"ContainerStarted","Data":"d42186dda997e7eea2ad8e2fe49e4065a0a7fa16abde74c6eb02667a3aad9b30"} Dec 10 23:09:50 crc kubenswrapper[4791]: I1210 23:09:50.648283 4791 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-e207-account-create-update-hlmck" event={"ID":"cf0595b8-0690-448d-9ef7-a64227bad653","Type":"ContainerDied","Data":"0adc35a87f508325498682c7c8f1a921d9cb89d48aa878fde128269d62043983"} Dec 10 23:09:50 crc kubenswrapper[4791]: I1210 23:09:50.648326 4791 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="0adc35a87f508325498682c7c8f1a921d9cb89d48aa878fde128269d62043983" Dec 10 23:09:50 crc kubenswrapper[4791]: I1210 23:09:50.648297 4791 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-e207-account-create-update-hlmck" Dec 10 23:09:50 crc kubenswrapper[4791]: I1210 23:09:50.651414 4791 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-8832-account-create-update-wp7p5" Dec 10 23:09:50 crc kubenswrapper[4791]: I1210 23:09:50.651563 4791 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-8832-account-create-update-wp7p5" event={"ID":"0a94efe1-86a8-4768-9199-f47051935bcf","Type":"ContainerDied","Data":"4586f22cc6cca6c7efa8850ef98273336200096b846dee143e924162d165de69"} Dec 10 23:09:50 crc kubenswrapper[4791]: I1210 23:09:50.651599 4791 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="4586f22cc6cca6c7efa8850ef98273336200096b846dee143e924162d165de69" Dec 10 23:09:51 crc kubenswrapper[4791]: I1210 23:09:51.661846 4791 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"053f7590-725f-4504-a204-be83d801bad2","Type":"ContainerStarted","Data":"b9fbe7eb299965c84c9556998893e0f9a21817ebbe4e48d8cd0d072a837e119a"} Dec 10 23:09:51 crc kubenswrapper[4791]: I1210 23:09:51.662073 4791 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="053f7590-725f-4504-a204-be83d801bad2" containerName="ceilometer-notification-agent" containerID="cri-o://1cd4657d399538b1cc65c711e3e11791d84c93aef6bed6e7009f412fb56b3453" gracePeriod=30 Dec 10 23:09:51 crc kubenswrapper[4791]: I1210 23:09:51.662047 4791 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="053f7590-725f-4504-a204-be83d801bad2" containerName="sg-core" containerID="cri-o://d42186dda997e7eea2ad8e2fe49e4065a0a7fa16abde74c6eb02667a3aad9b30" gracePeriod=30 Dec 10 23:09:51 crc kubenswrapper[4791]: I1210 23:09:51.662080 4791 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="053f7590-725f-4504-a204-be83d801bad2" containerName="proxy-httpd" containerID="cri-o://b9fbe7eb299965c84c9556998893e0f9a21817ebbe4e48d8cd0d072a837e119a" gracePeriod=30 Dec 10 23:09:51 crc kubenswrapper[4791]: I1210 23:09:51.661976 4791 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="053f7590-725f-4504-a204-be83d801bad2" containerName="ceilometer-central-agent" containerID="cri-o://3a843f3dd0a7f3fa2fe6719f2d7fcb8bb9ae41ef430830e5a912a792e2fe3c4a" gracePeriod=30 Dec 10 23:09:51 crc kubenswrapper[4791]: I1210 23:09:51.662528 4791 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/ceilometer-0" Dec 10 23:09:51 crc kubenswrapper[4791]: I1210 23:09:51.690888 4791 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ceilometer-0" podStartSLOduration=2.268982726 podStartE2EDuration="5.690872636s" podCreationTimestamp="2025-12-10 23:09:46 +0000 UTC" firstStartedPulling="2025-12-10 23:09:47.901728374 +0000 UTC m=+1222.331345987" lastFinishedPulling="2025-12-10 23:09:51.323618284 +0000 UTC m=+1225.753235897" observedRunningTime="2025-12-10 23:09:51.687871119 +0000 UTC m=+1226.117488732" watchObservedRunningTime="2025-12-10 23:09:51.690872636 +0000 UTC m=+1226.120490249" Dec 10 23:09:52 crc kubenswrapper[4791]: E1210 23:09:52.464228 4791 fsHandler.go:119] failed to collect filesystem stats - rootDiskErr: could not stat "/var/lib/containers/storage/overlay/327e7d6ba62cf115ae694867f41496f09719f1a2969198202c6d345e44ee3f6f/diff" to get inode usage: stat /var/lib/containers/storage/overlay/327e7d6ba62cf115ae694867f41496f09719f1a2969198202c6d345e44ee3f6f/diff: no such file or directory, extraDiskErr: Dec 10 23:09:52 crc kubenswrapper[4791]: I1210 23:09:52.672569 4791 generic.go:334] "Generic (PLEG): container finished" podID="053f7590-725f-4504-a204-be83d801bad2" containerID="d42186dda997e7eea2ad8e2fe49e4065a0a7fa16abde74c6eb02667a3aad9b30" exitCode=2 Dec 10 23:09:52 crc kubenswrapper[4791]: I1210 23:09:52.672599 4791 generic.go:334] "Generic (PLEG): container finished" podID="053f7590-725f-4504-a204-be83d801bad2" containerID="1cd4657d399538b1cc65c711e3e11791d84c93aef6bed6e7009f412fb56b3453" exitCode=0 Dec 10 23:09:52 crc kubenswrapper[4791]: I1210 23:09:52.672618 4791 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"053f7590-725f-4504-a204-be83d801bad2","Type":"ContainerDied","Data":"d42186dda997e7eea2ad8e2fe49e4065a0a7fa16abde74c6eb02667a3aad9b30"} Dec 10 23:09:52 crc kubenswrapper[4791]: I1210 23:09:52.672641 4791 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"053f7590-725f-4504-a204-be83d801bad2","Type":"ContainerDied","Data":"1cd4657d399538b1cc65c711e3e11791d84c93aef6bed6e7009f412fb56b3453"} Dec 10 23:09:52 crc kubenswrapper[4791]: E1210 23:09:52.869278 4791 fsHandler.go:119] failed to collect filesystem stats - rootDiskErr: could not stat "/var/lib/containers/storage/overlay/a447cfaf0bd74f3f95b3745b55ef52bf1bfd780b70ae876f5d65300d98440604/diff" to get inode usage: stat /var/lib/containers/storage/overlay/a447cfaf0bd74f3f95b3745b55ef52bf1bfd780b70ae876f5d65300d98440604/diff: no such file or directory, extraDiskErr: could not stat "/var/log/pods/openstack_neutron-85794c6bd6-ht8sz_75eb971a-04fc-40f5-9890-d5a0b4daccd1/neutron-api/0.log" to get inode usage: stat /var/log/pods/openstack_neutron-85794c6bd6-ht8sz_75eb971a-04fc-40f5-9890-d5a0b4daccd1/neutron-api/0.log: no such file or directory Dec 10 23:09:53 crc kubenswrapper[4791]: E1210 23:09:53.209300 4791 manager.go:1116] Failed to create existing container: /kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod75eb971a_04fc_40f5_9890_d5a0b4daccd1.slice/crio-dd2f20735a4b4cd7f75dfe9937d9cb51e2dc61038563f55ba640cae578951ee2: Error finding container dd2f20735a4b4cd7f75dfe9937d9cb51e2dc61038563f55ba640cae578951ee2: Status 404 returned error can't find the container with id dd2f20735a4b4cd7f75dfe9937d9cb51e2dc61038563f55ba640cae578951ee2 Dec 10 23:09:53 crc kubenswrapper[4791]: W1210 23:09:53.211084 4791 watcher.go:93] Error while processing event ("/sys/fs/cgroup/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod58167832_d8da_48de_8b8b_cf51fc54d92c.slice/crio-conmon-b8f078f30aed9ddb6d5815433c894f5ac4a00e3c7489055d52e0cb4879bb16cf.scope": 0x40000100 == IN_CREATE|IN_ISDIR): inotify_add_watch /sys/fs/cgroup/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod58167832_d8da_48de_8b8b_cf51fc54d92c.slice/crio-conmon-b8f078f30aed9ddb6d5815433c894f5ac4a00e3c7489055d52e0cb4879bb16cf.scope: no such file or directory Dec 10 23:09:53 crc kubenswrapper[4791]: W1210 23:09:53.212579 4791 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod58167832_d8da_48de_8b8b_cf51fc54d92c.slice/crio-ca930728dc321c5022ce2263f6f361be1de64fd0569818e4c3e7dca6b76ec3df WatchSource:0}: Error finding container ca930728dc321c5022ce2263f6f361be1de64fd0569818e4c3e7dca6b76ec3df: Status 404 returned error can't find the container with id ca930728dc321c5022ce2263f6f361be1de64fd0569818e4c3e7dca6b76ec3df Dec 10 23:09:53 crc kubenswrapper[4791]: W1210 23:09:53.212744 4791 watcher.go:93] Error while processing event ("/sys/fs/cgroup/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod58167832_d8da_48de_8b8b_cf51fc54d92c.slice/crio-b8f078f30aed9ddb6d5815433c894f5ac4a00e3c7489055d52e0cb4879bb16cf.scope": 0x40000100 == IN_CREATE|IN_ISDIR): inotify_add_watch /sys/fs/cgroup/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod58167832_d8da_48de_8b8b_cf51fc54d92c.slice/crio-b8f078f30aed9ddb6d5815433c894f5ac4a00e3c7489055d52e0cb4879bb16cf.scope: no such file or directory Dec 10 23:09:53 crc kubenswrapper[4791]: W1210 23:09:53.212864 4791 watcher.go:93] Error while processing event ("/sys/fs/cgroup/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod58167832_d8da_48de_8b8b_cf51fc54d92c.slice/crio-conmon-d7021f57af9ab5eef5206bcbbcf0c53acba5e7dfcef13ecaa64e06f7d4fe6fc2.scope": 0x40000100 == IN_CREATE|IN_ISDIR): inotify_add_watch /sys/fs/cgroup/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod58167832_d8da_48de_8b8b_cf51fc54d92c.slice/crio-conmon-d7021f57af9ab5eef5206bcbbcf0c53acba5e7dfcef13ecaa64e06f7d4fe6fc2.scope: no such file or directory Dec 10 23:09:53 crc kubenswrapper[4791]: W1210 23:09:53.212893 4791 watcher.go:93] Error while processing event ("/sys/fs/cgroup/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod58167832_d8da_48de_8b8b_cf51fc54d92c.slice/crio-d7021f57af9ab5eef5206bcbbcf0c53acba5e7dfcef13ecaa64e06f7d4fe6fc2.scope": 0x40000100 == IN_CREATE|IN_ISDIR): inotify_add_watch /sys/fs/cgroup/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod58167832_d8da_48de_8b8b_cf51fc54d92c.slice/crio-d7021f57af9ab5eef5206bcbbcf0c53acba5e7dfcef13ecaa64e06f7d4fe6fc2.scope: no such file or directory Dec 10 23:09:53 crc kubenswrapper[4791]: W1210 23:09:53.213335 4791 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod58167832_d8da_48de_8b8b_cf51fc54d92c.slice/crio-b32ac48681950cbc60cb9f131ea5fd50ebfc4c397b7e8f40d8325d077bbfcbcf.scope WatchSource:0}: Error finding container b32ac48681950cbc60cb9f131ea5fd50ebfc4c397b7e8f40d8325d077bbfcbcf: Status 404 returned error can't find the container with id b32ac48681950cbc60cb9f131ea5fd50ebfc4c397b7e8f40d8325d077bbfcbcf Dec 10 23:09:53 crc kubenswrapper[4791]: W1210 23:09:53.213418 4791 watcher.go:93] Error while processing event ("/sys/fs/cgroup/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod58167832_d8da_48de_8b8b_cf51fc54d92c.slice/crio-conmon-d3db4aa0578d2e480256777270f40b9150d19368ae770c08c87f98f42a115b57.scope": 0x40000100 == IN_CREATE|IN_ISDIR): inotify_add_watch /sys/fs/cgroup/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod58167832_d8da_48de_8b8b_cf51fc54d92c.slice/crio-conmon-d3db4aa0578d2e480256777270f40b9150d19368ae770c08c87f98f42a115b57.scope: no such file or directory Dec 10 23:09:53 crc kubenswrapper[4791]: W1210 23:09:53.213444 4791 watcher.go:93] Error while processing event ("/sys/fs/cgroup/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod58167832_d8da_48de_8b8b_cf51fc54d92c.slice/crio-d3db4aa0578d2e480256777270f40b9150d19368ae770c08c87f98f42a115b57.scope": 0x40000100 == IN_CREATE|IN_ISDIR): inotify_add_watch /sys/fs/cgroup/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod58167832_d8da_48de_8b8b_cf51fc54d92c.slice/crio-d3db4aa0578d2e480256777270f40b9150d19368ae770c08c87f98f42a115b57.scope: no such file or directory Dec 10 23:09:53 crc kubenswrapper[4791]: W1210 23:09:53.214212 4791 watcher.go:93] Error while processing event ("/sys/fs/cgroup/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod18c84e26_d0bb_4cd5_ba6a_49bf21155889.slice": 0x40000100 == IN_CREATE|IN_ISDIR): inotify_add_watch /sys/fs/cgroup/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod18c84e26_d0bb_4cd5_ba6a_49bf21155889.slice: no such file or directory Dec 10 23:09:53 crc kubenswrapper[4791]: W1210 23:09:53.214284 4791 watcher.go:93] Error while processing event ("/sys/fs/cgroup/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod0a94efe1_86a8_4768_9199_f47051935bcf.slice": 0x40000100 == IN_CREATE|IN_ISDIR): inotify_add_watch /sys/fs/cgroup/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod0a94efe1_86a8_4768_9199_f47051935bcf.slice: no such file or directory Dec 10 23:09:53 crc kubenswrapper[4791]: W1210 23:09:53.214309 4791 watcher.go:93] Error while processing event ("/sys/fs/cgroup/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podb8b6cd29_edfe_4177_bf14_14b8580d86c2.slice": 0x40000100 == IN_CREATE|IN_ISDIR): inotify_add_watch /sys/fs/cgroup/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podb8b6cd29_edfe_4177_bf14_14b8580d86c2.slice: no such file or directory Dec 10 23:09:53 crc kubenswrapper[4791]: W1210 23:09:53.214329 4791 watcher.go:93] Error while processing event ("/sys/fs/cgroup/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod100aa755_c120_4c5a_9135_23bb98661f3d.slice": 0x40000100 == IN_CREATE|IN_ISDIR): inotify_add_watch /sys/fs/cgroup/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod100aa755_c120_4c5a_9135_23bb98661f3d.slice: no such file or directory Dec 10 23:09:53 crc kubenswrapper[4791]: W1210 23:09:53.214362 4791 watcher.go:93] Error while processing event ("/sys/fs/cgroup/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podcf0595b8_0690_448d_9ef7_a64227bad653.slice": 0x40000100 == IN_CREATE|IN_ISDIR): inotify_add_watch /sys/fs/cgroup/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podcf0595b8_0690_448d_9ef7_a64227bad653.slice: no such file or directory Dec 10 23:09:53 crc kubenswrapper[4791]: W1210 23:09:53.214381 4791 watcher.go:93] Error while processing event ("/sys/fs/cgroup/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod83ddd1be_5119_45a6_a080_1b63f0c8f7a1.slice": 0x40000100 == IN_CREATE|IN_ISDIR): inotify_add_watch /sys/fs/cgroup/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod83ddd1be_5119_45a6_a080_1b63f0c8f7a1.slice: no such file or directory Dec 10 23:09:53 crc kubenswrapper[4791]: E1210 23:09:53.224696 4791 manager.go:1116] Failed to create existing container: /kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod9b340efd_bf7e_4caf_8c76_ddc500aa2085.slice/crio-8f95cb70af18d0c8b0d71d55cc909fb5e15dbe337f92069cd49742ac1b44c594: Error finding container 8f95cb70af18d0c8b0d71d55cc909fb5e15dbe337f92069cd49742ac1b44c594: Status 404 returned error can't find the container with id 8f95cb70af18d0c8b0d71d55cc909fb5e15dbe337f92069cd49742ac1b44c594 Dec 10 23:09:53 crc kubenswrapper[4791]: E1210 23:09:53.485698 4791 cadvisor_stats_provider.go:516] "Partial failure issuing cadvisor.ContainerInfoV2" err="partial failures: [\"/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod58167832_d8da_48de_8b8b_cf51fc54d92c.slice/crio-conmon-b32ac48681950cbc60cb9f131ea5fd50ebfc4c397b7e8f40d8325d077bbfcbcf.scope\": RecentStats: unable to find data in memory cache], [\"/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod9b340efd_bf7e_4caf_8c76_ddc500aa2085.slice/crio-31193dfd2f9e85871df3d79e4aef1d62fa1b1a1b269896d2e71dbdb4f4784a21.scope\": RecentStats: unable to find data in memory cache], [\"/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod9b340efd_bf7e_4caf_8c76_ddc500aa2085.slice\": RecentStats: unable to find data in memory cache], [\"/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod58167832_d8da_48de_8b8b_cf51fc54d92c.slice\": RecentStats: unable to find data in memory cache], [\"/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod9b340efd_bf7e_4caf_8c76_ddc500aa2085.slice/crio-conmon-31193dfd2f9e85871df3d79e4aef1d62fa1b1a1b269896d2e71dbdb4f4784a21.scope\": RecentStats: unable to find data in memory cache]" Dec 10 23:09:53 crc kubenswrapper[4791]: I1210 23:09:53.615641 4791 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/horizon-7cc66567bd-pwb2l" Dec 10 23:09:53 crc kubenswrapper[4791]: I1210 23:09:53.683721 4791 generic.go:334] "Generic (PLEG): container finished" podID="55139c7c-e9fa-4539-b7b2-b2be7ee43b44" containerID="7b1ada7642cab048e60fe308d453e086b82053b893d130082d56f6aaf472608a" exitCode=137 Dec 10 23:09:53 crc kubenswrapper[4791]: I1210 23:09:53.683760 4791 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/horizon-7cc66567bd-pwb2l" event={"ID":"55139c7c-e9fa-4539-b7b2-b2be7ee43b44","Type":"ContainerDied","Data":"7b1ada7642cab048e60fe308d453e086b82053b893d130082d56f6aaf472608a"} Dec 10 23:09:53 crc kubenswrapper[4791]: I1210 23:09:53.683783 4791 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/horizon-7cc66567bd-pwb2l" event={"ID":"55139c7c-e9fa-4539-b7b2-b2be7ee43b44","Type":"ContainerDied","Data":"d2883b26349a02a4c60df2f542814b6fa9be4acda940054df2fa1a9938da6439"} Dec 10 23:09:53 crc kubenswrapper[4791]: I1210 23:09:53.683798 4791 scope.go:117] "RemoveContainer" containerID="a441e9dd03ea293d73ed83e0ce82cd1f391a27dfc4d4c0f86b23bba57c39c2c5" Dec 10 23:09:53 crc kubenswrapper[4791]: I1210 23:09:53.683887 4791 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/horizon-7cc66567bd-pwb2l" Dec 10 23:09:53 crc kubenswrapper[4791]: I1210 23:09:53.776036 4791 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/55139c7c-e9fa-4539-b7b2-b2be7ee43b44-config-data\") pod \"55139c7c-e9fa-4539-b7b2-b2be7ee43b44\" (UID: \"55139c7c-e9fa-4539-b7b2-b2be7ee43b44\") " Dec 10 23:09:53 crc kubenswrapper[4791]: I1210 23:09:53.776143 4791 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"horizon-tls-certs\" (UniqueName: \"kubernetes.io/secret/55139c7c-e9fa-4539-b7b2-b2be7ee43b44-horizon-tls-certs\") pod \"55139c7c-e9fa-4539-b7b2-b2be7ee43b44\" (UID: \"55139c7c-e9fa-4539-b7b2-b2be7ee43b44\") " Dec 10 23:09:53 crc kubenswrapper[4791]: I1210 23:09:53.776182 4791 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/55139c7c-e9fa-4539-b7b2-b2be7ee43b44-combined-ca-bundle\") pod \"55139c7c-e9fa-4539-b7b2-b2be7ee43b44\" (UID: \"55139c7c-e9fa-4539-b7b2-b2be7ee43b44\") " Dec 10 23:09:53 crc kubenswrapper[4791]: I1210 23:09:53.776202 4791 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-qmqzj\" (UniqueName: \"kubernetes.io/projected/55139c7c-e9fa-4539-b7b2-b2be7ee43b44-kube-api-access-qmqzj\") pod \"55139c7c-e9fa-4539-b7b2-b2be7ee43b44\" (UID: \"55139c7c-e9fa-4539-b7b2-b2be7ee43b44\") " Dec 10 23:09:53 crc kubenswrapper[4791]: I1210 23:09:53.776273 4791 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/55139c7c-e9fa-4539-b7b2-b2be7ee43b44-logs\") pod \"55139c7c-e9fa-4539-b7b2-b2be7ee43b44\" (UID: \"55139c7c-e9fa-4539-b7b2-b2be7ee43b44\") " Dec 10 23:09:53 crc kubenswrapper[4791]: I1210 23:09:53.776335 4791 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/55139c7c-e9fa-4539-b7b2-b2be7ee43b44-scripts\") pod \"55139c7c-e9fa-4539-b7b2-b2be7ee43b44\" (UID: \"55139c7c-e9fa-4539-b7b2-b2be7ee43b44\") " Dec 10 23:09:53 crc kubenswrapper[4791]: I1210 23:09:53.776389 4791 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"horizon-secret-key\" (UniqueName: \"kubernetes.io/secret/55139c7c-e9fa-4539-b7b2-b2be7ee43b44-horizon-secret-key\") pod \"55139c7c-e9fa-4539-b7b2-b2be7ee43b44\" (UID: \"55139c7c-e9fa-4539-b7b2-b2be7ee43b44\") " Dec 10 23:09:53 crc kubenswrapper[4791]: I1210 23:09:53.777451 4791 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/55139c7c-e9fa-4539-b7b2-b2be7ee43b44-logs" (OuterVolumeSpecName: "logs") pod "55139c7c-e9fa-4539-b7b2-b2be7ee43b44" (UID: "55139c7c-e9fa-4539-b7b2-b2be7ee43b44"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 10 23:09:53 crc kubenswrapper[4791]: I1210 23:09:53.782659 4791 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/55139c7c-e9fa-4539-b7b2-b2be7ee43b44-horizon-secret-key" (OuterVolumeSpecName: "horizon-secret-key") pod "55139c7c-e9fa-4539-b7b2-b2be7ee43b44" (UID: "55139c7c-e9fa-4539-b7b2-b2be7ee43b44"). InnerVolumeSpecName "horizon-secret-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 10 23:09:53 crc kubenswrapper[4791]: I1210 23:09:53.785778 4791 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/55139c7c-e9fa-4539-b7b2-b2be7ee43b44-kube-api-access-qmqzj" (OuterVolumeSpecName: "kube-api-access-qmqzj") pod "55139c7c-e9fa-4539-b7b2-b2be7ee43b44" (UID: "55139c7c-e9fa-4539-b7b2-b2be7ee43b44"). InnerVolumeSpecName "kube-api-access-qmqzj". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 10 23:09:53 crc kubenswrapper[4791]: I1210 23:09:53.813158 4791 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/55139c7c-e9fa-4539-b7b2-b2be7ee43b44-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "55139c7c-e9fa-4539-b7b2-b2be7ee43b44" (UID: "55139c7c-e9fa-4539-b7b2-b2be7ee43b44"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 10 23:09:53 crc kubenswrapper[4791]: I1210 23:09:53.822104 4791 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/55139c7c-e9fa-4539-b7b2-b2be7ee43b44-scripts" (OuterVolumeSpecName: "scripts") pod "55139c7c-e9fa-4539-b7b2-b2be7ee43b44" (UID: "55139c7c-e9fa-4539-b7b2-b2be7ee43b44"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 10 23:09:53 crc kubenswrapper[4791]: I1210 23:09:53.824287 4791 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/55139c7c-e9fa-4539-b7b2-b2be7ee43b44-config-data" (OuterVolumeSpecName: "config-data") pod "55139c7c-e9fa-4539-b7b2-b2be7ee43b44" (UID: "55139c7c-e9fa-4539-b7b2-b2be7ee43b44"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 10 23:09:53 crc kubenswrapper[4791]: I1210 23:09:53.845938 4791 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/55139c7c-e9fa-4539-b7b2-b2be7ee43b44-horizon-tls-certs" (OuterVolumeSpecName: "horizon-tls-certs") pod "55139c7c-e9fa-4539-b7b2-b2be7ee43b44" (UID: "55139c7c-e9fa-4539-b7b2-b2be7ee43b44"). InnerVolumeSpecName "horizon-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 10 23:09:53 crc kubenswrapper[4791]: I1210 23:09:53.875983 4791 scope.go:117] "RemoveContainer" containerID="7b1ada7642cab048e60fe308d453e086b82053b893d130082d56f6aaf472608a" Dec 10 23:09:53 crc kubenswrapper[4791]: I1210 23:09:53.878420 4791 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/55139c7c-e9fa-4539-b7b2-b2be7ee43b44-logs\") on node \"crc\" DevicePath \"\"" Dec 10 23:09:53 crc kubenswrapper[4791]: I1210 23:09:53.878457 4791 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/55139c7c-e9fa-4539-b7b2-b2be7ee43b44-scripts\") on node \"crc\" DevicePath \"\"" Dec 10 23:09:53 crc kubenswrapper[4791]: I1210 23:09:53.878475 4791 reconciler_common.go:293] "Volume detached for volume \"horizon-secret-key\" (UniqueName: \"kubernetes.io/secret/55139c7c-e9fa-4539-b7b2-b2be7ee43b44-horizon-secret-key\") on node \"crc\" DevicePath \"\"" Dec 10 23:09:53 crc kubenswrapper[4791]: I1210 23:09:53.878491 4791 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/55139c7c-e9fa-4539-b7b2-b2be7ee43b44-config-data\") on node \"crc\" DevicePath \"\"" Dec 10 23:09:53 crc kubenswrapper[4791]: I1210 23:09:53.878503 4791 reconciler_common.go:293] "Volume detached for volume \"horizon-tls-certs\" (UniqueName: \"kubernetes.io/secret/55139c7c-e9fa-4539-b7b2-b2be7ee43b44-horizon-tls-certs\") on node \"crc\" DevicePath \"\"" Dec 10 23:09:53 crc kubenswrapper[4791]: I1210 23:09:53.878511 4791 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/55139c7c-e9fa-4539-b7b2-b2be7ee43b44-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 10 23:09:53 crc kubenswrapper[4791]: I1210 23:09:53.878519 4791 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-qmqzj\" (UniqueName: \"kubernetes.io/projected/55139c7c-e9fa-4539-b7b2-b2be7ee43b44-kube-api-access-qmqzj\") on node \"crc\" DevicePath \"\"" Dec 10 23:09:53 crc kubenswrapper[4791]: I1210 23:09:53.912815 4791 scope.go:117] "RemoveContainer" containerID="a441e9dd03ea293d73ed83e0ce82cd1f391a27dfc4d4c0f86b23bba57c39c2c5" Dec 10 23:09:53 crc kubenswrapper[4791]: E1210 23:09:53.913211 4791 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"a441e9dd03ea293d73ed83e0ce82cd1f391a27dfc4d4c0f86b23bba57c39c2c5\": container with ID starting with a441e9dd03ea293d73ed83e0ce82cd1f391a27dfc4d4c0f86b23bba57c39c2c5 not found: ID does not exist" containerID="a441e9dd03ea293d73ed83e0ce82cd1f391a27dfc4d4c0f86b23bba57c39c2c5" Dec 10 23:09:53 crc kubenswrapper[4791]: I1210 23:09:53.913246 4791 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"a441e9dd03ea293d73ed83e0ce82cd1f391a27dfc4d4c0f86b23bba57c39c2c5"} err="failed to get container status \"a441e9dd03ea293d73ed83e0ce82cd1f391a27dfc4d4c0f86b23bba57c39c2c5\": rpc error: code = NotFound desc = could not find container \"a441e9dd03ea293d73ed83e0ce82cd1f391a27dfc4d4c0f86b23bba57c39c2c5\": container with ID starting with a441e9dd03ea293d73ed83e0ce82cd1f391a27dfc4d4c0f86b23bba57c39c2c5 not found: ID does not exist" Dec 10 23:09:53 crc kubenswrapper[4791]: I1210 23:09:53.913278 4791 scope.go:117] "RemoveContainer" containerID="7b1ada7642cab048e60fe308d453e086b82053b893d130082d56f6aaf472608a" Dec 10 23:09:53 crc kubenswrapper[4791]: E1210 23:09:53.913535 4791 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"7b1ada7642cab048e60fe308d453e086b82053b893d130082d56f6aaf472608a\": container with ID starting with 7b1ada7642cab048e60fe308d453e086b82053b893d130082d56f6aaf472608a not found: ID does not exist" containerID="7b1ada7642cab048e60fe308d453e086b82053b893d130082d56f6aaf472608a" Dec 10 23:09:53 crc kubenswrapper[4791]: I1210 23:09:53.913569 4791 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"7b1ada7642cab048e60fe308d453e086b82053b893d130082d56f6aaf472608a"} err="failed to get container status \"7b1ada7642cab048e60fe308d453e086b82053b893d130082d56f6aaf472608a\": rpc error: code = NotFound desc = could not find container \"7b1ada7642cab048e60fe308d453e086b82053b893d130082d56f6aaf472608a\": container with ID starting with 7b1ada7642cab048e60fe308d453e086b82053b893d130082d56f6aaf472608a not found: ID does not exist" Dec 10 23:09:54 crc kubenswrapper[4791]: I1210 23:09:54.007533 4791 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/horizon-7cc66567bd-pwb2l"] Dec 10 23:09:54 crc kubenswrapper[4791]: I1210 23:09:54.016954 4791 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/horizon-7cc66567bd-pwb2l"] Dec 10 23:09:54 crc kubenswrapper[4791]: I1210 23:09:54.697361 4791 generic.go:334] "Generic (PLEG): container finished" podID="053f7590-725f-4504-a204-be83d801bad2" containerID="3a843f3dd0a7f3fa2fe6719f2d7fcb8bb9ae41ef430830e5a912a792e2fe3c4a" exitCode=0 Dec 10 23:09:54 crc kubenswrapper[4791]: I1210 23:09:54.697476 4791 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"053f7590-725f-4504-a204-be83d801bad2","Type":"ContainerDied","Data":"3a843f3dd0a7f3fa2fe6719f2d7fcb8bb9ae41ef430830e5a912a792e2fe3c4a"} Dec 10 23:09:55 crc kubenswrapper[4791]: I1210 23:09:55.813232 4791 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-cell0-conductor-db-sync-jff9b"] Dec 10 23:09:55 crc kubenswrapper[4791]: E1210 23:09:55.813703 4791 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="83ddd1be-5119-45a6-a080-1b63f0c8f7a1" containerName="mariadb-account-create-update" Dec 10 23:09:55 crc kubenswrapper[4791]: I1210 23:09:55.813721 4791 state_mem.go:107] "Deleted CPUSet assignment" podUID="83ddd1be-5119-45a6-a080-1b63f0c8f7a1" containerName="mariadb-account-create-update" Dec 10 23:09:55 crc kubenswrapper[4791]: E1210 23:09:55.813735 4791 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="55139c7c-e9fa-4539-b7b2-b2be7ee43b44" containerName="horizon-log" Dec 10 23:09:55 crc kubenswrapper[4791]: I1210 23:09:55.813746 4791 state_mem.go:107] "Deleted CPUSet assignment" podUID="55139c7c-e9fa-4539-b7b2-b2be7ee43b44" containerName="horizon-log" Dec 10 23:09:55 crc kubenswrapper[4791]: E1210 23:09:55.813767 4791 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="100aa755-c120-4c5a-9135-23bb98661f3d" containerName="mariadb-database-create" Dec 10 23:09:55 crc kubenswrapper[4791]: I1210 23:09:55.813776 4791 state_mem.go:107] "Deleted CPUSet assignment" podUID="100aa755-c120-4c5a-9135-23bb98661f3d" containerName="mariadb-database-create" Dec 10 23:09:55 crc kubenswrapper[4791]: E1210 23:09:55.813786 4791 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="cf0595b8-0690-448d-9ef7-a64227bad653" containerName="mariadb-account-create-update" Dec 10 23:09:55 crc kubenswrapper[4791]: I1210 23:09:55.813794 4791 state_mem.go:107] "Deleted CPUSet assignment" podUID="cf0595b8-0690-448d-9ef7-a64227bad653" containerName="mariadb-account-create-update" Dec 10 23:09:55 crc kubenswrapper[4791]: E1210 23:09:55.813817 4791 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="18c84e26-d0bb-4cd5-ba6a-49bf21155889" containerName="mariadb-database-create" Dec 10 23:09:55 crc kubenswrapper[4791]: I1210 23:09:55.813825 4791 state_mem.go:107] "Deleted CPUSet assignment" podUID="18c84e26-d0bb-4cd5-ba6a-49bf21155889" containerName="mariadb-database-create" Dec 10 23:09:55 crc kubenswrapper[4791]: E1210 23:09:55.813850 4791 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="b8b6cd29-edfe-4177-bf14-14b8580d86c2" containerName="mariadb-database-create" Dec 10 23:09:55 crc kubenswrapper[4791]: I1210 23:09:55.813858 4791 state_mem.go:107] "Deleted CPUSet assignment" podUID="b8b6cd29-edfe-4177-bf14-14b8580d86c2" containerName="mariadb-database-create" Dec 10 23:09:55 crc kubenswrapper[4791]: E1210 23:09:55.813865 4791 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="55139c7c-e9fa-4539-b7b2-b2be7ee43b44" containerName="horizon" Dec 10 23:09:55 crc kubenswrapper[4791]: I1210 23:09:55.813872 4791 state_mem.go:107] "Deleted CPUSet assignment" podUID="55139c7c-e9fa-4539-b7b2-b2be7ee43b44" containerName="horizon" Dec 10 23:09:55 crc kubenswrapper[4791]: E1210 23:09:55.813892 4791 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="0a94efe1-86a8-4768-9199-f47051935bcf" containerName="mariadb-account-create-update" Dec 10 23:09:55 crc kubenswrapper[4791]: I1210 23:09:55.813900 4791 state_mem.go:107] "Deleted CPUSet assignment" podUID="0a94efe1-86a8-4768-9199-f47051935bcf" containerName="mariadb-account-create-update" Dec 10 23:09:55 crc kubenswrapper[4791]: I1210 23:09:55.814141 4791 memory_manager.go:354] "RemoveStaleState removing state" podUID="55139c7c-e9fa-4539-b7b2-b2be7ee43b44" containerName="horizon" Dec 10 23:09:55 crc kubenswrapper[4791]: I1210 23:09:55.814163 4791 memory_manager.go:354] "RemoveStaleState removing state" podUID="0a94efe1-86a8-4768-9199-f47051935bcf" containerName="mariadb-account-create-update" Dec 10 23:09:55 crc kubenswrapper[4791]: I1210 23:09:55.814184 4791 memory_manager.go:354] "RemoveStaleState removing state" podUID="b8b6cd29-edfe-4177-bf14-14b8580d86c2" containerName="mariadb-database-create" Dec 10 23:09:55 crc kubenswrapper[4791]: I1210 23:09:55.814198 4791 memory_manager.go:354] "RemoveStaleState removing state" podUID="100aa755-c120-4c5a-9135-23bb98661f3d" containerName="mariadb-database-create" Dec 10 23:09:55 crc kubenswrapper[4791]: I1210 23:09:55.814210 4791 memory_manager.go:354] "RemoveStaleState removing state" podUID="18c84e26-d0bb-4cd5-ba6a-49bf21155889" containerName="mariadb-database-create" Dec 10 23:09:55 crc kubenswrapper[4791]: I1210 23:09:55.814223 4791 memory_manager.go:354] "RemoveStaleState removing state" podUID="cf0595b8-0690-448d-9ef7-a64227bad653" containerName="mariadb-account-create-update" Dec 10 23:09:55 crc kubenswrapper[4791]: I1210 23:09:55.814235 4791 memory_manager.go:354] "RemoveStaleState removing state" podUID="83ddd1be-5119-45a6-a080-1b63f0c8f7a1" containerName="mariadb-account-create-update" Dec 10 23:09:55 crc kubenswrapper[4791]: I1210 23:09:55.814249 4791 memory_manager.go:354] "RemoveStaleState removing state" podUID="55139c7c-e9fa-4539-b7b2-b2be7ee43b44" containerName="horizon-log" Dec 10 23:09:55 crc kubenswrapper[4791]: I1210 23:09:55.815020 4791 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-conductor-db-sync-jff9b" Dec 10 23:09:55 crc kubenswrapper[4791]: I1210 23:09:55.819947 4791 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell0-conductor-scripts" Dec 10 23:09:55 crc kubenswrapper[4791]: I1210 23:09:55.821119 4791 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-nova-dockercfg-vtzz5" Dec 10 23:09:55 crc kubenswrapper[4791]: I1210 23:09:55.822052 4791 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell0-conductor-config-data" Dec 10 23:09:55 crc kubenswrapper[4791]: I1210 23:09:55.826452 4791 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell0-conductor-db-sync-jff9b"] Dec 10 23:09:55 crc kubenswrapper[4791]: I1210 23:09:55.896291 4791 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="55139c7c-e9fa-4539-b7b2-b2be7ee43b44" path="/var/lib/kubelet/pods/55139c7c-e9fa-4539-b7b2-b2be7ee43b44/volumes" Dec 10 23:09:56 crc kubenswrapper[4791]: I1210 23:09:56.015502 4791 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/075728ce-048e-452d-adff-0e3d167d6673-scripts\") pod \"nova-cell0-conductor-db-sync-jff9b\" (UID: \"075728ce-048e-452d-adff-0e3d167d6673\") " pod="openstack/nova-cell0-conductor-db-sync-jff9b" Dec 10 23:09:56 crc kubenswrapper[4791]: I1210 23:09:56.015548 4791 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/075728ce-048e-452d-adff-0e3d167d6673-combined-ca-bundle\") pod \"nova-cell0-conductor-db-sync-jff9b\" (UID: \"075728ce-048e-452d-adff-0e3d167d6673\") " pod="openstack/nova-cell0-conductor-db-sync-jff9b" Dec 10 23:09:56 crc kubenswrapper[4791]: I1210 23:09:56.015600 4791 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/075728ce-048e-452d-adff-0e3d167d6673-config-data\") pod \"nova-cell0-conductor-db-sync-jff9b\" (UID: \"075728ce-048e-452d-adff-0e3d167d6673\") " pod="openstack/nova-cell0-conductor-db-sync-jff9b" Dec 10 23:09:56 crc kubenswrapper[4791]: I1210 23:09:56.015766 4791 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-9x2cf\" (UniqueName: \"kubernetes.io/projected/075728ce-048e-452d-adff-0e3d167d6673-kube-api-access-9x2cf\") pod \"nova-cell0-conductor-db-sync-jff9b\" (UID: \"075728ce-048e-452d-adff-0e3d167d6673\") " pod="openstack/nova-cell0-conductor-db-sync-jff9b" Dec 10 23:09:56 crc kubenswrapper[4791]: I1210 23:09:56.117449 4791 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-9x2cf\" (UniqueName: \"kubernetes.io/projected/075728ce-048e-452d-adff-0e3d167d6673-kube-api-access-9x2cf\") pod \"nova-cell0-conductor-db-sync-jff9b\" (UID: \"075728ce-048e-452d-adff-0e3d167d6673\") " pod="openstack/nova-cell0-conductor-db-sync-jff9b" Dec 10 23:09:56 crc kubenswrapper[4791]: I1210 23:09:56.117853 4791 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/075728ce-048e-452d-adff-0e3d167d6673-scripts\") pod \"nova-cell0-conductor-db-sync-jff9b\" (UID: \"075728ce-048e-452d-adff-0e3d167d6673\") " pod="openstack/nova-cell0-conductor-db-sync-jff9b" Dec 10 23:09:56 crc kubenswrapper[4791]: I1210 23:09:56.117888 4791 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/075728ce-048e-452d-adff-0e3d167d6673-combined-ca-bundle\") pod \"nova-cell0-conductor-db-sync-jff9b\" (UID: \"075728ce-048e-452d-adff-0e3d167d6673\") " pod="openstack/nova-cell0-conductor-db-sync-jff9b" Dec 10 23:09:56 crc kubenswrapper[4791]: I1210 23:09:56.117973 4791 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/075728ce-048e-452d-adff-0e3d167d6673-config-data\") pod \"nova-cell0-conductor-db-sync-jff9b\" (UID: \"075728ce-048e-452d-adff-0e3d167d6673\") " pod="openstack/nova-cell0-conductor-db-sync-jff9b" Dec 10 23:09:56 crc kubenswrapper[4791]: I1210 23:09:56.125593 4791 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/075728ce-048e-452d-adff-0e3d167d6673-config-data\") pod \"nova-cell0-conductor-db-sync-jff9b\" (UID: \"075728ce-048e-452d-adff-0e3d167d6673\") " pod="openstack/nova-cell0-conductor-db-sync-jff9b" Dec 10 23:09:56 crc kubenswrapper[4791]: I1210 23:09:56.132313 4791 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/075728ce-048e-452d-adff-0e3d167d6673-combined-ca-bundle\") pod \"nova-cell0-conductor-db-sync-jff9b\" (UID: \"075728ce-048e-452d-adff-0e3d167d6673\") " pod="openstack/nova-cell0-conductor-db-sync-jff9b" Dec 10 23:09:56 crc kubenswrapper[4791]: I1210 23:09:56.133583 4791 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/075728ce-048e-452d-adff-0e3d167d6673-scripts\") pod \"nova-cell0-conductor-db-sync-jff9b\" (UID: \"075728ce-048e-452d-adff-0e3d167d6673\") " pod="openstack/nova-cell0-conductor-db-sync-jff9b" Dec 10 23:09:56 crc kubenswrapper[4791]: I1210 23:09:56.135468 4791 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-9x2cf\" (UniqueName: \"kubernetes.io/projected/075728ce-048e-452d-adff-0e3d167d6673-kube-api-access-9x2cf\") pod \"nova-cell0-conductor-db-sync-jff9b\" (UID: \"075728ce-048e-452d-adff-0e3d167d6673\") " pod="openstack/nova-cell0-conductor-db-sync-jff9b" Dec 10 23:09:56 crc kubenswrapper[4791]: I1210 23:09:56.138474 4791 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-conductor-db-sync-jff9b" Dec 10 23:09:56 crc kubenswrapper[4791]: I1210 23:09:56.629237 4791 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell0-conductor-db-sync-jff9b"] Dec 10 23:09:56 crc kubenswrapper[4791]: I1210 23:09:56.716525 4791 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-conductor-db-sync-jff9b" event={"ID":"075728ce-048e-452d-adff-0e3d167d6673","Type":"ContainerStarted","Data":"1e95bcbbeac30c6337d15f2ef1bdaac77f18c46e4000b5c3fdf41d353afdf13a"} Dec 10 23:09:56 crc kubenswrapper[4791]: I1210 23:09:56.867509 4791 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/swift-proxy-c7bbc5ff-rlprn" Dec 10 23:09:56 crc kubenswrapper[4791]: I1210 23:09:56.877372 4791 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/swift-proxy-c7bbc5ff-rlprn" Dec 10 23:09:57 crc kubenswrapper[4791]: I1210 23:09:57.434300 4791 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/glance-default-external-api-0"] Dec 10 23:09:57 crc kubenswrapper[4791]: I1210 23:09:57.434889 4791 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/glance-default-external-api-0" podUID="e51944e5-b91a-4310-9693-5333a680b9ff" containerName="glance-log" containerID="cri-o://410370d8e7a6e6b4006e81059ec37dab14f19eb22582c9d4426ad2345c56d96e" gracePeriod=30 Dec 10 23:09:57 crc kubenswrapper[4791]: I1210 23:09:57.434975 4791 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/glance-default-external-api-0" podUID="e51944e5-b91a-4310-9693-5333a680b9ff" containerName="glance-httpd" containerID="cri-o://192f172550e5b82eaddebad8f678283f7d534c4c29b13ec6c9004f66e467599e" gracePeriod=30 Dec 10 23:09:57 crc kubenswrapper[4791]: I1210 23:09:57.730861 4791 generic.go:334] "Generic (PLEG): container finished" podID="e51944e5-b91a-4310-9693-5333a680b9ff" containerID="410370d8e7a6e6b4006e81059ec37dab14f19eb22582c9d4426ad2345c56d96e" exitCode=143 Dec 10 23:09:57 crc kubenswrapper[4791]: I1210 23:09:57.731776 4791 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"e51944e5-b91a-4310-9693-5333a680b9ff","Type":"ContainerDied","Data":"410370d8e7a6e6b4006e81059ec37dab14f19eb22582c9d4426ad2345c56d96e"} Dec 10 23:09:58 crc kubenswrapper[4791]: I1210 23:09:58.278706 4791 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/glance-default-internal-api-0"] Dec 10 23:09:58 crc kubenswrapper[4791]: I1210 23:09:58.279125 4791 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/glance-default-internal-api-0" podUID="9460828b-2713-404b-8c14-ca25566937f4" containerName="glance-log" containerID="cri-o://521f2aafeb6650480862ba7b822c1a7b87190962da901d6b233d82ced08d46f6" gracePeriod=30 Dec 10 23:09:58 crc kubenswrapper[4791]: I1210 23:09:58.279228 4791 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/glance-default-internal-api-0" podUID="9460828b-2713-404b-8c14-ca25566937f4" containerName="glance-httpd" containerID="cri-o://b4f9dd4411c75b64e8ac6561f42e13b64709dd0bfd4f266b0d540210a7768a27" gracePeriod=30 Dec 10 23:09:58 crc kubenswrapper[4791]: I1210 23:09:58.742609 4791 generic.go:334] "Generic (PLEG): container finished" podID="9460828b-2713-404b-8c14-ca25566937f4" containerID="521f2aafeb6650480862ba7b822c1a7b87190962da901d6b233d82ced08d46f6" exitCode=143 Dec 10 23:09:58 crc kubenswrapper[4791]: I1210 23:09:58.742698 4791 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-internal-api-0" event={"ID":"9460828b-2713-404b-8c14-ca25566937f4","Type":"ContainerDied","Data":"521f2aafeb6650480862ba7b822c1a7b87190962da901d6b233d82ced08d46f6"} Dec 10 23:10:00 crc kubenswrapper[4791]: I1210 23:10:00.766912 4791 generic.go:334] "Generic (PLEG): container finished" podID="e51944e5-b91a-4310-9693-5333a680b9ff" containerID="192f172550e5b82eaddebad8f678283f7d534c4c29b13ec6c9004f66e467599e" exitCode=0 Dec 10 23:10:00 crc kubenswrapper[4791]: I1210 23:10:00.766984 4791 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"e51944e5-b91a-4310-9693-5333a680b9ff","Type":"ContainerDied","Data":"192f172550e5b82eaddebad8f678283f7d534c4c29b13ec6c9004f66e467599e"} Dec 10 23:10:01 crc kubenswrapper[4791]: I1210 23:10:01.780935 4791 generic.go:334] "Generic (PLEG): container finished" podID="9460828b-2713-404b-8c14-ca25566937f4" containerID="b4f9dd4411c75b64e8ac6561f42e13b64709dd0bfd4f266b0d540210a7768a27" exitCode=0 Dec 10 23:10:01 crc kubenswrapper[4791]: I1210 23:10:01.780986 4791 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-internal-api-0" event={"ID":"9460828b-2713-404b-8c14-ca25566937f4","Type":"ContainerDied","Data":"b4f9dd4411c75b64e8ac6561f42e13b64709dd0bfd4f266b0d540210a7768a27"} Dec 10 23:10:04 crc kubenswrapper[4791]: I1210 23:10:04.761930 4791 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-external-api-0" Dec 10 23:10:04 crc kubenswrapper[4791]: I1210 23:10:04.807881 4791 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"e51944e5-b91a-4310-9693-5333a680b9ff","Type":"ContainerDied","Data":"0c8b58f66eadcfcb89b26e96d631131882ba4ffa15fe1e84d9225a05f54bbd28"} Dec 10 23:10:04 crc kubenswrapper[4791]: I1210 23:10:04.807938 4791 scope.go:117] "RemoveContainer" containerID="192f172550e5b82eaddebad8f678283f7d534c4c29b13ec6c9004f66e467599e" Dec 10 23:10:04 crc kubenswrapper[4791]: I1210 23:10:04.808048 4791 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-external-api-0" Dec 10 23:10:04 crc kubenswrapper[4791]: I1210 23:10:04.811643 4791 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-conductor-db-sync-jff9b" event={"ID":"075728ce-048e-452d-adff-0e3d167d6673","Type":"ContainerStarted","Data":"b94a8cbc00ff61fc1de29573492cf4b269b745351fa58d04b9a77143fae5b87d"} Dec 10 23:10:04 crc kubenswrapper[4791]: I1210 23:10:04.838897 4791 scope.go:117] "RemoveContainer" containerID="410370d8e7a6e6b4006e81059ec37dab14f19eb22582c9d4426ad2345c56d96e" Dec 10 23:10:04 crc kubenswrapper[4791]: I1210 23:10:04.853544 4791 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-cell0-conductor-db-sync-jff9b" podStartSLOduration=1.999225473 podStartE2EDuration="9.853523995s" podCreationTimestamp="2025-12-10 23:09:55 +0000 UTC" firstStartedPulling="2025-12-10 23:09:56.640106966 +0000 UTC m=+1231.069724579" lastFinishedPulling="2025-12-10 23:10:04.494405488 +0000 UTC m=+1238.924023101" observedRunningTime="2025-12-10 23:10:04.840145949 +0000 UTC m=+1239.269763562" watchObservedRunningTime="2025-12-10 23:10:04.853523995 +0000 UTC m=+1239.283141598" Dec 10 23:10:04 crc kubenswrapper[4791]: I1210 23:10:04.901810 4791 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/e51944e5-b91a-4310-9693-5333a680b9ff-config-data\") pod \"e51944e5-b91a-4310-9693-5333a680b9ff\" (UID: \"e51944e5-b91a-4310-9693-5333a680b9ff\") " Dec 10 23:10:04 crc kubenswrapper[4791]: I1210 23:10:04.901936 4791 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"glance\" (UniqueName: \"kubernetes.io/local-volume/local-storage06-crc\") pod \"e51944e5-b91a-4310-9693-5333a680b9ff\" (UID: \"e51944e5-b91a-4310-9693-5333a680b9ff\") " Dec 10 23:10:04 crc kubenswrapper[4791]: I1210 23:10:04.901978 4791 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/e51944e5-b91a-4310-9693-5333a680b9ff-combined-ca-bundle\") pod \"e51944e5-b91a-4310-9693-5333a680b9ff\" (UID: \"e51944e5-b91a-4310-9693-5333a680b9ff\") " Dec 10 23:10:04 crc kubenswrapper[4791]: I1210 23:10:04.902014 4791 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/e51944e5-b91a-4310-9693-5333a680b9ff-logs\") pod \"e51944e5-b91a-4310-9693-5333a680b9ff\" (UID: \"e51944e5-b91a-4310-9693-5333a680b9ff\") " Dec 10 23:10:04 crc kubenswrapper[4791]: I1210 23:10:04.902063 4791 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/e51944e5-b91a-4310-9693-5333a680b9ff-scripts\") pod \"e51944e5-b91a-4310-9693-5333a680b9ff\" (UID: \"e51944e5-b91a-4310-9693-5333a680b9ff\") " Dec 10 23:10:04 crc kubenswrapper[4791]: I1210 23:10:04.902165 4791 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-wsh77\" (UniqueName: \"kubernetes.io/projected/e51944e5-b91a-4310-9693-5333a680b9ff-kube-api-access-wsh77\") pod \"e51944e5-b91a-4310-9693-5333a680b9ff\" (UID: \"e51944e5-b91a-4310-9693-5333a680b9ff\") " Dec 10 23:10:04 crc kubenswrapper[4791]: I1210 23:10:04.902202 4791 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/e51944e5-b91a-4310-9693-5333a680b9ff-httpd-run\") pod \"e51944e5-b91a-4310-9693-5333a680b9ff\" (UID: \"e51944e5-b91a-4310-9693-5333a680b9ff\") " Dec 10 23:10:04 crc kubenswrapper[4791]: I1210 23:10:04.902239 4791 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/e51944e5-b91a-4310-9693-5333a680b9ff-public-tls-certs\") pod \"e51944e5-b91a-4310-9693-5333a680b9ff\" (UID: \"e51944e5-b91a-4310-9693-5333a680b9ff\") " Dec 10 23:10:04 crc kubenswrapper[4791]: I1210 23:10:04.904827 4791 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/e51944e5-b91a-4310-9693-5333a680b9ff-logs" (OuterVolumeSpecName: "logs") pod "e51944e5-b91a-4310-9693-5333a680b9ff" (UID: "e51944e5-b91a-4310-9693-5333a680b9ff"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 10 23:10:04 crc kubenswrapper[4791]: I1210 23:10:04.907079 4791 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/e51944e5-b91a-4310-9693-5333a680b9ff-httpd-run" (OuterVolumeSpecName: "httpd-run") pod "e51944e5-b91a-4310-9693-5333a680b9ff" (UID: "e51944e5-b91a-4310-9693-5333a680b9ff"). InnerVolumeSpecName "httpd-run". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 10 23:10:04 crc kubenswrapper[4791]: I1210 23:10:04.909330 4791 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/e51944e5-b91a-4310-9693-5333a680b9ff-kube-api-access-wsh77" (OuterVolumeSpecName: "kube-api-access-wsh77") pod "e51944e5-b91a-4310-9693-5333a680b9ff" (UID: "e51944e5-b91a-4310-9693-5333a680b9ff"). InnerVolumeSpecName "kube-api-access-wsh77". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 10 23:10:04 crc kubenswrapper[4791]: I1210 23:10:04.918504 4791 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/e51944e5-b91a-4310-9693-5333a680b9ff-scripts" (OuterVolumeSpecName: "scripts") pod "e51944e5-b91a-4310-9693-5333a680b9ff" (UID: "e51944e5-b91a-4310-9693-5333a680b9ff"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 10 23:10:04 crc kubenswrapper[4791]: I1210 23:10:04.920238 4791 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/local-volume/local-storage06-crc" (OuterVolumeSpecName: "glance") pod "e51944e5-b91a-4310-9693-5333a680b9ff" (UID: "e51944e5-b91a-4310-9693-5333a680b9ff"). InnerVolumeSpecName "local-storage06-crc". PluginName "kubernetes.io/local-volume", VolumeGidValue "" Dec 10 23:10:04 crc kubenswrapper[4791]: I1210 23:10:04.950315 4791 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/e51944e5-b91a-4310-9693-5333a680b9ff-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "e51944e5-b91a-4310-9693-5333a680b9ff" (UID: "e51944e5-b91a-4310-9693-5333a680b9ff"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 10 23:10:04 crc kubenswrapper[4791]: I1210 23:10:04.975494 4791 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/e51944e5-b91a-4310-9693-5333a680b9ff-config-data" (OuterVolumeSpecName: "config-data") pod "e51944e5-b91a-4310-9693-5333a680b9ff" (UID: "e51944e5-b91a-4310-9693-5333a680b9ff"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 10 23:10:04 crc kubenswrapper[4791]: I1210 23:10:04.978575 4791 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/e51944e5-b91a-4310-9693-5333a680b9ff-public-tls-certs" (OuterVolumeSpecName: "public-tls-certs") pod "e51944e5-b91a-4310-9693-5333a680b9ff" (UID: "e51944e5-b91a-4310-9693-5333a680b9ff"). InnerVolumeSpecName "public-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 10 23:10:05 crc kubenswrapper[4791]: I1210 23:10:05.004413 4791 reconciler_common.go:286] "operationExecutor.UnmountDevice started for volume \"local-storage06-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage06-crc\") on node \"crc\" " Dec 10 23:10:05 crc kubenswrapper[4791]: I1210 23:10:05.004452 4791 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/e51944e5-b91a-4310-9693-5333a680b9ff-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 10 23:10:05 crc kubenswrapper[4791]: I1210 23:10:05.004464 4791 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/e51944e5-b91a-4310-9693-5333a680b9ff-logs\") on node \"crc\" DevicePath \"\"" Dec 10 23:10:05 crc kubenswrapper[4791]: I1210 23:10:05.004472 4791 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/e51944e5-b91a-4310-9693-5333a680b9ff-scripts\") on node \"crc\" DevicePath \"\"" Dec 10 23:10:05 crc kubenswrapper[4791]: I1210 23:10:05.004482 4791 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-wsh77\" (UniqueName: \"kubernetes.io/projected/e51944e5-b91a-4310-9693-5333a680b9ff-kube-api-access-wsh77\") on node \"crc\" DevicePath \"\"" Dec 10 23:10:05 crc kubenswrapper[4791]: I1210 23:10:05.004492 4791 reconciler_common.go:293] "Volume detached for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/e51944e5-b91a-4310-9693-5333a680b9ff-httpd-run\") on node \"crc\" DevicePath \"\"" Dec 10 23:10:05 crc kubenswrapper[4791]: I1210 23:10:05.004517 4791 reconciler_common.go:293] "Volume detached for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/e51944e5-b91a-4310-9693-5333a680b9ff-public-tls-certs\") on node \"crc\" DevicePath \"\"" Dec 10 23:10:05 crc kubenswrapper[4791]: I1210 23:10:05.004846 4791 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/e51944e5-b91a-4310-9693-5333a680b9ff-config-data\") on node \"crc\" DevicePath \"\"" Dec 10 23:10:05 crc kubenswrapper[4791]: I1210 23:10:05.024228 4791 operation_generator.go:917] UnmountDevice succeeded for volume "local-storage06-crc" (UniqueName: "kubernetes.io/local-volume/local-storage06-crc") on node "crc" Dec 10 23:10:05 crc kubenswrapper[4791]: I1210 23:10:05.107527 4791 reconciler_common.go:293] "Volume detached for volume \"local-storage06-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage06-crc\") on node \"crc\" DevicePath \"\"" Dec 10 23:10:05 crc kubenswrapper[4791]: I1210 23:10:05.139255 4791 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/glance-default-external-api-0"] Dec 10 23:10:05 crc kubenswrapper[4791]: I1210 23:10:05.152688 4791 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/glance-default-external-api-0"] Dec 10 23:10:05 crc kubenswrapper[4791]: I1210 23:10:05.163420 4791 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/glance-default-external-api-0"] Dec 10 23:10:05 crc kubenswrapper[4791]: E1210 23:10:05.164107 4791 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="e51944e5-b91a-4310-9693-5333a680b9ff" containerName="glance-log" Dec 10 23:10:05 crc kubenswrapper[4791]: I1210 23:10:05.164133 4791 state_mem.go:107] "Deleted CPUSet assignment" podUID="e51944e5-b91a-4310-9693-5333a680b9ff" containerName="glance-log" Dec 10 23:10:05 crc kubenswrapper[4791]: E1210 23:10:05.164151 4791 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="e51944e5-b91a-4310-9693-5333a680b9ff" containerName="glance-httpd" Dec 10 23:10:05 crc kubenswrapper[4791]: I1210 23:10:05.164162 4791 state_mem.go:107] "Deleted CPUSet assignment" podUID="e51944e5-b91a-4310-9693-5333a680b9ff" containerName="glance-httpd" Dec 10 23:10:05 crc kubenswrapper[4791]: I1210 23:10:05.164447 4791 memory_manager.go:354] "RemoveStaleState removing state" podUID="e51944e5-b91a-4310-9693-5333a680b9ff" containerName="glance-log" Dec 10 23:10:05 crc kubenswrapper[4791]: I1210 23:10:05.164480 4791 memory_manager.go:354] "RemoveStaleState removing state" podUID="e51944e5-b91a-4310-9693-5333a680b9ff" containerName="glance-httpd" Dec 10 23:10:05 crc kubenswrapper[4791]: I1210 23:10:05.165951 4791 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-external-api-0" Dec 10 23:10:05 crc kubenswrapper[4791]: I1210 23:10:05.170535 4791 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"glance-default-external-config-data" Dec 10 23:10:05 crc kubenswrapper[4791]: I1210 23:10:05.171292 4791 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-glance-default-public-svc" Dec 10 23:10:05 crc kubenswrapper[4791]: I1210 23:10:05.195954 4791 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-default-external-api-0"] Dec 10 23:10:05 crc kubenswrapper[4791]: I1210 23:10:05.314360 4791 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage06-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage06-crc\") pod \"glance-default-external-api-0\" (UID: \"9655fe55-b388-4296-b9e4-198ea42c70cd\") " pod="openstack/glance-default-external-api-0" Dec 10 23:10:05 crc kubenswrapper[4791]: I1210 23:10:05.314757 4791 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/9655fe55-b388-4296-b9e4-198ea42c70cd-httpd-run\") pod \"glance-default-external-api-0\" (UID: \"9655fe55-b388-4296-b9e4-198ea42c70cd\") " pod="openstack/glance-default-external-api-0" Dec 10 23:10:05 crc kubenswrapper[4791]: I1210 23:10:05.314866 4791 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/9655fe55-b388-4296-b9e4-198ea42c70cd-scripts\") pod \"glance-default-external-api-0\" (UID: \"9655fe55-b388-4296-b9e4-198ea42c70cd\") " pod="openstack/glance-default-external-api-0" Dec 10 23:10:05 crc kubenswrapper[4791]: I1210 23:10:05.314923 4791 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-pwpqm\" (UniqueName: \"kubernetes.io/projected/9655fe55-b388-4296-b9e4-198ea42c70cd-kube-api-access-pwpqm\") pod \"glance-default-external-api-0\" (UID: \"9655fe55-b388-4296-b9e4-198ea42c70cd\") " pod="openstack/glance-default-external-api-0" Dec 10 23:10:05 crc kubenswrapper[4791]: I1210 23:10:05.315032 4791 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/9655fe55-b388-4296-b9e4-198ea42c70cd-logs\") pod \"glance-default-external-api-0\" (UID: \"9655fe55-b388-4296-b9e4-198ea42c70cd\") " pod="openstack/glance-default-external-api-0" Dec 10 23:10:05 crc kubenswrapper[4791]: I1210 23:10:05.315080 4791 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/9655fe55-b388-4296-b9e4-198ea42c70cd-public-tls-certs\") pod \"glance-default-external-api-0\" (UID: \"9655fe55-b388-4296-b9e4-198ea42c70cd\") " pod="openstack/glance-default-external-api-0" Dec 10 23:10:05 crc kubenswrapper[4791]: I1210 23:10:05.315130 4791 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/9655fe55-b388-4296-b9e4-198ea42c70cd-config-data\") pod \"glance-default-external-api-0\" (UID: \"9655fe55-b388-4296-b9e4-198ea42c70cd\") " pod="openstack/glance-default-external-api-0" Dec 10 23:10:05 crc kubenswrapper[4791]: I1210 23:10:05.315192 4791 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/9655fe55-b388-4296-b9e4-198ea42c70cd-combined-ca-bundle\") pod \"glance-default-external-api-0\" (UID: \"9655fe55-b388-4296-b9e4-198ea42c70cd\") " pod="openstack/glance-default-external-api-0" Dec 10 23:10:05 crc kubenswrapper[4791]: I1210 23:10:05.417637 4791 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/9655fe55-b388-4296-b9e4-198ea42c70cd-logs\") pod \"glance-default-external-api-0\" (UID: \"9655fe55-b388-4296-b9e4-198ea42c70cd\") " pod="openstack/glance-default-external-api-0" Dec 10 23:10:05 crc kubenswrapper[4791]: I1210 23:10:05.417720 4791 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/9655fe55-b388-4296-b9e4-198ea42c70cd-public-tls-certs\") pod \"glance-default-external-api-0\" (UID: \"9655fe55-b388-4296-b9e4-198ea42c70cd\") " pod="openstack/glance-default-external-api-0" Dec 10 23:10:05 crc kubenswrapper[4791]: I1210 23:10:05.417809 4791 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/9655fe55-b388-4296-b9e4-198ea42c70cd-config-data\") pod \"glance-default-external-api-0\" (UID: \"9655fe55-b388-4296-b9e4-198ea42c70cd\") " pod="openstack/glance-default-external-api-0" Dec 10 23:10:05 crc kubenswrapper[4791]: I1210 23:10:05.417867 4791 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/9655fe55-b388-4296-b9e4-198ea42c70cd-combined-ca-bundle\") pod \"glance-default-external-api-0\" (UID: \"9655fe55-b388-4296-b9e4-198ea42c70cd\") " pod="openstack/glance-default-external-api-0" Dec 10 23:10:05 crc kubenswrapper[4791]: I1210 23:10:05.417925 4791 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage06-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage06-crc\") pod \"glance-default-external-api-0\" (UID: \"9655fe55-b388-4296-b9e4-198ea42c70cd\") " pod="openstack/glance-default-external-api-0" Dec 10 23:10:05 crc kubenswrapper[4791]: I1210 23:10:05.417964 4791 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/9655fe55-b388-4296-b9e4-198ea42c70cd-httpd-run\") pod \"glance-default-external-api-0\" (UID: \"9655fe55-b388-4296-b9e4-198ea42c70cd\") " pod="openstack/glance-default-external-api-0" Dec 10 23:10:05 crc kubenswrapper[4791]: I1210 23:10:05.418035 4791 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/9655fe55-b388-4296-b9e4-198ea42c70cd-scripts\") pod \"glance-default-external-api-0\" (UID: \"9655fe55-b388-4296-b9e4-198ea42c70cd\") " pod="openstack/glance-default-external-api-0" Dec 10 23:10:05 crc kubenswrapper[4791]: I1210 23:10:05.418066 4791 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-pwpqm\" (UniqueName: \"kubernetes.io/projected/9655fe55-b388-4296-b9e4-198ea42c70cd-kube-api-access-pwpqm\") pod \"glance-default-external-api-0\" (UID: \"9655fe55-b388-4296-b9e4-198ea42c70cd\") " pod="openstack/glance-default-external-api-0" Dec 10 23:10:05 crc kubenswrapper[4791]: I1210 23:10:05.418422 4791 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage06-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage06-crc\") pod \"glance-default-external-api-0\" (UID: \"9655fe55-b388-4296-b9e4-198ea42c70cd\") device mount path \"/mnt/openstack/pv06\"" pod="openstack/glance-default-external-api-0" Dec 10 23:10:05 crc kubenswrapper[4791]: I1210 23:10:05.418484 4791 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/9655fe55-b388-4296-b9e4-198ea42c70cd-httpd-run\") pod \"glance-default-external-api-0\" (UID: \"9655fe55-b388-4296-b9e4-198ea42c70cd\") " pod="openstack/glance-default-external-api-0" Dec 10 23:10:05 crc kubenswrapper[4791]: I1210 23:10:05.421517 4791 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/9655fe55-b388-4296-b9e4-198ea42c70cd-logs\") pod \"glance-default-external-api-0\" (UID: \"9655fe55-b388-4296-b9e4-198ea42c70cd\") " pod="openstack/glance-default-external-api-0" Dec 10 23:10:05 crc kubenswrapper[4791]: I1210 23:10:05.422567 4791 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/9655fe55-b388-4296-b9e4-198ea42c70cd-public-tls-certs\") pod \"glance-default-external-api-0\" (UID: \"9655fe55-b388-4296-b9e4-198ea42c70cd\") " pod="openstack/glance-default-external-api-0" Dec 10 23:10:05 crc kubenswrapper[4791]: I1210 23:10:05.423192 4791 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/9655fe55-b388-4296-b9e4-198ea42c70cd-scripts\") pod \"glance-default-external-api-0\" (UID: \"9655fe55-b388-4296-b9e4-198ea42c70cd\") " pod="openstack/glance-default-external-api-0" Dec 10 23:10:05 crc kubenswrapper[4791]: I1210 23:10:05.424604 4791 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/9655fe55-b388-4296-b9e4-198ea42c70cd-config-data\") pod \"glance-default-external-api-0\" (UID: \"9655fe55-b388-4296-b9e4-198ea42c70cd\") " pod="openstack/glance-default-external-api-0" Dec 10 23:10:05 crc kubenswrapper[4791]: I1210 23:10:05.424989 4791 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/9655fe55-b388-4296-b9e4-198ea42c70cd-combined-ca-bundle\") pod \"glance-default-external-api-0\" (UID: \"9655fe55-b388-4296-b9e4-198ea42c70cd\") " pod="openstack/glance-default-external-api-0" Dec 10 23:10:05 crc kubenswrapper[4791]: I1210 23:10:05.440533 4791 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-pwpqm\" (UniqueName: \"kubernetes.io/projected/9655fe55-b388-4296-b9e4-198ea42c70cd-kube-api-access-pwpqm\") pod \"glance-default-external-api-0\" (UID: \"9655fe55-b388-4296-b9e4-198ea42c70cd\") " pod="openstack/glance-default-external-api-0" Dec 10 23:10:05 crc kubenswrapper[4791]: I1210 23:10:05.449811 4791 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage06-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage06-crc\") pod \"glance-default-external-api-0\" (UID: \"9655fe55-b388-4296-b9e4-198ea42c70cd\") " pod="openstack/glance-default-external-api-0" Dec 10 23:10:05 crc kubenswrapper[4791]: I1210 23:10:05.488171 4791 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-external-api-0" Dec 10 23:10:05 crc kubenswrapper[4791]: I1210 23:10:05.852612 4791 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-internal-api-0" event={"ID":"9460828b-2713-404b-8c14-ca25566937f4","Type":"ContainerDied","Data":"49081a7073620d4096b556ad3cf34629657b4fce3626e6ec53cb34cca27a990c"} Dec 10 23:10:05 crc kubenswrapper[4791]: I1210 23:10:05.853060 4791 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="49081a7073620d4096b556ad3cf34629657b4fce3626e6ec53cb34cca27a990c" Dec 10 23:10:05 crc kubenswrapper[4791]: I1210 23:10:05.879011 4791 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-internal-api-0" Dec 10 23:10:05 crc kubenswrapper[4791]: I1210 23:10:05.894615 4791 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="e51944e5-b91a-4310-9693-5333a680b9ff" path="/var/lib/kubelet/pods/e51944e5-b91a-4310-9693-5333a680b9ff/volumes" Dec 10 23:10:06 crc kubenswrapper[4791]: I1210 23:10:06.033603 4791 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/9460828b-2713-404b-8c14-ca25566937f4-scripts\") pod \"9460828b-2713-404b-8c14-ca25566937f4\" (UID: \"9460828b-2713-404b-8c14-ca25566937f4\") " Dec 10 23:10:06 crc kubenswrapper[4791]: I1210 23:10:06.033681 4791 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/9460828b-2713-404b-8c14-ca25566937f4-config-data\") pod \"9460828b-2713-404b-8c14-ca25566937f4\" (UID: \"9460828b-2713-404b-8c14-ca25566937f4\") " Dec 10 23:10:06 crc kubenswrapper[4791]: I1210 23:10:06.033784 4791 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-9v6tz\" (UniqueName: \"kubernetes.io/projected/9460828b-2713-404b-8c14-ca25566937f4-kube-api-access-9v6tz\") pod \"9460828b-2713-404b-8c14-ca25566937f4\" (UID: \"9460828b-2713-404b-8c14-ca25566937f4\") " Dec 10 23:10:06 crc kubenswrapper[4791]: I1210 23:10:06.033842 4791 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"glance\" (UniqueName: \"kubernetes.io/local-volume/local-storage03-crc\") pod \"9460828b-2713-404b-8c14-ca25566937f4\" (UID: \"9460828b-2713-404b-8c14-ca25566937f4\") " Dec 10 23:10:06 crc kubenswrapper[4791]: I1210 23:10:06.033957 4791 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/9460828b-2713-404b-8c14-ca25566937f4-internal-tls-certs\") pod \"9460828b-2713-404b-8c14-ca25566937f4\" (UID: \"9460828b-2713-404b-8c14-ca25566937f4\") " Dec 10 23:10:06 crc kubenswrapper[4791]: I1210 23:10:06.034006 4791 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/9460828b-2713-404b-8c14-ca25566937f4-logs\") pod \"9460828b-2713-404b-8c14-ca25566937f4\" (UID: \"9460828b-2713-404b-8c14-ca25566937f4\") " Dec 10 23:10:06 crc kubenswrapper[4791]: I1210 23:10:06.034050 4791 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/9460828b-2713-404b-8c14-ca25566937f4-httpd-run\") pod \"9460828b-2713-404b-8c14-ca25566937f4\" (UID: \"9460828b-2713-404b-8c14-ca25566937f4\") " Dec 10 23:10:06 crc kubenswrapper[4791]: I1210 23:10:06.034175 4791 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/9460828b-2713-404b-8c14-ca25566937f4-combined-ca-bundle\") pod \"9460828b-2713-404b-8c14-ca25566937f4\" (UID: \"9460828b-2713-404b-8c14-ca25566937f4\") " Dec 10 23:10:06 crc kubenswrapper[4791]: I1210 23:10:06.034982 4791 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/9460828b-2713-404b-8c14-ca25566937f4-httpd-run" (OuterVolumeSpecName: "httpd-run") pod "9460828b-2713-404b-8c14-ca25566937f4" (UID: "9460828b-2713-404b-8c14-ca25566937f4"). InnerVolumeSpecName "httpd-run". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 10 23:10:06 crc kubenswrapper[4791]: I1210 23:10:06.035404 4791 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/9460828b-2713-404b-8c14-ca25566937f4-logs" (OuterVolumeSpecName: "logs") pod "9460828b-2713-404b-8c14-ca25566937f4" (UID: "9460828b-2713-404b-8c14-ca25566937f4"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 10 23:10:06 crc kubenswrapper[4791]: I1210 23:10:06.045476 4791 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/9460828b-2713-404b-8c14-ca25566937f4-scripts" (OuterVolumeSpecName: "scripts") pod "9460828b-2713-404b-8c14-ca25566937f4" (UID: "9460828b-2713-404b-8c14-ca25566937f4"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 10 23:10:06 crc kubenswrapper[4791]: I1210 23:10:06.045670 4791 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/local-volume/local-storage03-crc" (OuterVolumeSpecName: "glance") pod "9460828b-2713-404b-8c14-ca25566937f4" (UID: "9460828b-2713-404b-8c14-ca25566937f4"). InnerVolumeSpecName "local-storage03-crc". PluginName "kubernetes.io/local-volume", VolumeGidValue "" Dec 10 23:10:06 crc kubenswrapper[4791]: I1210 23:10:06.048191 4791 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/9460828b-2713-404b-8c14-ca25566937f4-kube-api-access-9v6tz" (OuterVolumeSpecName: "kube-api-access-9v6tz") pod "9460828b-2713-404b-8c14-ca25566937f4" (UID: "9460828b-2713-404b-8c14-ca25566937f4"). InnerVolumeSpecName "kube-api-access-9v6tz". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 10 23:10:06 crc kubenswrapper[4791]: I1210 23:10:06.103186 4791 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/9460828b-2713-404b-8c14-ca25566937f4-config-data" (OuterVolumeSpecName: "config-data") pod "9460828b-2713-404b-8c14-ca25566937f4" (UID: "9460828b-2713-404b-8c14-ca25566937f4"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 10 23:10:06 crc kubenswrapper[4791]: I1210 23:10:06.120692 4791 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/9460828b-2713-404b-8c14-ca25566937f4-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "9460828b-2713-404b-8c14-ca25566937f4" (UID: "9460828b-2713-404b-8c14-ca25566937f4"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 10 23:10:06 crc kubenswrapper[4791]: I1210 23:10:06.121034 4791 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/9460828b-2713-404b-8c14-ca25566937f4-internal-tls-certs" (OuterVolumeSpecName: "internal-tls-certs") pod "9460828b-2713-404b-8c14-ca25566937f4" (UID: "9460828b-2713-404b-8c14-ca25566937f4"). InnerVolumeSpecName "internal-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 10 23:10:06 crc kubenswrapper[4791]: I1210 23:10:06.137521 4791 reconciler_common.go:293] "Volume detached for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/9460828b-2713-404b-8c14-ca25566937f4-internal-tls-certs\") on node \"crc\" DevicePath \"\"" Dec 10 23:10:06 crc kubenswrapper[4791]: I1210 23:10:06.137580 4791 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/9460828b-2713-404b-8c14-ca25566937f4-logs\") on node \"crc\" DevicePath \"\"" Dec 10 23:10:06 crc kubenswrapper[4791]: I1210 23:10:06.137598 4791 reconciler_common.go:293] "Volume detached for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/9460828b-2713-404b-8c14-ca25566937f4-httpd-run\") on node \"crc\" DevicePath \"\"" Dec 10 23:10:06 crc kubenswrapper[4791]: I1210 23:10:06.137613 4791 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/9460828b-2713-404b-8c14-ca25566937f4-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 10 23:10:06 crc kubenswrapper[4791]: I1210 23:10:06.137623 4791 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/9460828b-2713-404b-8c14-ca25566937f4-scripts\") on node \"crc\" DevicePath \"\"" Dec 10 23:10:06 crc kubenswrapper[4791]: I1210 23:10:06.137632 4791 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/9460828b-2713-404b-8c14-ca25566937f4-config-data\") on node \"crc\" DevicePath \"\"" Dec 10 23:10:06 crc kubenswrapper[4791]: I1210 23:10:06.137641 4791 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-9v6tz\" (UniqueName: \"kubernetes.io/projected/9460828b-2713-404b-8c14-ca25566937f4-kube-api-access-9v6tz\") on node \"crc\" DevicePath \"\"" Dec 10 23:10:06 crc kubenswrapper[4791]: I1210 23:10:06.137688 4791 reconciler_common.go:286] "operationExecutor.UnmountDevice started for volume \"local-storage03-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage03-crc\") on node \"crc\" " Dec 10 23:10:06 crc kubenswrapper[4791]: I1210 23:10:06.161705 4791 operation_generator.go:917] UnmountDevice succeeded for volume "local-storage03-crc" (UniqueName: "kubernetes.io/local-volume/local-storage03-crc") on node "crc" Dec 10 23:10:06 crc kubenswrapper[4791]: I1210 23:10:06.169169 4791 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-default-external-api-0"] Dec 10 23:10:06 crc kubenswrapper[4791]: I1210 23:10:06.238365 4791 reconciler_common.go:293] "Volume detached for volume \"local-storage03-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage03-crc\") on node \"crc\" DevicePath \"\"" Dec 10 23:10:06 crc kubenswrapper[4791]: I1210 23:10:06.872777 4791 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"9655fe55-b388-4296-b9e4-198ea42c70cd","Type":"ContainerStarted","Data":"adb2f6cd0c139330be95683714dd660eecbc9da6312f7801e297dd2b0f08bf28"} Dec 10 23:10:06 crc kubenswrapper[4791]: I1210 23:10:06.873090 4791 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"9655fe55-b388-4296-b9e4-198ea42c70cd","Type":"ContainerStarted","Data":"09a1f931a260317c19fb3b447194bfa1a61f3eff2abdd634806c27f647e85bdd"} Dec 10 23:10:06 crc kubenswrapper[4791]: I1210 23:10:06.872790 4791 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-internal-api-0" Dec 10 23:10:06 crc kubenswrapper[4791]: I1210 23:10:06.913939 4791 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/glance-default-internal-api-0"] Dec 10 23:10:06 crc kubenswrapper[4791]: I1210 23:10:06.925830 4791 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/glance-default-internal-api-0"] Dec 10 23:10:06 crc kubenswrapper[4791]: I1210 23:10:06.968185 4791 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/glance-default-internal-api-0"] Dec 10 23:10:06 crc kubenswrapper[4791]: E1210 23:10:06.968769 4791 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="9460828b-2713-404b-8c14-ca25566937f4" containerName="glance-httpd" Dec 10 23:10:06 crc kubenswrapper[4791]: I1210 23:10:06.968795 4791 state_mem.go:107] "Deleted CPUSet assignment" podUID="9460828b-2713-404b-8c14-ca25566937f4" containerName="glance-httpd" Dec 10 23:10:06 crc kubenswrapper[4791]: E1210 23:10:06.968862 4791 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="9460828b-2713-404b-8c14-ca25566937f4" containerName="glance-log" Dec 10 23:10:06 crc kubenswrapper[4791]: I1210 23:10:06.968869 4791 state_mem.go:107] "Deleted CPUSet assignment" podUID="9460828b-2713-404b-8c14-ca25566937f4" containerName="glance-log" Dec 10 23:10:06 crc kubenswrapper[4791]: I1210 23:10:06.969063 4791 memory_manager.go:354] "RemoveStaleState removing state" podUID="9460828b-2713-404b-8c14-ca25566937f4" containerName="glance-httpd" Dec 10 23:10:06 crc kubenswrapper[4791]: I1210 23:10:06.969082 4791 memory_manager.go:354] "RemoveStaleState removing state" podUID="9460828b-2713-404b-8c14-ca25566937f4" containerName="glance-log" Dec 10 23:10:06 crc kubenswrapper[4791]: I1210 23:10:06.995212 4791 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-default-internal-api-0"] Dec 10 23:10:06 crc kubenswrapper[4791]: I1210 23:10:06.995421 4791 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-internal-api-0" Dec 10 23:10:07 crc kubenswrapper[4791]: I1210 23:10:07.003321 4791 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"glance-default-internal-config-data" Dec 10 23:10:07 crc kubenswrapper[4791]: I1210 23:10:07.003670 4791 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-glance-default-internal-svc" Dec 10 23:10:07 crc kubenswrapper[4791]: I1210 23:10:07.157060 4791 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/d3317b5c-5c3b-4138-8f93-1bd9a159f8dd-internal-tls-certs\") pod \"glance-default-internal-api-0\" (UID: \"d3317b5c-5c3b-4138-8f93-1bd9a159f8dd\") " pod="openstack/glance-default-internal-api-0" Dec 10 23:10:07 crc kubenswrapper[4791]: I1210 23:10:07.157114 4791 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/d3317b5c-5c3b-4138-8f93-1bd9a159f8dd-scripts\") pod \"glance-default-internal-api-0\" (UID: \"d3317b5c-5c3b-4138-8f93-1bd9a159f8dd\") " pod="openstack/glance-default-internal-api-0" Dec 10 23:10:07 crc kubenswrapper[4791]: I1210 23:10:07.157137 4791 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/d3317b5c-5c3b-4138-8f93-1bd9a159f8dd-logs\") pod \"glance-default-internal-api-0\" (UID: \"d3317b5c-5c3b-4138-8f93-1bd9a159f8dd\") " pod="openstack/glance-default-internal-api-0" Dec 10 23:10:07 crc kubenswrapper[4791]: I1210 23:10:07.157227 4791 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-m7dms\" (UniqueName: \"kubernetes.io/projected/d3317b5c-5c3b-4138-8f93-1bd9a159f8dd-kube-api-access-m7dms\") pod \"glance-default-internal-api-0\" (UID: \"d3317b5c-5c3b-4138-8f93-1bd9a159f8dd\") " pod="openstack/glance-default-internal-api-0" Dec 10 23:10:07 crc kubenswrapper[4791]: I1210 23:10:07.157330 4791 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d3317b5c-5c3b-4138-8f93-1bd9a159f8dd-combined-ca-bundle\") pod \"glance-default-internal-api-0\" (UID: \"d3317b5c-5c3b-4138-8f93-1bd9a159f8dd\") " pod="openstack/glance-default-internal-api-0" Dec 10 23:10:07 crc kubenswrapper[4791]: I1210 23:10:07.157376 4791 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/d3317b5c-5c3b-4138-8f93-1bd9a159f8dd-config-data\") pod \"glance-default-internal-api-0\" (UID: \"d3317b5c-5c3b-4138-8f93-1bd9a159f8dd\") " pod="openstack/glance-default-internal-api-0" Dec 10 23:10:07 crc kubenswrapper[4791]: I1210 23:10:07.157456 4791 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/d3317b5c-5c3b-4138-8f93-1bd9a159f8dd-httpd-run\") pod \"glance-default-internal-api-0\" (UID: \"d3317b5c-5c3b-4138-8f93-1bd9a159f8dd\") " pod="openstack/glance-default-internal-api-0" Dec 10 23:10:07 crc kubenswrapper[4791]: I1210 23:10:07.157481 4791 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage03-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage03-crc\") pod \"glance-default-internal-api-0\" (UID: \"d3317b5c-5c3b-4138-8f93-1bd9a159f8dd\") " pod="openstack/glance-default-internal-api-0" Dec 10 23:10:07 crc kubenswrapper[4791]: I1210 23:10:07.260396 4791 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d3317b5c-5c3b-4138-8f93-1bd9a159f8dd-combined-ca-bundle\") pod \"glance-default-internal-api-0\" (UID: \"d3317b5c-5c3b-4138-8f93-1bd9a159f8dd\") " pod="openstack/glance-default-internal-api-0" Dec 10 23:10:07 crc kubenswrapper[4791]: I1210 23:10:07.260933 4791 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/d3317b5c-5c3b-4138-8f93-1bd9a159f8dd-config-data\") pod \"glance-default-internal-api-0\" (UID: \"d3317b5c-5c3b-4138-8f93-1bd9a159f8dd\") " pod="openstack/glance-default-internal-api-0" Dec 10 23:10:07 crc kubenswrapper[4791]: I1210 23:10:07.261023 4791 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/d3317b5c-5c3b-4138-8f93-1bd9a159f8dd-httpd-run\") pod \"glance-default-internal-api-0\" (UID: \"d3317b5c-5c3b-4138-8f93-1bd9a159f8dd\") " pod="openstack/glance-default-internal-api-0" Dec 10 23:10:07 crc kubenswrapper[4791]: I1210 23:10:07.261060 4791 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage03-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage03-crc\") pod \"glance-default-internal-api-0\" (UID: \"d3317b5c-5c3b-4138-8f93-1bd9a159f8dd\") " pod="openstack/glance-default-internal-api-0" Dec 10 23:10:07 crc kubenswrapper[4791]: I1210 23:10:07.261115 4791 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/d3317b5c-5c3b-4138-8f93-1bd9a159f8dd-internal-tls-certs\") pod \"glance-default-internal-api-0\" (UID: \"d3317b5c-5c3b-4138-8f93-1bd9a159f8dd\") " pod="openstack/glance-default-internal-api-0" Dec 10 23:10:07 crc kubenswrapper[4791]: I1210 23:10:07.261153 4791 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/d3317b5c-5c3b-4138-8f93-1bd9a159f8dd-scripts\") pod \"glance-default-internal-api-0\" (UID: \"d3317b5c-5c3b-4138-8f93-1bd9a159f8dd\") " pod="openstack/glance-default-internal-api-0" Dec 10 23:10:07 crc kubenswrapper[4791]: I1210 23:10:07.261184 4791 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/d3317b5c-5c3b-4138-8f93-1bd9a159f8dd-logs\") pod \"glance-default-internal-api-0\" (UID: \"d3317b5c-5c3b-4138-8f93-1bd9a159f8dd\") " pod="openstack/glance-default-internal-api-0" Dec 10 23:10:07 crc kubenswrapper[4791]: I1210 23:10:07.261245 4791 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-m7dms\" (UniqueName: \"kubernetes.io/projected/d3317b5c-5c3b-4138-8f93-1bd9a159f8dd-kube-api-access-m7dms\") pod \"glance-default-internal-api-0\" (UID: \"d3317b5c-5c3b-4138-8f93-1bd9a159f8dd\") " pod="openstack/glance-default-internal-api-0" Dec 10 23:10:07 crc kubenswrapper[4791]: I1210 23:10:07.262015 4791 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage03-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage03-crc\") pod \"glance-default-internal-api-0\" (UID: \"d3317b5c-5c3b-4138-8f93-1bd9a159f8dd\") device mount path \"/mnt/openstack/pv03\"" pod="openstack/glance-default-internal-api-0" Dec 10 23:10:07 crc kubenswrapper[4791]: I1210 23:10:07.262264 4791 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/d3317b5c-5c3b-4138-8f93-1bd9a159f8dd-logs\") pod \"glance-default-internal-api-0\" (UID: \"d3317b5c-5c3b-4138-8f93-1bd9a159f8dd\") " pod="openstack/glance-default-internal-api-0" Dec 10 23:10:07 crc kubenswrapper[4791]: I1210 23:10:07.262521 4791 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/d3317b5c-5c3b-4138-8f93-1bd9a159f8dd-httpd-run\") pod \"glance-default-internal-api-0\" (UID: \"d3317b5c-5c3b-4138-8f93-1bd9a159f8dd\") " pod="openstack/glance-default-internal-api-0" Dec 10 23:10:07 crc kubenswrapper[4791]: I1210 23:10:07.266814 4791 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/d3317b5c-5c3b-4138-8f93-1bd9a159f8dd-scripts\") pod \"glance-default-internal-api-0\" (UID: \"d3317b5c-5c3b-4138-8f93-1bd9a159f8dd\") " pod="openstack/glance-default-internal-api-0" Dec 10 23:10:07 crc kubenswrapper[4791]: I1210 23:10:07.270506 4791 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/d3317b5c-5c3b-4138-8f93-1bd9a159f8dd-internal-tls-certs\") pod \"glance-default-internal-api-0\" (UID: \"d3317b5c-5c3b-4138-8f93-1bd9a159f8dd\") " pod="openstack/glance-default-internal-api-0" Dec 10 23:10:07 crc kubenswrapper[4791]: I1210 23:10:07.271301 4791 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d3317b5c-5c3b-4138-8f93-1bd9a159f8dd-combined-ca-bundle\") pod \"glance-default-internal-api-0\" (UID: \"d3317b5c-5c3b-4138-8f93-1bd9a159f8dd\") " pod="openstack/glance-default-internal-api-0" Dec 10 23:10:07 crc kubenswrapper[4791]: I1210 23:10:07.275692 4791 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/d3317b5c-5c3b-4138-8f93-1bd9a159f8dd-config-data\") pod \"glance-default-internal-api-0\" (UID: \"d3317b5c-5c3b-4138-8f93-1bd9a159f8dd\") " pod="openstack/glance-default-internal-api-0" Dec 10 23:10:07 crc kubenswrapper[4791]: I1210 23:10:07.281913 4791 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-m7dms\" (UniqueName: \"kubernetes.io/projected/d3317b5c-5c3b-4138-8f93-1bd9a159f8dd-kube-api-access-m7dms\") pod \"glance-default-internal-api-0\" (UID: \"d3317b5c-5c3b-4138-8f93-1bd9a159f8dd\") " pod="openstack/glance-default-internal-api-0" Dec 10 23:10:07 crc kubenswrapper[4791]: I1210 23:10:07.298527 4791 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage03-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage03-crc\") pod \"glance-default-internal-api-0\" (UID: \"d3317b5c-5c3b-4138-8f93-1bd9a159f8dd\") " pod="openstack/glance-default-internal-api-0" Dec 10 23:10:07 crc kubenswrapper[4791]: I1210 23:10:07.327368 4791 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-internal-api-0" Dec 10 23:10:07 crc kubenswrapper[4791]: I1210 23:10:07.898852 4791 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="9460828b-2713-404b-8c14-ca25566937f4" path="/var/lib/kubelet/pods/9460828b-2713-404b-8c14-ca25566937f4/volumes" Dec 10 23:10:07 crc kubenswrapper[4791]: I1210 23:10:07.899802 4791 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"9655fe55-b388-4296-b9e4-198ea42c70cd","Type":"ContainerStarted","Data":"dfd82d6613e334401987ceb6680c7b6bcb97e544cf49cd11314b680c6cd40cda"} Dec 10 23:10:07 crc kubenswrapper[4791]: I1210 23:10:07.899836 4791 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-default-internal-api-0"] Dec 10 23:10:07 crc kubenswrapper[4791]: I1210 23:10:07.919227 4791 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/glance-default-external-api-0" podStartSLOduration=2.919208714 podStartE2EDuration="2.919208714s" podCreationTimestamp="2025-12-10 23:10:05 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-10 23:10:07.912208472 +0000 UTC m=+1242.341826085" watchObservedRunningTime="2025-12-10 23:10:07.919208714 +0000 UTC m=+1242.348826327" Dec 10 23:10:08 crc kubenswrapper[4791]: I1210 23:10:08.902620 4791 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-internal-api-0" event={"ID":"d3317b5c-5c3b-4138-8f93-1bd9a159f8dd","Type":"ContainerStarted","Data":"cf91c623231cc2ca9c9b6260d0840a61c2464f7b32765e4be892f0e52b8d7995"} Dec 10 23:10:08 crc kubenswrapper[4791]: I1210 23:10:08.903436 4791 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-internal-api-0" event={"ID":"d3317b5c-5c3b-4138-8f93-1bd9a159f8dd","Type":"ContainerStarted","Data":"6364181bb7c3477d64426443dd7d2978a3de2f908375799092d734167adefc57"} Dec 10 23:10:09 crc kubenswrapper[4791]: I1210 23:10:09.913374 4791 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-internal-api-0" event={"ID":"d3317b5c-5c3b-4138-8f93-1bd9a159f8dd","Type":"ContainerStarted","Data":"b7792d4b75508c386ed1929ca4cfa62bb93a9e2281ea7b856dc8dac783a5d481"} Dec 10 23:10:12 crc kubenswrapper[4791]: I1210 23:10:12.981268 4791 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/glance-default-internal-api-0" podStartSLOduration=6.981247981 podStartE2EDuration="6.981247981s" podCreationTimestamp="2025-12-10 23:10:06 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-10 23:10:12.972534339 +0000 UTC m=+1247.402151952" watchObservedRunningTime="2025-12-10 23:10:12.981247981 +0000 UTC m=+1247.410865604" Dec 10 23:10:15 crc kubenswrapper[4791]: I1210 23:10:15.488956 4791 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/glance-default-external-api-0" Dec 10 23:10:15 crc kubenswrapper[4791]: I1210 23:10:15.489022 4791 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/glance-default-external-api-0" Dec 10 23:10:15 crc kubenswrapper[4791]: I1210 23:10:15.552047 4791 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/glance-default-external-api-0" Dec 10 23:10:15 crc kubenswrapper[4791]: I1210 23:10:15.559931 4791 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/glance-default-external-api-0" Dec 10 23:10:15 crc kubenswrapper[4791]: I1210 23:10:15.982584 4791 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/glance-default-external-api-0" Dec 10 23:10:15 crc kubenswrapper[4791]: I1210 23:10:15.982636 4791 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/glance-default-external-api-0" Dec 10 23:10:17 crc kubenswrapper[4791]: I1210 23:10:17.363313 4791 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/glance-default-internal-api-0" Dec 10 23:10:17 crc kubenswrapper[4791]: I1210 23:10:17.363806 4791 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/glance-default-internal-api-0" Dec 10 23:10:17 crc kubenswrapper[4791]: I1210 23:10:17.371696 4791 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/ceilometer-0" podUID="053f7590-725f-4504-a204-be83d801bad2" containerName="proxy-httpd" probeResult="failure" output="HTTP probe failed with statuscode: 503" Dec 10 23:10:17 crc kubenswrapper[4791]: I1210 23:10:17.399242 4791 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/glance-default-internal-api-0" Dec 10 23:10:17 crc kubenswrapper[4791]: I1210 23:10:17.419293 4791 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/glance-default-internal-api-0" Dec 10 23:10:17 crc kubenswrapper[4791]: I1210 23:10:17.996646 4791 prober_manager.go:312] "Failed to trigger a manual run" probe="Readiness" Dec 10 23:10:17 crc kubenswrapper[4791]: I1210 23:10:17.996688 4791 prober_manager.go:312] "Failed to trigger a manual run" probe="Readiness" Dec 10 23:10:17 crc kubenswrapper[4791]: I1210 23:10:17.997067 4791 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/glance-default-internal-api-0" Dec 10 23:10:17 crc kubenswrapper[4791]: I1210 23:10:17.997114 4791 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/glance-default-internal-api-0" Dec 10 23:10:18 crc kubenswrapper[4791]: I1210 23:10:18.207075 4791 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/glance-default-external-api-0" Dec 10 23:10:18 crc kubenswrapper[4791]: I1210 23:10:18.215736 4791 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/glance-default-external-api-0" Dec 10 23:10:20 crc kubenswrapper[4791]: I1210 23:10:20.290226 4791 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/glance-default-internal-api-0" Dec 10 23:10:20 crc kubenswrapper[4791]: I1210 23:10:20.290622 4791 prober_manager.go:312] "Failed to trigger a manual run" probe="Readiness" Dec 10 23:10:20 crc kubenswrapper[4791]: I1210 23:10:20.301055 4791 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/glance-default-internal-api-0" Dec 10 23:10:21 crc kubenswrapper[4791]: I1210 23:10:21.028583 4791 generic.go:334] "Generic (PLEG): container finished" podID="075728ce-048e-452d-adff-0e3d167d6673" containerID="b94a8cbc00ff61fc1de29573492cf4b269b745351fa58d04b9a77143fae5b87d" exitCode=0 Dec 10 23:10:21 crc kubenswrapper[4791]: I1210 23:10:21.029969 4791 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-conductor-db-sync-jff9b" event={"ID":"075728ce-048e-452d-adff-0e3d167d6673","Type":"ContainerDied","Data":"b94a8cbc00ff61fc1de29573492cf4b269b745351fa58d04b9a77143fae5b87d"} Dec 10 23:10:21 crc kubenswrapper[4791]: W1210 23:10:21.703295 4791 watcher.go:93] Error while processing event ("/sys/fs/cgroup/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod075728ce_048e_452d_adff_0e3d167d6673.slice/crio-conmon-b94a8cbc00ff61fc1de29573492cf4b269b745351fa58d04b9a77143fae5b87d.scope": 0x40000100 == IN_CREATE|IN_ISDIR): inotify_add_watch /sys/fs/cgroup/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod075728ce_048e_452d_adff_0e3d167d6673.slice/crio-conmon-b94a8cbc00ff61fc1de29573492cf4b269b745351fa58d04b9a77143fae5b87d.scope: no such file or directory Dec 10 23:10:21 crc kubenswrapper[4791]: W1210 23:10:21.703403 4791 watcher.go:93] Error while processing event ("/sys/fs/cgroup/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod075728ce_048e_452d_adff_0e3d167d6673.slice/crio-b94a8cbc00ff61fc1de29573492cf4b269b745351fa58d04b9a77143fae5b87d.scope": 0x40000100 == IN_CREATE|IN_ISDIR): inotify_add_watch /sys/fs/cgroup/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod075728ce_048e_452d_adff_0e3d167d6673.slice/crio-b94a8cbc00ff61fc1de29573492cf4b269b745351fa58d04b9a77143fae5b87d.scope: no such file or directory Dec 10 23:10:22 crc kubenswrapper[4791]: I1210 23:10:22.046634 4791 generic.go:334] "Generic (PLEG): container finished" podID="053f7590-725f-4504-a204-be83d801bad2" containerID="b9fbe7eb299965c84c9556998893e0f9a21817ebbe4e48d8cd0d072a837e119a" exitCode=137 Dec 10 23:10:22 crc kubenswrapper[4791]: I1210 23:10:22.047236 4791 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"053f7590-725f-4504-a204-be83d801bad2","Type":"ContainerDied","Data":"b9fbe7eb299965c84c9556998893e0f9a21817ebbe4e48d8cd0d072a837e119a"} Dec 10 23:10:22 crc kubenswrapper[4791]: I1210 23:10:22.047278 4791 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"053f7590-725f-4504-a204-be83d801bad2","Type":"ContainerDied","Data":"b599db3cab1becab15d15a46369973f0d8c5999cd6e208610c4736b0264bab56"} Dec 10 23:10:22 crc kubenswrapper[4791]: I1210 23:10:22.047296 4791 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="b599db3cab1becab15d15a46369973f0d8c5999cd6e208610c4736b0264bab56" Dec 10 23:10:22 crc kubenswrapper[4791]: I1210 23:10:22.108033 4791 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Dec 10 23:10:22 crc kubenswrapper[4791]: I1210 23:10:22.173522 4791 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/053f7590-725f-4504-a204-be83d801bad2-log-httpd\") pod \"053f7590-725f-4504-a204-be83d801bad2\" (UID: \"053f7590-725f-4504-a204-be83d801bad2\") " Dec 10 23:10:22 crc kubenswrapper[4791]: I1210 23:10:22.173559 4791 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/053f7590-725f-4504-a204-be83d801bad2-scripts\") pod \"053f7590-725f-4504-a204-be83d801bad2\" (UID: \"053f7590-725f-4504-a204-be83d801bad2\") " Dec 10 23:10:22 crc kubenswrapper[4791]: I1210 23:10:22.173642 4791 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/053f7590-725f-4504-a204-be83d801bad2-combined-ca-bundle\") pod \"053f7590-725f-4504-a204-be83d801bad2\" (UID: \"053f7590-725f-4504-a204-be83d801bad2\") " Dec 10 23:10:22 crc kubenswrapper[4791]: I1210 23:10:22.173662 4791 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/053f7590-725f-4504-a204-be83d801bad2-run-httpd\") pod \"053f7590-725f-4504-a204-be83d801bad2\" (UID: \"053f7590-725f-4504-a204-be83d801bad2\") " Dec 10 23:10:22 crc kubenswrapper[4791]: I1210 23:10:22.173757 4791 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/053f7590-725f-4504-a204-be83d801bad2-config-data\") pod \"053f7590-725f-4504-a204-be83d801bad2\" (UID: \"053f7590-725f-4504-a204-be83d801bad2\") " Dec 10 23:10:22 crc kubenswrapper[4791]: I1210 23:10:22.173787 4791 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/053f7590-725f-4504-a204-be83d801bad2-sg-core-conf-yaml\") pod \"053f7590-725f-4504-a204-be83d801bad2\" (UID: \"053f7590-725f-4504-a204-be83d801bad2\") " Dec 10 23:10:22 crc kubenswrapper[4791]: I1210 23:10:22.173848 4791 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-wrsg2\" (UniqueName: \"kubernetes.io/projected/053f7590-725f-4504-a204-be83d801bad2-kube-api-access-wrsg2\") pod \"053f7590-725f-4504-a204-be83d801bad2\" (UID: \"053f7590-725f-4504-a204-be83d801bad2\") " Dec 10 23:10:22 crc kubenswrapper[4791]: I1210 23:10:22.174058 4791 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/053f7590-725f-4504-a204-be83d801bad2-log-httpd" (OuterVolumeSpecName: "log-httpd") pod "053f7590-725f-4504-a204-be83d801bad2" (UID: "053f7590-725f-4504-a204-be83d801bad2"). InnerVolumeSpecName "log-httpd". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 10 23:10:22 crc kubenswrapper[4791]: I1210 23:10:22.174519 4791 reconciler_common.go:293] "Volume detached for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/053f7590-725f-4504-a204-be83d801bad2-log-httpd\") on node \"crc\" DevicePath \"\"" Dec 10 23:10:22 crc kubenswrapper[4791]: I1210 23:10:22.174903 4791 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/053f7590-725f-4504-a204-be83d801bad2-run-httpd" (OuterVolumeSpecName: "run-httpd") pod "053f7590-725f-4504-a204-be83d801bad2" (UID: "053f7590-725f-4504-a204-be83d801bad2"). InnerVolumeSpecName "run-httpd". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 10 23:10:22 crc kubenswrapper[4791]: I1210 23:10:22.181448 4791 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/053f7590-725f-4504-a204-be83d801bad2-scripts" (OuterVolumeSpecName: "scripts") pod "053f7590-725f-4504-a204-be83d801bad2" (UID: "053f7590-725f-4504-a204-be83d801bad2"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 10 23:10:22 crc kubenswrapper[4791]: I1210 23:10:22.192599 4791 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/053f7590-725f-4504-a204-be83d801bad2-kube-api-access-wrsg2" (OuterVolumeSpecName: "kube-api-access-wrsg2") pod "053f7590-725f-4504-a204-be83d801bad2" (UID: "053f7590-725f-4504-a204-be83d801bad2"). InnerVolumeSpecName "kube-api-access-wrsg2". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 10 23:10:22 crc kubenswrapper[4791]: I1210 23:10:22.230699 4791 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/053f7590-725f-4504-a204-be83d801bad2-sg-core-conf-yaml" (OuterVolumeSpecName: "sg-core-conf-yaml") pod "053f7590-725f-4504-a204-be83d801bad2" (UID: "053f7590-725f-4504-a204-be83d801bad2"). InnerVolumeSpecName "sg-core-conf-yaml". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 10 23:10:22 crc kubenswrapper[4791]: I1210 23:10:22.276466 4791 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-wrsg2\" (UniqueName: \"kubernetes.io/projected/053f7590-725f-4504-a204-be83d801bad2-kube-api-access-wrsg2\") on node \"crc\" DevicePath \"\"" Dec 10 23:10:22 crc kubenswrapper[4791]: I1210 23:10:22.276502 4791 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/053f7590-725f-4504-a204-be83d801bad2-scripts\") on node \"crc\" DevicePath \"\"" Dec 10 23:10:22 crc kubenswrapper[4791]: I1210 23:10:22.276515 4791 reconciler_common.go:293] "Volume detached for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/053f7590-725f-4504-a204-be83d801bad2-run-httpd\") on node \"crc\" DevicePath \"\"" Dec 10 23:10:22 crc kubenswrapper[4791]: I1210 23:10:22.276527 4791 reconciler_common.go:293] "Volume detached for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/053f7590-725f-4504-a204-be83d801bad2-sg-core-conf-yaml\") on node \"crc\" DevicePath \"\"" Dec 10 23:10:22 crc kubenswrapper[4791]: I1210 23:10:22.299968 4791 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/053f7590-725f-4504-a204-be83d801bad2-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "053f7590-725f-4504-a204-be83d801bad2" (UID: "053f7590-725f-4504-a204-be83d801bad2"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 10 23:10:22 crc kubenswrapper[4791]: I1210 23:10:22.308497 4791 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/053f7590-725f-4504-a204-be83d801bad2-config-data" (OuterVolumeSpecName: "config-data") pod "053f7590-725f-4504-a204-be83d801bad2" (UID: "053f7590-725f-4504-a204-be83d801bad2"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 10 23:10:22 crc kubenswrapper[4791]: I1210 23:10:22.342212 4791 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-conductor-db-sync-jff9b" Dec 10 23:10:22 crc kubenswrapper[4791]: I1210 23:10:22.378472 4791 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/053f7590-725f-4504-a204-be83d801bad2-config-data\") on node \"crc\" DevicePath \"\"" Dec 10 23:10:22 crc kubenswrapper[4791]: I1210 23:10:22.378504 4791 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/053f7590-725f-4504-a204-be83d801bad2-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 10 23:10:22 crc kubenswrapper[4791]: I1210 23:10:22.479436 4791 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/075728ce-048e-452d-adff-0e3d167d6673-combined-ca-bundle\") pod \"075728ce-048e-452d-adff-0e3d167d6673\" (UID: \"075728ce-048e-452d-adff-0e3d167d6673\") " Dec 10 23:10:22 crc kubenswrapper[4791]: I1210 23:10:22.479657 4791 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/075728ce-048e-452d-adff-0e3d167d6673-config-data\") pod \"075728ce-048e-452d-adff-0e3d167d6673\" (UID: \"075728ce-048e-452d-adff-0e3d167d6673\") " Dec 10 23:10:22 crc kubenswrapper[4791]: I1210 23:10:22.479701 4791 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/075728ce-048e-452d-adff-0e3d167d6673-scripts\") pod \"075728ce-048e-452d-adff-0e3d167d6673\" (UID: \"075728ce-048e-452d-adff-0e3d167d6673\") " Dec 10 23:10:22 crc kubenswrapper[4791]: I1210 23:10:22.479770 4791 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-9x2cf\" (UniqueName: \"kubernetes.io/projected/075728ce-048e-452d-adff-0e3d167d6673-kube-api-access-9x2cf\") pod \"075728ce-048e-452d-adff-0e3d167d6673\" (UID: \"075728ce-048e-452d-adff-0e3d167d6673\") " Dec 10 23:10:22 crc kubenswrapper[4791]: I1210 23:10:22.482857 4791 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/075728ce-048e-452d-adff-0e3d167d6673-scripts" (OuterVolumeSpecName: "scripts") pod "075728ce-048e-452d-adff-0e3d167d6673" (UID: "075728ce-048e-452d-adff-0e3d167d6673"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 10 23:10:22 crc kubenswrapper[4791]: I1210 23:10:22.484525 4791 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/075728ce-048e-452d-adff-0e3d167d6673-kube-api-access-9x2cf" (OuterVolumeSpecName: "kube-api-access-9x2cf") pod "075728ce-048e-452d-adff-0e3d167d6673" (UID: "075728ce-048e-452d-adff-0e3d167d6673"). InnerVolumeSpecName "kube-api-access-9x2cf". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 10 23:10:22 crc kubenswrapper[4791]: I1210 23:10:22.512404 4791 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/075728ce-048e-452d-adff-0e3d167d6673-config-data" (OuterVolumeSpecName: "config-data") pod "075728ce-048e-452d-adff-0e3d167d6673" (UID: "075728ce-048e-452d-adff-0e3d167d6673"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 10 23:10:22 crc kubenswrapper[4791]: I1210 23:10:22.518093 4791 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/075728ce-048e-452d-adff-0e3d167d6673-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "075728ce-048e-452d-adff-0e3d167d6673" (UID: "075728ce-048e-452d-adff-0e3d167d6673"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 10 23:10:22 crc kubenswrapper[4791]: I1210 23:10:22.582698 4791 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/075728ce-048e-452d-adff-0e3d167d6673-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 10 23:10:22 crc kubenswrapper[4791]: I1210 23:10:22.582736 4791 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/075728ce-048e-452d-adff-0e3d167d6673-config-data\") on node \"crc\" DevicePath \"\"" Dec 10 23:10:22 crc kubenswrapper[4791]: I1210 23:10:22.582750 4791 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/075728ce-048e-452d-adff-0e3d167d6673-scripts\") on node \"crc\" DevicePath \"\"" Dec 10 23:10:22 crc kubenswrapper[4791]: I1210 23:10:22.582762 4791 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-9x2cf\" (UniqueName: \"kubernetes.io/projected/075728ce-048e-452d-adff-0e3d167d6673-kube-api-access-9x2cf\") on node \"crc\" DevicePath \"\"" Dec 10 23:10:23 crc kubenswrapper[4791]: I1210 23:10:23.062979 4791 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Dec 10 23:10:23 crc kubenswrapper[4791]: I1210 23:10:23.066383 4791 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-conductor-db-sync-jff9b" event={"ID":"075728ce-048e-452d-adff-0e3d167d6673","Type":"ContainerDied","Data":"1e95bcbbeac30c6337d15f2ef1bdaac77f18c46e4000b5c3fdf41d353afdf13a"} Dec 10 23:10:23 crc kubenswrapper[4791]: I1210 23:10:23.066443 4791 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="1e95bcbbeac30c6337d15f2ef1bdaac77f18c46e4000b5c3fdf41d353afdf13a" Dec 10 23:10:23 crc kubenswrapper[4791]: I1210 23:10:23.066400 4791 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-conductor-db-sync-jff9b" Dec 10 23:10:23 crc kubenswrapper[4791]: I1210 23:10:23.138090 4791 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ceilometer-0"] Dec 10 23:10:23 crc kubenswrapper[4791]: I1210 23:10:23.151594 4791 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/ceilometer-0"] Dec 10 23:10:23 crc kubenswrapper[4791]: I1210 23:10:23.167046 4791 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ceilometer-0"] Dec 10 23:10:23 crc kubenswrapper[4791]: E1210 23:10:23.167541 4791 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="053f7590-725f-4504-a204-be83d801bad2" containerName="sg-core" Dec 10 23:10:23 crc kubenswrapper[4791]: I1210 23:10:23.167564 4791 state_mem.go:107] "Deleted CPUSet assignment" podUID="053f7590-725f-4504-a204-be83d801bad2" containerName="sg-core" Dec 10 23:10:23 crc kubenswrapper[4791]: E1210 23:10:23.167578 4791 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="053f7590-725f-4504-a204-be83d801bad2" containerName="proxy-httpd" Dec 10 23:10:23 crc kubenswrapper[4791]: I1210 23:10:23.167587 4791 state_mem.go:107] "Deleted CPUSet assignment" podUID="053f7590-725f-4504-a204-be83d801bad2" containerName="proxy-httpd" Dec 10 23:10:23 crc kubenswrapper[4791]: E1210 23:10:23.167616 4791 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="053f7590-725f-4504-a204-be83d801bad2" containerName="ceilometer-central-agent" Dec 10 23:10:23 crc kubenswrapper[4791]: I1210 23:10:23.167625 4791 state_mem.go:107] "Deleted CPUSet assignment" podUID="053f7590-725f-4504-a204-be83d801bad2" containerName="ceilometer-central-agent" Dec 10 23:10:23 crc kubenswrapper[4791]: E1210 23:10:23.167642 4791 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="053f7590-725f-4504-a204-be83d801bad2" containerName="ceilometer-notification-agent" Dec 10 23:10:23 crc kubenswrapper[4791]: I1210 23:10:23.167650 4791 state_mem.go:107] "Deleted CPUSet assignment" podUID="053f7590-725f-4504-a204-be83d801bad2" containerName="ceilometer-notification-agent" Dec 10 23:10:23 crc kubenswrapper[4791]: E1210 23:10:23.167662 4791 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="075728ce-048e-452d-adff-0e3d167d6673" containerName="nova-cell0-conductor-db-sync" Dec 10 23:10:23 crc kubenswrapper[4791]: I1210 23:10:23.167670 4791 state_mem.go:107] "Deleted CPUSet assignment" podUID="075728ce-048e-452d-adff-0e3d167d6673" containerName="nova-cell0-conductor-db-sync" Dec 10 23:10:23 crc kubenswrapper[4791]: I1210 23:10:23.167891 4791 memory_manager.go:354] "RemoveStaleState removing state" podUID="075728ce-048e-452d-adff-0e3d167d6673" containerName="nova-cell0-conductor-db-sync" Dec 10 23:10:23 crc kubenswrapper[4791]: I1210 23:10:23.167915 4791 memory_manager.go:354] "RemoveStaleState removing state" podUID="053f7590-725f-4504-a204-be83d801bad2" containerName="proxy-httpd" Dec 10 23:10:23 crc kubenswrapper[4791]: I1210 23:10:23.167939 4791 memory_manager.go:354] "RemoveStaleState removing state" podUID="053f7590-725f-4504-a204-be83d801bad2" containerName="sg-core" Dec 10 23:10:23 crc kubenswrapper[4791]: I1210 23:10:23.167949 4791 memory_manager.go:354] "RemoveStaleState removing state" podUID="053f7590-725f-4504-a204-be83d801bad2" containerName="ceilometer-central-agent" Dec 10 23:10:23 crc kubenswrapper[4791]: I1210 23:10:23.167972 4791 memory_manager.go:354] "RemoveStaleState removing state" podUID="053f7590-725f-4504-a204-be83d801bad2" containerName="ceilometer-notification-agent" Dec 10 23:10:23 crc kubenswrapper[4791]: I1210 23:10:23.170207 4791 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Dec 10 23:10:23 crc kubenswrapper[4791]: I1210 23:10:23.173500 4791 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-scripts" Dec 10 23:10:23 crc kubenswrapper[4791]: I1210 23:10:23.175979 4791 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-config-data" Dec 10 23:10:23 crc kubenswrapper[4791]: I1210 23:10:23.189672 4791 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceilometer-0"] Dec 10 23:10:23 crc kubenswrapper[4791]: I1210 23:10:23.248398 4791 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-cell0-conductor-0"] Dec 10 23:10:23 crc kubenswrapper[4791]: I1210 23:10:23.249709 4791 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-conductor-0" Dec 10 23:10:23 crc kubenswrapper[4791]: I1210 23:10:23.252038 4791 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-nova-dockercfg-vtzz5" Dec 10 23:10:23 crc kubenswrapper[4791]: I1210 23:10:23.252178 4791 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell0-conductor-config-data" Dec 10 23:10:23 crc kubenswrapper[4791]: I1210 23:10:23.275987 4791 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell0-conductor-0"] Dec 10 23:10:23 crc kubenswrapper[4791]: I1210 23:10:23.301052 4791 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/65afcf68-16b8-4f9d-ab77-ab910ca9469b-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"65afcf68-16b8-4f9d-ab77-ab910ca9469b\") " pod="openstack/ceilometer-0" Dec 10 23:10:23 crc kubenswrapper[4791]: I1210 23:10:23.301225 4791 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-t8gkz\" (UniqueName: \"kubernetes.io/projected/65afcf68-16b8-4f9d-ab77-ab910ca9469b-kube-api-access-t8gkz\") pod \"ceilometer-0\" (UID: \"65afcf68-16b8-4f9d-ab77-ab910ca9469b\") " pod="openstack/ceilometer-0" Dec 10 23:10:23 crc kubenswrapper[4791]: I1210 23:10:23.301272 4791 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/65afcf68-16b8-4f9d-ab77-ab910ca9469b-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"65afcf68-16b8-4f9d-ab77-ab910ca9469b\") " pod="openstack/ceilometer-0" Dec 10 23:10:23 crc kubenswrapper[4791]: I1210 23:10:23.301306 4791 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/65afcf68-16b8-4f9d-ab77-ab910ca9469b-scripts\") pod \"ceilometer-0\" (UID: \"65afcf68-16b8-4f9d-ab77-ab910ca9469b\") " pod="openstack/ceilometer-0" Dec 10 23:10:23 crc kubenswrapper[4791]: I1210 23:10:23.301421 4791 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/65afcf68-16b8-4f9d-ab77-ab910ca9469b-run-httpd\") pod \"ceilometer-0\" (UID: \"65afcf68-16b8-4f9d-ab77-ab910ca9469b\") " pod="openstack/ceilometer-0" Dec 10 23:10:23 crc kubenswrapper[4791]: I1210 23:10:23.301453 4791 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/65afcf68-16b8-4f9d-ab77-ab910ca9469b-log-httpd\") pod \"ceilometer-0\" (UID: \"65afcf68-16b8-4f9d-ab77-ab910ca9469b\") " pod="openstack/ceilometer-0" Dec 10 23:10:23 crc kubenswrapper[4791]: I1210 23:10:23.301511 4791 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/65afcf68-16b8-4f9d-ab77-ab910ca9469b-config-data\") pod \"ceilometer-0\" (UID: \"65afcf68-16b8-4f9d-ab77-ab910ca9469b\") " pod="openstack/ceilometer-0" Dec 10 23:10:23 crc kubenswrapper[4791]: I1210 23:10:23.403224 4791 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/65afcf68-16b8-4f9d-ab77-ab910ca9469b-config-data\") pod \"ceilometer-0\" (UID: \"65afcf68-16b8-4f9d-ab77-ab910ca9469b\") " pod="openstack/ceilometer-0" Dec 10 23:10:23 crc kubenswrapper[4791]: I1210 23:10:23.403299 4791 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-l4mpc\" (UniqueName: \"kubernetes.io/projected/1082e5c5-2e93-4950-93b4-c66ee224c711-kube-api-access-l4mpc\") pod \"nova-cell0-conductor-0\" (UID: \"1082e5c5-2e93-4950-93b4-c66ee224c711\") " pod="openstack/nova-cell0-conductor-0" Dec 10 23:10:23 crc kubenswrapper[4791]: I1210 23:10:23.403360 4791 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/1082e5c5-2e93-4950-93b4-c66ee224c711-config-data\") pod \"nova-cell0-conductor-0\" (UID: \"1082e5c5-2e93-4950-93b4-c66ee224c711\") " pod="openstack/nova-cell0-conductor-0" Dec 10 23:10:23 crc kubenswrapper[4791]: I1210 23:10:23.403380 4791 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/1082e5c5-2e93-4950-93b4-c66ee224c711-combined-ca-bundle\") pod \"nova-cell0-conductor-0\" (UID: \"1082e5c5-2e93-4950-93b4-c66ee224c711\") " pod="openstack/nova-cell0-conductor-0" Dec 10 23:10:23 crc kubenswrapper[4791]: I1210 23:10:23.403410 4791 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/65afcf68-16b8-4f9d-ab77-ab910ca9469b-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"65afcf68-16b8-4f9d-ab77-ab910ca9469b\") " pod="openstack/ceilometer-0" Dec 10 23:10:23 crc kubenswrapper[4791]: I1210 23:10:23.403691 4791 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-t8gkz\" (UniqueName: \"kubernetes.io/projected/65afcf68-16b8-4f9d-ab77-ab910ca9469b-kube-api-access-t8gkz\") pod \"ceilometer-0\" (UID: \"65afcf68-16b8-4f9d-ab77-ab910ca9469b\") " pod="openstack/ceilometer-0" Dec 10 23:10:23 crc kubenswrapper[4791]: I1210 23:10:23.403739 4791 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/65afcf68-16b8-4f9d-ab77-ab910ca9469b-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"65afcf68-16b8-4f9d-ab77-ab910ca9469b\") " pod="openstack/ceilometer-0" Dec 10 23:10:23 crc kubenswrapper[4791]: I1210 23:10:23.403776 4791 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/65afcf68-16b8-4f9d-ab77-ab910ca9469b-scripts\") pod \"ceilometer-0\" (UID: \"65afcf68-16b8-4f9d-ab77-ab910ca9469b\") " pod="openstack/ceilometer-0" Dec 10 23:10:23 crc kubenswrapper[4791]: I1210 23:10:23.403819 4791 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/65afcf68-16b8-4f9d-ab77-ab910ca9469b-run-httpd\") pod \"ceilometer-0\" (UID: \"65afcf68-16b8-4f9d-ab77-ab910ca9469b\") " pod="openstack/ceilometer-0" Dec 10 23:10:23 crc kubenswrapper[4791]: I1210 23:10:23.403847 4791 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/65afcf68-16b8-4f9d-ab77-ab910ca9469b-log-httpd\") pod \"ceilometer-0\" (UID: \"65afcf68-16b8-4f9d-ab77-ab910ca9469b\") " pod="openstack/ceilometer-0" Dec 10 23:10:23 crc kubenswrapper[4791]: I1210 23:10:23.404330 4791 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/65afcf68-16b8-4f9d-ab77-ab910ca9469b-log-httpd\") pod \"ceilometer-0\" (UID: \"65afcf68-16b8-4f9d-ab77-ab910ca9469b\") " pod="openstack/ceilometer-0" Dec 10 23:10:23 crc kubenswrapper[4791]: I1210 23:10:23.404533 4791 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/65afcf68-16b8-4f9d-ab77-ab910ca9469b-run-httpd\") pod \"ceilometer-0\" (UID: \"65afcf68-16b8-4f9d-ab77-ab910ca9469b\") " pod="openstack/ceilometer-0" Dec 10 23:10:23 crc kubenswrapper[4791]: I1210 23:10:23.409153 4791 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/65afcf68-16b8-4f9d-ab77-ab910ca9469b-scripts\") pod \"ceilometer-0\" (UID: \"65afcf68-16b8-4f9d-ab77-ab910ca9469b\") " pod="openstack/ceilometer-0" Dec 10 23:10:23 crc kubenswrapper[4791]: I1210 23:10:23.421264 4791 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/65afcf68-16b8-4f9d-ab77-ab910ca9469b-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"65afcf68-16b8-4f9d-ab77-ab910ca9469b\") " pod="openstack/ceilometer-0" Dec 10 23:10:23 crc kubenswrapper[4791]: I1210 23:10:23.422520 4791 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/65afcf68-16b8-4f9d-ab77-ab910ca9469b-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"65afcf68-16b8-4f9d-ab77-ab910ca9469b\") " pod="openstack/ceilometer-0" Dec 10 23:10:23 crc kubenswrapper[4791]: I1210 23:10:23.422988 4791 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/65afcf68-16b8-4f9d-ab77-ab910ca9469b-config-data\") pod \"ceilometer-0\" (UID: \"65afcf68-16b8-4f9d-ab77-ab910ca9469b\") " pod="openstack/ceilometer-0" Dec 10 23:10:23 crc kubenswrapper[4791]: I1210 23:10:23.425129 4791 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-t8gkz\" (UniqueName: \"kubernetes.io/projected/65afcf68-16b8-4f9d-ab77-ab910ca9469b-kube-api-access-t8gkz\") pod \"ceilometer-0\" (UID: \"65afcf68-16b8-4f9d-ab77-ab910ca9469b\") " pod="openstack/ceilometer-0" Dec 10 23:10:23 crc kubenswrapper[4791]: I1210 23:10:23.495543 4791 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Dec 10 23:10:23 crc kubenswrapper[4791]: I1210 23:10:23.506226 4791 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-l4mpc\" (UniqueName: \"kubernetes.io/projected/1082e5c5-2e93-4950-93b4-c66ee224c711-kube-api-access-l4mpc\") pod \"nova-cell0-conductor-0\" (UID: \"1082e5c5-2e93-4950-93b4-c66ee224c711\") " pod="openstack/nova-cell0-conductor-0" Dec 10 23:10:23 crc kubenswrapper[4791]: I1210 23:10:23.506348 4791 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/1082e5c5-2e93-4950-93b4-c66ee224c711-config-data\") pod \"nova-cell0-conductor-0\" (UID: \"1082e5c5-2e93-4950-93b4-c66ee224c711\") " pod="openstack/nova-cell0-conductor-0" Dec 10 23:10:23 crc kubenswrapper[4791]: I1210 23:10:23.506425 4791 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/1082e5c5-2e93-4950-93b4-c66ee224c711-combined-ca-bundle\") pod \"nova-cell0-conductor-0\" (UID: \"1082e5c5-2e93-4950-93b4-c66ee224c711\") " pod="openstack/nova-cell0-conductor-0" Dec 10 23:10:23 crc kubenswrapper[4791]: I1210 23:10:23.511876 4791 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/1082e5c5-2e93-4950-93b4-c66ee224c711-combined-ca-bundle\") pod \"nova-cell0-conductor-0\" (UID: \"1082e5c5-2e93-4950-93b4-c66ee224c711\") " pod="openstack/nova-cell0-conductor-0" Dec 10 23:10:23 crc kubenswrapper[4791]: I1210 23:10:23.515448 4791 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/1082e5c5-2e93-4950-93b4-c66ee224c711-config-data\") pod \"nova-cell0-conductor-0\" (UID: \"1082e5c5-2e93-4950-93b4-c66ee224c711\") " pod="openstack/nova-cell0-conductor-0" Dec 10 23:10:23 crc kubenswrapper[4791]: I1210 23:10:23.528610 4791 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-l4mpc\" (UniqueName: \"kubernetes.io/projected/1082e5c5-2e93-4950-93b4-c66ee224c711-kube-api-access-l4mpc\") pod \"nova-cell0-conductor-0\" (UID: \"1082e5c5-2e93-4950-93b4-c66ee224c711\") " pod="openstack/nova-cell0-conductor-0" Dec 10 23:10:23 crc kubenswrapper[4791]: I1210 23:10:23.579836 4791 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-conductor-0" Dec 10 23:10:23 crc kubenswrapper[4791]: I1210 23:10:23.895050 4791 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="053f7590-725f-4504-a204-be83d801bad2" path="/var/lib/kubelet/pods/053f7590-725f-4504-a204-be83d801bad2/volumes" Dec 10 23:10:23 crc kubenswrapper[4791]: I1210 23:10:23.978455 4791 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceilometer-0"] Dec 10 23:10:24 crc kubenswrapper[4791]: W1210 23:10:24.069911 4791 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod1082e5c5_2e93_4950_93b4_c66ee224c711.slice/crio-980f977f0c5b2cfa4a5898d517787139a113fea46dcfe9d332c67f4a68b7aa7a WatchSource:0}: Error finding container 980f977f0c5b2cfa4a5898d517787139a113fea46dcfe9d332c67f4a68b7aa7a: Status 404 returned error can't find the container with id 980f977f0c5b2cfa4a5898d517787139a113fea46dcfe9d332c67f4a68b7aa7a Dec 10 23:10:24 crc kubenswrapper[4791]: I1210 23:10:24.072045 4791 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell0-conductor-0"] Dec 10 23:10:24 crc kubenswrapper[4791]: I1210 23:10:24.074715 4791 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"65afcf68-16b8-4f9d-ab77-ab910ca9469b","Type":"ContainerStarted","Data":"4ae58e4f34c7896f75bcd7f3998c4819f2accc012ff8d2bc1d9ae544e732ab92"} Dec 10 23:10:25 crc kubenswrapper[4791]: I1210 23:10:25.085806 4791 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"65afcf68-16b8-4f9d-ab77-ab910ca9469b","Type":"ContainerStarted","Data":"3a5559784c422a1e8dd3f804e260a986d9b1c7dac2ccb1aea0c905532562cfd4"} Dec 10 23:10:25 crc kubenswrapper[4791]: I1210 23:10:25.087020 4791 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-conductor-0" event={"ID":"1082e5c5-2e93-4950-93b4-c66ee224c711","Type":"ContainerStarted","Data":"e92b745c98135d453431be6c8149b77b3918d36029c35d4548a0a5c13917c55a"} Dec 10 23:10:25 crc kubenswrapper[4791]: I1210 23:10:25.087045 4791 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-conductor-0" event={"ID":"1082e5c5-2e93-4950-93b4-c66ee224c711","Type":"ContainerStarted","Data":"980f977f0c5b2cfa4a5898d517787139a113fea46dcfe9d332c67f4a68b7aa7a"} Dec 10 23:10:25 crc kubenswrapper[4791]: I1210 23:10:25.087288 4791 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-cell0-conductor-0" Dec 10 23:10:25 crc kubenswrapper[4791]: I1210 23:10:25.107668 4791 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-cell0-conductor-0" podStartSLOduration=2.107650285 podStartE2EDuration="2.107650285s" podCreationTimestamp="2025-12-10 23:10:23 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-10 23:10:25.104586657 +0000 UTC m=+1259.534204280" watchObservedRunningTime="2025-12-10 23:10:25.107650285 +0000 UTC m=+1259.537267898" Dec 10 23:10:26 crc kubenswrapper[4791]: I1210 23:10:26.108556 4791 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"65afcf68-16b8-4f9d-ab77-ab910ca9469b","Type":"ContainerStarted","Data":"6f1983af270707fe501a3d4c71341f8755d1b9f230aebf1ddfe48385bd734154"} Dec 10 23:10:27 crc kubenswrapper[4791]: I1210 23:10:27.121827 4791 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"65afcf68-16b8-4f9d-ab77-ab910ca9469b","Type":"ContainerStarted","Data":"bd8db88f95d6197edaa97ac8dc83a26e1f4e74cecfab63880dba48e7f4142325"} Dec 10 23:10:29 crc kubenswrapper[4791]: I1210 23:10:29.151410 4791 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"65afcf68-16b8-4f9d-ab77-ab910ca9469b","Type":"ContainerStarted","Data":"bfb1f419fcadf3ec57c071c7d7e0d51159099c674bd52ff80933b2cba66a8da4"} Dec 10 23:10:29 crc kubenswrapper[4791]: I1210 23:10:29.152081 4791 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/ceilometer-0" Dec 10 23:10:29 crc kubenswrapper[4791]: I1210 23:10:29.185527 4791 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ceilometer-0" podStartSLOduration=2.25396247 podStartE2EDuration="6.185499862s" podCreationTimestamp="2025-12-10 23:10:23 +0000 UTC" firstStartedPulling="2025-12-10 23:10:23.992305049 +0000 UTC m=+1258.421922662" lastFinishedPulling="2025-12-10 23:10:27.923842431 +0000 UTC m=+1262.353460054" observedRunningTime="2025-12-10 23:10:29.182466704 +0000 UTC m=+1263.612084407" watchObservedRunningTime="2025-12-10 23:10:29.185499862 +0000 UTC m=+1263.615117495" Dec 10 23:10:33 crc kubenswrapper[4791]: I1210 23:10:33.615694 4791 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-cell0-conductor-0" Dec 10 23:10:34 crc kubenswrapper[4791]: I1210 23:10:34.365424 4791 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-cell0-cell-mapping-k64vq"] Dec 10 23:10:34 crc kubenswrapper[4791]: I1210 23:10:34.368220 4791 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-cell-mapping-k64vq" Dec 10 23:10:34 crc kubenswrapper[4791]: I1210 23:10:34.376212 4791 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell0-manage-scripts" Dec 10 23:10:34 crc kubenswrapper[4791]: I1210 23:10:34.376312 4791 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell0-manage-config-data" Dec 10 23:10:34 crc kubenswrapper[4791]: I1210 23:10:34.389612 4791 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell0-cell-mapping-k64vq"] Dec 10 23:10:34 crc kubenswrapper[4791]: I1210 23:10:34.518571 4791 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/da9f584f-c289-4d67-ae91-8d6bf7507fb6-combined-ca-bundle\") pod \"nova-cell0-cell-mapping-k64vq\" (UID: \"da9f584f-c289-4d67-ae91-8d6bf7507fb6\") " pod="openstack/nova-cell0-cell-mapping-k64vq" Dec 10 23:10:34 crc kubenswrapper[4791]: I1210 23:10:34.518698 4791 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/da9f584f-c289-4d67-ae91-8d6bf7507fb6-config-data\") pod \"nova-cell0-cell-mapping-k64vq\" (UID: \"da9f584f-c289-4d67-ae91-8d6bf7507fb6\") " pod="openstack/nova-cell0-cell-mapping-k64vq" Dec 10 23:10:34 crc kubenswrapper[4791]: I1210 23:10:34.518966 4791 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/da9f584f-c289-4d67-ae91-8d6bf7507fb6-scripts\") pod \"nova-cell0-cell-mapping-k64vq\" (UID: \"da9f584f-c289-4d67-ae91-8d6bf7507fb6\") " pod="openstack/nova-cell0-cell-mapping-k64vq" Dec 10 23:10:34 crc kubenswrapper[4791]: I1210 23:10:34.519124 4791 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-xdpx4\" (UniqueName: \"kubernetes.io/projected/da9f584f-c289-4d67-ae91-8d6bf7507fb6-kube-api-access-xdpx4\") pod \"nova-cell0-cell-mapping-k64vq\" (UID: \"da9f584f-c289-4d67-ae91-8d6bf7507fb6\") " pod="openstack/nova-cell0-cell-mapping-k64vq" Dec 10 23:10:34 crc kubenswrapper[4791]: I1210 23:10:34.548825 4791 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-api-0"] Dec 10 23:10:34 crc kubenswrapper[4791]: I1210 23:10:34.555302 4791 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Dec 10 23:10:34 crc kubenswrapper[4791]: I1210 23:10:34.565545 4791 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-api-config-data" Dec 10 23:10:34 crc kubenswrapper[4791]: I1210 23:10:34.571494 4791 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-api-0"] Dec 10 23:10:34 crc kubenswrapper[4791]: I1210 23:10:34.625241 4791 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-xdpx4\" (UniqueName: \"kubernetes.io/projected/da9f584f-c289-4d67-ae91-8d6bf7507fb6-kube-api-access-xdpx4\") pod \"nova-cell0-cell-mapping-k64vq\" (UID: \"da9f584f-c289-4d67-ae91-8d6bf7507fb6\") " pod="openstack/nova-cell0-cell-mapping-k64vq" Dec 10 23:10:34 crc kubenswrapper[4791]: I1210 23:10:34.625413 4791 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/739a67cd-2a88-4c1d-bf76-5aac9b4b9e46-config-data\") pod \"nova-api-0\" (UID: \"739a67cd-2a88-4c1d-bf76-5aac9b4b9e46\") " pod="openstack/nova-api-0" Dec 10 23:10:34 crc kubenswrapper[4791]: I1210 23:10:34.625441 4791 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/739a67cd-2a88-4c1d-bf76-5aac9b4b9e46-combined-ca-bundle\") pod \"nova-api-0\" (UID: \"739a67cd-2a88-4c1d-bf76-5aac9b4b9e46\") " pod="openstack/nova-api-0" Dec 10 23:10:34 crc kubenswrapper[4791]: I1210 23:10:34.625474 4791 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/da9f584f-c289-4d67-ae91-8d6bf7507fb6-combined-ca-bundle\") pod \"nova-cell0-cell-mapping-k64vq\" (UID: \"da9f584f-c289-4d67-ae91-8d6bf7507fb6\") " pod="openstack/nova-cell0-cell-mapping-k64vq" Dec 10 23:10:34 crc kubenswrapper[4791]: I1210 23:10:34.625526 4791 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/739a67cd-2a88-4c1d-bf76-5aac9b4b9e46-logs\") pod \"nova-api-0\" (UID: \"739a67cd-2a88-4c1d-bf76-5aac9b4b9e46\") " pod="openstack/nova-api-0" Dec 10 23:10:34 crc kubenswrapper[4791]: I1210 23:10:34.625560 4791 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/da9f584f-c289-4d67-ae91-8d6bf7507fb6-config-data\") pod \"nova-cell0-cell-mapping-k64vq\" (UID: \"da9f584f-c289-4d67-ae91-8d6bf7507fb6\") " pod="openstack/nova-cell0-cell-mapping-k64vq" Dec 10 23:10:34 crc kubenswrapper[4791]: I1210 23:10:34.625702 4791 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-9mj6d\" (UniqueName: \"kubernetes.io/projected/739a67cd-2a88-4c1d-bf76-5aac9b4b9e46-kube-api-access-9mj6d\") pod \"nova-api-0\" (UID: \"739a67cd-2a88-4c1d-bf76-5aac9b4b9e46\") " pod="openstack/nova-api-0" Dec 10 23:10:34 crc kubenswrapper[4791]: I1210 23:10:34.625781 4791 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/da9f584f-c289-4d67-ae91-8d6bf7507fb6-scripts\") pod \"nova-cell0-cell-mapping-k64vq\" (UID: \"da9f584f-c289-4d67-ae91-8d6bf7507fb6\") " pod="openstack/nova-cell0-cell-mapping-k64vq" Dec 10 23:10:34 crc kubenswrapper[4791]: I1210 23:10:34.654149 4791 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/da9f584f-c289-4d67-ae91-8d6bf7507fb6-combined-ca-bundle\") pod \"nova-cell0-cell-mapping-k64vq\" (UID: \"da9f584f-c289-4d67-ae91-8d6bf7507fb6\") " pod="openstack/nova-cell0-cell-mapping-k64vq" Dec 10 23:10:34 crc kubenswrapper[4791]: I1210 23:10:34.655693 4791 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/da9f584f-c289-4d67-ae91-8d6bf7507fb6-scripts\") pod \"nova-cell0-cell-mapping-k64vq\" (UID: \"da9f584f-c289-4d67-ae91-8d6bf7507fb6\") " pod="openstack/nova-cell0-cell-mapping-k64vq" Dec 10 23:10:34 crc kubenswrapper[4791]: I1210 23:10:34.658373 4791 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/da9f584f-c289-4d67-ae91-8d6bf7507fb6-config-data\") pod \"nova-cell0-cell-mapping-k64vq\" (UID: \"da9f584f-c289-4d67-ae91-8d6bf7507fb6\") " pod="openstack/nova-cell0-cell-mapping-k64vq" Dec 10 23:10:34 crc kubenswrapper[4791]: I1210 23:10:34.683711 4791 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-cell1-novncproxy-0"] Dec 10 23:10:34 crc kubenswrapper[4791]: I1210 23:10:34.695537 4791 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-novncproxy-0" Dec 10 23:10:34 crc kubenswrapper[4791]: I1210 23:10:34.697953 4791 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell1-novncproxy-config-data" Dec 10 23:10:34 crc kubenswrapper[4791]: I1210 23:10:34.699730 4791 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-xdpx4\" (UniqueName: \"kubernetes.io/projected/da9f584f-c289-4d67-ae91-8d6bf7507fb6-kube-api-access-xdpx4\") pod \"nova-cell0-cell-mapping-k64vq\" (UID: \"da9f584f-c289-4d67-ae91-8d6bf7507fb6\") " pod="openstack/nova-cell0-cell-mapping-k64vq" Dec 10 23:10:34 crc kubenswrapper[4791]: I1210 23:10:34.729382 4791 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/739a67cd-2a88-4c1d-bf76-5aac9b4b9e46-config-data\") pod \"nova-api-0\" (UID: \"739a67cd-2a88-4c1d-bf76-5aac9b4b9e46\") " pod="openstack/nova-api-0" Dec 10 23:10:34 crc kubenswrapper[4791]: I1210 23:10:34.729427 4791 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/739a67cd-2a88-4c1d-bf76-5aac9b4b9e46-combined-ca-bundle\") pod \"nova-api-0\" (UID: \"739a67cd-2a88-4c1d-bf76-5aac9b4b9e46\") " pod="openstack/nova-api-0" Dec 10 23:10:34 crc kubenswrapper[4791]: I1210 23:10:34.729462 4791 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/739a67cd-2a88-4c1d-bf76-5aac9b4b9e46-logs\") pod \"nova-api-0\" (UID: \"739a67cd-2a88-4c1d-bf76-5aac9b4b9e46\") " pod="openstack/nova-api-0" Dec 10 23:10:34 crc kubenswrapper[4791]: I1210 23:10:34.729560 4791 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-9mj6d\" (UniqueName: \"kubernetes.io/projected/739a67cd-2a88-4c1d-bf76-5aac9b4b9e46-kube-api-access-9mj6d\") pod \"nova-api-0\" (UID: \"739a67cd-2a88-4c1d-bf76-5aac9b4b9e46\") " pod="openstack/nova-api-0" Dec 10 23:10:34 crc kubenswrapper[4791]: I1210 23:10:34.731801 4791 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/739a67cd-2a88-4c1d-bf76-5aac9b4b9e46-logs\") pod \"nova-api-0\" (UID: \"739a67cd-2a88-4c1d-bf76-5aac9b4b9e46\") " pod="openstack/nova-api-0" Dec 10 23:10:34 crc kubenswrapper[4791]: I1210 23:10:34.740708 4791 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/739a67cd-2a88-4c1d-bf76-5aac9b4b9e46-combined-ca-bundle\") pod \"nova-api-0\" (UID: \"739a67cd-2a88-4c1d-bf76-5aac9b4b9e46\") " pod="openstack/nova-api-0" Dec 10 23:10:34 crc kubenswrapper[4791]: I1210 23:10:34.741190 4791 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/739a67cd-2a88-4c1d-bf76-5aac9b4b9e46-config-data\") pod \"nova-api-0\" (UID: \"739a67cd-2a88-4c1d-bf76-5aac9b4b9e46\") " pod="openstack/nova-api-0" Dec 10 23:10:34 crc kubenswrapper[4791]: I1210 23:10:34.745625 4791 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-scheduler-0"] Dec 10 23:10:34 crc kubenswrapper[4791]: I1210 23:10:34.746887 4791 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-scheduler-0" Dec 10 23:10:34 crc kubenswrapper[4791]: I1210 23:10:34.751246 4791 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-scheduler-config-data" Dec 10 23:10:34 crc kubenswrapper[4791]: I1210 23:10:34.760695 4791 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-9mj6d\" (UniqueName: \"kubernetes.io/projected/739a67cd-2a88-4c1d-bf76-5aac9b4b9e46-kube-api-access-9mj6d\") pod \"nova-api-0\" (UID: \"739a67cd-2a88-4c1d-bf76-5aac9b4b9e46\") " pod="openstack/nova-api-0" Dec 10 23:10:34 crc kubenswrapper[4791]: I1210 23:10:34.785409 4791 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-novncproxy-0"] Dec 10 23:10:34 crc kubenswrapper[4791]: I1210 23:10:34.810009 4791 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-scheduler-0"] Dec 10 23:10:34 crc kubenswrapper[4791]: I1210 23:10:34.831236 4791 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/e2d1527a-aac6-4ee0-9c41-60ef0784aa2a-combined-ca-bundle\") pod \"nova-cell1-novncproxy-0\" (UID: \"e2d1527a-aac6-4ee0-9c41-60ef0784aa2a\") " pod="openstack/nova-cell1-novncproxy-0" Dec 10 23:10:34 crc kubenswrapper[4791]: I1210 23:10:34.831306 4791 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-844r7\" (UniqueName: \"kubernetes.io/projected/83b6abc9-aa7f-4ddb-8b9c-b38fcb5a9145-kube-api-access-844r7\") pod \"nova-scheduler-0\" (UID: \"83b6abc9-aa7f-4ddb-8b9c-b38fcb5a9145\") " pod="openstack/nova-scheduler-0" Dec 10 23:10:34 crc kubenswrapper[4791]: I1210 23:10:34.831401 4791 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-gdgcq\" (UniqueName: \"kubernetes.io/projected/e2d1527a-aac6-4ee0-9c41-60ef0784aa2a-kube-api-access-gdgcq\") pod \"nova-cell1-novncproxy-0\" (UID: \"e2d1527a-aac6-4ee0-9c41-60ef0784aa2a\") " pod="openstack/nova-cell1-novncproxy-0" Dec 10 23:10:34 crc kubenswrapper[4791]: I1210 23:10:34.831446 4791 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/83b6abc9-aa7f-4ddb-8b9c-b38fcb5a9145-config-data\") pod \"nova-scheduler-0\" (UID: \"83b6abc9-aa7f-4ddb-8b9c-b38fcb5a9145\") " pod="openstack/nova-scheduler-0" Dec 10 23:10:34 crc kubenswrapper[4791]: I1210 23:10:34.831488 4791 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/e2d1527a-aac6-4ee0-9c41-60ef0784aa2a-config-data\") pod \"nova-cell1-novncproxy-0\" (UID: \"e2d1527a-aac6-4ee0-9c41-60ef0784aa2a\") " pod="openstack/nova-cell1-novncproxy-0" Dec 10 23:10:34 crc kubenswrapper[4791]: I1210 23:10:34.831521 4791 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/83b6abc9-aa7f-4ddb-8b9c-b38fcb5a9145-combined-ca-bundle\") pod \"nova-scheduler-0\" (UID: \"83b6abc9-aa7f-4ddb-8b9c-b38fcb5a9145\") " pod="openstack/nova-scheduler-0" Dec 10 23:10:34 crc kubenswrapper[4791]: I1210 23:10:34.859951 4791 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-metadata-0"] Dec 10 23:10:34 crc kubenswrapper[4791]: I1210 23:10:34.862162 4791 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-metadata-0" Dec 10 23:10:34 crc kubenswrapper[4791]: I1210 23:10:34.868320 4791 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-metadata-config-data" Dec 10 23:10:34 crc kubenswrapper[4791]: I1210 23:10:34.879411 4791 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-metadata-0"] Dec 10 23:10:34 crc kubenswrapper[4791]: I1210 23:10:34.889471 4791 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Dec 10 23:10:34 crc kubenswrapper[4791]: I1210 23:10:34.935168 4791 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/b010c513-02c4-4cf1-9751-4b1447588b1a-logs\") pod \"nova-metadata-0\" (UID: \"b010c513-02c4-4cf1-9751-4b1447588b1a\") " pod="openstack/nova-metadata-0" Dec 10 23:10:34 crc kubenswrapper[4791]: I1210 23:10:34.935303 4791 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-kbbvd\" (UniqueName: \"kubernetes.io/projected/b010c513-02c4-4cf1-9751-4b1447588b1a-kube-api-access-kbbvd\") pod \"nova-metadata-0\" (UID: \"b010c513-02c4-4cf1-9751-4b1447588b1a\") " pod="openstack/nova-metadata-0" Dec 10 23:10:34 crc kubenswrapper[4791]: I1210 23:10:34.935432 4791 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-gdgcq\" (UniqueName: \"kubernetes.io/projected/e2d1527a-aac6-4ee0-9c41-60ef0784aa2a-kube-api-access-gdgcq\") pod \"nova-cell1-novncproxy-0\" (UID: \"e2d1527a-aac6-4ee0-9c41-60ef0784aa2a\") " pod="openstack/nova-cell1-novncproxy-0" Dec 10 23:10:34 crc kubenswrapper[4791]: I1210 23:10:34.935471 4791 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/b010c513-02c4-4cf1-9751-4b1447588b1a-combined-ca-bundle\") pod \"nova-metadata-0\" (UID: \"b010c513-02c4-4cf1-9751-4b1447588b1a\") " pod="openstack/nova-metadata-0" Dec 10 23:10:34 crc kubenswrapper[4791]: I1210 23:10:34.935551 4791 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/83b6abc9-aa7f-4ddb-8b9c-b38fcb5a9145-config-data\") pod \"nova-scheduler-0\" (UID: \"83b6abc9-aa7f-4ddb-8b9c-b38fcb5a9145\") " pod="openstack/nova-scheduler-0" Dec 10 23:10:34 crc kubenswrapper[4791]: I1210 23:10:34.935624 4791 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/e2d1527a-aac6-4ee0-9c41-60ef0784aa2a-config-data\") pod \"nova-cell1-novncproxy-0\" (UID: \"e2d1527a-aac6-4ee0-9c41-60ef0784aa2a\") " pod="openstack/nova-cell1-novncproxy-0" Dec 10 23:10:34 crc kubenswrapper[4791]: I1210 23:10:34.935686 4791 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/83b6abc9-aa7f-4ddb-8b9c-b38fcb5a9145-combined-ca-bundle\") pod \"nova-scheduler-0\" (UID: \"83b6abc9-aa7f-4ddb-8b9c-b38fcb5a9145\") " pod="openstack/nova-scheduler-0" Dec 10 23:10:34 crc kubenswrapper[4791]: I1210 23:10:34.935720 4791 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/b010c513-02c4-4cf1-9751-4b1447588b1a-config-data\") pod \"nova-metadata-0\" (UID: \"b010c513-02c4-4cf1-9751-4b1447588b1a\") " pod="openstack/nova-metadata-0" Dec 10 23:10:34 crc kubenswrapper[4791]: I1210 23:10:34.935756 4791 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/e2d1527a-aac6-4ee0-9c41-60ef0784aa2a-combined-ca-bundle\") pod \"nova-cell1-novncproxy-0\" (UID: \"e2d1527a-aac6-4ee0-9c41-60ef0784aa2a\") " pod="openstack/nova-cell1-novncproxy-0" Dec 10 23:10:34 crc kubenswrapper[4791]: I1210 23:10:34.935816 4791 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-844r7\" (UniqueName: \"kubernetes.io/projected/83b6abc9-aa7f-4ddb-8b9c-b38fcb5a9145-kube-api-access-844r7\") pod \"nova-scheduler-0\" (UID: \"83b6abc9-aa7f-4ddb-8b9c-b38fcb5a9145\") " pod="openstack/nova-scheduler-0" Dec 10 23:10:34 crc kubenswrapper[4791]: I1210 23:10:34.947196 4791 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/e2d1527a-aac6-4ee0-9c41-60ef0784aa2a-config-data\") pod \"nova-cell1-novncproxy-0\" (UID: \"e2d1527a-aac6-4ee0-9c41-60ef0784aa2a\") " pod="openstack/nova-cell1-novncproxy-0" Dec 10 23:10:34 crc kubenswrapper[4791]: I1210 23:10:34.948060 4791 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-845d6d6f59-d6v8c"] Dec 10 23:10:34 crc kubenswrapper[4791]: I1210 23:10:34.950741 4791 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-845d6d6f59-d6v8c" Dec 10 23:10:34 crc kubenswrapper[4791]: I1210 23:10:34.962245 4791 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/e2d1527a-aac6-4ee0-9c41-60ef0784aa2a-combined-ca-bundle\") pod \"nova-cell1-novncproxy-0\" (UID: \"e2d1527a-aac6-4ee0-9c41-60ef0784aa2a\") " pod="openstack/nova-cell1-novncproxy-0" Dec 10 23:10:34 crc kubenswrapper[4791]: I1210 23:10:34.973880 4791 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-844r7\" (UniqueName: \"kubernetes.io/projected/83b6abc9-aa7f-4ddb-8b9c-b38fcb5a9145-kube-api-access-844r7\") pod \"nova-scheduler-0\" (UID: \"83b6abc9-aa7f-4ddb-8b9c-b38fcb5a9145\") " pod="openstack/nova-scheduler-0" Dec 10 23:10:34 crc kubenswrapper[4791]: I1210 23:10:34.974454 4791 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/83b6abc9-aa7f-4ddb-8b9c-b38fcb5a9145-combined-ca-bundle\") pod \"nova-scheduler-0\" (UID: \"83b6abc9-aa7f-4ddb-8b9c-b38fcb5a9145\") " pod="openstack/nova-scheduler-0" Dec 10 23:10:34 crc kubenswrapper[4791]: I1210 23:10:34.983757 4791 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-gdgcq\" (UniqueName: \"kubernetes.io/projected/e2d1527a-aac6-4ee0-9c41-60ef0784aa2a-kube-api-access-gdgcq\") pod \"nova-cell1-novncproxy-0\" (UID: \"e2d1527a-aac6-4ee0-9c41-60ef0784aa2a\") " pod="openstack/nova-cell1-novncproxy-0" Dec 10 23:10:34 crc kubenswrapper[4791]: I1210 23:10:34.985894 4791 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/83b6abc9-aa7f-4ddb-8b9c-b38fcb5a9145-config-data\") pod \"nova-scheduler-0\" (UID: \"83b6abc9-aa7f-4ddb-8b9c-b38fcb5a9145\") " pod="openstack/nova-scheduler-0" Dec 10 23:10:34 crc kubenswrapper[4791]: I1210 23:10:34.993296 4791 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-cell-mapping-k64vq" Dec 10 23:10:35 crc kubenswrapper[4791]: I1210 23:10:35.009428 4791 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-845d6d6f59-d6v8c"] Dec 10 23:10:35 crc kubenswrapper[4791]: I1210 23:10:35.038671 4791 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/6202cb66-3ca7-4988-a800-2682cba44d72-ovsdbserver-nb\") pod \"dnsmasq-dns-845d6d6f59-d6v8c\" (UID: \"6202cb66-3ca7-4988-a800-2682cba44d72\") " pod="openstack/dnsmasq-dns-845d6d6f59-d6v8c" Dec 10 23:10:35 crc kubenswrapper[4791]: I1210 23:10:35.039163 4791 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/b010c513-02c4-4cf1-9751-4b1447588b1a-logs\") pod \"nova-metadata-0\" (UID: \"b010c513-02c4-4cf1-9751-4b1447588b1a\") " pod="openstack/nova-metadata-0" Dec 10 23:10:35 crc kubenswrapper[4791]: I1210 23:10:35.039252 4791 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/6202cb66-3ca7-4988-a800-2682cba44d72-dns-svc\") pod \"dnsmasq-dns-845d6d6f59-d6v8c\" (UID: \"6202cb66-3ca7-4988-a800-2682cba44d72\") " pod="openstack/dnsmasq-dns-845d6d6f59-d6v8c" Dec 10 23:10:35 crc kubenswrapper[4791]: I1210 23:10:35.039499 4791 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-kbbvd\" (UniqueName: \"kubernetes.io/projected/b010c513-02c4-4cf1-9751-4b1447588b1a-kube-api-access-kbbvd\") pod \"nova-metadata-0\" (UID: \"b010c513-02c4-4cf1-9751-4b1447588b1a\") " pod="openstack/nova-metadata-0" Dec 10 23:10:35 crc kubenswrapper[4791]: I1210 23:10:35.039589 4791 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-blchs\" (UniqueName: \"kubernetes.io/projected/6202cb66-3ca7-4988-a800-2682cba44d72-kube-api-access-blchs\") pod \"dnsmasq-dns-845d6d6f59-d6v8c\" (UID: \"6202cb66-3ca7-4988-a800-2682cba44d72\") " pod="openstack/dnsmasq-dns-845d6d6f59-d6v8c" Dec 10 23:10:35 crc kubenswrapper[4791]: I1210 23:10:35.039669 4791 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/b010c513-02c4-4cf1-9751-4b1447588b1a-combined-ca-bundle\") pod \"nova-metadata-0\" (UID: \"b010c513-02c4-4cf1-9751-4b1447588b1a\") " pod="openstack/nova-metadata-0" Dec 10 23:10:35 crc kubenswrapper[4791]: I1210 23:10:35.039754 4791 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/6202cb66-3ca7-4988-a800-2682cba44d72-config\") pod \"dnsmasq-dns-845d6d6f59-d6v8c\" (UID: \"6202cb66-3ca7-4988-a800-2682cba44d72\") " pod="openstack/dnsmasq-dns-845d6d6f59-d6v8c" Dec 10 23:10:35 crc kubenswrapper[4791]: I1210 23:10:35.039842 4791 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/6202cb66-3ca7-4988-a800-2682cba44d72-dns-swift-storage-0\") pod \"dnsmasq-dns-845d6d6f59-d6v8c\" (UID: \"6202cb66-3ca7-4988-a800-2682cba44d72\") " pod="openstack/dnsmasq-dns-845d6d6f59-d6v8c" Dec 10 23:10:35 crc kubenswrapper[4791]: I1210 23:10:35.039987 4791 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/6202cb66-3ca7-4988-a800-2682cba44d72-ovsdbserver-sb\") pod \"dnsmasq-dns-845d6d6f59-d6v8c\" (UID: \"6202cb66-3ca7-4988-a800-2682cba44d72\") " pod="openstack/dnsmasq-dns-845d6d6f59-d6v8c" Dec 10 23:10:35 crc kubenswrapper[4791]: I1210 23:10:35.040173 4791 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/b010c513-02c4-4cf1-9751-4b1447588b1a-config-data\") pod \"nova-metadata-0\" (UID: \"b010c513-02c4-4cf1-9751-4b1447588b1a\") " pod="openstack/nova-metadata-0" Dec 10 23:10:35 crc kubenswrapper[4791]: I1210 23:10:35.043383 4791 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/b010c513-02c4-4cf1-9751-4b1447588b1a-logs\") pod \"nova-metadata-0\" (UID: \"b010c513-02c4-4cf1-9751-4b1447588b1a\") " pod="openstack/nova-metadata-0" Dec 10 23:10:35 crc kubenswrapper[4791]: I1210 23:10:35.048631 4791 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/b010c513-02c4-4cf1-9751-4b1447588b1a-config-data\") pod \"nova-metadata-0\" (UID: \"b010c513-02c4-4cf1-9751-4b1447588b1a\") " pod="openstack/nova-metadata-0" Dec 10 23:10:35 crc kubenswrapper[4791]: I1210 23:10:35.051904 4791 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/b010c513-02c4-4cf1-9751-4b1447588b1a-combined-ca-bundle\") pod \"nova-metadata-0\" (UID: \"b010c513-02c4-4cf1-9751-4b1447588b1a\") " pod="openstack/nova-metadata-0" Dec 10 23:10:35 crc kubenswrapper[4791]: I1210 23:10:35.065044 4791 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-kbbvd\" (UniqueName: \"kubernetes.io/projected/b010c513-02c4-4cf1-9751-4b1447588b1a-kube-api-access-kbbvd\") pod \"nova-metadata-0\" (UID: \"b010c513-02c4-4cf1-9751-4b1447588b1a\") " pod="openstack/nova-metadata-0" Dec 10 23:10:35 crc kubenswrapper[4791]: I1210 23:10:35.112282 4791 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-novncproxy-0" Dec 10 23:10:35 crc kubenswrapper[4791]: I1210 23:10:35.124034 4791 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-scheduler-0" Dec 10 23:10:35 crc kubenswrapper[4791]: I1210 23:10:35.141768 4791 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/6202cb66-3ca7-4988-a800-2682cba44d72-ovsdbserver-nb\") pod \"dnsmasq-dns-845d6d6f59-d6v8c\" (UID: \"6202cb66-3ca7-4988-a800-2682cba44d72\") " pod="openstack/dnsmasq-dns-845d6d6f59-d6v8c" Dec 10 23:10:35 crc kubenswrapper[4791]: I1210 23:10:35.141847 4791 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/6202cb66-3ca7-4988-a800-2682cba44d72-dns-svc\") pod \"dnsmasq-dns-845d6d6f59-d6v8c\" (UID: \"6202cb66-3ca7-4988-a800-2682cba44d72\") " pod="openstack/dnsmasq-dns-845d6d6f59-d6v8c" Dec 10 23:10:35 crc kubenswrapper[4791]: I1210 23:10:35.141917 4791 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-blchs\" (UniqueName: \"kubernetes.io/projected/6202cb66-3ca7-4988-a800-2682cba44d72-kube-api-access-blchs\") pod \"dnsmasq-dns-845d6d6f59-d6v8c\" (UID: \"6202cb66-3ca7-4988-a800-2682cba44d72\") " pod="openstack/dnsmasq-dns-845d6d6f59-d6v8c" Dec 10 23:10:35 crc kubenswrapper[4791]: I1210 23:10:35.141941 4791 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/6202cb66-3ca7-4988-a800-2682cba44d72-config\") pod \"dnsmasq-dns-845d6d6f59-d6v8c\" (UID: \"6202cb66-3ca7-4988-a800-2682cba44d72\") " pod="openstack/dnsmasq-dns-845d6d6f59-d6v8c" Dec 10 23:10:35 crc kubenswrapper[4791]: I1210 23:10:35.141963 4791 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/6202cb66-3ca7-4988-a800-2682cba44d72-dns-swift-storage-0\") pod \"dnsmasq-dns-845d6d6f59-d6v8c\" (UID: \"6202cb66-3ca7-4988-a800-2682cba44d72\") " pod="openstack/dnsmasq-dns-845d6d6f59-d6v8c" Dec 10 23:10:35 crc kubenswrapper[4791]: I1210 23:10:35.141994 4791 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/6202cb66-3ca7-4988-a800-2682cba44d72-ovsdbserver-sb\") pod \"dnsmasq-dns-845d6d6f59-d6v8c\" (UID: \"6202cb66-3ca7-4988-a800-2682cba44d72\") " pod="openstack/dnsmasq-dns-845d6d6f59-d6v8c" Dec 10 23:10:35 crc kubenswrapper[4791]: I1210 23:10:35.142774 4791 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/6202cb66-3ca7-4988-a800-2682cba44d72-ovsdbserver-nb\") pod \"dnsmasq-dns-845d6d6f59-d6v8c\" (UID: \"6202cb66-3ca7-4988-a800-2682cba44d72\") " pod="openstack/dnsmasq-dns-845d6d6f59-d6v8c" Dec 10 23:10:35 crc kubenswrapper[4791]: I1210 23:10:35.142806 4791 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/6202cb66-3ca7-4988-a800-2682cba44d72-ovsdbserver-sb\") pod \"dnsmasq-dns-845d6d6f59-d6v8c\" (UID: \"6202cb66-3ca7-4988-a800-2682cba44d72\") " pod="openstack/dnsmasq-dns-845d6d6f59-d6v8c" Dec 10 23:10:35 crc kubenswrapper[4791]: I1210 23:10:35.143649 4791 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/6202cb66-3ca7-4988-a800-2682cba44d72-dns-svc\") pod \"dnsmasq-dns-845d6d6f59-d6v8c\" (UID: \"6202cb66-3ca7-4988-a800-2682cba44d72\") " pod="openstack/dnsmasq-dns-845d6d6f59-d6v8c" Dec 10 23:10:35 crc kubenswrapper[4791]: I1210 23:10:35.143676 4791 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/6202cb66-3ca7-4988-a800-2682cba44d72-config\") pod \"dnsmasq-dns-845d6d6f59-d6v8c\" (UID: \"6202cb66-3ca7-4988-a800-2682cba44d72\") " pod="openstack/dnsmasq-dns-845d6d6f59-d6v8c" Dec 10 23:10:35 crc kubenswrapper[4791]: I1210 23:10:35.143734 4791 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/6202cb66-3ca7-4988-a800-2682cba44d72-dns-swift-storage-0\") pod \"dnsmasq-dns-845d6d6f59-d6v8c\" (UID: \"6202cb66-3ca7-4988-a800-2682cba44d72\") " pod="openstack/dnsmasq-dns-845d6d6f59-d6v8c" Dec 10 23:10:35 crc kubenswrapper[4791]: I1210 23:10:35.173695 4791 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-blchs\" (UniqueName: \"kubernetes.io/projected/6202cb66-3ca7-4988-a800-2682cba44d72-kube-api-access-blchs\") pod \"dnsmasq-dns-845d6d6f59-d6v8c\" (UID: \"6202cb66-3ca7-4988-a800-2682cba44d72\") " pod="openstack/dnsmasq-dns-845d6d6f59-d6v8c" Dec 10 23:10:35 crc kubenswrapper[4791]: I1210 23:10:35.203101 4791 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-metadata-0" Dec 10 23:10:35 crc kubenswrapper[4791]: I1210 23:10:35.377810 4791 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-845d6d6f59-d6v8c" Dec 10 23:10:35 crc kubenswrapper[4791]: I1210 23:10:35.535009 4791 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-api-0"] Dec 10 23:10:35 crc kubenswrapper[4791]: W1210 23:10:35.547670 4791 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod739a67cd_2a88_4c1d_bf76_5aac9b4b9e46.slice/crio-acdedd13a96fb8b18d9dc85527f97cf6678f43930fd183d20cf7220c7d07524e WatchSource:0}: Error finding container acdedd13a96fb8b18d9dc85527f97cf6678f43930fd183d20cf7220c7d07524e: Status 404 returned error can't find the container with id acdedd13a96fb8b18d9dc85527f97cf6678f43930fd183d20cf7220c7d07524e Dec 10 23:10:35 crc kubenswrapper[4791]: I1210 23:10:35.597421 4791 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-cell1-conductor-db-sync-brjck"] Dec 10 23:10:35 crc kubenswrapper[4791]: I1210 23:10:35.599540 4791 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-conductor-db-sync-brjck" Dec 10 23:10:35 crc kubenswrapper[4791]: I1210 23:10:35.603150 4791 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell1-conductor-config-data" Dec 10 23:10:35 crc kubenswrapper[4791]: I1210 23:10:35.603528 4791 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell1-conductor-scripts" Dec 10 23:10:35 crc kubenswrapper[4791]: I1210 23:10:35.623042 4791 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-conductor-db-sync-brjck"] Dec 10 23:10:35 crc kubenswrapper[4791]: I1210 23:10:35.652893 4791 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/660af60d-b638-42d4-86d1-3a7a07b1b936-combined-ca-bundle\") pod \"nova-cell1-conductor-db-sync-brjck\" (UID: \"660af60d-b638-42d4-86d1-3a7a07b1b936\") " pod="openstack/nova-cell1-conductor-db-sync-brjck" Dec 10 23:10:35 crc kubenswrapper[4791]: I1210 23:10:35.653007 4791 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/660af60d-b638-42d4-86d1-3a7a07b1b936-scripts\") pod \"nova-cell1-conductor-db-sync-brjck\" (UID: \"660af60d-b638-42d4-86d1-3a7a07b1b936\") " pod="openstack/nova-cell1-conductor-db-sync-brjck" Dec 10 23:10:35 crc kubenswrapper[4791]: I1210 23:10:35.653129 4791 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/660af60d-b638-42d4-86d1-3a7a07b1b936-config-data\") pod \"nova-cell1-conductor-db-sync-brjck\" (UID: \"660af60d-b638-42d4-86d1-3a7a07b1b936\") " pod="openstack/nova-cell1-conductor-db-sync-brjck" Dec 10 23:10:35 crc kubenswrapper[4791]: I1210 23:10:35.653177 4791 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-8dfkk\" (UniqueName: \"kubernetes.io/projected/660af60d-b638-42d4-86d1-3a7a07b1b936-kube-api-access-8dfkk\") pod \"nova-cell1-conductor-db-sync-brjck\" (UID: \"660af60d-b638-42d4-86d1-3a7a07b1b936\") " pod="openstack/nova-cell1-conductor-db-sync-brjck" Dec 10 23:10:35 crc kubenswrapper[4791]: I1210 23:10:35.717644 4791 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell0-cell-mapping-k64vq"] Dec 10 23:10:35 crc kubenswrapper[4791]: I1210 23:10:35.754503 4791 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/660af60d-b638-42d4-86d1-3a7a07b1b936-config-data\") pod \"nova-cell1-conductor-db-sync-brjck\" (UID: \"660af60d-b638-42d4-86d1-3a7a07b1b936\") " pod="openstack/nova-cell1-conductor-db-sync-brjck" Dec 10 23:10:35 crc kubenswrapper[4791]: I1210 23:10:35.754805 4791 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-8dfkk\" (UniqueName: \"kubernetes.io/projected/660af60d-b638-42d4-86d1-3a7a07b1b936-kube-api-access-8dfkk\") pod \"nova-cell1-conductor-db-sync-brjck\" (UID: \"660af60d-b638-42d4-86d1-3a7a07b1b936\") " pod="openstack/nova-cell1-conductor-db-sync-brjck" Dec 10 23:10:35 crc kubenswrapper[4791]: I1210 23:10:35.754884 4791 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/660af60d-b638-42d4-86d1-3a7a07b1b936-combined-ca-bundle\") pod \"nova-cell1-conductor-db-sync-brjck\" (UID: \"660af60d-b638-42d4-86d1-3a7a07b1b936\") " pod="openstack/nova-cell1-conductor-db-sync-brjck" Dec 10 23:10:35 crc kubenswrapper[4791]: I1210 23:10:35.754914 4791 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/660af60d-b638-42d4-86d1-3a7a07b1b936-scripts\") pod \"nova-cell1-conductor-db-sync-brjck\" (UID: \"660af60d-b638-42d4-86d1-3a7a07b1b936\") " pod="openstack/nova-cell1-conductor-db-sync-brjck" Dec 10 23:10:35 crc kubenswrapper[4791]: I1210 23:10:35.758850 4791 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/660af60d-b638-42d4-86d1-3a7a07b1b936-config-data\") pod \"nova-cell1-conductor-db-sync-brjck\" (UID: \"660af60d-b638-42d4-86d1-3a7a07b1b936\") " pod="openstack/nova-cell1-conductor-db-sync-brjck" Dec 10 23:10:35 crc kubenswrapper[4791]: I1210 23:10:35.759438 4791 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/660af60d-b638-42d4-86d1-3a7a07b1b936-combined-ca-bundle\") pod \"nova-cell1-conductor-db-sync-brjck\" (UID: \"660af60d-b638-42d4-86d1-3a7a07b1b936\") " pod="openstack/nova-cell1-conductor-db-sync-brjck" Dec 10 23:10:35 crc kubenswrapper[4791]: I1210 23:10:35.770398 4791 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/660af60d-b638-42d4-86d1-3a7a07b1b936-scripts\") pod \"nova-cell1-conductor-db-sync-brjck\" (UID: \"660af60d-b638-42d4-86d1-3a7a07b1b936\") " pod="openstack/nova-cell1-conductor-db-sync-brjck" Dec 10 23:10:35 crc kubenswrapper[4791]: I1210 23:10:35.773881 4791 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-8dfkk\" (UniqueName: \"kubernetes.io/projected/660af60d-b638-42d4-86d1-3a7a07b1b936-kube-api-access-8dfkk\") pod \"nova-cell1-conductor-db-sync-brjck\" (UID: \"660af60d-b638-42d4-86d1-3a7a07b1b936\") " pod="openstack/nova-cell1-conductor-db-sync-brjck" Dec 10 23:10:35 crc kubenswrapper[4791]: I1210 23:10:35.799640 4791 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-novncproxy-0"] Dec 10 23:10:35 crc kubenswrapper[4791]: I1210 23:10:35.815327 4791 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-scheduler-0"] Dec 10 23:10:35 crc kubenswrapper[4791]: I1210 23:10:35.933797 4791 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-conductor-db-sync-brjck" Dec 10 23:10:36 crc kubenswrapper[4791]: I1210 23:10:36.008191 4791 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-metadata-0"] Dec 10 23:10:36 crc kubenswrapper[4791]: I1210 23:10:36.057594 4791 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-845d6d6f59-d6v8c"] Dec 10 23:10:36 crc kubenswrapper[4791]: W1210 23:10:36.063406 4791 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod6202cb66_3ca7_4988_a800_2682cba44d72.slice/crio-3190dbd7cddd5578a9e13db8ddfc19754d10d685bddb95bc7772fb7786e3334a WatchSource:0}: Error finding container 3190dbd7cddd5578a9e13db8ddfc19754d10d685bddb95bc7772fb7786e3334a: Status 404 returned error can't find the container with id 3190dbd7cddd5578a9e13db8ddfc19754d10d685bddb95bc7772fb7786e3334a Dec 10 23:10:36 crc kubenswrapper[4791]: I1210 23:10:36.279111 4791 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-conductor-db-sync-brjck"] Dec 10 23:10:36 crc kubenswrapper[4791]: I1210 23:10:36.280582 4791 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-cell-mapping-k64vq" event={"ID":"da9f584f-c289-4d67-ae91-8d6bf7507fb6","Type":"ContainerStarted","Data":"ccf60cdfa54857941fadb4022be27fb785d133c0ce8f11676f66323e635c7b71"} Dec 10 23:10:36 crc kubenswrapper[4791]: I1210 23:10:36.280657 4791 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-cell-mapping-k64vq" event={"ID":"da9f584f-c289-4d67-ae91-8d6bf7507fb6","Type":"ContainerStarted","Data":"3c7542d32480373b1cee85c021b42ffe5b2cbb8f402c8084948eeb72f19c8700"} Dec 10 23:10:36 crc kubenswrapper[4791]: I1210 23:10:36.295896 4791 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-845d6d6f59-d6v8c" event={"ID":"6202cb66-3ca7-4988-a800-2682cba44d72","Type":"ContainerStarted","Data":"3190dbd7cddd5578a9e13db8ddfc19754d10d685bddb95bc7772fb7786e3334a"} Dec 10 23:10:36 crc kubenswrapper[4791]: I1210 23:10:36.306610 4791 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-scheduler-0" event={"ID":"83b6abc9-aa7f-4ddb-8b9c-b38fcb5a9145","Type":"ContainerStarted","Data":"03be616aec5c91eea6fa4690564c719ddb65089216635dcfb713674db11da81f"} Dec 10 23:10:36 crc kubenswrapper[4791]: I1210 23:10:36.315908 4791 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"739a67cd-2a88-4c1d-bf76-5aac9b4b9e46","Type":"ContainerStarted","Data":"acdedd13a96fb8b18d9dc85527f97cf6678f43930fd183d20cf7220c7d07524e"} Dec 10 23:10:36 crc kubenswrapper[4791]: I1210 23:10:36.324893 4791 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-novncproxy-0" event={"ID":"e2d1527a-aac6-4ee0-9c41-60ef0784aa2a","Type":"ContainerStarted","Data":"1010cab3c0202dd1630d3aaf7a72e0d3e82b2e33edbfd1d44a507070c7b291b4"} Dec 10 23:10:36 crc kubenswrapper[4791]: I1210 23:10:36.337201 4791 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"b010c513-02c4-4cf1-9751-4b1447588b1a","Type":"ContainerStarted","Data":"e48b599333ea5cb479afa2a013a9591a0af5f45ddf9dc5191fcfde815c9a8963"} Dec 10 23:10:36 crc kubenswrapper[4791]: I1210 23:10:36.362163 4791 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-cell0-cell-mapping-k64vq" podStartSLOduration=2.357315632 podStartE2EDuration="2.357315632s" podCreationTimestamp="2025-12-10 23:10:34 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-10 23:10:36.329063646 +0000 UTC m=+1270.758681259" watchObservedRunningTime="2025-12-10 23:10:36.357315632 +0000 UTC m=+1270.786933245" Dec 10 23:10:37 crc kubenswrapper[4791]: I1210 23:10:37.376237 4791 generic.go:334] "Generic (PLEG): container finished" podID="6202cb66-3ca7-4988-a800-2682cba44d72" containerID="6559890b469fbfa46c5812c5c8a37e419f0707fab3dc84b65e7d1f9a42fdd86e" exitCode=0 Dec 10 23:10:37 crc kubenswrapper[4791]: I1210 23:10:37.376614 4791 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-845d6d6f59-d6v8c" event={"ID":"6202cb66-3ca7-4988-a800-2682cba44d72","Type":"ContainerDied","Data":"6559890b469fbfa46c5812c5c8a37e419f0707fab3dc84b65e7d1f9a42fdd86e"} Dec 10 23:10:37 crc kubenswrapper[4791]: I1210 23:10:37.393421 4791 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-conductor-db-sync-brjck" event={"ID":"660af60d-b638-42d4-86d1-3a7a07b1b936","Type":"ContainerStarted","Data":"51879f033bb3a1c8fd84acf13c94ff25334fae72bdf0c18287e1e1c3aeda56a7"} Dec 10 23:10:37 crc kubenswrapper[4791]: I1210 23:10:37.393477 4791 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-conductor-db-sync-brjck" event={"ID":"660af60d-b638-42d4-86d1-3a7a07b1b936","Type":"ContainerStarted","Data":"ad353c4652105c4569ebe7e1bd36575c22b7dcc1500c78ea7bfeb8e73d12546f"} Dec 10 23:10:38 crc kubenswrapper[4791]: I1210 23:10:38.390561 4791 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-cell1-conductor-db-sync-brjck" podStartSLOduration=3.3905342960000002 podStartE2EDuration="3.390534296s" podCreationTimestamp="2025-12-10 23:10:35 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-10 23:10:37.430737029 +0000 UTC m=+1271.860354642" watchObservedRunningTime="2025-12-10 23:10:38.390534296 +0000 UTC m=+1272.820151909" Dec 10 23:10:38 crc kubenswrapper[4791]: I1210 23:10:38.405587 4791 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-cell1-novncproxy-0"] Dec 10 23:10:38 crc kubenswrapper[4791]: I1210 23:10:38.421817 4791 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-845d6d6f59-d6v8c" event={"ID":"6202cb66-3ca7-4988-a800-2682cba44d72","Type":"ContainerStarted","Data":"23c8115e63afae1fa7a5d92c7142b2a8f9b026f2cd7fdba3e8bd355ae6b7935f"} Dec 10 23:10:38 crc kubenswrapper[4791]: I1210 23:10:38.421893 4791 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/dnsmasq-dns-845d6d6f59-d6v8c" Dec 10 23:10:38 crc kubenswrapper[4791]: I1210 23:10:38.426380 4791 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-metadata-0"] Dec 10 23:10:38 crc kubenswrapper[4791]: I1210 23:10:38.453190 4791 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/dnsmasq-dns-845d6d6f59-d6v8c" podStartSLOduration=4.453166874 podStartE2EDuration="4.453166874s" podCreationTimestamp="2025-12-10 23:10:34 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-10 23:10:38.446216713 +0000 UTC m=+1272.875834326" watchObservedRunningTime="2025-12-10 23:10:38.453166874 +0000 UTC m=+1272.882784487" Dec 10 23:10:41 crc kubenswrapper[4791]: I1210 23:10:41.466964 4791 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-novncproxy-0" event={"ID":"e2d1527a-aac6-4ee0-9c41-60ef0784aa2a","Type":"ContainerStarted","Data":"825d750a33ded08f66510a0c44cef8a3f541452ccad46a74bd1a0a7e72513c5d"} Dec 10 23:10:41 crc kubenswrapper[4791]: I1210 23:10:41.467057 4791 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-cell1-novncproxy-0" podUID="e2d1527a-aac6-4ee0-9c41-60ef0784aa2a" containerName="nova-cell1-novncproxy-novncproxy" containerID="cri-o://825d750a33ded08f66510a0c44cef8a3f541452ccad46a74bd1a0a7e72513c5d" gracePeriod=30 Dec 10 23:10:41 crc kubenswrapper[4791]: I1210 23:10:41.469772 4791 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"b010c513-02c4-4cf1-9751-4b1447588b1a","Type":"ContainerStarted","Data":"d5303b42805b7c27da72680839b093b2c716650a6c2f528d6f6666c4db2b9acd"} Dec 10 23:10:41 crc kubenswrapper[4791]: I1210 23:10:41.469836 4791 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"b010c513-02c4-4cf1-9751-4b1447588b1a","Type":"ContainerStarted","Data":"ab35dc970dfe36c5b9556f9cab59b7916ddbd22a4e1f125f3f8544c2505c6efe"} Dec 10 23:10:41 crc kubenswrapper[4791]: I1210 23:10:41.469978 4791 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-metadata-0" podUID="b010c513-02c4-4cf1-9751-4b1447588b1a" containerName="nova-metadata-log" containerID="cri-o://ab35dc970dfe36c5b9556f9cab59b7916ddbd22a4e1f125f3f8544c2505c6efe" gracePeriod=30 Dec 10 23:10:41 crc kubenswrapper[4791]: I1210 23:10:41.470105 4791 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-metadata-0" podUID="b010c513-02c4-4cf1-9751-4b1447588b1a" containerName="nova-metadata-metadata" containerID="cri-o://d5303b42805b7c27da72680839b093b2c716650a6c2f528d6f6666c4db2b9acd" gracePeriod=30 Dec 10 23:10:41 crc kubenswrapper[4791]: I1210 23:10:41.472297 4791 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-scheduler-0" event={"ID":"83b6abc9-aa7f-4ddb-8b9c-b38fcb5a9145","Type":"ContainerStarted","Data":"4eb9f1cc7aae957c4598d0c9a8860aedc87158afb8efbea24833107d43e12aea"} Dec 10 23:10:41 crc kubenswrapper[4791]: I1210 23:10:41.480515 4791 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"739a67cd-2a88-4c1d-bf76-5aac9b4b9e46","Type":"ContainerStarted","Data":"a8354ff62f46c0a1cbbcd0e6eb42b520e0b04bcc5960c67755b39312a1f45c3c"} Dec 10 23:10:41 crc kubenswrapper[4791]: I1210 23:10:41.480555 4791 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"739a67cd-2a88-4c1d-bf76-5aac9b4b9e46","Type":"ContainerStarted","Data":"7cdac15bae335278abc24d833001eafda5d44d0849475b593b93f231aff6eb2e"} Dec 10 23:10:41 crc kubenswrapper[4791]: I1210 23:10:41.491555 4791 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-cell1-novncproxy-0" podStartSLOduration=2.973288924 podStartE2EDuration="7.491535624s" podCreationTimestamp="2025-12-10 23:10:34 +0000 UTC" firstStartedPulling="2025-12-10 23:10:35.83312062 +0000 UTC m=+1270.262738233" lastFinishedPulling="2025-12-10 23:10:40.35136732 +0000 UTC m=+1274.780984933" observedRunningTime="2025-12-10 23:10:41.491078751 +0000 UTC m=+1275.920696374" watchObservedRunningTime="2025-12-10 23:10:41.491535624 +0000 UTC m=+1275.921153237" Dec 10 23:10:41 crc kubenswrapper[4791]: I1210 23:10:41.546677 4791 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-scheduler-0" podStartSLOduration=3.00744247 podStartE2EDuration="7.546654425s" podCreationTimestamp="2025-12-10 23:10:34 +0000 UTC" firstStartedPulling="2025-12-10 23:10:35.819376603 +0000 UTC m=+1270.248994206" lastFinishedPulling="2025-12-10 23:10:40.358588548 +0000 UTC m=+1274.788206161" observedRunningTime="2025-12-10 23:10:41.527703298 +0000 UTC m=+1275.957320911" watchObservedRunningTime="2025-12-10 23:10:41.546654425 +0000 UTC m=+1275.976272038" Dec 10 23:10:41 crc kubenswrapper[4791]: I1210 23:10:41.550067 4791 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-metadata-0" podStartSLOduration=3.205112756 podStartE2EDuration="7.550051103s" podCreationTimestamp="2025-12-10 23:10:34 +0000 UTC" firstStartedPulling="2025-12-10 23:10:36.014849196 +0000 UTC m=+1270.444466809" lastFinishedPulling="2025-12-10 23:10:40.359787543 +0000 UTC m=+1274.789405156" observedRunningTime="2025-12-10 23:10:41.546039387 +0000 UTC m=+1275.975657010" watchObservedRunningTime="2025-12-10 23:10:41.550051103 +0000 UTC m=+1275.979668716" Dec 10 23:10:41 crc kubenswrapper[4791]: I1210 23:10:41.571451 4791 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-api-0" podStartSLOduration=2.765230548 podStartE2EDuration="7.57142952s" podCreationTimestamp="2025-12-10 23:10:34 +0000 UTC" firstStartedPulling="2025-12-10 23:10:35.559860472 +0000 UTC m=+1269.989478085" lastFinishedPulling="2025-12-10 23:10:40.366059424 +0000 UTC m=+1274.795677057" observedRunningTime="2025-12-10 23:10:41.566529269 +0000 UTC m=+1275.996146882" watchObservedRunningTime="2025-12-10 23:10:41.57142952 +0000 UTC m=+1276.001047133" Dec 10 23:10:42 crc kubenswrapper[4791]: I1210 23:10:42.126711 4791 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-metadata-0" Dec 10 23:10:42 crc kubenswrapper[4791]: I1210 23:10:42.204884 4791 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/b010c513-02c4-4cf1-9751-4b1447588b1a-config-data\") pod \"b010c513-02c4-4cf1-9751-4b1447588b1a\" (UID: \"b010c513-02c4-4cf1-9751-4b1447588b1a\") " Dec 10 23:10:42 crc kubenswrapper[4791]: I1210 23:10:42.204977 4791 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/b010c513-02c4-4cf1-9751-4b1447588b1a-combined-ca-bundle\") pod \"b010c513-02c4-4cf1-9751-4b1447588b1a\" (UID: \"b010c513-02c4-4cf1-9751-4b1447588b1a\") " Dec 10 23:10:42 crc kubenswrapper[4791]: I1210 23:10:42.205020 4791 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/b010c513-02c4-4cf1-9751-4b1447588b1a-logs\") pod \"b010c513-02c4-4cf1-9751-4b1447588b1a\" (UID: \"b010c513-02c4-4cf1-9751-4b1447588b1a\") " Dec 10 23:10:42 crc kubenswrapper[4791]: I1210 23:10:42.205101 4791 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-kbbvd\" (UniqueName: \"kubernetes.io/projected/b010c513-02c4-4cf1-9751-4b1447588b1a-kube-api-access-kbbvd\") pod \"b010c513-02c4-4cf1-9751-4b1447588b1a\" (UID: \"b010c513-02c4-4cf1-9751-4b1447588b1a\") " Dec 10 23:10:42 crc kubenswrapper[4791]: I1210 23:10:42.206362 4791 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/b010c513-02c4-4cf1-9751-4b1447588b1a-logs" (OuterVolumeSpecName: "logs") pod "b010c513-02c4-4cf1-9751-4b1447588b1a" (UID: "b010c513-02c4-4cf1-9751-4b1447588b1a"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 10 23:10:42 crc kubenswrapper[4791]: I1210 23:10:42.211427 4791 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/b010c513-02c4-4cf1-9751-4b1447588b1a-kube-api-access-kbbvd" (OuterVolumeSpecName: "kube-api-access-kbbvd") pod "b010c513-02c4-4cf1-9751-4b1447588b1a" (UID: "b010c513-02c4-4cf1-9751-4b1447588b1a"). InnerVolumeSpecName "kube-api-access-kbbvd". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 10 23:10:42 crc kubenswrapper[4791]: I1210 23:10:42.245938 4791 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/b010c513-02c4-4cf1-9751-4b1447588b1a-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "b010c513-02c4-4cf1-9751-4b1447588b1a" (UID: "b010c513-02c4-4cf1-9751-4b1447588b1a"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 10 23:10:42 crc kubenswrapper[4791]: I1210 23:10:42.249160 4791 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/b010c513-02c4-4cf1-9751-4b1447588b1a-config-data" (OuterVolumeSpecName: "config-data") pod "b010c513-02c4-4cf1-9751-4b1447588b1a" (UID: "b010c513-02c4-4cf1-9751-4b1447588b1a"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 10 23:10:42 crc kubenswrapper[4791]: I1210 23:10:42.318311 4791 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-kbbvd\" (UniqueName: \"kubernetes.io/projected/b010c513-02c4-4cf1-9751-4b1447588b1a-kube-api-access-kbbvd\") on node \"crc\" DevicePath \"\"" Dec 10 23:10:42 crc kubenswrapper[4791]: I1210 23:10:42.318426 4791 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/b010c513-02c4-4cf1-9751-4b1447588b1a-config-data\") on node \"crc\" DevicePath \"\"" Dec 10 23:10:42 crc kubenswrapper[4791]: I1210 23:10:42.318462 4791 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/b010c513-02c4-4cf1-9751-4b1447588b1a-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 10 23:10:42 crc kubenswrapper[4791]: I1210 23:10:42.318481 4791 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/b010c513-02c4-4cf1-9751-4b1447588b1a-logs\") on node \"crc\" DevicePath \"\"" Dec 10 23:10:42 crc kubenswrapper[4791]: I1210 23:10:42.492282 4791 generic.go:334] "Generic (PLEG): container finished" podID="b010c513-02c4-4cf1-9751-4b1447588b1a" containerID="d5303b42805b7c27da72680839b093b2c716650a6c2f528d6f6666c4db2b9acd" exitCode=0 Dec 10 23:10:42 crc kubenswrapper[4791]: I1210 23:10:42.492329 4791 generic.go:334] "Generic (PLEG): container finished" podID="b010c513-02c4-4cf1-9751-4b1447588b1a" containerID="ab35dc970dfe36c5b9556f9cab59b7916ddbd22a4e1f125f3f8544c2505c6efe" exitCode=143 Dec 10 23:10:42 crc kubenswrapper[4791]: I1210 23:10:42.492354 4791 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-metadata-0" Dec 10 23:10:42 crc kubenswrapper[4791]: I1210 23:10:42.492390 4791 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"b010c513-02c4-4cf1-9751-4b1447588b1a","Type":"ContainerDied","Data":"d5303b42805b7c27da72680839b093b2c716650a6c2f528d6f6666c4db2b9acd"} Dec 10 23:10:42 crc kubenswrapper[4791]: I1210 23:10:42.492423 4791 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"b010c513-02c4-4cf1-9751-4b1447588b1a","Type":"ContainerDied","Data":"ab35dc970dfe36c5b9556f9cab59b7916ddbd22a4e1f125f3f8544c2505c6efe"} Dec 10 23:10:42 crc kubenswrapper[4791]: I1210 23:10:42.492438 4791 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"b010c513-02c4-4cf1-9751-4b1447588b1a","Type":"ContainerDied","Data":"e48b599333ea5cb479afa2a013a9591a0af5f45ddf9dc5191fcfde815c9a8963"} Dec 10 23:10:42 crc kubenswrapper[4791]: I1210 23:10:42.492457 4791 scope.go:117] "RemoveContainer" containerID="d5303b42805b7c27da72680839b093b2c716650a6c2f528d6f6666c4db2b9acd" Dec 10 23:10:42 crc kubenswrapper[4791]: I1210 23:10:42.525384 4791 scope.go:117] "RemoveContainer" containerID="ab35dc970dfe36c5b9556f9cab59b7916ddbd22a4e1f125f3f8544c2505c6efe" Dec 10 23:10:42 crc kubenswrapper[4791]: I1210 23:10:42.546299 4791 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-metadata-0"] Dec 10 23:10:42 crc kubenswrapper[4791]: I1210 23:10:42.557137 4791 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-metadata-0"] Dec 10 23:10:42 crc kubenswrapper[4791]: I1210 23:10:42.566393 4791 scope.go:117] "RemoveContainer" containerID="d5303b42805b7c27da72680839b093b2c716650a6c2f528d6f6666c4db2b9acd" Dec 10 23:10:42 crc kubenswrapper[4791]: I1210 23:10:42.567786 4791 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-metadata-0"] Dec 10 23:10:42 crc kubenswrapper[4791]: E1210 23:10:42.568207 4791 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="b010c513-02c4-4cf1-9751-4b1447588b1a" containerName="nova-metadata-metadata" Dec 10 23:10:42 crc kubenswrapper[4791]: I1210 23:10:42.568226 4791 state_mem.go:107] "Deleted CPUSet assignment" podUID="b010c513-02c4-4cf1-9751-4b1447588b1a" containerName="nova-metadata-metadata" Dec 10 23:10:42 crc kubenswrapper[4791]: E1210 23:10:42.568243 4791 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="b010c513-02c4-4cf1-9751-4b1447588b1a" containerName="nova-metadata-log" Dec 10 23:10:42 crc kubenswrapper[4791]: I1210 23:10:42.568250 4791 state_mem.go:107] "Deleted CPUSet assignment" podUID="b010c513-02c4-4cf1-9751-4b1447588b1a" containerName="nova-metadata-log" Dec 10 23:10:42 crc kubenswrapper[4791]: I1210 23:10:42.568471 4791 memory_manager.go:354] "RemoveStaleState removing state" podUID="b010c513-02c4-4cf1-9751-4b1447588b1a" containerName="nova-metadata-log" Dec 10 23:10:42 crc kubenswrapper[4791]: I1210 23:10:42.568503 4791 memory_manager.go:354] "RemoveStaleState removing state" podUID="b010c513-02c4-4cf1-9751-4b1447588b1a" containerName="nova-metadata-metadata" Dec 10 23:10:42 crc kubenswrapper[4791]: I1210 23:10:42.569562 4791 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-metadata-0" Dec 10 23:10:42 crc kubenswrapper[4791]: E1210 23:10:42.570085 4791 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"d5303b42805b7c27da72680839b093b2c716650a6c2f528d6f6666c4db2b9acd\": container with ID starting with d5303b42805b7c27da72680839b093b2c716650a6c2f528d6f6666c4db2b9acd not found: ID does not exist" containerID="d5303b42805b7c27da72680839b093b2c716650a6c2f528d6f6666c4db2b9acd" Dec 10 23:10:42 crc kubenswrapper[4791]: I1210 23:10:42.570217 4791 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"d5303b42805b7c27da72680839b093b2c716650a6c2f528d6f6666c4db2b9acd"} err="failed to get container status \"d5303b42805b7c27da72680839b093b2c716650a6c2f528d6f6666c4db2b9acd\": rpc error: code = NotFound desc = could not find container \"d5303b42805b7c27da72680839b093b2c716650a6c2f528d6f6666c4db2b9acd\": container with ID starting with d5303b42805b7c27da72680839b093b2c716650a6c2f528d6f6666c4db2b9acd not found: ID does not exist" Dec 10 23:10:42 crc kubenswrapper[4791]: I1210 23:10:42.570503 4791 scope.go:117] "RemoveContainer" containerID="ab35dc970dfe36c5b9556f9cab59b7916ddbd22a4e1f125f3f8544c2505c6efe" Dec 10 23:10:42 crc kubenswrapper[4791]: I1210 23:10:42.574163 4791 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-metadata-config-data" Dec 10 23:10:42 crc kubenswrapper[4791]: I1210 23:10:42.574421 4791 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-nova-metadata-internal-svc" Dec 10 23:10:42 crc kubenswrapper[4791]: E1210 23:10:42.582324 4791 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"ab35dc970dfe36c5b9556f9cab59b7916ddbd22a4e1f125f3f8544c2505c6efe\": container with ID starting with ab35dc970dfe36c5b9556f9cab59b7916ddbd22a4e1f125f3f8544c2505c6efe not found: ID does not exist" containerID="ab35dc970dfe36c5b9556f9cab59b7916ddbd22a4e1f125f3f8544c2505c6efe" Dec 10 23:10:42 crc kubenswrapper[4791]: I1210 23:10:42.582412 4791 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"ab35dc970dfe36c5b9556f9cab59b7916ddbd22a4e1f125f3f8544c2505c6efe"} err="failed to get container status \"ab35dc970dfe36c5b9556f9cab59b7916ddbd22a4e1f125f3f8544c2505c6efe\": rpc error: code = NotFound desc = could not find container \"ab35dc970dfe36c5b9556f9cab59b7916ddbd22a4e1f125f3f8544c2505c6efe\": container with ID starting with ab35dc970dfe36c5b9556f9cab59b7916ddbd22a4e1f125f3f8544c2505c6efe not found: ID does not exist" Dec 10 23:10:42 crc kubenswrapper[4791]: I1210 23:10:42.582452 4791 scope.go:117] "RemoveContainer" containerID="d5303b42805b7c27da72680839b093b2c716650a6c2f528d6f6666c4db2b9acd" Dec 10 23:10:42 crc kubenswrapper[4791]: I1210 23:10:42.586195 4791 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"d5303b42805b7c27da72680839b093b2c716650a6c2f528d6f6666c4db2b9acd"} err="failed to get container status \"d5303b42805b7c27da72680839b093b2c716650a6c2f528d6f6666c4db2b9acd\": rpc error: code = NotFound desc = could not find container \"d5303b42805b7c27da72680839b093b2c716650a6c2f528d6f6666c4db2b9acd\": container with ID starting with d5303b42805b7c27da72680839b093b2c716650a6c2f528d6f6666c4db2b9acd not found: ID does not exist" Dec 10 23:10:42 crc kubenswrapper[4791]: I1210 23:10:42.586243 4791 scope.go:117] "RemoveContainer" containerID="ab35dc970dfe36c5b9556f9cab59b7916ddbd22a4e1f125f3f8544c2505c6efe" Dec 10 23:10:42 crc kubenswrapper[4791]: I1210 23:10:42.588414 4791 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"ab35dc970dfe36c5b9556f9cab59b7916ddbd22a4e1f125f3f8544c2505c6efe"} err="failed to get container status \"ab35dc970dfe36c5b9556f9cab59b7916ddbd22a4e1f125f3f8544c2505c6efe\": rpc error: code = NotFound desc = could not find container \"ab35dc970dfe36c5b9556f9cab59b7916ddbd22a4e1f125f3f8544c2505c6efe\": container with ID starting with ab35dc970dfe36c5b9556f9cab59b7916ddbd22a4e1f125f3f8544c2505c6efe not found: ID does not exist" Dec 10 23:10:42 crc kubenswrapper[4791]: I1210 23:10:42.596181 4791 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-metadata-0"] Dec 10 23:10:42 crc kubenswrapper[4791]: I1210 23:10:42.626230 4791 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-w58mp\" (UniqueName: \"kubernetes.io/projected/5d7b44d4-1936-4588-af50-442e04f245fc-kube-api-access-w58mp\") pod \"nova-metadata-0\" (UID: \"5d7b44d4-1936-4588-af50-442e04f245fc\") " pod="openstack/nova-metadata-0" Dec 10 23:10:42 crc kubenswrapper[4791]: I1210 23:10:42.626768 4791 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/5d7b44d4-1936-4588-af50-442e04f245fc-combined-ca-bundle\") pod \"nova-metadata-0\" (UID: \"5d7b44d4-1936-4588-af50-442e04f245fc\") " pod="openstack/nova-metadata-0" Dec 10 23:10:42 crc kubenswrapper[4791]: I1210 23:10:42.627233 4791 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/5d7b44d4-1936-4588-af50-442e04f245fc-logs\") pod \"nova-metadata-0\" (UID: \"5d7b44d4-1936-4588-af50-442e04f245fc\") " pod="openstack/nova-metadata-0" Dec 10 23:10:42 crc kubenswrapper[4791]: I1210 23:10:42.627305 4791 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"nova-metadata-tls-certs\" (UniqueName: \"kubernetes.io/secret/5d7b44d4-1936-4588-af50-442e04f245fc-nova-metadata-tls-certs\") pod \"nova-metadata-0\" (UID: \"5d7b44d4-1936-4588-af50-442e04f245fc\") " pod="openstack/nova-metadata-0" Dec 10 23:10:42 crc kubenswrapper[4791]: I1210 23:10:42.627406 4791 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/5d7b44d4-1936-4588-af50-442e04f245fc-config-data\") pod \"nova-metadata-0\" (UID: \"5d7b44d4-1936-4588-af50-442e04f245fc\") " pod="openstack/nova-metadata-0" Dec 10 23:10:42 crc kubenswrapper[4791]: I1210 23:10:42.729182 4791 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/5d7b44d4-1936-4588-af50-442e04f245fc-config-data\") pod \"nova-metadata-0\" (UID: \"5d7b44d4-1936-4588-af50-442e04f245fc\") " pod="openstack/nova-metadata-0" Dec 10 23:10:42 crc kubenswrapper[4791]: I1210 23:10:42.729452 4791 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-w58mp\" (UniqueName: \"kubernetes.io/projected/5d7b44d4-1936-4588-af50-442e04f245fc-kube-api-access-w58mp\") pod \"nova-metadata-0\" (UID: \"5d7b44d4-1936-4588-af50-442e04f245fc\") " pod="openstack/nova-metadata-0" Dec 10 23:10:42 crc kubenswrapper[4791]: I1210 23:10:42.729538 4791 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/5d7b44d4-1936-4588-af50-442e04f245fc-combined-ca-bundle\") pod \"nova-metadata-0\" (UID: \"5d7b44d4-1936-4588-af50-442e04f245fc\") " pod="openstack/nova-metadata-0" Dec 10 23:10:42 crc kubenswrapper[4791]: I1210 23:10:42.729627 4791 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/5d7b44d4-1936-4588-af50-442e04f245fc-logs\") pod \"nova-metadata-0\" (UID: \"5d7b44d4-1936-4588-af50-442e04f245fc\") " pod="openstack/nova-metadata-0" Dec 10 23:10:42 crc kubenswrapper[4791]: I1210 23:10:42.729657 4791 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nova-metadata-tls-certs\" (UniqueName: \"kubernetes.io/secret/5d7b44d4-1936-4588-af50-442e04f245fc-nova-metadata-tls-certs\") pod \"nova-metadata-0\" (UID: \"5d7b44d4-1936-4588-af50-442e04f245fc\") " pod="openstack/nova-metadata-0" Dec 10 23:10:42 crc kubenswrapper[4791]: I1210 23:10:42.730255 4791 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/5d7b44d4-1936-4588-af50-442e04f245fc-logs\") pod \"nova-metadata-0\" (UID: \"5d7b44d4-1936-4588-af50-442e04f245fc\") " pod="openstack/nova-metadata-0" Dec 10 23:10:42 crc kubenswrapper[4791]: I1210 23:10:42.735319 4791 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"nova-metadata-tls-certs\" (UniqueName: \"kubernetes.io/secret/5d7b44d4-1936-4588-af50-442e04f245fc-nova-metadata-tls-certs\") pod \"nova-metadata-0\" (UID: \"5d7b44d4-1936-4588-af50-442e04f245fc\") " pod="openstack/nova-metadata-0" Dec 10 23:10:42 crc kubenswrapper[4791]: I1210 23:10:42.736102 4791 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/5d7b44d4-1936-4588-af50-442e04f245fc-combined-ca-bundle\") pod \"nova-metadata-0\" (UID: \"5d7b44d4-1936-4588-af50-442e04f245fc\") " pod="openstack/nova-metadata-0" Dec 10 23:10:42 crc kubenswrapper[4791]: I1210 23:10:42.744018 4791 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/5d7b44d4-1936-4588-af50-442e04f245fc-config-data\") pod \"nova-metadata-0\" (UID: \"5d7b44d4-1936-4588-af50-442e04f245fc\") " pod="openstack/nova-metadata-0" Dec 10 23:10:42 crc kubenswrapper[4791]: I1210 23:10:42.759095 4791 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-w58mp\" (UniqueName: \"kubernetes.io/projected/5d7b44d4-1936-4588-af50-442e04f245fc-kube-api-access-w58mp\") pod \"nova-metadata-0\" (UID: \"5d7b44d4-1936-4588-af50-442e04f245fc\") " pod="openstack/nova-metadata-0" Dec 10 23:10:42 crc kubenswrapper[4791]: I1210 23:10:42.888126 4791 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-metadata-0" Dec 10 23:10:43 crc kubenswrapper[4791]: I1210 23:10:43.418849 4791 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-metadata-0"] Dec 10 23:10:43 crc kubenswrapper[4791]: I1210 23:10:43.591112 4791 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"5d7b44d4-1936-4588-af50-442e04f245fc","Type":"ContainerStarted","Data":"6f03f3d2cd8adb3c9c0e887946e97b5decff87ff3451ba631aac7b33c8870dff"} Dec 10 23:10:43 crc kubenswrapper[4791]: I1210 23:10:43.898713 4791 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="b010c513-02c4-4cf1-9751-4b1447588b1a" path="/var/lib/kubelet/pods/b010c513-02c4-4cf1-9751-4b1447588b1a/volumes" Dec 10 23:10:44 crc kubenswrapper[4791]: I1210 23:10:44.614777 4791 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"5d7b44d4-1936-4588-af50-442e04f245fc","Type":"ContainerStarted","Data":"d5f65d5be988d066eebaad7dc5f9b030442a2ed929476f704d133fc049905317"} Dec 10 23:10:44 crc kubenswrapper[4791]: I1210 23:10:44.614821 4791 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"5d7b44d4-1936-4588-af50-442e04f245fc","Type":"ContainerStarted","Data":"452e27ffea76e2ade3d7bf72c2abc8c40cca428c422ffe73d9e7aa831c8647d1"} Dec 10 23:10:44 crc kubenswrapper[4791]: I1210 23:10:44.653288 4791 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-metadata-0" podStartSLOduration=2.653263394 podStartE2EDuration="2.653263394s" podCreationTimestamp="2025-12-10 23:10:42 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-10 23:10:44.637587742 +0000 UTC m=+1279.067205365" watchObservedRunningTime="2025-12-10 23:10:44.653263394 +0000 UTC m=+1279.082881027" Dec 10 23:10:44 crc kubenswrapper[4791]: I1210 23:10:44.890936 4791 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-api-0" Dec 10 23:10:44 crc kubenswrapper[4791]: I1210 23:10:44.890995 4791 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-api-0" Dec 10 23:10:45 crc kubenswrapper[4791]: I1210 23:10:45.114799 4791 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-cell1-novncproxy-0" Dec 10 23:10:45 crc kubenswrapper[4791]: I1210 23:10:45.128390 4791 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-scheduler-0" Dec 10 23:10:45 crc kubenswrapper[4791]: I1210 23:10:45.128456 4791 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-scheduler-0" Dec 10 23:10:45 crc kubenswrapper[4791]: I1210 23:10:45.161167 4791 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/nova-scheduler-0" Dec 10 23:10:45 crc kubenswrapper[4791]: I1210 23:10:45.381085 4791 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/dnsmasq-dns-845d6d6f59-d6v8c" Dec 10 23:10:45 crc kubenswrapper[4791]: I1210 23:10:45.477119 4791 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-5784cf869f-ngjqj"] Dec 10 23:10:45 crc kubenswrapper[4791]: I1210 23:10:45.480453 4791 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/dnsmasq-dns-5784cf869f-ngjqj" podUID="224a8e50-4270-42e8-a1df-268bc144b1d5" containerName="dnsmasq-dns" containerID="cri-o://d0b7aa50e7609ff3f59cefa955b1e1e05f6d96ae4a3e4143c8e51f4065c45be1" gracePeriod=10 Dec 10 23:10:45 crc kubenswrapper[4791]: I1210 23:10:45.648862 4791 generic.go:334] "Generic (PLEG): container finished" podID="da9f584f-c289-4d67-ae91-8d6bf7507fb6" containerID="ccf60cdfa54857941fadb4022be27fb785d133c0ce8f11676f66323e635c7b71" exitCode=0 Dec 10 23:10:45 crc kubenswrapper[4791]: I1210 23:10:45.648915 4791 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-cell-mapping-k64vq" event={"ID":"da9f584f-c289-4d67-ae91-8d6bf7507fb6","Type":"ContainerDied","Data":"ccf60cdfa54857941fadb4022be27fb785d133c0ce8f11676f66323e635c7b71"} Dec 10 23:10:45 crc kubenswrapper[4791]: I1210 23:10:45.654412 4791 generic.go:334] "Generic (PLEG): container finished" podID="660af60d-b638-42d4-86d1-3a7a07b1b936" containerID="51879f033bb3a1c8fd84acf13c94ff25334fae72bdf0c18287e1e1c3aeda56a7" exitCode=0 Dec 10 23:10:45 crc kubenswrapper[4791]: I1210 23:10:45.654463 4791 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-conductor-db-sync-brjck" event={"ID":"660af60d-b638-42d4-86d1-3a7a07b1b936","Type":"ContainerDied","Data":"51879f033bb3a1c8fd84acf13c94ff25334fae72bdf0c18287e1e1c3aeda56a7"} Dec 10 23:10:45 crc kubenswrapper[4791]: I1210 23:10:45.656322 4791 generic.go:334] "Generic (PLEG): container finished" podID="224a8e50-4270-42e8-a1df-268bc144b1d5" containerID="d0b7aa50e7609ff3f59cefa955b1e1e05f6d96ae4a3e4143c8e51f4065c45be1" exitCode=0 Dec 10 23:10:45 crc kubenswrapper[4791]: I1210 23:10:45.656673 4791 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-5784cf869f-ngjqj" event={"ID":"224a8e50-4270-42e8-a1df-268bc144b1d5","Type":"ContainerDied","Data":"d0b7aa50e7609ff3f59cefa955b1e1e05f6d96ae4a3e4143c8e51f4065c45be1"} Dec 10 23:10:45 crc kubenswrapper[4791]: I1210 23:10:45.698770 4791 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-scheduler-0" Dec 10 23:10:45 crc kubenswrapper[4791]: I1210 23:10:45.974616 4791 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/nova-api-0" podUID="739a67cd-2a88-4c1d-bf76-5aac9b4b9e46" containerName="nova-api-log" probeResult="failure" output="Get \"http://10.217.0.183:8774/\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" Dec 10 23:10:45 crc kubenswrapper[4791]: I1210 23:10:45.974623 4791 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/nova-api-0" podUID="739a67cd-2a88-4c1d-bf76-5aac9b4b9e46" containerName="nova-api-api" probeResult="failure" output="Get \"http://10.217.0.183:8774/\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" Dec 10 23:10:45 crc kubenswrapper[4791]: I1210 23:10:45.997930 4791 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-5784cf869f-ngjqj" Dec 10 23:10:46 crc kubenswrapper[4791]: I1210 23:10:46.116454 4791 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/224a8e50-4270-42e8-a1df-268bc144b1d5-dns-svc\") pod \"224a8e50-4270-42e8-a1df-268bc144b1d5\" (UID: \"224a8e50-4270-42e8-a1df-268bc144b1d5\") " Dec 10 23:10:46 crc kubenswrapper[4791]: I1210 23:10:46.116550 4791 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/224a8e50-4270-42e8-a1df-268bc144b1d5-ovsdbserver-sb\") pod \"224a8e50-4270-42e8-a1df-268bc144b1d5\" (UID: \"224a8e50-4270-42e8-a1df-268bc144b1d5\") " Dec 10 23:10:46 crc kubenswrapper[4791]: I1210 23:10:46.116633 4791 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-xfqw8\" (UniqueName: \"kubernetes.io/projected/224a8e50-4270-42e8-a1df-268bc144b1d5-kube-api-access-xfqw8\") pod \"224a8e50-4270-42e8-a1df-268bc144b1d5\" (UID: \"224a8e50-4270-42e8-a1df-268bc144b1d5\") " Dec 10 23:10:46 crc kubenswrapper[4791]: I1210 23:10:46.116683 4791 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/224a8e50-4270-42e8-a1df-268bc144b1d5-dns-swift-storage-0\") pod \"224a8e50-4270-42e8-a1df-268bc144b1d5\" (UID: \"224a8e50-4270-42e8-a1df-268bc144b1d5\") " Dec 10 23:10:46 crc kubenswrapper[4791]: I1210 23:10:46.116790 4791 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/224a8e50-4270-42e8-a1df-268bc144b1d5-config\") pod \"224a8e50-4270-42e8-a1df-268bc144b1d5\" (UID: \"224a8e50-4270-42e8-a1df-268bc144b1d5\") " Dec 10 23:10:46 crc kubenswrapper[4791]: I1210 23:10:46.116916 4791 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/224a8e50-4270-42e8-a1df-268bc144b1d5-ovsdbserver-nb\") pod \"224a8e50-4270-42e8-a1df-268bc144b1d5\" (UID: \"224a8e50-4270-42e8-a1df-268bc144b1d5\") " Dec 10 23:10:46 crc kubenswrapper[4791]: I1210 23:10:46.124425 4791 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/224a8e50-4270-42e8-a1df-268bc144b1d5-kube-api-access-xfqw8" (OuterVolumeSpecName: "kube-api-access-xfqw8") pod "224a8e50-4270-42e8-a1df-268bc144b1d5" (UID: "224a8e50-4270-42e8-a1df-268bc144b1d5"). InnerVolumeSpecName "kube-api-access-xfqw8". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 10 23:10:46 crc kubenswrapper[4791]: I1210 23:10:46.174594 4791 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/224a8e50-4270-42e8-a1df-268bc144b1d5-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "224a8e50-4270-42e8-a1df-268bc144b1d5" (UID: "224a8e50-4270-42e8-a1df-268bc144b1d5"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 10 23:10:46 crc kubenswrapper[4791]: I1210 23:10:46.189389 4791 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/224a8e50-4270-42e8-a1df-268bc144b1d5-ovsdbserver-sb" (OuterVolumeSpecName: "ovsdbserver-sb") pod "224a8e50-4270-42e8-a1df-268bc144b1d5" (UID: "224a8e50-4270-42e8-a1df-268bc144b1d5"). InnerVolumeSpecName "ovsdbserver-sb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 10 23:10:46 crc kubenswrapper[4791]: I1210 23:10:46.192080 4791 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/224a8e50-4270-42e8-a1df-268bc144b1d5-ovsdbserver-nb" (OuterVolumeSpecName: "ovsdbserver-nb") pod "224a8e50-4270-42e8-a1df-268bc144b1d5" (UID: "224a8e50-4270-42e8-a1df-268bc144b1d5"). InnerVolumeSpecName "ovsdbserver-nb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 10 23:10:46 crc kubenswrapper[4791]: I1210 23:10:46.193783 4791 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/224a8e50-4270-42e8-a1df-268bc144b1d5-config" (OuterVolumeSpecName: "config") pod "224a8e50-4270-42e8-a1df-268bc144b1d5" (UID: "224a8e50-4270-42e8-a1df-268bc144b1d5"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 10 23:10:46 crc kubenswrapper[4791]: I1210 23:10:46.217557 4791 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/224a8e50-4270-42e8-a1df-268bc144b1d5-dns-swift-storage-0" (OuterVolumeSpecName: "dns-swift-storage-0") pod "224a8e50-4270-42e8-a1df-268bc144b1d5" (UID: "224a8e50-4270-42e8-a1df-268bc144b1d5"). InnerVolumeSpecName "dns-swift-storage-0". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 10 23:10:46 crc kubenswrapper[4791]: I1210 23:10:46.219023 4791 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/224a8e50-4270-42e8-a1df-268bc144b1d5-ovsdbserver-nb\") on node \"crc\" DevicePath \"\"" Dec 10 23:10:46 crc kubenswrapper[4791]: I1210 23:10:46.219051 4791 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/224a8e50-4270-42e8-a1df-268bc144b1d5-dns-svc\") on node \"crc\" DevicePath \"\"" Dec 10 23:10:46 crc kubenswrapper[4791]: I1210 23:10:46.219061 4791 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/224a8e50-4270-42e8-a1df-268bc144b1d5-ovsdbserver-sb\") on node \"crc\" DevicePath \"\"" Dec 10 23:10:46 crc kubenswrapper[4791]: I1210 23:10:46.219073 4791 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-xfqw8\" (UniqueName: \"kubernetes.io/projected/224a8e50-4270-42e8-a1df-268bc144b1d5-kube-api-access-xfqw8\") on node \"crc\" DevicePath \"\"" Dec 10 23:10:46 crc kubenswrapper[4791]: I1210 23:10:46.219084 4791 reconciler_common.go:293] "Volume detached for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/224a8e50-4270-42e8-a1df-268bc144b1d5-dns-swift-storage-0\") on node \"crc\" DevicePath \"\"" Dec 10 23:10:46 crc kubenswrapper[4791]: I1210 23:10:46.219094 4791 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/224a8e50-4270-42e8-a1df-268bc144b1d5-config\") on node \"crc\" DevicePath \"\"" Dec 10 23:10:46 crc kubenswrapper[4791]: I1210 23:10:46.676630 4791 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-5784cf869f-ngjqj" event={"ID":"224a8e50-4270-42e8-a1df-268bc144b1d5","Type":"ContainerDied","Data":"7bacd3841f62b678498a49df871b9e42dccba73bee49d48be648aeb6a7d1abbb"} Dec 10 23:10:46 crc kubenswrapper[4791]: I1210 23:10:46.677035 4791 scope.go:117] "RemoveContainer" containerID="d0b7aa50e7609ff3f59cefa955b1e1e05f6d96ae4a3e4143c8e51f4065c45be1" Dec 10 23:10:46 crc kubenswrapper[4791]: I1210 23:10:46.677154 4791 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-5784cf869f-ngjqj" Dec 10 23:10:46 crc kubenswrapper[4791]: I1210 23:10:46.723170 4791 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-5784cf869f-ngjqj"] Dec 10 23:10:46 crc kubenswrapper[4791]: I1210 23:10:46.727400 4791 scope.go:117] "RemoveContainer" containerID="930aefa334c3a39fdcf6c92de149a6110c9b76856c19a726d39b0c5feff817ad" Dec 10 23:10:46 crc kubenswrapper[4791]: I1210 23:10:46.733012 4791 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-5784cf869f-ngjqj"] Dec 10 23:10:47 crc kubenswrapper[4791]: I1210 23:10:47.195946 4791 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-conductor-db-sync-brjck" Dec 10 23:10:47 crc kubenswrapper[4791]: I1210 23:10:47.204414 4791 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-cell-mapping-k64vq" Dec 10 23:10:47 crc kubenswrapper[4791]: I1210 23:10:47.347260 4791 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-8dfkk\" (UniqueName: \"kubernetes.io/projected/660af60d-b638-42d4-86d1-3a7a07b1b936-kube-api-access-8dfkk\") pod \"660af60d-b638-42d4-86d1-3a7a07b1b936\" (UID: \"660af60d-b638-42d4-86d1-3a7a07b1b936\") " Dec 10 23:10:47 crc kubenswrapper[4791]: I1210 23:10:47.347426 4791 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/da9f584f-c289-4d67-ae91-8d6bf7507fb6-scripts\") pod \"da9f584f-c289-4d67-ae91-8d6bf7507fb6\" (UID: \"da9f584f-c289-4d67-ae91-8d6bf7507fb6\") " Dec 10 23:10:47 crc kubenswrapper[4791]: I1210 23:10:47.347452 4791 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/da9f584f-c289-4d67-ae91-8d6bf7507fb6-combined-ca-bundle\") pod \"da9f584f-c289-4d67-ae91-8d6bf7507fb6\" (UID: \"da9f584f-c289-4d67-ae91-8d6bf7507fb6\") " Dec 10 23:10:47 crc kubenswrapper[4791]: I1210 23:10:47.347512 4791 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-xdpx4\" (UniqueName: \"kubernetes.io/projected/da9f584f-c289-4d67-ae91-8d6bf7507fb6-kube-api-access-xdpx4\") pod \"da9f584f-c289-4d67-ae91-8d6bf7507fb6\" (UID: \"da9f584f-c289-4d67-ae91-8d6bf7507fb6\") " Dec 10 23:10:47 crc kubenswrapper[4791]: I1210 23:10:47.347544 4791 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/660af60d-b638-42d4-86d1-3a7a07b1b936-scripts\") pod \"660af60d-b638-42d4-86d1-3a7a07b1b936\" (UID: \"660af60d-b638-42d4-86d1-3a7a07b1b936\") " Dec 10 23:10:47 crc kubenswrapper[4791]: I1210 23:10:47.347605 4791 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/660af60d-b638-42d4-86d1-3a7a07b1b936-config-data\") pod \"660af60d-b638-42d4-86d1-3a7a07b1b936\" (UID: \"660af60d-b638-42d4-86d1-3a7a07b1b936\") " Dec 10 23:10:47 crc kubenswrapper[4791]: I1210 23:10:47.347664 4791 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/da9f584f-c289-4d67-ae91-8d6bf7507fb6-config-data\") pod \"da9f584f-c289-4d67-ae91-8d6bf7507fb6\" (UID: \"da9f584f-c289-4d67-ae91-8d6bf7507fb6\") " Dec 10 23:10:47 crc kubenswrapper[4791]: I1210 23:10:47.347781 4791 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/660af60d-b638-42d4-86d1-3a7a07b1b936-combined-ca-bundle\") pod \"660af60d-b638-42d4-86d1-3a7a07b1b936\" (UID: \"660af60d-b638-42d4-86d1-3a7a07b1b936\") " Dec 10 23:10:47 crc kubenswrapper[4791]: I1210 23:10:47.354720 4791 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/660af60d-b638-42d4-86d1-3a7a07b1b936-scripts" (OuterVolumeSpecName: "scripts") pod "660af60d-b638-42d4-86d1-3a7a07b1b936" (UID: "660af60d-b638-42d4-86d1-3a7a07b1b936"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 10 23:10:47 crc kubenswrapper[4791]: I1210 23:10:47.360717 4791 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/da9f584f-c289-4d67-ae91-8d6bf7507fb6-kube-api-access-xdpx4" (OuterVolumeSpecName: "kube-api-access-xdpx4") pod "da9f584f-c289-4d67-ae91-8d6bf7507fb6" (UID: "da9f584f-c289-4d67-ae91-8d6bf7507fb6"). InnerVolumeSpecName "kube-api-access-xdpx4". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 10 23:10:47 crc kubenswrapper[4791]: I1210 23:10:47.360757 4791 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/da9f584f-c289-4d67-ae91-8d6bf7507fb6-scripts" (OuterVolumeSpecName: "scripts") pod "da9f584f-c289-4d67-ae91-8d6bf7507fb6" (UID: "da9f584f-c289-4d67-ae91-8d6bf7507fb6"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 10 23:10:47 crc kubenswrapper[4791]: I1210 23:10:47.360822 4791 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/660af60d-b638-42d4-86d1-3a7a07b1b936-kube-api-access-8dfkk" (OuterVolumeSpecName: "kube-api-access-8dfkk") pod "660af60d-b638-42d4-86d1-3a7a07b1b936" (UID: "660af60d-b638-42d4-86d1-3a7a07b1b936"). InnerVolumeSpecName "kube-api-access-8dfkk". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 10 23:10:47 crc kubenswrapper[4791]: I1210 23:10:47.378452 4791 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/660af60d-b638-42d4-86d1-3a7a07b1b936-config-data" (OuterVolumeSpecName: "config-data") pod "660af60d-b638-42d4-86d1-3a7a07b1b936" (UID: "660af60d-b638-42d4-86d1-3a7a07b1b936"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 10 23:10:47 crc kubenswrapper[4791]: I1210 23:10:47.384135 4791 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/da9f584f-c289-4d67-ae91-8d6bf7507fb6-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "da9f584f-c289-4d67-ae91-8d6bf7507fb6" (UID: "da9f584f-c289-4d67-ae91-8d6bf7507fb6"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 10 23:10:47 crc kubenswrapper[4791]: I1210 23:10:47.386554 4791 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/660af60d-b638-42d4-86d1-3a7a07b1b936-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "660af60d-b638-42d4-86d1-3a7a07b1b936" (UID: "660af60d-b638-42d4-86d1-3a7a07b1b936"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 10 23:10:47 crc kubenswrapper[4791]: I1210 23:10:47.401548 4791 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/da9f584f-c289-4d67-ae91-8d6bf7507fb6-config-data" (OuterVolumeSpecName: "config-data") pod "da9f584f-c289-4d67-ae91-8d6bf7507fb6" (UID: "da9f584f-c289-4d67-ae91-8d6bf7507fb6"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 10 23:10:47 crc kubenswrapper[4791]: I1210 23:10:47.453916 4791 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/da9f584f-c289-4d67-ae91-8d6bf7507fb6-config-data\") on node \"crc\" DevicePath \"\"" Dec 10 23:10:47 crc kubenswrapper[4791]: I1210 23:10:47.453955 4791 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/660af60d-b638-42d4-86d1-3a7a07b1b936-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 10 23:10:47 crc kubenswrapper[4791]: I1210 23:10:47.453966 4791 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-8dfkk\" (UniqueName: \"kubernetes.io/projected/660af60d-b638-42d4-86d1-3a7a07b1b936-kube-api-access-8dfkk\") on node \"crc\" DevicePath \"\"" Dec 10 23:10:47 crc kubenswrapper[4791]: I1210 23:10:47.453975 4791 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/da9f584f-c289-4d67-ae91-8d6bf7507fb6-scripts\") on node \"crc\" DevicePath \"\"" Dec 10 23:10:47 crc kubenswrapper[4791]: I1210 23:10:47.453983 4791 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/da9f584f-c289-4d67-ae91-8d6bf7507fb6-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 10 23:10:47 crc kubenswrapper[4791]: I1210 23:10:47.454009 4791 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-xdpx4\" (UniqueName: \"kubernetes.io/projected/da9f584f-c289-4d67-ae91-8d6bf7507fb6-kube-api-access-xdpx4\") on node \"crc\" DevicePath \"\"" Dec 10 23:10:47 crc kubenswrapper[4791]: I1210 23:10:47.454018 4791 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/660af60d-b638-42d4-86d1-3a7a07b1b936-scripts\") on node \"crc\" DevicePath \"\"" Dec 10 23:10:47 crc kubenswrapper[4791]: I1210 23:10:47.454026 4791 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/660af60d-b638-42d4-86d1-3a7a07b1b936-config-data\") on node \"crc\" DevicePath \"\"" Dec 10 23:10:47 crc kubenswrapper[4791]: I1210 23:10:47.688076 4791 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-cell-mapping-k64vq" Dec 10 23:10:47 crc kubenswrapper[4791]: I1210 23:10:47.688293 4791 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-cell-mapping-k64vq" event={"ID":"da9f584f-c289-4d67-ae91-8d6bf7507fb6","Type":"ContainerDied","Data":"3c7542d32480373b1cee85c021b42ffe5b2cbb8f402c8084948eeb72f19c8700"} Dec 10 23:10:47 crc kubenswrapper[4791]: I1210 23:10:47.689278 4791 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="3c7542d32480373b1cee85c021b42ffe5b2cbb8f402c8084948eeb72f19c8700" Dec 10 23:10:47 crc kubenswrapper[4791]: I1210 23:10:47.695182 4791 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-conductor-db-sync-brjck" Dec 10 23:10:47 crc kubenswrapper[4791]: I1210 23:10:47.702405 4791 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-conductor-db-sync-brjck" event={"ID":"660af60d-b638-42d4-86d1-3a7a07b1b936","Type":"ContainerDied","Data":"ad353c4652105c4569ebe7e1bd36575c22b7dcc1500c78ea7bfeb8e73d12546f"} Dec 10 23:10:47 crc kubenswrapper[4791]: I1210 23:10:47.702496 4791 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="ad353c4652105c4569ebe7e1bd36575c22b7dcc1500c78ea7bfeb8e73d12546f" Dec 10 23:10:47 crc kubenswrapper[4791]: I1210 23:10:47.777831 4791 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-cell1-conductor-0"] Dec 10 23:10:47 crc kubenswrapper[4791]: E1210 23:10:47.778310 4791 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="224a8e50-4270-42e8-a1df-268bc144b1d5" containerName="dnsmasq-dns" Dec 10 23:10:47 crc kubenswrapper[4791]: I1210 23:10:47.778328 4791 state_mem.go:107] "Deleted CPUSet assignment" podUID="224a8e50-4270-42e8-a1df-268bc144b1d5" containerName="dnsmasq-dns" Dec 10 23:10:47 crc kubenswrapper[4791]: E1210 23:10:47.778360 4791 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="660af60d-b638-42d4-86d1-3a7a07b1b936" containerName="nova-cell1-conductor-db-sync" Dec 10 23:10:47 crc kubenswrapper[4791]: I1210 23:10:47.778370 4791 state_mem.go:107] "Deleted CPUSet assignment" podUID="660af60d-b638-42d4-86d1-3a7a07b1b936" containerName="nova-cell1-conductor-db-sync" Dec 10 23:10:47 crc kubenswrapper[4791]: E1210 23:10:47.778400 4791 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="224a8e50-4270-42e8-a1df-268bc144b1d5" containerName="init" Dec 10 23:10:47 crc kubenswrapper[4791]: I1210 23:10:47.778406 4791 state_mem.go:107] "Deleted CPUSet assignment" podUID="224a8e50-4270-42e8-a1df-268bc144b1d5" containerName="init" Dec 10 23:10:47 crc kubenswrapper[4791]: E1210 23:10:47.778415 4791 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="da9f584f-c289-4d67-ae91-8d6bf7507fb6" containerName="nova-manage" Dec 10 23:10:47 crc kubenswrapper[4791]: I1210 23:10:47.778421 4791 state_mem.go:107] "Deleted CPUSet assignment" podUID="da9f584f-c289-4d67-ae91-8d6bf7507fb6" containerName="nova-manage" Dec 10 23:10:47 crc kubenswrapper[4791]: I1210 23:10:47.778653 4791 memory_manager.go:354] "RemoveStaleState removing state" podUID="da9f584f-c289-4d67-ae91-8d6bf7507fb6" containerName="nova-manage" Dec 10 23:10:47 crc kubenswrapper[4791]: I1210 23:10:47.778671 4791 memory_manager.go:354] "RemoveStaleState removing state" podUID="224a8e50-4270-42e8-a1df-268bc144b1d5" containerName="dnsmasq-dns" Dec 10 23:10:47 crc kubenswrapper[4791]: I1210 23:10:47.778683 4791 memory_manager.go:354] "RemoveStaleState removing state" podUID="660af60d-b638-42d4-86d1-3a7a07b1b936" containerName="nova-cell1-conductor-db-sync" Dec 10 23:10:47 crc kubenswrapper[4791]: I1210 23:10:47.779395 4791 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-conductor-0" Dec 10 23:10:47 crc kubenswrapper[4791]: I1210 23:10:47.784729 4791 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell1-conductor-config-data" Dec 10 23:10:47 crc kubenswrapper[4791]: I1210 23:10:47.787330 4791 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-conductor-0"] Dec 10 23:10:47 crc kubenswrapper[4791]: I1210 23:10:47.861912 4791 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/2dff7039-7fed-41ac-8703-8e6023146318-combined-ca-bundle\") pod \"nova-cell1-conductor-0\" (UID: \"2dff7039-7fed-41ac-8703-8e6023146318\") " pod="openstack/nova-cell1-conductor-0" Dec 10 23:10:47 crc kubenswrapper[4791]: I1210 23:10:47.861993 4791 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-6wqr7\" (UniqueName: \"kubernetes.io/projected/2dff7039-7fed-41ac-8703-8e6023146318-kube-api-access-6wqr7\") pod \"nova-cell1-conductor-0\" (UID: \"2dff7039-7fed-41ac-8703-8e6023146318\") " pod="openstack/nova-cell1-conductor-0" Dec 10 23:10:47 crc kubenswrapper[4791]: I1210 23:10:47.862027 4791 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/2dff7039-7fed-41ac-8703-8e6023146318-config-data\") pod \"nova-cell1-conductor-0\" (UID: \"2dff7039-7fed-41ac-8703-8e6023146318\") " pod="openstack/nova-cell1-conductor-0" Dec 10 23:10:47 crc kubenswrapper[4791]: I1210 23:10:47.934529 4791 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="224a8e50-4270-42e8-a1df-268bc144b1d5" path="/var/lib/kubelet/pods/224a8e50-4270-42e8-a1df-268bc144b1d5/volumes" Dec 10 23:10:47 crc kubenswrapper[4791]: I1210 23:10:47.935547 4791 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-metadata-0" Dec 10 23:10:47 crc kubenswrapper[4791]: I1210 23:10:47.935588 4791 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-metadata-0" Dec 10 23:10:47 crc kubenswrapper[4791]: I1210 23:10:47.954294 4791 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-api-0"] Dec 10 23:10:47 crc kubenswrapper[4791]: I1210 23:10:47.954601 4791 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-api-0" podUID="739a67cd-2a88-4c1d-bf76-5aac9b4b9e46" containerName="nova-api-log" containerID="cri-o://7cdac15bae335278abc24d833001eafda5d44d0849475b593b93f231aff6eb2e" gracePeriod=30 Dec 10 23:10:47 crc kubenswrapper[4791]: I1210 23:10:47.955033 4791 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-api-0" podUID="739a67cd-2a88-4c1d-bf76-5aac9b4b9e46" containerName="nova-api-api" containerID="cri-o://a8354ff62f46c0a1cbbcd0e6eb42b520e0b04bcc5960c67755b39312a1f45c3c" gracePeriod=30 Dec 10 23:10:47 crc kubenswrapper[4791]: I1210 23:10:47.963317 4791 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/2dff7039-7fed-41ac-8703-8e6023146318-combined-ca-bundle\") pod \"nova-cell1-conductor-0\" (UID: \"2dff7039-7fed-41ac-8703-8e6023146318\") " pod="openstack/nova-cell1-conductor-0" Dec 10 23:10:47 crc kubenswrapper[4791]: I1210 23:10:47.963398 4791 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-6wqr7\" (UniqueName: \"kubernetes.io/projected/2dff7039-7fed-41ac-8703-8e6023146318-kube-api-access-6wqr7\") pod \"nova-cell1-conductor-0\" (UID: \"2dff7039-7fed-41ac-8703-8e6023146318\") " pod="openstack/nova-cell1-conductor-0" Dec 10 23:10:47 crc kubenswrapper[4791]: I1210 23:10:47.963417 4791 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/2dff7039-7fed-41ac-8703-8e6023146318-config-data\") pod \"nova-cell1-conductor-0\" (UID: \"2dff7039-7fed-41ac-8703-8e6023146318\") " pod="openstack/nova-cell1-conductor-0" Dec 10 23:10:47 crc kubenswrapper[4791]: I1210 23:10:47.969777 4791 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/2dff7039-7fed-41ac-8703-8e6023146318-combined-ca-bundle\") pod \"nova-cell1-conductor-0\" (UID: \"2dff7039-7fed-41ac-8703-8e6023146318\") " pod="openstack/nova-cell1-conductor-0" Dec 10 23:10:47 crc kubenswrapper[4791]: I1210 23:10:47.970157 4791 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/2dff7039-7fed-41ac-8703-8e6023146318-config-data\") pod \"nova-cell1-conductor-0\" (UID: \"2dff7039-7fed-41ac-8703-8e6023146318\") " pod="openstack/nova-cell1-conductor-0" Dec 10 23:10:47 crc kubenswrapper[4791]: I1210 23:10:47.979992 4791 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-scheduler-0"] Dec 10 23:10:47 crc kubenswrapper[4791]: I1210 23:10:47.980247 4791 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-scheduler-0" podUID="83b6abc9-aa7f-4ddb-8b9c-b38fcb5a9145" containerName="nova-scheduler-scheduler" containerID="cri-o://4eb9f1cc7aae957c4598d0c9a8860aedc87158afb8efbea24833107d43e12aea" gracePeriod=30 Dec 10 23:10:47 crc kubenswrapper[4791]: I1210 23:10:47.993613 4791 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-metadata-0"] Dec 10 23:10:47 crc kubenswrapper[4791]: I1210 23:10:47.998680 4791 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-6wqr7\" (UniqueName: \"kubernetes.io/projected/2dff7039-7fed-41ac-8703-8e6023146318-kube-api-access-6wqr7\") pod \"nova-cell1-conductor-0\" (UID: \"2dff7039-7fed-41ac-8703-8e6023146318\") " pod="openstack/nova-cell1-conductor-0" Dec 10 23:10:48 crc kubenswrapper[4791]: I1210 23:10:48.102756 4791 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-conductor-0" Dec 10 23:10:48 crc kubenswrapper[4791]: I1210 23:10:48.601678 4791 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-conductor-0"] Dec 10 23:10:48 crc kubenswrapper[4791]: I1210 23:10:48.720221 4791 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-conductor-0" event={"ID":"2dff7039-7fed-41ac-8703-8e6023146318","Type":"ContainerStarted","Data":"5eb6edf2623c7cb04e50ddc1760a43e08d60a8848da21f69fc1062818ec44e60"} Dec 10 23:10:48 crc kubenswrapper[4791]: I1210 23:10:48.735524 4791 generic.go:334] "Generic (PLEG): container finished" podID="83b6abc9-aa7f-4ddb-8b9c-b38fcb5a9145" containerID="4eb9f1cc7aae957c4598d0c9a8860aedc87158afb8efbea24833107d43e12aea" exitCode=0 Dec 10 23:10:48 crc kubenswrapper[4791]: I1210 23:10:48.735712 4791 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-scheduler-0" event={"ID":"83b6abc9-aa7f-4ddb-8b9c-b38fcb5a9145","Type":"ContainerDied","Data":"4eb9f1cc7aae957c4598d0c9a8860aedc87158afb8efbea24833107d43e12aea"} Dec 10 23:10:48 crc kubenswrapper[4791]: I1210 23:10:48.740323 4791 generic.go:334] "Generic (PLEG): container finished" podID="739a67cd-2a88-4c1d-bf76-5aac9b4b9e46" containerID="7cdac15bae335278abc24d833001eafda5d44d0849475b593b93f231aff6eb2e" exitCode=143 Dec 10 23:10:48 crc kubenswrapper[4791]: I1210 23:10:48.740508 4791 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"739a67cd-2a88-4c1d-bf76-5aac9b4b9e46","Type":"ContainerDied","Data":"7cdac15bae335278abc24d833001eafda5d44d0849475b593b93f231aff6eb2e"} Dec 10 23:10:48 crc kubenswrapper[4791]: I1210 23:10:48.740709 4791 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-metadata-0" podUID="5d7b44d4-1936-4588-af50-442e04f245fc" containerName="nova-metadata-log" containerID="cri-o://452e27ffea76e2ade3d7bf72c2abc8c40cca428c422ffe73d9e7aa831c8647d1" gracePeriod=30 Dec 10 23:10:48 crc kubenswrapper[4791]: I1210 23:10:48.740799 4791 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-metadata-0" podUID="5d7b44d4-1936-4588-af50-442e04f245fc" containerName="nova-metadata-metadata" containerID="cri-o://d5f65d5be988d066eebaad7dc5f9b030442a2ed929476f704d133fc049905317" gracePeriod=30 Dec 10 23:10:48 crc kubenswrapper[4791]: I1210 23:10:48.887431 4791 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-scheduler-0" Dec 10 23:10:48 crc kubenswrapper[4791]: I1210 23:10:48.986597 4791 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-844r7\" (UniqueName: \"kubernetes.io/projected/83b6abc9-aa7f-4ddb-8b9c-b38fcb5a9145-kube-api-access-844r7\") pod \"83b6abc9-aa7f-4ddb-8b9c-b38fcb5a9145\" (UID: \"83b6abc9-aa7f-4ddb-8b9c-b38fcb5a9145\") " Dec 10 23:10:48 crc kubenswrapper[4791]: I1210 23:10:48.986698 4791 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/83b6abc9-aa7f-4ddb-8b9c-b38fcb5a9145-combined-ca-bundle\") pod \"83b6abc9-aa7f-4ddb-8b9c-b38fcb5a9145\" (UID: \"83b6abc9-aa7f-4ddb-8b9c-b38fcb5a9145\") " Dec 10 23:10:48 crc kubenswrapper[4791]: I1210 23:10:48.986782 4791 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/83b6abc9-aa7f-4ddb-8b9c-b38fcb5a9145-config-data\") pod \"83b6abc9-aa7f-4ddb-8b9c-b38fcb5a9145\" (UID: \"83b6abc9-aa7f-4ddb-8b9c-b38fcb5a9145\") " Dec 10 23:10:48 crc kubenswrapper[4791]: I1210 23:10:48.995017 4791 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/83b6abc9-aa7f-4ddb-8b9c-b38fcb5a9145-kube-api-access-844r7" (OuterVolumeSpecName: "kube-api-access-844r7") pod "83b6abc9-aa7f-4ddb-8b9c-b38fcb5a9145" (UID: "83b6abc9-aa7f-4ddb-8b9c-b38fcb5a9145"). InnerVolumeSpecName "kube-api-access-844r7". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 10 23:10:49 crc kubenswrapper[4791]: I1210 23:10:49.021871 4791 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/83b6abc9-aa7f-4ddb-8b9c-b38fcb5a9145-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "83b6abc9-aa7f-4ddb-8b9c-b38fcb5a9145" (UID: "83b6abc9-aa7f-4ddb-8b9c-b38fcb5a9145"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 10 23:10:49 crc kubenswrapper[4791]: I1210 23:10:49.023558 4791 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/83b6abc9-aa7f-4ddb-8b9c-b38fcb5a9145-config-data" (OuterVolumeSpecName: "config-data") pod "83b6abc9-aa7f-4ddb-8b9c-b38fcb5a9145" (UID: "83b6abc9-aa7f-4ddb-8b9c-b38fcb5a9145"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 10 23:10:49 crc kubenswrapper[4791]: I1210 23:10:49.089097 4791 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/83b6abc9-aa7f-4ddb-8b9c-b38fcb5a9145-config-data\") on node \"crc\" DevicePath \"\"" Dec 10 23:10:49 crc kubenswrapper[4791]: I1210 23:10:49.089144 4791 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-844r7\" (UniqueName: \"kubernetes.io/projected/83b6abc9-aa7f-4ddb-8b9c-b38fcb5a9145-kube-api-access-844r7\") on node \"crc\" DevicePath \"\"" Dec 10 23:10:49 crc kubenswrapper[4791]: I1210 23:10:49.089157 4791 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/83b6abc9-aa7f-4ddb-8b9c-b38fcb5a9145-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 10 23:10:49 crc kubenswrapper[4791]: I1210 23:10:49.211452 4791 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-metadata-0" Dec 10 23:10:49 crc kubenswrapper[4791]: I1210 23:10:49.292101 4791 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/5d7b44d4-1936-4588-af50-442e04f245fc-config-data\") pod \"5d7b44d4-1936-4588-af50-442e04f245fc\" (UID: \"5d7b44d4-1936-4588-af50-442e04f245fc\") " Dec 10 23:10:49 crc kubenswrapper[4791]: I1210 23:10:49.292154 4791 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/5d7b44d4-1936-4588-af50-442e04f245fc-logs\") pod \"5d7b44d4-1936-4588-af50-442e04f245fc\" (UID: \"5d7b44d4-1936-4588-af50-442e04f245fc\") " Dec 10 23:10:49 crc kubenswrapper[4791]: I1210 23:10:49.292174 4791 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"nova-metadata-tls-certs\" (UniqueName: \"kubernetes.io/secret/5d7b44d4-1936-4588-af50-442e04f245fc-nova-metadata-tls-certs\") pod \"5d7b44d4-1936-4588-af50-442e04f245fc\" (UID: \"5d7b44d4-1936-4588-af50-442e04f245fc\") " Dec 10 23:10:49 crc kubenswrapper[4791]: I1210 23:10:49.292313 4791 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-w58mp\" (UniqueName: \"kubernetes.io/projected/5d7b44d4-1936-4588-af50-442e04f245fc-kube-api-access-w58mp\") pod \"5d7b44d4-1936-4588-af50-442e04f245fc\" (UID: \"5d7b44d4-1936-4588-af50-442e04f245fc\") " Dec 10 23:10:49 crc kubenswrapper[4791]: I1210 23:10:49.292774 4791 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/5d7b44d4-1936-4588-af50-442e04f245fc-logs" (OuterVolumeSpecName: "logs") pod "5d7b44d4-1936-4588-af50-442e04f245fc" (UID: "5d7b44d4-1936-4588-af50-442e04f245fc"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 10 23:10:49 crc kubenswrapper[4791]: I1210 23:10:49.292968 4791 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/5d7b44d4-1936-4588-af50-442e04f245fc-combined-ca-bundle\") pod \"5d7b44d4-1936-4588-af50-442e04f245fc\" (UID: \"5d7b44d4-1936-4588-af50-442e04f245fc\") " Dec 10 23:10:49 crc kubenswrapper[4791]: I1210 23:10:49.293456 4791 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/5d7b44d4-1936-4588-af50-442e04f245fc-logs\") on node \"crc\" DevicePath \"\"" Dec 10 23:10:49 crc kubenswrapper[4791]: I1210 23:10:49.299537 4791 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/5d7b44d4-1936-4588-af50-442e04f245fc-kube-api-access-w58mp" (OuterVolumeSpecName: "kube-api-access-w58mp") pod "5d7b44d4-1936-4588-af50-442e04f245fc" (UID: "5d7b44d4-1936-4588-af50-442e04f245fc"). InnerVolumeSpecName "kube-api-access-w58mp". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 10 23:10:49 crc kubenswrapper[4791]: I1210 23:10:49.337493 4791 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/5d7b44d4-1936-4588-af50-442e04f245fc-config-data" (OuterVolumeSpecName: "config-data") pod "5d7b44d4-1936-4588-af50-442e04f245fc" (UID: "5d7b44d4-1936-4588-af50-442e04f245fc"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 10 23:10:49 crc kubenswrapper[4791]: I1210 23:10:49.357540 4791 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/5d7b44d4-1936-4588-af50-442e04f245fc-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "5d7b44d4-1936-4588-af50-442e04f245fc" (UID: "5d7b44d4-1936-4588-af50-442e04f245fc"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 10 23:10:49 crc kubenswrapper[4791]: I1210 23:10:49.400551 4791 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/5d7b44d4-1936-4588-af50-442e04f245fc-nova-metadata-tls-certs" (OuterVolumeSpecName: "nova-metadata-tls-certs") pod "5d7b44d4-1936-4588-af50-442e04f245fc" (UID: "5d7b44d4-1936-4588-af50-442e04f245fc"). InnerVolumeSpecName "nova-metadata-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 10 23:10:49 crc kubenswrapper[4791]: I1210 23:10:49.401544 4791 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/5d7b44d4-1936-4588-af50-442e04f245fc-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 10 23:10:49 crc kubenswrapper[4791]: I1210 23:10:49.401581 4791 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/5d7b44d4-1936-4588-af50-442e04f245fc-config-data\") on node \"crc\" DevicePath \"\"" Dec 10 23:10:49 crc kubenswrapper[4791]: I1210 23:10:49.401591 4791 reconciler_common.go:293] "Volume detached for volume \"nova-metadata-tls-certs\" (UniqueName: \"kubernetes.io/secret/5d7b44d4-1936-4588-af50-442e04f245fc-nova-metadata-tls-certs\") on node \"crc\" DevicePath \"\"" Dec 10 23:10:49 crc kubenswrapper[4791]: I1210 23:10:49.401603 4791 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-w58mp\" (UniqueName: \"kubernetes.io/projected/5d7b44d4-1936-4588-af50-442e04f245fc-kube-api-access-w58mp\") on node \"crc\" DevicePath \"\"" Dec 10 23:10:49 crc kubenswrapper[4791]: I1210 23:10:49.758148 4791 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-conductor-0" event={"ID":"2dff7039-7fed-41ac-8703-8e6023146318","Type":"ContainerStarted","Data":"556a1b610e47e58b8fb748c5e56c72d5c097cad46c2b3b75227a56412cd981a7"} Dec 10 23:10:49 crc kubenswrapper[4791]: I1210 23:10:49.760568 4791 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-cell1-conductor-0" Dec 10 23:10:49 crc kubenswrapper[4791]: I1210 23:10:49.764608 4791 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-scheduler-0" Dec 10 23:10:49 crc kubenswrapper[4791]: I1210 23:10:49.764656 4791 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-scheduler-0" event={"ID":"83b6abc9-aa7f-4ddb-8b9c-b38fcb5a9145","Type":"ContainerDied","Data":"03be616aec5c91eea6fa4690564c719ddb65089216635dcfb713674db11da81f"} Dec 10 23:10:49 crc kubenswrapper[4791]: I1210 23:10:49.764695 4791 scope.go:117] "RemoveContainer" containerID="4eb9f1cc7aae957c4598d0c9a8860aedc87158afb8efbea24833107d43e12aea" Dec 10 23:10:49 crc kubenswrapper[4791]: I1210 23:10:49.770919 4791 generic.go:334] "Generic (PLEG): container finished" podID="5d7b44d4-1936-4588-af50-442e04f245fc" containerID="d5f65d5be988d066eebaad7dc5f9b030442a2ed929476f704d133fc049905317" exitCode=0 Dec 10 23:10:49 crc kubenswrapper[4791]: I1210 23:10:49.770950 4791 generic.go:334] "Generic (PLEG): container finished" podID="5d7b44d4-1936-4588-af50-442e04f245fc" containerID="452e27ffea76e2ade3d7bf72c2abc8c40cca428c422ffe73d9e7aa831c8647d1" exitCode=143 Dec 10 23:10:49 crc kubenswrapper[4791]: I1210 23:10:49.770970 4791 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"5d7b44d4-1936-4588-af50-442e04f245fc","Type":"ContainerDied","Data":"d5f65d5be988d066eebaad7dc5f9b030442a2ed929476f704d133fc049905317"} Dec 10 23:10:49 crc kubenswrapper[4791]: I1210 23:10:49.770990 4791 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"5d7b44d4-1936-4588-af50-442e04f245fc","Type":"ContainerDied","Data":"452e27ffea76e2ade3d7bf72c2abc8c40cca428c422ffe73d9e7aa831c8647d1"} Dec 10 23:10:49 crc kubenswrapper[4791]: I1210 23:10:49.771000 4791 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"5d7b44d4-1936-4588-af50-442e04f245fc","Type":"ContainerDied","Data":"6f03f3d2cd8adb3c9c0e887946e97b5decff87ff3451ba631aac7b33c8870dff"} Dec 10 23:10:49 crc kubenswrapper[4791]: I1210 23:10:49.771077 4791 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-metadata-0" Dec 10 23:10:49 crc kubenswrapper[4791]: I1210 23:10:49.796658 4791 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-cell1-conductor-0" podStartSLOduration=2.796630008 podStartE2EDuration="2.796630008s" podCreationTimestamp="2025-12-10 23:10:47 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-10 23:10:49.790774209 +0000 UTC m=+1284.220391822" watchObservedRunningTime="2025-12-10 23:10:49.796630008 +0000 UTC m=+1284.226247621" Dec 10 23:10:49 crc kubenswrapper[4791]: I1210 23:10:49.830626 4791 scope.go:117] "RemoveContainer" containerID="d5f65d5be988d066eebaad7dc5f9b030442a2ed929476f704d133fc049905317" Dec 10 23:10:49 crc kubenswrapper[4791]: I1210 23:10:49.833220 4791 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-metadata-0"] Dec 10 23:10:49 crc kubenswrapper[4791]: I1210 23:10:49.844697 4791 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-metadata-0"] Dec 10 23:10:49 crc kubenswrapper[4791]: I1210 23:10:49.867162 4791 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-scheduler-0"] Dec 10 23:10:49 crc kubenswrapper[4791]: I1210 23:10:49.896461 4791 scope.go:117] "RemoveContainer" containerID="452e27ffea76e2ade3d7bf72c2abc8c40cca428c422ffe73d9e7aa831c8647d1" Dec 10 23:10:49 crc kubenswrapper[4791]: I1210 23:10:49.902543 4791 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="5d7b44d4-1936-4588-af50-442e04f245fc" path="/var/lib/kubelet/pods/5d7b44d4-1936-4588-af50-442e04f245fc/volumes" Dec 10 23:10:49 crc kubenswrapper[4791]: I1210 23:10:49.903314 4791 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-scheduler-0"] Dec 10 23:10:49 crc kubenswrapper[4791]: I1210 23:10:49.905782 4791 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-metadata-0"] Dec 10 23:10:49 crc kubenswrapper[4791]: E1210 23:10:49.906299 4791 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="5d7b44d4-1936-4588-af50-442e04f245fc" containerName="nova-metadata-log" Dec 10 23:10:49 crc kubenswrapper[4791]: I1210 23:10:49.906318 4791 state_mem.go:107] "Deleted CPUSet assignment" podUID="5d7b44d4-1936-4588-af50-442e04f245fc" containerName="nova-metadata-log" Dec 10 23:10:49 crc kubenswrapper[4791]: E1210 23:10:49.906330 4791 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="83b6abc9-aa7f-4ddb-8b9c-b38fcb5a9145" containerName="nova-scheduler-scheduler" Dec 10 23:10:49 crc kubenswrapper[4791]: I1210 23:10:49.906362 4791 state_mem.go:107] "Deleted CPUSet assignment" podUID="83b6abc9-aa7f-4ddb-8b9c-b38fcb5a9145" containerName="nova-scheduler-scheduler" Dec 10 23:10:49 crc kubenswrapper[4791]: E1210 23:10:49.906380 4791 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="5d7b44d4-1936-4588-af50-442e04f245fc" containerName="nova-metadata-metadata" Dec 10 23:10:49 crc kubenswrapper[4791]: I1210 23:10:49.906386 4791 state_mem.go:107] "Deleted CPUSet assignment" podUID="5d7b44d4-1936-4588-af50-442e04f245fc" containerName="nova-metadata-metadata" Dec 10 23:10:49 crc kubenswrapper[4791]: I1210 23:10:49.906579 4791 memory_manager.go:354] "RemoveStaleState removing state" podUID="83b6abc9-aa7f-4ddb-8b9c-b38fcb5a9145" containerName="nova-scheduler-scheduler" Dec 10 23:10:49 crc kubenswrapper[4791]: I1210 23:10:49.906609 4791 memory_manager.go:354] "RemoveStaleState removing state" podUID="5d7b44d4-1936-4588-af50-442e04f245fc" containerName="nova-metadata-metadata" Dec 10 23:10:49 crc kubenswrapper[4791]: I1210 23:10:49.906623 4791 memory_manager.go:354] "RemoveStaleState removing state" podUID="5d7b44d4-1936-4588-af50-442e04f245fc" containerName="nova-metadata-log" Dec 10 23:10:49 crc kubenswrapper[4791]: I1210 23:10:49.907664 4791 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-metadata-0" Dec 10 23:10:49 crc kubenswrapper[4791]: I1210 23:10:49.911715 4791 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-nova-metadata-internal-svc" Dec 10 23:10:49 crc kubenswrapper[4791]: I1210 23:10:49.916242 4791 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-metadata-config-data" Dec 10 23:10:49 crc kubenswrapper[4791]: I1210 23:10:49.925415 4791 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-metadata-0"] Dec 10 23:10:49 crc kubenswrapper[4791]: I1210 23:10:49.937968 4791 scope.go:117] "RemoveContainer" containerID="d5f65d5be988d066eebaad7dc5f9b030442a2ed929476f704d133fc049905317" Dec 10 23:10:49 crc kubenswrapper[4791]: E1210 23:10:49.938435 4791 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"d5f65d5be988d066eebaad7dc5f9b030442a2ed929476f704d133fc049905317\": container with ID starting with d5f65d5be988d066eebaad7dc5f9b030442a2ed929476f704d133fc049905317 not found: ID does not exist" containerID="d5f65d5be988d066eebaad7dc5f9b030442a2ed929476f704d133fc049905317" Dec 10 23:10:49 crc kubenswrapper[4791]: I1210 23:10:49.938504 4791 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"d5f65d5be988d066eebaad7dc5f9b030442a2ed929476f704d133fc049905317"} err="failed to get container status \"d5f65d5be988d066eebaad7dc5f9b030442a2ed929476f704d133fc049905317\": rpc error: code = NotFound desc = could not find container \"d5f65d5be988d066eebaad7dc5f9b030442a2ed929476f704d133fc049905317\": container with ID starting with d5f65d5be988d066eebaad7dc5f9b030442a2ed929476f704d133fc049905317 not found: ID does not exist" Dec 10 23:10:49 crc kubenswrapper[4791]: I1210 23:10:49.938525 4791 scope.go:117] "RemoveContainer" containerID="452e27ffea76e2ade3d7bf72c2abc8c40cca428c422ffe73d9e7aa831c8647d1" Dec 10 23:10:49 crc kubenswrapper[4791]: E1210 23:10:49.938925 4791 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"452e27ffea76e2ade3d7bf72c2abc8c40cca428c422ffe73d9e7aa831c8647d1\": container with ID starting with 452e27ffea76e2ade3d7bf72c2abc8c40cca428c422ffe73d9e7aa831c8647d1 not found: ID does not exist" containerID="452e27ffea76e2ade3d7bf72c2abc8c40cca428c422ffe73d9e7aa831c8647d1" Dec 10 23:10:49 crc kubenswrapper[4791]: I1210 23:10:49.938948 4791 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"452e27ffea76e2ade3d7bf72c2abc8c40cca428c422ffe73d9e7aa831c8647d1"} err="failed to get container status \"452e27ffea76e2ade3d7bf72c2abc8c40cca428c422ffe73d9e7aa831c8647d1\": rpc error: code = NotFound desc = could not find container \"452e27ffea76e2ade3d7bf72c2abc8c40cca428c422ffe73d9e7aa831c8647d1\": container with ID starting with 452e27ffea76e2ade3d7bf72c2abc8c40cca428c422ffe73d9e7aa831c8647d1 not found: ID does not exist" Dec 10 23:10:49 crc kubenswrapper[4791]: I1210 23:10:49.938963 4791 scope.go:117] "RemoveContainer" containerID="d5f65d5be988d066eebaad7dc5f9b030442a2ed929476f704d133fc049905317" Dec 10 23:10:49 crc kubenswrapper[4791]: I1210 23:10:49.939312 4791 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"d5f65d5be988d066eebaad7dc5f9b030442a2ed929476f704d133fc049905317"} err="failed to get container status \"d5f65d5be988d066eebaad7dc5f9b030442a2ed929476f704d133fc049905317\": rpc error: code = NotFound desc = could not find container \"d5f65d5be988d066eebaad7dc5f9b030442a2ed929476f704d133fc049905317\": container with ID starting with d5f65d5be988d066eebaad7dc5f9b030442a2ed929476f704d133fc049905317 not found: ID does not exist" Dec 10 23:10:49 crc kubenswrapper[4791]: I1210 23:10:49.939329 4791 scope.go:117] "RemoveContainer" containerID="452e27ffea76e2ade3d7bf72c2abc8c40cca428c422ffe73d9e7aa831c8647d1" Dec 10 23:10:49 crc kubenswrapper[4791]: I1210 23:10:49.939571 4791 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"452e27ffea76e2ade3d7bf72c2abc8c40cca428c422ffe73d9e7aa831c8647d1"} err="failed to get container status \"452e27ffea76e2ade3d7bf72c2abc8c40cca428c422ffe73d9e7aa831c8647d1\": rpc error: code = NotFound desc = could not find container \"452e27ffea76e2ade3d7bf72c2abc8c40cca428c422ffe73d9e7aa831c8647d1\": container with ID starting with 452e27ffea76e2ade3d7bf72c2abc8c40cca428c422ffe73d9e7aa831c8647d1 not found: ID does not exist" Dec 10 23:10:49 crc kubenswrapper[4791]: I1210 23:10:49.944460 4791 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-scheduler-0"] Dec 10 23:10:49 crc kubenswrapper[4791]: I1210 23:10:49.945754 4791 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-scheduler-0" Dec 10 23:10:49 crc kubenswrapper[4791]: I1210 23:10:49.948435 4791 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-scheduler-config-data" Dec 10 23:10:49 crc kubenswrapper[4791]: I1210 23:10:49.966237 4791 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-scheduler-0"] Dec 10 23:10:50 crc kubenswrapper[4791]: I1210 23:10:50.013389 4791 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-vkc9h\" (UniqueName: \"kubernetes.io/projected/f947977e-b1aa-43c5-b987-cea516e6a154-kube-api-access-vkc9h\") pod \"nova-scheduler-0\" (UID: \"f947977e-b1aa-43c5-b987-cea516e6a154\") " pod="openstack/nova-scheduler-0" Dec 10 23:10:50 crc kubenswrapper[4791]: I1210 23:10:50.013466 4791 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/1fab60e1-51e0-4101-8c37-c23d8f259cf2-logs\") pod \"nova-metadata-0\" (UID: \"1fab60e1-51e0-4101-8c37-c23d8f259cf2\") " pod="openstack/nova-metadata-0" Dec 10 23:10:50 crc kubenswrapper[4791]: I1210 23:10:50.013505 4791 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f947977e-b1aa-43c5-b987-cea516e6a154-combined-ca-bundle\") pod \"nova-scheduler-0\" (UID: \"f947977e-b1aa-43c5-b987-cea516e6a154\") " pod="openstack/nova-scheduler-0" Dec 10 23:10:50 crc kubenswrapper[4791]: I1210 23:10:50.013586 4791 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"nova-metadata-tls-certs\" (UniqueName: \"kubernetes.io/secret/1fab60e1-51e0-4101-8c37-c23d8f259cf2-nova-metadata-tls-certs\") pod \"nova-metadata-0\" (UID: \"1fab60e1-51e0-4101-8c37-c23d8f259cf2\") " pod="openstack/nova-metadata-0" Dec 10 23:10:50 crc kubenswrapper[4791]: I1210 23:10:50.013618 4791 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/1fab60e1-51e0-4101-8c37-c23d8f259cf2-combined-ca-bundle\") pod \"nova-metadata-0\" (UID: \"1fab60e1-51e0-4101-8c37-c23d8f259cf2\") " pod="openstack/nova-metadata-0" Dec 10 23:10:50 crc kubenswrapper[4791]: I1210 23:10:50.013669 4791 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/1fab60e1-51e0-4101-8c37-c23d8f259cf2-config-data\") pod \"nova-metadata-0\" (UID: \"1fab60e1-51e0-4101-8c37-c23d8f259cf2\") " pod="openstack/nova-metadata-0" Dec 10 23:10:50 crc kubenswrapper[4791]: I1210 23:10:50.013694 4791 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-mrhgz\" (UniqueName: \"kubernetes.io/projected/1fab60e1-51e0-4101-8c37-c23d8f259cf2-kube-api-access-mrhgz\") pod \"nova-metadata-0\" (UID: \"1fab60e1-51e0-4101-8c37-c23d8f259cf2\") " pod="openstack/nova-metadata-0" Dec 10 23:10:50 crc kubenswrapper[4791]: I1210 23:10:50.013895 4791 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/f947977e-b1aa-43c5-b987-cea516e6a154-config-data\") pod \"nova-scheduler-0\" (UID: \"f947977e-b1aa-43c5-b987-cea516e6a154\") " pod="openstack/nova-scheduler-0" Dec 10 23:10:50 crc kubenswrapper[4791]: I1210 23:10:50.115319 4791 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nova-metadata-tls-certs\" (UniqueName: \"kubernetes.io/secret/1fab60e1-51e0-4101-8c37-c23d8f259cf2-nova-metadata-tls-certs\") pod \"nova-metadata-0\" (UID: \"1fab60e1-51e0-4101-8c37-c23d8f259cf2\") " pod="openstack/nova-metadata-0" Dec 10 23:10:50 crc kubenswrapper[4791]: I1210 23:10:50.115387 4791 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/1fab60e1-51e0-4101-8c37-c23d8f259cf2-combined-ca-bundle\") pod \"nova-metadata-0\" (UID: \"1fab60e1-51e0-4101-8c37-c23d8f259cf2\") " pod="openstack/nova-metadata-0" Dec 10 23:10:50 crc kubenswrapper[4791]: I1210 23:10:50.115425 4791 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/1fab60e1-51e0-4101-8c37-c23d8f259cf2-config-data\") pod \"nova-metadata-0\" (UID: \"1fab60e1-51e0-4101-8c37-c23d8f259cf2\") " pod="openstack/nova-metadata-0" Dec 10 23:10:50 crc kubenswrapper[4791]: I1210 23:10:50.115449 4791 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-mrhgz\" (UniqueName: \"kubernetes.io/projected/1fab60e1-51e0-4101-8c37-c23d8f259cf2-kube-api-access-mrhgz\") pod \"nova-metadata-0\" (UID: \"1fab60e1-51e0-4101-8c37-c23d8f259cf2\") " pod="openstack/nova-metadata-0" Dec 10 23:10:50 crc kubenswrapper[4791]: I1210 23:10:50.115564 4791 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/f947977e-b1aa-43c5-b987-cea516e6a154-config-data\") pod \"nova-scheduler-0\" (UID: \"f947977e-b1aa-43c5-b987-cea516e6a154\") " pod="openstack/nova-scheduler-0" Dec 10 23:10:50 crc kubenswrapper[4791]: I1210 23:10:50.115611 4791 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-vkc9h\" (UniqueName: \"kubernetes.io/projected/f947977e-b1aa-43c5-b987-cea516e6a154-kube-api-access-vkc9h\") pod \"nova-scheduler-0\" (UID: \"f947977e-b1aa-43c5-b987-cea516e6a154\") " pod="openstack/nova-scheduler-0" Dec 10 23:10:50 crc kubenswrapper[4791]: I1210 23:10:50.115639 4791 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/1fab60e1-51e0-4101-8c37-c23d8f259cf2-logs\") pod \"nova-metadata-0\" (UID: \"1fab60e1-51e0-4101-8c37-c23d8f259cf2\") " pod="openstack/nova-metadata-0" Dec 10 23:10:50 crc kubenswrapper[4791]: I1210 23:10:50.115663 4791 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f947977e-b1aa-43c5-b987-cea516e6a154-combined-ca-bundle\") pod \"nova-scheduler-0\" (UID: \"f947977e-b1aa-43c5-b987-cea516e6a154\") " pod="openstack/nova-scheduler-0" Dec 10 23:10:50 crc kubenswrapper[4791]: I1210 23:10:50.117007 4791 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/1fab60e1-51e0-4101-8c37-c23d8f259cf2-logs\") pod \"nova-metadata-0\" (UID: \"1fab60e1-51e0-4101-8c37-c23d8f259cf2\") " pod="openstack/nova-metadata-0" Dec 10 23:10:50 crc kubenswrapper[4791]: I1210 23:10:50.120198 4791 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/1fab60e1-51e0-4101-8c37-c23d8f259cf2-combined-ca-bundle\") pod \"nova-metadata-0\" (UID: \"1fab60e1-51e0-4101-8c37-c23d8f259cf2\") " pod="openstack/nova-metadata-0" Dec 10 23:10:50 crc kubenswrapper[4791]: I1210 23:10:50.120873 4791 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f947977e-b1aa-43c5-b987-cea516e6a154-combined-ca-bundle\") pod \"nova-scheduler-0\" (UID: \"f947977e-b1aa-43c5-b987-cea516e6a154\") " pod="openstack/nova-scheduler-0" Dec 10 23:10:50 crc kubenswrapper[4791]: I1210 23:10:50.122073 4791 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/1fab60e1-51e0-4101-8c37-c23d8f259cf2-config-data\") pod \"nova-metadata-0\" (UID: \"1fab60e1-51e0-4101-8c37-c23d8f259cf2\") " pod="openstack/nova-metadata-0" Dec 10 23:10:50 crc kubenswrapper[4791]: I1210 23:10:50.123068 4791 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/f947977e-b1aa-43c5-b987-cea516e6a154-config-data\") pod \"nova-scheduler-0\" (UID: \"f947977e-b1aa-43c5-b987-cea516e6a154\") " pod="openstack/nova-scheduler-0" Dec 10 23:10:50 crc kubenswrapper[4791]: I1210 23:10:50.136689 4791 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"nova-metadata-tls-certs\" (UniqueName: \"kubernetes.io/secret/1fab60e1-51e0-4101-8c37-c23d8f259cf2-nova-metadata-tls-certs\") pod \"nova-metadata-0\" (UID: \"1fab60e1-51e0-4101-8c37-c23d8f259cf2\") " pod="openstack/nova-metadata-0" Dec 10 23:10:50 crc kubenswrapper[4791]: I1210 23:10:50.137401 4791 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-vkc9h\" (UniqueName: \"kubernetes.io/projected/f947977e-b1aa-43c5-b987-cea516e6a154-kube-api-access-vkc9h\") pod \"nova-scheduler-0\" (UID: \"f947977e-b1aa-43c5-b987-cea516e6a154\") " pod="openstack/nova-scheduler-0" Dec 10 23:10:50 crc kubenswrapper[4791]: I1210 23:10:50.139982 4791 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-mrhgz\" (UniqueName: \"kubernetes.io/projected/1fab60e1-51e0-4101-8c37-c23d8f259cf2-kube-api-access-mrhgz\") pod \"nova-metadata-0\" (UID: \"1fab60e1-51e0-4101-8c37-c23d8f259cf2\") " pod="openstack/nova-metadata-0" Dec 10 23:10:50 crc kubenswrapper[4791]: I1210 23:10:50.241514 4791 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-metadata-0" Dec 10 23:10:50 crc kubenswrapper[4791]: I1210 23:10:50.264585 4791 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-scheduler-0" Dec 10 23:10:50 crc kubenswrapper[4791]: I1210 23:10:50.769055 4791 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-metadata-0"] Dec 10 23:10:50 crc kubenswrapper[4791]: I1210 23:10:50.776804 4791 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/dnsmasq-dns-5784cf869f-ngjqj" podUID="224a8e50-4270-42e8-a1df-268bc144b1d5" containerName="dnsmasq-dns" probeResult="failure" output="dial tcp 10.217.0.162:5353: i/o timeout" Dec 10 23:10:50 crc kubenswrapper[4791]: I1210 23:10:50.790652 4791 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"1fab60e1-51e0-4101-8c37-c23d8f259cf2","Type":"ContainerStarted","Data":"12a745ff744ef1b5f57d94629dd336396569e9fbc02be11d7305eb4edd2d4c83"} Dec 10 23:10:50 crc kubenswrapper[4791]: W1210 23:10:50.850949 4791 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podf947977e_b1aa_43c5_b987_cea516e6a154.slice/crio-d41024783f0519ec8b3a38b1ab62cebe3f7c2d789f1d7ab6a7610be4985ae070 WatchSource:0}: Error finding container d41024783f0519ec8b3a38b1ab62cebe3f7c2d789f1d7ab6a7610be4985ae070: Status 404 returned error can't find the container with id d41024783f0519ec8b3a38b1ab62cebe3f7c2d789f1d7ab6a7610be4985ae070 Dec 10 23:10:50 crc kubenswrapper[4791]: I1210 23:10:50.854093 4791 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-scheduler-0"] Dec 10 23:10:51 crc kubenswrapper[4791]: I1210 23:10:51.504327 4791 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Dec 10 23:10:51 crc kubenswrapper[4791]: I1210 23:10:51.564874 4791 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-9mj6d\" (UniqueName: \"kubernetes.io/projected/739a67cd-2a88-4c1d-bf76-5aac9b4b9e46-kube-api-access-9mj6d\") pod \"739a67cd-2a88-4c1d-bf76-5aac9b4b9e46\" (UID: \"739a67cd-2a88-4c1d-bf76-5aac9b4b9e46\") " Dec 10 23:10:51 crc kubenswrapper[4791]: I1210 23:10:51.565050 4791 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/739a67cd-2a88-4c1d-bf76-5aac9b4b9e46-config-data\") pod \"739a67cd-2a88-4c1d-bf76-5aac9b4b9e46\" (UID: \"739a67cd-2a88-4c1d-bf76-5aac9b4b9e46\") " Dec 10 23:10:51 crc kubenswrapper[4791]: I1210 23:10:51.565195 4791 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/739a67cd-2a88-4c1d-bf76-5aac9b4b9e46-logs\") pod \"739a67cd-2a88-4c1d-bf76-5aac9b4b9e46\" (UID: \"739a67cd-2a88-4c1d-bf76-5aac9b4b9e46\") " Dec 10 23:10:51 crc kubenswrapper[4791]: I1210 23:10:51.565239 4791 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/739a67cd-2a88-4c1d-bf76-5aac9b4b9e46-combined-ca-bundle\") pod \"739a67cd-2a88-4c1d-bf76-5aac9b4b9e46\" (UID: \"739a67cd-2a88-4c1d-bf76-5aac9b4b9e46\") " Dec 10 23:10:51 crc kubenswrapper[4791]: I1210 23:10:51.567934 4791 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/739a67cd-2a88-4c1d-bf76-5aac9b4b9e46-logs" (OuterVolumeSpecName: "logs") pod "739a67cd-2a88-4c1d-bf76-5aac9b4b9e46" (UID: "739a67cd-2a88-4c1d-bf76-5aac9b4b9e46"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 10 23:10:51 crc kubenswrapper[4791]: I1210 23:10:51.572891 4791 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/739a67cd-2a88-4c1d-bf76-5aac9b4b9e46-kube-api-access-9mj6d" (OuterVolumeSpecName: "kube-api-access-9mj6d") pod "739a67cd-2a88-4c1d-bf76-5aac9b4b9e46" (UID: "739a67cd-2a88-4c1d-bf76-5aac9b4b9e46"). InnerVolumeSpecName "kube-api-access-9mj6d". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 10 23:10:51 crc kubenswrapper[4791]: I1210 23:10:51.597479 4791 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/739a67cd-2a88-4c1d-bf76-5aac9b4b9e46-config-data" (OuterVolumeSpecName: "config-data") pod "739a67cd-2a88-4c1d-bf76-5aac9b4b9e46" (UID: "739a67cd-2a88-4c1d-bf76-5aac9b4b9e46"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 10 23:10:51 crc kubenswrapper[4791]: I1210 23:10:51.600531 4791 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/739a67cd-2a88-4c1d-bf76-5aac9b4b9e46-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "739a67cd-2a88-4c1d-bf76-5aac9b4b9e46" (UID: "739a67cd-2a88-4c1d-bf76-5aac9b4b9e46"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 10 23:10:51 crc kubenswrapper[4791]: I1210 23:10:51.667551 4791 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/739a67cd-2a88-4c1d-bf76-5aac9b4b9e46-config-data\") on node \"crc\" DevicePath \"\"" Dec 10 23:10:51 crc kubenswrapper[4791]: I1210 23:10:51.667591 4791 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/739a67cd-2a88-4c1d-bf76-5aac9b4b9e46-logs\") on node \"crc\" DevicePath \"\"" Dec 10 23:10:51 crc kubenswrapper[4791]: I1210 23:10:51.667605 4791 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/739a67cd-2a88-4c1d-bf76-5aac9b4b9e46-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 10 23:10:51 crc kubenswrapper[4791]: I1210 23:10:51.667618 4791 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-9mj6d\" (UniqueName: \"kubernetes.io/projected/739a67cd-2a88-4c1d-bf76-5aac9b4b9e46-kube-api-access-9mj6d\") on node \"crc\" DevicePath \"\"" Dec 10 23:10:51 crc kubenswrapper[4791]: I1210 23:10:51.804849 4791 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-scheduler-0" event={"ID":"f947977e-b1aa-43c5-b987-cea516e6a154","Type":"ContainerStarted","Data":"d5fbff44eb0dca42788f629ce6e1217af29146cb334e533f0e3c8327af2d22b4"} Dec 10 23:10:51 crc kubenswrapper[4791]: I1210 23:10:51.806147 4791 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-scheduler-0" event={"ID":"f947977e-b1aa-43c5-b987-cea516e6a154","Type":"ContainerStarted","Data":"d41024783f0519ec8b3a38b1ab62cebe3f7c2d789f1d7ab6a7610be4985ae070"} Dec 10 23:10:51 crc kubenswrapper[4791]: I1210 23:10:51.807549 4791 generic.go:334] "Generic (PLEG): container finished" podID="739a67cd-2a88-4c1d-bf76-5aac9b4b9e46" containerID="a8354ff62f46c0a1cbbcd0e6eb42b520e0b04bcc5960c67755b39312a1f45c3c" exitCode=0 Dec 10 23:10:51 crc kubenswrapper[4791]: I1210 23:10:51.807603 4791 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Dec 10 23:10:51 crc kubenswrapper[4791]: I1210 23:10:51.807686 4791 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"739a67cd-2a88-4c1d-bf76-5aac9b4b9e46","Type":"ContainerDied","Data":"a8354ff62f46c0a1cbbcd0e6eb42b520e0b04bcc5960c67755b39312a1f45c3c"} Dec 10 23:10:51 crc kubenswrapper[4791]: I1210 23:10:51.807769 4791 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"739a67cd-2a88-4c1d-bf76-5aac9b4b9e46","Type":"ContainerDied","Data":"acdedd13a96fb8b18d9dc85527f97cf6678f43930fd183d20cf7220c7d07524e"} Dec 10 23:10:51 crc kubenswrapper[4791]: I1210 23:10:51.807863 4791 scope.go:117] "RemoveContainer" containerID="a8354ff62f46c0a1cbbcd0e6eb42b520e0b04bcc5960c67755b39312a1f45c3c" Dec 10 23:10:51 crc kubenswrapper[4791]: I1210 23:10:51.813848 4791 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"1fab60e1-51e0-4101-8c37-c23d8f259cf2","Type":"ContainerStarted","Data":"ea7f48b4a18e81bcdaa4b7d4980a08c16e8da208d46b06b9f6ae598a2931ea9d"} Dec 10 23:10:51 crc kubenswrapper[4791]: I1210 23:10:51.813903 4791 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"1fab60e1-51e0-4101-8c37-c23d8f259cf2","Type":"ContainerStarted","Data":"1a13d9936e0cd13b45bac08f1af7f835c53e17adc738f8d92e00a0fd7f301300"} Dec 10 23:10:51 crc kubenswrapper[4791]: I1210 23:10:51.844973 4791 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-scheduler-0" podStartSLOduration=2.844953128 podStartE2EDuration="2.844953128s" podCreationTimestamp="2025-12-10 23:10:49 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-10 23:10:51.835043872 +0000 UTC m=+1286.264661515" watchObservedRunningTime="2025-12-10 23:10:51.844953128 +0000 UTC m=+1286.274570741" Dec 10 23:10:51 crc kubenswrapper[4791]: I1210 23:10:51.846482 4791 scope.go:117] "RemoveContainer" containerID="7cdac15bae335278abc24d833001eafda5d44d0849475b593b93f231aff6eb2e" Dec 10 23:10:51 crc kubenswrapper[4791]: I1210 23:10:51.883061 4791 scope.go:117] "RemoveContainer" containerID="a8354ff62f46c0a1cbbcd0e6eb42b520e0b04bcc5960c67755b39312a1f45c3c" Dec 10 23:10:51 crc kubenswrapper[4791]: E1210 23:10:51.883641 4791 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"a8354ff62f46c0a1cbbcd0e6eb42b520e0b04bcc5960c67755b39312a1f45c3c\": container with ID starting with a8354ff62f46c0a1cbbcd0e6eb42b520e0b04bcc5960c67755b39312a1f45c3c not found: ID does not exist" containerID="a8354ff62f46c0a1cbbcd0e6eb42b520e0b04bcc5960c67755b39312a1f45c3c" Dec 10 23:10:51 crc kubenswrapper[4791]: I1210 23:10:51.883700 4791 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"a8354ff62f46c0a1cbbcd0e6eb42b520e0b04bcc5960c67755b39312a1f45c3c"} err="failed to get container status \"a8354ff62f46c0a1cbbcd0e6eb42b520e0b04bcc5960c67755b39312a1f45c3c\": rpc error: code = NotFound desc = could not find container \"a8354ff62f46c0a1cbbcd0e6eb42b520e0b04bcc5960c67755b39312a1f45c3c\": container with ID starting with a8354ff62f46c0a1cbbcd0e6eb42b520e0b04bcc5960c67755b39312a1f45c3c not found: ID does not exist" Dec 10 23:10:51 crc kubenswrapper[4791]: I1210 23:10:51.883745 4791 scope.go:117] "RemoveContainer" containerID="7cdac15bae335278abc24d833001eafda5d44d0849475b593b93f231aff6eb2e" Dec 10 23:10:51 crc kubenswrapper[4791]: I1210 23:10:51.884939 4791 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-metadata-0" podStartSLOduration=2.884925082 podStartE2EDuration="2.884925082s" podCreationTimestamp="2025-12-10 23:10:49 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-10 23:10:51.856496691 +0000 UTC m=+1286.286114304" watchObservedRunningTime="2025-12-10 23:10:51.884925082 +0000 UTC m=+1286.314542695" Dec 10 23:10:51 crc kubenswrapper[4791]: E1210 23:10:51.887060 4791 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"7cdac15bae335278abc24d833001eafda5d44d0849475b593b93f231aff6eb2e\": container with ID starting with 7cdac15bae335278abc24d833001eafda5d44d0849475b593b93f231aff6eb2e not found: ID does not exist" containerID="7cdac15bae335278abc24d833001eafda5d44d0849475b593b93f231aff6eb2e" Dec 10 23:10:51 crc kubenswrapper[4791]: I1210 23:10:51.887235 4791 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"7cdac15bae335278abc24d833001eafda5d44d0849475b593b93f231aff6eb2e"} err="failed to get container status \"7cdac15bae335278abc24d833001eafda5d44d0849475b593b93f231aff6eb2e\": rpc error: code = NotFound desc = could not find container \"7cdac15bae335278abc24d833001eafda5d44d0849475b593b93f231aff6eb2e\": container with ID starting with 7cdac15bae335278abc24d833001eafda5d44d0849475b593b93f231aff6eb2e not found: ID does not exist" Dec 10 23:10:51 crc kubenswrapper[4791]: I1210 23:10:51.901722 4791 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="83b6abc9-aa7f-4ddb-8b9c-b38fcb5a9145" path="/var/lib/kubelet/pods/83b6abc9-aa7f-4ddb-8b9c-b38fcb5a9145/volumes" Dec 10 23:10:51 crc kubenswrapper[4791]: I1210 23:10:51.902404 4791 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-api-0"] Dec 10 23:10:51 crc kubenswrapper[4791]: I1210 23:10:51.911778 4791 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-api-0"] Dec 10 23:10:51 crc kubenswrapper[4791]: I1210 23:10:51.924055 4791 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-api-0"] Dec 10 23:10:51 crc kubenswrapper[4791]: E1210 23:10:51.924756 4791 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="739a67cd-2a88-4c1d-bf76-5aac9b4b9e46" containerName="nova-api-log" Dec 10 23:10:51 crc kubenswrapper[4791]: I1210 23:10:51.924775 4791 state_mem.go:107] "Deleted CPUSet assignment" podUID="739a67cd-2a88-4c1d-bf76-5aac9b4b9e46" containerName="nova-api-log" Dec 10 23:10:51 crc kubenswrapper[4791]: E1210 23:10:51.924818 4791 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="739a67cd-2a88-4c1d-bf76-5aac9b4b9e46" containerName="nova-api-api" Dec 10 23:10:51 crc kubenswrapper[4791]: I1210 23:10:51.924827 4791 state_mem.go:107] "Deleted CPUSet assignment" podUID="739a67cd-2a88-4c1d-bf76-5aac9b4b9e46" containerName="nova-api-api" Dec 10 23:10:51 crc kubenswrapper[4791]: I1210 23:10:51.955389 4791 memory_manager.go:354] "RemoveStaleState removing state" podUID="739a67cd-2a88-4c1d-bf76-5aac9b4b9e46" containerName="nova-api-log" Dec 10 23:10:51 crc kubenswrapper[4791]: I1210 23:10:51.955467 4791 memory_manager.go:354] "RemoveStaleState removing state" podUID="739a67cd-2a88-4c1d-bf76-5aac9b4b9e46" containerName="nova-api-api" Dec 10 23:10:51 crc kubenswrapper[4791]: I1210 23:10:51.956869 4791 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Dec 10 23:10:51 crc kubenswrapper[4791]: I1210 23:10:51.960244 4791 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-api-config-data" Dec 10 23:10:51 crc kubenswrapper[4791]: I1210 23:10:51.978678 4791 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-api-0"] Dec 10 23:10:52 crc kubenswrapper[4791]: I1210 23:10:52.078771 4791 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/c435d469-c6b0-45a0-b61b-fd638540bf51-config-data\") pod \"nova-api-0\" (UID: \"c435d469-c6b0-45a0-b61b-fd638540bf51\") " pod="openstack/nova-api-0" Dec 10 23:10:52 crc kubenswrapper[4791]: I1210 23:10:52.078875 4791 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-5ggxm\" (UniqueName: \"kubernetes.io/projected/c435d469-c6b0-45a0-b61b-fd638540bf51-kube-api-access-5ggxm\") pod \"nova-api-0\" (UID: \"c435d469-c6b0-45a0-b61b-fd638540bf51\") " pod="openstack/nova-api-0" Dec 10 23:10:52 crc kubenswrapper[4791]: I1210 23:10:52.078916 4791 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/c435d469-c6b0-45a0-b61b-fd638540bf51-logs\") pod \"nova-api-0\" (UID: \"c435d469-c6b0-45a0-b61b-fd638540bf51\") " pod="openstack/nova-api-0" Dec 10 23:10:52 crc kubenswrapper[4791]: I1210 23:10:52.078967 4791 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c435d469-c6b0-45a0-b61b-fd638540bf51-combined-ca-bundle\") pod \"nova-api-0\" (UID: \"c435d469-c6b0-45a0-b61b-fd638540bf51\") " pod="openstack/nova-api-0" Dec 10 23:10:52 crc kubenswrapper[4791]: I1210 23:10:52.180434 4791 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/c435d469-c6b0-45a0-b61b-fd638540bf51-config-data\") pod \"nova-api-0\" (UID: \"c435d469-c6b0-45a0-b61b-fd638540bf51\") " pod="openstack/nova-api-0" Dec 10 23:10:52 crc kubenswrapper[4791]: I1210 23:10:52.181282 4791 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-5ggxm\" (UniqueName: \"kubernetes.io/projected/c435d469-c6b0-45a0-b61b-fd638540bf51-kube-api-access-5ggxm\") pod \"nova-api-0\" (UID: \"c435d469-c6b0-45a0-b61b-fd638540bf51\") " pod="openstack/nova-api-0" Dec 10 23:10:52 crc kubenswrapper[4791]: I1210 23:10:52.181331 4791 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/c435d469-c6b0-45a0-b61b-fd638540bf51-logs\") pod \"nova-api-0\" (UID: \"c435d469-c6b0-45a0-b61b-fd638540bf51\") " pod="openstack/nova-api-0" Dec 10 23:10:52 crc kubenswrapper[4791]: I1210 23:10:52.181398 4791 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c435d469-c6b0-45a0-b61b-fd638540bf51-combined-ca-bundle\") pod \"nova-api-0\" (UID: \"c435d469-c6b0-45a0-b61b-fd638540bf51\") " pod="openstack/nova-api-0" Dec 10 23:10:52 crc kubenswrapper[4791]: I1210 23:10:52.182092 4791 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/c435d469-c6b0-45a0-b61b-fd638540bf51-logs\") pod \"nova-api-0\" (UID: \"c435d469-c6b0-45a0-b61b-fd638540bf51\") " pod="openstack/nova-api-0" Dec 10 23:10:52 crc kubenswrapper[4791]: I1210 23:10:52.186124 4791 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c435d469-c6b0-45a0-b61b-fd638540bf51-combined-ca-bundle\") pod \"nova-api-0\" (UID: \"c435d469-c6b0-45a0-b61b-fd638540bf51\") " pod="openstack/nova-api-0" Dec 10 23:10:52 crc kubenswrapper[4791]: I1210 23:10:52.194761 4791 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/c435d469-c6b0-45a0-b61b-fd638540bf51-config-data\") pod \"nova-api-0\" (UID: \"c435d469-c6b0-45a0-b61b-fd638540bf51\") " pod="openstack/nova-api-0" Dec 10 23:10:52 crc kubenswrapper[4791]: I1210 23:10:52.199502 4791 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-5ggxm\" (UniqueName: \"kubernetes.io/projected/c435d469-c6b0-45a0-b61b-fd638540bf51-kube-api-access-5ggxm\") pod \"nova-api-0\" (UID: \"c435d469-c6b0-45a0-b61b-fd638540bf51\") " pod="openstack/nova-api-0" Dec 10 23:10:52 crc kubenswrapper[4791]: I1210 23:10:52.282459 4791 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Dec 10 23:10:52 crc kubenswrapper[4791]: I1210 23:10:52.787945 4791 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-api-0"] Dec 10 23:10:52 crc kubenswrapper[4791]: I1210 23:10:52.826704 4791 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"c435d469-c6b0-45a0-b61b-fd638540bf51","Type":"ContainerStarted","Data":"b3b5c917af99a1aaf2f7aac51b126d7ab683f8bfff1792bf9133276410a36beb"} Dec 10 23:10:53 crc kubenswrapper[4791]: I1210 23:10:53.144358 4791 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-cell1-conductor-0" Dec 10 23:10:53 crc kubenswrapper[4791]: I1210 23:10:53.502655 4791 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/ceilometer-0" Dec 10 23:10:53 crc kubenswrapper[4791]: I1210 23:10:53.837971 4791 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"c435d469-c6b0-45a0-b61b-fd638540bf51","Type":"ContainerStarted","Data":"c2bd07aa7c52e85c622f844ee2cd9b8761a2f8e9220b0e3214c3a421060ea90d"} Dec 10 23:10:53 crc kubenswrapper[4791]: I1210 23:10:53.838215 4791 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"c435d469-c6b0-45a0-b61b-fd638540bf51","Type":"ContainerStarted","Data":"940bdefb6fe251888b9f0d4224ac2c43268c7a2c548cd5eec01678307a80317d"} Dec 10 23:10:53 crc kubenswrapper[4791]: I1210 23:10:53.862852 4791 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-api-0" podStartSLOduration=2.862835188 podStartE2EDuration="2.862835188s" podCreationTimestamp="2025-12-10 23:10:51 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-10 23:10:53.856616159 +0000 UTC m=+1288.286233762" watchObservedRunningTime="2025-12-10 23:10:53.862835188 +0000 UTC m=+1288.292452801" Dec 10 23:10:53 crc kubenswrapper[4791]: I1210 23:10:53.895298 4791 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="739a67cd-2a88-4c1d-bf76-5aac9b4b9e46" path="/var/lib/kubelet/pods/739a67cd-2a88-4c1d-bf76-5aac9b4b9e46/volumes" Dec 10 23:10:55 crc kubenswrapper[4791]: I1210 23:10:55.242553 4791 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-metadata-0" Dec 10 23:10:55 crc kubenswrapper[4791]: I1210 23:10:55.242630 4791 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-metadata-0" Dec 10 23:10:55 crc kubenswrapper[4791]: I1210 23:10:55.265412 4791 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-scheduler-0" Dec 10 23:10:57 crc kubenswrapper[4791]: I1210 23:10:57.039201 4791 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/kube-state-metrics-0"] Dec 10 23:10:57 crc kubenswrapper[4791]: I1210 23:10:57.040818 4791 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/kube-state-metrics-0" podUID="247efe24-6d59-4c93-ab04-6e249e1005a8" containerName="kube-state-metrics" containerID="cri-o://98e8f3bcbc5efc1d138b8c561352d547944c2c4e43d105c7bd3827406ac310dd" gracePeriod=30 Dec 10 23:10:57 crc kubenswrapper[4791]: I1210 23:10:57.577049 4791 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/kube-state-metrics-0" Dec 10 23:10:57 crc kubenswrapper[4791]: I1210 23:10:57.606369 4791 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-tlj7g\" (UniqueName: \"kubernetes.io/projected/247efe24-6d59-4c93-ab04-6e249e1005a8-kube-api-access-tlj7g\") pod \"247efe24-6d59-4c93-ab04-6e249e1005a8\" (UID: \"247efe24-6d59-4c93-ab04-6e249e1005a8\") " Dec 10 23:10:57 crc kubenswrapper[4791]: I1210 23:10:57.622307 4791 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/247efe24-6d59-4c93-ab04-6e249e1005a8-kube-api-access-tlj7g" (OuterVolumeSpecName: "kube-api-access-tlj7g") pod "247efe24-6d59-4c93-ab04-6e249e1005a8" (UID: "247efe24-6d59-4c93-ab04-6e249e1005a8"). InnerVolumeSpecName "kube-api-access-tlj7g". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 10 23:10:57 crc kubenswrapper[4791]: I1210 23:10:57.708081 4791 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-tlj7g\" (UniqueName: \"kubernetes.io/projected/247efe24-6d59-4c93-ab04-6e249e1005a8-kube-api-access-tlj7g\") on node \"crc\" DevicePath \"\"" Dec 10 23:10:57 crc kubenswrapper[4791]: I1210 23:10:57.906124 4791 generic.go:334] "Generic (PLEG): container finished" podID="247efe24-6d59-4c93-ab04-6e249e1005a8" containerID="98e8f3bcbc5efc1d138b8c561352d547944c2c4e43d105c7bd3827406ac310dd" exitCode=2 Dec 10 23:10:57 crc kubenswrapper[4791]: I1210 23:10:57.906196 4791 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/kube-state-metrics-0" event={"ID":"247efe24-6d59-4c93-ab04-6e249e1005a8","Type":"ContainerDied","Data":"98e8f3bcbc5efc1d138b8c561352d547944c2c4e43d105c7bd3827406ac310dd"} Dec 10 23:10:57 crc kubenswrapper[4791]: I1210 23:10:57.906252 4791 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/kube-state-metrics-0" event={"ID":"247efe24-6d59-4c93-ab04-6e249e1005a8","Type":"ContainerDied","Data":"675f475dd72ea2082913c5c6f79644c0372948ff1f0345a38a944d833d3ece2c"} Dec 10 23:10:57 crc kubenswrapper[4791]: I1210 23:10:57.906278 4791 scope.go:117] "RemoveContainer" containerID="98e8f3bcbc5efc1d138b8c561352d547944c2c4e43d105c7bd3827406ac310dd" Dec 10 23:10:57 crc kubenswrapper[4791]: I1210 23:10:57.906788 4791 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/kube-state-metrics-0" Dec 10 23:10:57 crc kubenswrapper[4791]: I1210 23:10:57.943731 4791 scope.go:117] "RemoveContainer" containerID="98e8f3bcbc5efc1d138b8c561352d547944c2c4e43d105c7bd3827406ac310dd" Dec 10 23:10:57 crc kubenswrapper[4791]: E1210 23:10:57.945835 4791 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"98e8f3bcbc5efc1d138b8c561352d547944c2c4e43d105c7bd3827406ac310dd\": container with ID starting with 98e8f3bcbc5efc1d138b8c561352d547944c2c4e43d105c7bd3827406ac310dd not found: ID does not exist" containerID="98e8f3bcbc5efc1d138b8c561352d547944c2c4e43d105c7bd3827406ac310dd" Dec 10 23:10:57 crc kubenswrapper[4791]: I1210 23:10:57.945908 4791 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"98e8f3bcbc5efc1d138b8c561352d547944c2c4e43d105c7bd3827406ac310dd"} err="failed to get container status \"98e8f3bcbc5efc1d138b8c561352d547944c2c4e43d105c7bd3827406ac310dd\": rpc error: code = NotFound desc = could not find container \"98e8f3bcbc5efc1d138b8c561352d547944c2c4e43d105c7bd3827406ac310dd\": container with ID starting with 98e8f3bcbc5efc1d138b8c561352d547944c2c4e43d105c7bd3827406ac310dd not found: ID does not exist" Dec 10 23:10:57 crc kubenswrapper[4791]: I1210 23:10:57.947165 4791 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/kube-state-metrics-0"] Dec 10 23:10:57 crc kubenswrapper[4791]: I1210 23:10:57.954790 4791 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/kube-state-metrics-0"] Dec 10 23:10:57 crc kubenswrapper[4791]: I1210 23:10:57.966834 4791 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/kube-state-metrics-0"] Dec 10 23:10:57 crc kubenswrapper[4791]: E1210 23:10:57.967581 4791 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="247efe24-6d59-4c93-ab04-6e249e1005a8" containerName="kube-state-metrics" Dec 10 23:10:57 crc kubenswrapper[4791]: I1210 23:10:57.967632 4791 state_mem.go:107] "Deleted CPUSet assignment" podUID="247efe24-6d59-4c93-ab04-6e249e1005a8" containerName="kube-state-metrics" Dec 10 23:10:57 crc kubenswrapper[4791]: I1210 23:10:57.968117 4791 memory_manager.go:354] "RemoveStaleState removing state" podUID="247efe24-6d59-4c93-ab04-6e249e1005a8" containerName="kube-state-metrics" Dec 10 23:10:57 crc kubenswrapper[4791]: I1210 23:10:57.969432 4791 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/kube-state-metrics-0" Dec 10 23:10:57 crc kubenswrapper[4791]: I1210 23:10:57.973018 4791 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"kube-state-metrics-tls-config" Dec 10 23:10:57 crc kubenswrapper[4791]: I1210 23:10:57.975864 4791 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/kube-state-metrics-0"] Dec 10 23:10:57 crc kubenswrapper[4791]: I1210 23:10:57.981070 4791 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-kube-state-metrics-svc" Dec 10 23:10:58 crc kubenswrapper[4791]: I1210 23:10:58.013193 4791 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-4s4tl\" (UniqueName: \"kubernetes.io/projected/a5c489de-39fe-42e1-963b-10a99cb531b6-kube-api-access-4s4tl\") pod \"kube-state-metrics-0\" (UID: \"a5c489de-39fe-42e1-963b-10a99cb531b6\") " pod="openstack/kube-state-metrics-0" Dec 10 23:10:58 crc kubenswrapper[4791]: I1210 23:10:58.013296 4791 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-state-metrics-tls-certs\" (UniqueName: \"kubernetes.io/secret/a5c489de-39fe-42e1-963b-10a99cb531b6-kube-state-metrics-tls-certs\") pod \"kube-state-metrics-0\" (UID: \"a5c489de-39fe-42e1-963b-10a99cb531b6\") " pod="openstack/kube-state-metrics-0" Dec 10 23:10:58 crc kubenswrapper[4791]: I1210 23:10:58.013378 4791 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-state-metrics-tls-config\" (UniqueName: \"kubernetes.io/secret/a5c489de-39fe-42e1-963b-10a99cb531b6-kube-state-metrics-tls-config\") pod \"kube-state-metrics-0\" (UID: \"a5c489de-39fe-42e1-963b-10a99cb531b6\") " pod="openstack/kube-state-metrics-0" Dec 10 23:10:58 crc kubenswrapper[4791]: I1210 23:10:58.013548 4791 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a5c489de-39fe-42e1-963b-10a99cb531b6-combined-ca-bundle\") pod \"kube-state-metrics-0\" (UID: \"a5c489de-39fe-42e1-963b-10a99cb531b6\") " pod="openstack/kube-state-metrics-0" Dec 10 23:10:58 crc kubenswrapper[4791]: I1210 23:10:58.115565 4791 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-state-metrics-tls-certs\" (UniqueName: \"kubernetes.io/secret/a5c489de-39fe-42e1-963b-10a99cb531b6-kube-state-metrics-tls-certs\") pod \"kube-state-metrics-0\" (UID: \"a5c489de-39fe-42e1-963b-10a99cb531b6\") " pod="openstack/kube-state-metrics-0" Dec 10 23:10:58 crc kubenswrapper[4791]: I1210 23:10:58.115644 4791 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-state-metrics-tls-config\" (UniqueName: \"kubernetes.io/secret/a5c489de-39fe-42e1-963b-10a99cb531b6-kube-state-metrics-tls-config\") pod \"kube-state-metrics-0\" (UID: \"a5c489de-39fe-42e1-963b-10a99cb531b6\") " pod="openstack/kube-state-metrics-0" Dec 10 23:10:58 crc kubenswrapper[4791]: I1210 23:10:58.115774 4791 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a5c489de-39fe-42e1-963b-10a99cb531b6-combined-ca-bundle\") pod \"kube-state-metrics-0\" (UID: \"a5c489de-39fe-42e1-963b-10a99cb531b6\") " pod="openstack/kube-state-metrics-0" Dec 10 23:10:58 crc kubenswrapper[4791]: I1210 23:10:58.115882 4791 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-4s4tl\" (UniqueName: \"kubernetes.io/projected/a5c489de-39fe-42e1-963b-10a99cb531b6-kube-api-access-4s4tl\") pod \"kube-state-metrics-0\" (UID: \"a5c489de-39fe-42e1-963b-10a99cb531b6\") " pod="openstack/kube-state-metrics-0" Dec 10 23:10:58 crc kubenswrapper[4791]: I1210 23:10:58.120941 4791 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-state-metrics-tls-config\" (UniqueName: \"kubernetes.io/secret/a5c489de-39fe-42e1-963b-10a99cb531b6-kube-state-metrics-tls-config\") pod \"kube-state-metrics-0\" (UID: \"a5c489de-39fe-42e1-963b-10a99cb531b6\") " pod="openstack/kube-state-metrics-0" Dec 10 23:10:58 crc kubenswrapper[4791]: I1210 23:10:58.121093 4791 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-state-metrics-tls-certs\" (UniqueName: \"kubernetes.io/secret/a5c489de-39fe-42e1-963b-10a99cb531b6-kube-state-metrics-tls-certs\") pod \"kube-state-metrics-0\" (UID: \"a5c489de-39fe-42e1-963b-10a99cb531b6\") " pod="openstack/kube-state-metrics-0" Dec 10 23:10:58 crc kubenswrapper[4791]: I1210 23:10:58.131607 4791 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a5c489de-39fe-42e1-963b-10a99cb531b6-combined-ca-bundle\") pod \"kube-state-metrics-0\" (UID: \"a5c489de-39fe-42e1-963b-10a99cb531b6\") " pod="openstack/kube-state-metrics-0" Dec 10 23:10:58 crc kubenswrapper[4791]: I1210 23:10:58.132257 4791 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-4s4tl\" (UniqueName: \"kubernetes.io/projected/a5c489de-39fe-42e1-963b-10a99cb531b6-kube-api-access-4s4tl\") pod \"kube-state-metrics-0\" (UID: \"a5c489de-39fe-42e1-963b-10a99cb531b6\") " pod="openstack/kube-state-metrics-0" Dec 10 23:10:58 crc kubenswrapper[4791]: I1210 23:10:58.285143 4791 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/kube-state-metrics-0" Dec 10 23:10:58 crc kubenswrapper[4791]: I1210 23:10:58.743323 4791 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ceilometer-0"] Dec 10 23:10:58 crc kubenswrapper[4791]: I1210 23:10:58.743960 4791 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="65afcf68-16b8-4f9d-ab77-ab910ca9469b" containerName="ceilometer-central-agent" containerID="cri-o://3a5559784c422a1e8dd3f804e260a986d9b1c7dac2ccb1aea0c905532562cfd4" gracePeriod=30 Dec 10 23:10:58 crc kubenswrapper[4791]: I1210 23:10:58.743996 4791 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="65afcf68-16b8-4f9d-ab77-ab910ca9469b" containerName="proxy-httpd" containerID="cri-o://bfb1f419fcadf3ec57c071c7d7e0d51159099c674bd52ff80933b2cba66a8da4" gracePeriod=30 Dec 10 23:10:58 crc kubenswrapper[4791]: I1210 23:10:58.744294 4791 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="65afcf68-16b8-4f9d-ab77-ab910ca9469b" containerName="sg-core" containerID="cri-o://bd8db88f95d6197edaa97ac8dc83a26e1f4e74cecfab63880dba48e7f4142325" gracePeriod=30 Dec 10 23:10:58 crc kubenswrapper[4791]: I1210 23:10:58.744330 4791 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="65afcf68-16b8-4f9d-ab77-ab910ca9469b" containerName="ceilometer-notification-agent" containerID="cri-o://6f1983af270707fe501a3d4c71341f8755d1b9f230aebf1ddfe48385bd734154" gracePeriod=30 Dec 10 23:10:58 crc kubenswrapper[4791]: I1210 23:10:58.767125 4791 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/kube-state-metrics-0"] Dec 10 23:10:58 crc kubenswrapper[4791]: W1210 23:10:58.787539 4791 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-poda5c489de_39fe_42e1_963b_10a99cb531b6.slice/crio-18cad32593ea09fdb0e00546c768a7b3809af43a4001cb3f6db3ee6fe8829837 WatchSource:0}: Error finding container 18cad32593ea09fdb0e00546c768a7b3809af43a4001cb3f6db3ee6fe8829837: Status 404 returned error can't find the container with id 18cad32593ea09fdb0e00546c768a7b3809af43a4001cb3f6db3ee6fe8829837 Dec 10 23:10:58 crc kubenswrapper[4791]: I1210 23:10:58.790319 4791 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Dec 10 23:10:58 crc kubenswrapper[4791]: I1210 23:10:58.917571 4791 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/kube-state-metrics-0" event={"ID":"a5c489de-39fe-42e1-963b-10a99cb531b6","Type":"ContainerStarted","Data":"18cad32593ea09fdb0e00546c768a7b3809af43a4001cb3f6db3ee6fe8829837"} Dec 10 23:10:58 crc kubenswrapper[4791]: I1210 23:10:58.920562 4791 generic.go:334] "Generic (PLEG): container finished" podID="65afcf68-16b8-4f9d-ab77-ab910ca9469b" containerID="bd8db88f95d6197edaa97ac8dc83a26e1f4e74cecfab63880dba48e7f4142325" exitCode=2 Dec 10 23:10:58 crc kubenswrapper[4791]: I1210 23:10:58.920610 4791 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"65afcf68-16b8-4f9d-ab77-ab910ca9469b","Type":"ContainerDied","Data":"bd8db88f95d6197edaa97ac8dc83a26e1f4e74cecfab63880dba48e7f4142325"} Dec 10 23:10:59 crc kubenswrapper[4791]: I1210 23:10:59.907593 4791 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="247efe24-6d59-4c93-ab04-6e249e1005a8" path="/var/lib/kubelet/pods/247efe24-6d59-4c93-ab04-6e249e1005a8/volumes" Dec 10 23:10:59 crc kubenswrapper[4791]: I1210 23:10:59.931215 4791 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/kube-state-metrics-0" event={"ID":"a5c489de-39fe-42e1-963b-10a99cb531b6","Type":"ContainerStarted","Data":"8b2259e8c8cabcbab69d60dbed4da9afaeb2131e02116b53cdf25427b3d53aae"} Dec 10 23:10:59 crc kubenswrapper[4791]: I1210 23:10:59.931637 4791 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/kube-state-metrics-0" Dec 10 23:10:59 crc kubenswrapper[4791]: I1210 23:10:59.933767 4791 generic.go:334] "Generic (PLEG): container finished" podID="65afcf68-16b8-4f9d-ab77-ab910ca9469b" containerID="bfb1f419fcadf3ec57c071c7d7e0d51159099c674bd52ff80933b2cba66a8da4" exitCode=0 Dec 10 23:10:59 crc kubenswrapper[4791]: I1210 23:10:59.933793 4791 generic.go:334] "Generic (PLEG): container finished" podID="65afcf68-16b8-4f9d-ab77-ab910ca9469b" containerID="3a5559784c422a1e8dd3f804e260a986d9b1c7dac2ccb1aea0c905532562cfd4" exitCode=0 Dec 10 23:10:59 crc kubenswrapper[4791]: I1210 23:10:59.933812 4791 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"65afcf68-16b8-4f9d-ab77-ab910ca9469b","Type":"ContainerDied","Data":"bfb1f419fcadf3ec57c071c7d7e0d51159099c674bd52ff80933b2cba66a8da4"} Dec 10 23:10:59 crc kubenswrapper[4791]: I1210 23:10:59.933835 4791 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"65afcf68-16b8-4f9d-ab77-ab910ca9469b","Type":"ContainerDied","Data":"3a5559784c422a1e8dd3f804e260a986d9b1c7dac2ccb1aea0c905532562cfd4"} Dec 10 23:10:59 crc kubenswrapper[4791]: I1210 23:10:59.955806 4791 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/kube-state-metrics-0" podStartSLOduration=2.585465056 podStartE2EDuration="2.955787456s" podCreationTimestamp="2025-12-10 23:10:57 +0000 UTC" firstStartedPulling="2025-12-10 23:10:58.790130726 +0000 UTC m=+1293.219748329" lastFinishedPulling="2025-12-10 23:10:59.160453116 +0000 UTC m=+1293.590070729" observedRunningTime="2025-12-10 23:10:59.946679273 +0000 UTC m=+1294.376296886" watchObservedRunningTime="2025-12-10 23:10:59.955787456 +0000 UTC m=+1294.385405069" Dec 10 23:11:00 crc kubenswrapper[4791]: I1210 23:11:00.242380 4791 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-metadata-0" Dec 10 23:11:00 crc kubenswrapper[4791]: I1210 23:11:00.242438 4791 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-metadata-0" Dec 10 23:11:00 crc kubenswrapper[4791]: I1210 23:11:00.265993 4791 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-scheduler-0" Dec 10 23:11:00 crc kubenswrapper[4791]: I1210 23:11:00.300361 4791 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/nova-scheduler-0" Dec 10 23:11:00 crc kubenswrapper[4791]: I1210 23:11:00.987691 4791 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-scheduler-0" Dec 10 23:11:01 crc kubenswrapper[4791]: I1210 23:11:01.258582 4791 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/nova-metadata-0" podUID="1fab60e1-51e0-4101-8c37-c23d8f259cf2" containerName="nova-metadata-metadata" probeResult="failure" output="Get \"https://10.217.0.191:8775/\": net/http: request canceled (Client.Timeout exceeded while awaiting headers)" Dec 10 23:11:01 crc kubenswrapper[4791]: I1210 23:11:01.258582 4791 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/nova-metadata-0" podUID="1fab60e1-51e0-4101-8c37-c23d8f259cf2" containerName="nova-metadata-log" probeResult="failure" output="Get \"https://10.217.0.191:8775/\": net/http: request canceled (Client.Timeout exceeded while awaiting headers)" Dec 10 23:11:02 crc kubenswrapper[4791]: I1210 23:11:02.283692 4791 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-api-0" Dec 10 23:11:02 crc kubenswrapper[4791]: I1210 23:11:02.284019 4791 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-api-0" Dec 10 23:11:03 crc kubenswrapper[4791]: I1210 23:11:03.324625 4791 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/nova-api-0" podUID="c435d469-c6b0-45a0-b61b-fd638540bf51" containerName="nova-api-log" probeResult="failure" output="Get \"http://10.217.0.193:8774/\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" Dec 10 23:11:03 crc kubenswrapper[4791]: I1210 23:11:03.365601 4791 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/nova-api-0" podUID="c435d469-c6b0-45a0-b61b-fd638540bf51" containerName="nova-api-api" probeResult="failure" output="Get \"http://10.217.0.193:8774/\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" Dec 10 23:11:08 crc kubenswrapper[4791]: I1210 23:11:08.031995 4791 generic.go:334] "Generic (PLEG): container finished" podID="65afcf68-16b8-4f9d-ab77-ab910ca9469b" containerID="6f1983af270707fe501a3d4c71341f8755d1b9f230aebf1ddfe48385bd734154" exitCode=0 Dec 10 23:11:08 crc kubenswrapper[4791]: I1210 23:11:08.032081 4791 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"65afcf68-16b8-4f9d-ab77-ab910ca9469b","Type":"ContainerDied","Data":"6f1983af270707fe501a3d4c71341f8755d1b9f230aebf1ddfe48385bd734154"} Dec 10 23:11:08 crc kubenswrapper[4791]: I1210 23:11:08.302655 4791 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/kube-state-metrics-0" Dec 10 23:11:08 crc kubenswrapper[4791]: I1210 23:11:08.397196 4791 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Dec 10 23:11:08 crc kubenswrapper[4791]: I1210 23:11:08.422907 4791 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/65afcf68-16b8-4f9d-ab77-ab910ca9469b-combined-ca-bundle\") pod \"65afcf68-16b8-4f9d-ab77-ab910ca9469b\" (UID: \"65afcf68-16b8-4f9d-ab77-ab910ca9469b\") " Dec 10 23:11:08 crc kubenswrapper[4791]: I1210 23:11:08.423263 4791 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/65afcf68-16b8-4f9d-ab77-ab910ca9469b-log-httpd\") pod \"65afcf68-16b8-4f9d-ab77-ab910ca9469b\" (UID: \"65afcf68-16b8-4f9d-ab77-ab910ca9469b\") " Dec 10 23:11:08 crc kubenswrapper[4791]: I1210 23:11:08.423452 4791 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/65afcf68-16b8-4f9d-ab77-ab910ca9469b-config-data\") pod \"65afcf68-16b8-4f9d-ab77-ab910ca9469b\" (UID: \"65afcf68-16b8-4f9d-ab77-ab910ca9469b\") " Dec 10 23:11:08 crc kubenswrapper[4791]: I1210 23:11:08.423623 4791 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-t8gkz\" (UniqueName: \"kubernetes.io/projected/65afcf68-16b8-4f9d-ab77-ab910ca9469b-kube-api-access-t8gkz\") pod \"65afcf68-16b8-4f9d-ab77-ab910ca9469b\" (UID: \"65afcf68-16b8-4f9d-ab77-ab910ca9469b\") " Dec 10 23:11:08 crc kubenswrapper[4791]: I1210 23:11:08.423738 4791 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/65afcf68-16b8-4f9d-ab77-ab910ca9469b-run-httpd\") pod \"65afcf68-16b8-4f9d-ab77-ab910ca9469b\" (UID: \"65afcf68-16b8-4f9d-ab77-ab910ca9469b\") " Dec 10 23:11:08 crc kubenswrapper[4791]: I1210 23:11:08.424391 4791 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/65afcf68-16b8-4f9d-ab77-ab910ca9469b-sg-core-conf-yaml\") pod \"65afcf68-16b8-4f9d-ab77-ab910ca9469b\" (UID: \"65afcf68-16b8-4f9d-ab77-ab910ca9469b\") " Dec 10 23:11:08 crc kubenswrapper[4791]: I1210 23:11:08.424939 4791 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/65afcf68-16b8-4f9d-ab77-ab910ca9469b-scripts\") pod \"65afcf68-16b8-4f9d-ab77-ab910ca9469b\" (UID: \"65afcf68-16b8-4f9d-ab77-ab910ca9469b\") " Dec 10 23:11:08 crc kubenswrapper[4791]: I1210 23:11:08.423765 4791 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/65afcf68-16b8-4f9d-ab77-ab910ca9469b-log-httpd" (OuterVolumeSpecName: "log-httpd") pod "65afcf68-16b8-4f9d-ab77-ab910ca9469b" (UID: "65afcf68-16b8-4f9d-ab77-ab910ca9469b"). InnerVolumeSpecName "log-httpd". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 10 23:11:08 crc kubenswrapper[4791]: I1210 23:11:08.424080 4791 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/65afcf68-16b8-4f9d-ab77-ab910ca9469b-run-httpd" (OuterVolumeSpecName: "run-httpd") pod "65afcf68-16b8-4f9d-ab77-ab910ca9469b" (UID: "65afcf68-16b8-4f9d-ab77-ab910ca9469b"). InnerVolumeSpecName "run-httpd". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 10 23:11:08 crc kubenswrapper[4791]: I1210 23:11:08.432503 4791 reconciler_common.go:293] "Volume detached for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/65afcf68-16b8-4f9d-ab77-ab910ca9469b-log-httpd\") on node \"crc\" DevicePath \"\"" Dec 10 23:11:08 crc kubenswrapper[4791]: I1210 23:11:08.432835 4791 reconciler_common.go:293] "Volume detached for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/65afcf68-16b8-4f9d-ab77-ab910ca9469b-run-httpd\") on node \"crc\" DevicePath \"\"" Dec 10 23:11:08 crc kubenswrapper[4791]: I1210 23:11:08.432562 4791 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/65afcf68-16b8-4f9d-ab77-ab910ca9469b-scripts" (OuterVolumeSpecName: "scripts") pod "65afcf68-16b8-4f9d-ab77-ab910ca9469b" (UID: "65afcf68-16b8-4f9d-ab77-ab910ca9469b"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 10 23:11:08 crc kubenswrapper[4791]: I1210 23:11:08.443849 4791 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/65afcf68-16b8-4f9d-ab77-ab910ca9469b-kube-api-access-t8gkz" (OuterVolumeSpecName: "kube-api-access-t8gkz") pod "65afcf68-16b8-4f9d-ab77-ab910ca9469b" (UID: "65afcf68-16b8-4f9d-ab77-ab910ca9469b"). InnerVolumeSpecName "kube-api-access-t8gkz". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 10 23:11:08 crc kubenswrapper[4791]: I1210 23:11:08.451938 4791 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/65afcf68-16b8-4f9d-ab77-ab910ca9469b-sg-core-conf-yaml" (OuterVolumeSpecName: "sg-core-conf-yaml") pod "65afcf68-16b8-4f9d-ab77-ab910ca9469b" (UID: "65afcf68-16b8-4f9d-ab77-ab910ca9469b"). InnerVolumeSpecName "sg-core-conf-yaml". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 10 23:11:08 crc kubenswrapper[4791]: I1210 23:11:08.526921 4791 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/65afcf68-16b8-4f9d-ab77-ab910ca9469b-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "65afcf68-16b8-4f9d-ab77-ab910ca9469b" (UID: "65afcf68-16b8-4f9d-ab77-ab910ca9469b"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 10 23:11:08 crc kubenswrapper[4791]: I1210 23:11:08.534945 4791 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-t8gkz\" (UniqueName: \"kubernetes.io/projected/65afcf68-16b8-4f9d-ab77-ab910ca9469b-kube-api-access-t8gkz\") on node \"crc\" DevicePath \"\"" Dec 10 23:11:08 crc kubenswrapper[4791]: I1210 23:11:08.534976 4791 reconciler_common.go:293] "Volume detached for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/65afcf68-16b8-4f9d-ab77-ab910ca9469b-sg-core-conf-yaml\") on node \"crc\" DevicePath \"\"" Dec 10 23:11:08 crc kubenswrapper[4791]: I1210 23:11:08.534987 4791 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/65afcf68-16b8-4f9d-ab77-ab910ca9469b-scripts\") on node \"crc\" DevicePath \"\"" Dec 10 23:11:08 crc kubenswrapper[4791]: I1210 23:11:08.534997 4791 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/65afcf68-16b8-4f9d-ab77-ab910ca9469b-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 10 23:11:08 crc kubenswrapper[4791]: I1210 23:11:08.555895 4791 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/65afcf68-16b8-4f9d-ab77-ab910ca9469b-config-data" (OuterVolumeSpecName: "config-data") pod "65afcf68-16b8-4f9d-ab77-ab910ca9469b" (UID: "65afcf68-16b8-4f9d-ab77-ab910ca9469b"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 10 23:11:08 crc kubenswrapper[4791]: I1210 23:11:08.637045 4791 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/65afcf68-16b8-4f9d-ab77-ab910ca9469b-config-data\") on node \"crc\" DevicePath \"\"" Dec 10 23:11:09 crc kubenswrapper[4791]: I1210 23:11:09.047330 4791 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"65afcf68-16b8-4f9d-ab77-ab910ca9469b","Type":"ContainerDied","Data":"4ae58e4f34c7896f75bcd7f3998c4819f2accc012ff8d2bc1d9ae544e732ab92"} Dec 10 23:11:09 crc kubenswrapper[4791]: I1210 23:11:09.047607 4791 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Dec 10 23:11:09 crc kubenswrapper[4791]: I1210 23:11:09.047647 4791 scope.go:117] "RemoveContainer" containerID="bfb1f419fcadf3ec57c071c7d7e0d51159099c674bd52ff80933b2cba66a8da4" Dec 10 23:11:09 crc kubenswrapper[4791]: I1210 23:11:09.103395 4791 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ceilometer-0"] Dec 10 23:11:09 crc kubenswrapper[4791]: I1210 23:11:09.112211 4791 scope.go:117] "RemoveContainer" containerID="bd8db88f95d6197edaa97ac8dc83a26e1f4e74cecfab63880dba48e7f4142325" Dec 10 23:11:09 crc kubenswrapper[4791]: I1210 23:11:09.122279 4791 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/ceilometer-0"] Dec 10 23:11:09 crc kubenswrapper[4791]: I1210 23:11:09.131022 4791 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ceilometer-0"] Dec 10 23:11:09 crc kubenswrapper[4791]: E1210 23:11:09.131563 4791 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="65afcf68-16b8-4f9d-ab77-ab910ca9469b" containerName="ceilometer-central-agent" Dec 10 23:11:09 crc kubenswrapper[4791]: I1210 23:11:09.131592 4791 state_mem.go:107] "Deleted CPUSet assignment" podUID="65afcf68-16b8-4f9d-ab77-ab910ca9469b" containerName="ceilometer-central-agent" Dec 10 23:11:09 crc kubenswrapper[4791]: E1210 23:11:09.131604 4791 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="65afcf68-16b8-4f9d-ab77-ab910ca9469b" containerName="sg-core" Dec 10 23:11:09 crc kubenswrapper[4791]: I1210 23:11:09.131615 4791 state_mem.go:107] "Deleted CPUSet assignment" podUID="65afcf68-16b8-4f9d-ab77-ab910ca9469b" containerName="sg-core" Dec 10 23:11:09 crc kubenswrapper[4791]: E1210 23:11:09.131632 4791 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="65afcf68-16b8-4f9d-ab77-ab910ca9469b" containerName="proxy-httpd" Dec 10 23:11:09 crc kubenswrapper[4791]: I1210 23:11:09.131641 4791 state_mem.go:107] "Deleted CPUSet assignment" podUID="65afcf68-16b8-4f9d-ab77-ab910ca9469b" containerName="proxy-httpd" Dec 10 23:11:09 crc kubenswrapper[4791]: E1210 23:11:09.131653 4791 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="65afcf68-16b8-4f9d-ab77-ab910ca9469b" containerName="ceilometer-notification-agent" Dec 10 23:11:09 crc kubenswrapper[4791]: I1210 23:11:09.131659 4791 state_mem.go:107] "Deleted CPUSet assignment" podUID="65afcf68-16b8-4f9d-ab77-ab910ca9469b" containerName="ceilometer-notification-agent" Dec 10 23:11:09 crc kubenswrapper[4791]: I1210 23:11:09.131854 4791 memory_manager.go:354] "RemoveStaleState removing state" podUID="65afcf68-16b8-4f9d-ab77-ab910ca9469b" containerName="ceilometer-notification-agent" Dec 10 23:11:09 crc kubenswrapper[4791]: I1210 23:11:09.131872 4791 memory_manager.go:354] "RemoveStaleState removing state" podUID="65afcf68-16b8-4f9d-ab77-ab910ca9469b" containerName="sg-core" Dec 10 23:11:09 crc kubenswrapper[4791]: I1210 23:11:09.131888 4791 memory_manager.go:354] "RemoveStaleState removing state" podUID="65afcf68-16b8-4f9d-ab77-ab910ca9469b" containerName="proxy-httpd" Dec 10 23:11:09 crc kubenswrapper[4791]: I1210 23:11:09.131904 4791 memory_manager.go:354] "RemoveStaleState removing state" podUID="65afcf68-16b8-4f9d-ab77-ab910ca9469b" containerName="ceilometer-central-agent" Dec 10 23:11:09 crc kubenswrapper[4791]: I1210 23:11:09.133818 4791 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Dec 10 23:11:09 crc kubenswrapper[4791]: I1210 23:11:09.136815 4791 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-scripts" Dec 10 23:11:09 crc kubenswrapper[4791]: I1210 23:11:09.136928 4791 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-ceilometer-internal-svc" Dec 10 23:11:09 crc kubenswrapper[4791]: I1210 23:11:09.137090 4791 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-config-data" Dec 10 23:11:09 crc kubenswrapper[4791]: I1210 23:11:09.138939 4791 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceilometer-0"] Dec 10 23:11:09 crc kubenswrapper[4791]: I1210 23:11:09.147007 4791 scope.go:117] "RemoveContainer" containerID="6f1983af270707fe501a3d4c71341f8755d1b9f230aebf1ddfe48385bd734154" Dec 10 23:11:09 crc kubenswrapper[4791]: I1210 23:11:09.196718 4791 scope.go:117] "RemoveContainer" containerID="3a5559784c422a1e8dd3f804e260a986d9b1c7dac2ccb1aea0c905532562cfd4" Dec 10 23:11:09 crc kubenswrapper[4791]: I1210 23:11:09.248122 4791 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/36c779d8-1b0f-4d7e-94ca-a05bda0b4ee8-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"36c779d8-1b0f-4d7e-94ca-a05bda0b4ee8\") " pod="openstack/ceilometer-0" Dec 10 23:11:09 crc kubenswrapper[4791]: I1210 23:11:09.248375 4791 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ceilometer-tls-certs\" (UniqueName: \"kubernetes.io/secret/36c779d8-1b0f-4d7e-94ca-a05bda0b4ee8-ceilometer-tls-certs\") pod \"ceilometer-0\" (UID: \"36c779d8-1b0f-4d7e-94ca-a05bda0b4ee8\") " pod="openstack/ceilometer-0" Dec 10 23:11:09 crc kubenswrapper[4791]: I1210 23:11:09.248446 4791 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-rhtwc\" (UniqueName: \"kubernetes.io/projected/36c779d8-1b0f-4d7e-94ca-a05bda0b4ee8-kube-api-access-rhtwc\") pod \"ceilometer-0\" (UID: \"36c779d8-1b0f-4d7e-94ca-a05bda0b4ee8\") " pod="openstack/ceilometer-0" Dec 10 23:11:09 crc kubenswrapper[4791]: I1210 23:11:09.248682 4791 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/36c779d8-1b0f-4d7e-94ca-a05bda0b4ee8-scripts\") pod \"ceilometer-0\" (UID: \"36c779d8-1b0f-4d7e-94ca-a05bda0b4ee8\") " pod="openstack/ceilometer-0" Dec 10 23:11:09 crc kubenswrapper[4791]: I1210 23:11:09.248716 4791 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/36c779d8-1b0f-4d7e-94ca-a05bda0b4ee8-log-httpd\") pod \"ceilometer-0\" (UID: \"36c779d8-1b0f-4d7e-94ca-a05bda0b4ee8\") " pod="openstack/ceilometer-0" Dec 10 23:11:09 crc kubenswrapper[4791]: I1210 23:11:09.248796 4791 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/36c779d8-1b0f-4d7e-94ca-a05bda0b4ee8-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"36c779d8-1b0f-4d7e-94ca-a05bda0b4ee8\") " pod="openstack/ceilometer-0" Dec 10 23:11:09 crc kubenswrapper[4791]: I1210 23:11:09.248999 4791 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/36c779d8-1b0f-4d7e-94ca-a05bda0b4ee8-config-data\") pod \"ceilometer-0\" (UID: \"36c779d8-1b0f-4d7e-94ca-a05bda0b4ee8\") " pod="openstack/ceilometer-0" Dec 10 23:11:09 crc kubenswrapper[4791]: I1210 23:11:09.249068 4791 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/36c779d8-1b0f-4d7e-94ca-a05bda0b4ee8-run-httpd\") pod \"ceilometer-0\" (UID: \"36c779d8-1b0f-4d7e-94ca-a05bda0b4ee8\") " pod="openstack/ceilometer-0" Dec 10 23:11:09 crc kubenswrapper[4791]: I1210 23:11:09.350689 4791 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-rhtwc\" (UniqueName: \"kubernetes.io/projected/36c779d8-1b0f-4d7e-94ca-a05bda0b4ee8-kube-api-access-rhtwc\") pod \"ceilometer-0\" (UID: \"36c779d8-1b0f-4d7e-94ca-a05bda0b4ee8\") " pod="openstack/ceilometer-0" Dec 10 23:11:09 crc kubenswrapper[4791]: I1210 23:11:09.350785 4791 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/36c779d8-1b0f-4d7e-94ca-a05bda0b4ee8-scripts\") pod \"ceilometer-0\" (UID: \"36c779d8-1b0f-4d7e-94ca-a05bda0b4ee8\") " pod="openstack/ceilometer-0" Dec 10 23:11:09 crc kubenswrapper[4791]: I1210 23:11:09.350834 4791 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/36c779d8-1b0f-4d7e-94ca-a05bda0b4ee8-log-httpd\") pod \"ceilometer-0\" (UID: \"36c779d8-1b0f-4d7e-94ca-a05bda0b4ee8\") " pod="openstack/ceilometer-0" Dec 10 23:11:09 crc kubenswrapper[4791]: I1210 23:11:09.350866 4791 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/36c779d8-1b0f-4d7e-94ca-a05bda0b4ee8-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"36c779d8-1b0f-4d7e-94ca-a05bda0b4ee8\") " pod="openstack/ceilometer-0" Dec 10 23:11:09 crc kubenswrapper[4791]: I1210 23:11:09.350931 4791 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/36c779d8-1b0f-4d7e-94ca-a05bda0b4ee8-config-data\") pod \"ceilometer-0\" (UID: \"36c779d8-1b0f-4d7e-94ca-a05bda0b4ee8\") " pod="openstack/ceilometer-0" Dec 10 23:11:09 crc kubenswrapper[4791]: I1210 23:11:09.350956 4791 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/36c779d8-1b0f-4d7e-94ca-a05bda0b4ee8-run-httpd\") pod \"ceilometer-0\" (UID: \"36c779d8-1b0f-4d7e-94ca-a05bda0b4ee8\") " pod="openstack/ceilometer-0" Dec 10 23:11:09 crc kubenswrapper[4791]: I1210 23:11:09.351009 4791 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/36c779d8-1b0f-4d7e-94ca-a05bda0b4ee8-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"36c779d8-1b0f-4d7e-94ca-a05bda0b4ee8\") " pod="openstack/ceilometer-0" Dec 10 23:11:09 crc kubenswrapper[4791]: I1210 23:11:09.351087 4791 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ceilometer-tls-certs\" (UniqueName: \"kubernetes.io/secret/36c779d8-1b0f-4d7e-94ca-a05bda0b4ee8-ceilometer-tls-certs\") pod \"ceilometer-0\" (UID: \"36c779d8-1b0f-4d7e-94ca-a05bda0b4ee8\") " pod="openstack/ceilometer-0" Dec 10 23:11:09 crc kubenswrapper[4791]: I1210 23:11:09.351381 4791 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/36c779d8-1b0f-4d7e-94ca-a05bda0b4ee8-log-httpd\") pod \"ceilometer-0\" (UID: \"36c779d8-1b0f-4d7e-94ca-a05bda0b4ee8\") " pod="openstack/ceilometer-0" Dec 10 23:11:09 crc kubenswrapper[4791]: I1210 23:11:09.352122 4791 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/36c779d8-1b0f-4d7e-94ca-a05bda0b4ee8-run-httpd\") pod \"ceilometer-0\" (UID: \"36c779d8-1b0f-4d7e-94ca-a05bda0b4ee8\") " pod="openstack/ceilometer-0" Dec 10 23:11:09 crc kubenswrapper[4791]: I1210 23:11:09.355908 4791 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/36c779d8-1b0f-4d7e-94ca-a05bda0b4ee8-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"36c779d8-1b0f-4d7e-94ca-a05bda0b4ee8\") " pod="openstack/ceilometer-0" Dec 10 23:11:09 crc kubenswrapper[4791]: I1210 23:11:09.356952 4791 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ceilometer-tls-certs\" (UniqueName: \"kubernetes.io/secret/36c779d8-1b0f-4d7e-94ca-a05bda0b4ee8-ceilometer-tls-certs\") pod \"ceilometer-0\" (UID: \"36c779d8-1b0f-4d7e-94ca-a05bda0b4ee8\") " pod="openstack/ceilometer-0" Dec 10 23:11:09 crc kubenswrapper[4791]: I1210 23:11:09.357126 4791 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/36c779d8-1b0f-4d7e-94ca-a05bda0b4ee8-config-data\") pod \"ceilometer-0\" (UID: \"36c779d8-1b0f-4d7e-94ca-a05bda0b4ee8\") " pod="openstack/ceilometer-0" Dec 10 23:11:09 crc kubenswrapper[4791]: I1210 23:11:09.363909 4791 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/36c779d8-1b0f-4d7e-94ca-a05bda0b4ee8-scripts\") pod \"ceilometer-0\" (UID: \"36c779d8-1b0f-4d7e-94ca-a05bda0b4ee8\") " pod="openstack/ceilometer-0" Dec 10 23:11:09 crc kubenswrapper[4791]: I1210 23:11:09.364378 4791 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/36c779d8-1b0f-4d7e-94ca-a05bda0b4ee8-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"36c779d8-1b0f-4d7e-94ca-a05bda0b4ee8\") " pod="openstack/ceilometer-0" Dec 10 23:11:09 crc kubenswrapper[4791]: I1210 23:11:09.375560 4791 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-rhtwc\" (UniqueName: \"kubernetes.io/projected/36c779d8-1b0f-4d7e-94ca-a05bda0b4ee8-kube-api-access-rhtwc\") pod \"ceilometer-0\" (UID: \"36c779d8-1b0f-4d7e-94ca-a05bda0b4ee8\") " pod="openstack/ceilometer-0" Dec 10 23:11:09 crc kubenswrapper[4791]: I1210 23:11:09.465698 4791 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Dec 10 23:11:09 crc kubenswrapper[4791]: I1210 23:11:09.895830 4791 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="65afcf68-16b8-4f9d-ab77-ab910ca9469b" path="/var/lib/kubelet/pods/65afcf68-16b8-4f9d-ab77-ab910ca9469b/volumes" Dec 10 23:11:09 crc kubenswrapper[4791]: I1210 23:11:09.913215 4791 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceilometer-0"] Dec 10 23:11:09 crc kubenswrapper[4791]: W1210 23:11:09.916391 4791 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod36c779d8_1b0f_4d7e_94ca_a05bda0b4ee8.slice/crio-5c9ac83bdac520a40246ebf6e5c87a217e53983f738e170e9314eebc7b756041 WatchSource:0}: Error finding container 5c9ac83bdac520a40246ebf6e5c87a217e53983f738e170e9314eebc7b756041: Status 404 returned error can't find the container with id 5c9ac83bdac520a40246ebf6e5c87a217e53983f738e170e9314eebc7b756041 Dec 10 23:11:10 crc kubenswrapper[4791]: I1210 23:11:10.059606 4791 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"36c779d8-1b0f-4d7e-94ca-a05bda0b4ee8","Type":"ContainerStarted","Data":"5c9ac83bdac520a40246ebf6e5c87a217e53983f738e170e9314eebc7b756041"} Dec 10 23:11:10 crc kubenswrapper[4791]: I1210 23:11:10.248614 4791 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/nova-metadata-0" Dec 10 23:11:10 crc kubenswrapper[4791]: I1210 23:11:10.250102 4791 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/nova-metadata-0" Dec 10 23:11:10 crc kubenswrapper[4791]: I1210 23:11:10.256553 4791 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-metadata-0" Dec 10 23:11:11 crc kubenswrapper[4791]: I1210 23:11:11.071110 4791 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"36c779d8-1b0f-4d7e-94ca-a05bda0b4ee8","Type":"ContainerStarted","Data":"bd7013d70d1fb7fb35a7924c0c7006151aa72907faa8474540bc5912b7de09a2"} Dec 10 23:11:11 crc kubenswrapper[4791]: I1210 23:11:11.075697 4791 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-metadata-0" Dec 10 23:11:11 crc kubenswrapper[4791]: I1210 23:11:11.837929 4791 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-novncproxy-0" Dec 10 23:11:11 crc kubenswrapper[4791]: I1210 23:11:11.910209 4791 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/e2d1527a-aac6-4ee0-9c41-60ef0784aa2a-combined-ca-bundle\") pod \"e2d1527a-aac6-4ee0-9c41-60ef0784aa2a\" (UID: \"e2d1527a-aac6-4ee0-9c41-60ef0784aa2a\") " Dec 10 23:11:11 crc kubenswrapper[4791]: I1210 23:11:11.910294 4791 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-gdgcq\" (UniqueName: \"kubernetes.io/projected/e2d1527a-aac6-4ee0-9c41-60ef0784aa2a-kube-api-access-gdgcq\") pod \"e2d1527a-aac6-4ee0-9c41-60ef0784aa2a\" (UID: \"e2d1527a-aac6-4ee0-9c41-60ef0784aa2a\") " Dec 10 23:11:11 crc kubenswrapper[4791]: I1210 23:11:11.910624 4791 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/e2d1527a-aac6-4ee0-9c41-60ef0784aa2a-config-data\") pod \"e2d1527a-aac6-4ee0-9c41-60ef0784aa2a\" (UID: \"e2d1527a-aac6-4ee0-9c41-60ef0784aa2a\") " Dec 10 23:11:11 crc kubenswrapper[4791]: I1210 23:11:11.917916 4791 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/e2d1527a-aac6-4ee0-9c41-60ef0784aa2a-kube-api-access-gdgcq" (OuterVolumeSpecName: "kube-api-access-gdgcq") pod "e2d1527a-aac6-4ee0-9c41-60ef0784aa2a" (UID: "e2d1527a-aac6-4ee0-9c41-60ef0784aa2a"). InnerVolumeSpecName "kube-api-access-gdgcq". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 10 23:11:11 crc kubenswrapper[4791]: I1210 23:11:11.947974 4791 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/e2d1527a-aac6-4ee0-9c41-60ef0784aa2a-config-data" (OuterVolumeSpecName: "config-data") pod "e2d1527a-aac6-4ee0-9c41-60ef0784aa2a" (UID: "e2d1527a-aac6-4ee0-9c41-60ef0784aa2a"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 10 23:11:11 crc kubenswrapper[4791]: I1210 23:11:11.969090 4791 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/e2d1527a-aac6-4ee0-9c41-60ef0784aa2a-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "e2d1527a-aac6-4ee0-9c41-60ef0784aa2a" (UID: "e2d1527a-aac6-4ee0-9c41-60ef0784aa2a"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 10 23:11:12 crc kubenswrapper[4791]: I1210 23:11:12.013215 4791 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/e2d1527a-aac6-4ee0-9c41-60ef0784aa2a-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 10 23:11:12 crc kubenswrapper[4791]: I1210 23:11:12.013444 4791 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-gdgcq\" (UniqueName: \"kubernetes.io/projected/e2d1527a-aac6-4ee0-9c41-60ef0784aa2a-kube-api-access-gdgcq\") on node \"crc\" DevicePath \"\"" Dec 10 23:11:12 crc kubenswrapper[4791]: I1210 23:11:12.013583 4791 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/e2d1527a-aac6-4ee0-9c41-60ef0784aa2a-config-data\") on node \"crc\" DevicePath \"\"" Dec 10 23:11:12 crc kubenswrapper[4791]: I1210 23:11:12.080375 4791 generic.go:334] "Generic (PLEG): container finished" podID="e2d1527a-aac6-4ee0-9c41-60ef0784aa2a" containerID="825d750a33ded08f66510a0c44cef8a3f541452ccad46a74bd1a0a7e72513c5d" exitCode=137 Dec 10 23:11:12 crc kubenswrapper[4791]: I1210 23:11:12.080421 4791 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-novncproxy-0" Dec 10 23:11:12 crc kubenswrapper[4791]: I1210 23:11:12.080446 4791 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-novncproxy-0" event={"ID":"e2d1527a-aac6-4ee0-9c41-60ef0784aa2a","Type":"ContainerDied","Data":"825d750a33ded08f66510a0c44cef8a3f541452ccad46a74bd1a0a7e72513c5d"} Dec 10 23:11:12 crc kubenswrapper[4791]: I1210 23:11:12.080993 4791 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-novncproxy-0" event={"ID":"e2d1527a-aac6-4ee0-9c41-60ef0784aa2a","Type":"ContainerDied","Data":"1010cab3c0202dd1630d3aaf7a72e0d3e82b2e33edbfd1d44a507070c7b291b4"} Dec 10 23:11:12 crc kubenswrapper[4791]: I1210 23:11:12.081012 4791 scope.go:117] "RemoveContainer" containerID="825d750a33ded08f66510a0c44cef8a3f541452ccad46a74bd1a0a7e72513c5d" Dec 10 23:11:12 crc kubenswrapper[4791]: I1210 23:11:12.084298 4791 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"36c779d8-1b0f-4d7e-94ca-a05bda0b4ee8","Type":"ContainerStarted","Data":"517279520b8563ef48642d1a1f8a4fcc59abeb2393de710719a9f748d5ab9a24"} Dec 10 23:11:12 crc kubenswrapper[4791]: I1210 23:11:12.120261 4791 scope.go:117] "RemoveContainer" containerID="825d750a33ded08f66510a0c44cef8a3f541452ccad46a74bd1a0a7e72513c5d" Dec 10 23:11:12 crc kubenswrapper[4791]: E1210 23:11:12.120978 4791 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"825d750a33ded08f66510a0c44cef8a3f541452ccad46a74bd1a0a7e72513c5d\": container with ID starting with 825d750a33ded08f66510a0c44cef8a3f541452ccad46a74bd1a0a7e72513c5d not found: ID does not exist" containerID="825d750a33ded08f66510a0c44cef8a3f541452ccad46a74bd1a0a7e72513c5d" Dec 10 23:11:12 crc kubenswrapper[4791]: I1210 23:11:12.121011 4791 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"825d750a33ded08f66510a0c44cef8a3f541452ccad46a74bd1a0a7e72513c5d"} err="failed to get container status \"825d750a33ded08f66510a0c44cef8a3f541452ccad46a74bd1a0a7e72513c5d\": rpc error: code = NotFound desc = could not find container \"825d750a33ded08f66510a0c44cef8a3f541452ccad46a74bd1a0a7e72513c5d\": container with ID starting with 825d750a33ded08f66510a0c44cef8a3f541452ccad46a74bd1a0a7e72513c5d not found: ID does not exist" Dec 10 23:11:12 crc kubenswrapper[4791]: I1210 23:11:12.129439 4791 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-cell1-novncproxy-0"] Dec 10 23:11:12 crc kubenswrapper[4791]: I1210 23:11:12.142763 4791 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-cell1-novncproxy-0"] Dec 10 23:11:12 crc kubenswrapper[4791]: I1210 23:11:12.164319 4791 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-cell1-novncproxy-0"] Dec 10 23:11:12 crc kubenswrapper[4791]: E1210 23:11:12.164811 4791 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="e2d1527a-aac6-4ee0-9c41-60ef0784aa2a" containerName="nova-cell1-novncproxy-novncproxy" Dec 10 23:11:12 crc kubenswrapper[4791]: I1210 23:11:12.164827 4791 state_mem.go:107] "Deleted CPUSet assignment" podUID="e2d1527a-aac6-4ee0-9c41-60ef0784aa2a" containerName="nova-cell1-novncproxy-novncproxy" Dec 10 23:11:12 crc kubenswrapper[4791]: I1210 23:11:12.165094 4791 memory_manager.go:354] "RemoveStaleState removing state" podUID="e2d1527a-aac6-4ee0-9c41-60ef0784aa2a" containerName="nova-cell1-novncproxy-novncproxy" Dec 10 23:11:12 crc kubenswrapper[4791]: I1210 23:11:12.165853 4791 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-novncproxy-0" Dec 10 23:11:12 crc kubenswrapper[4791]: I1210 23:11:12.167456 4791 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-nova-novncproxy-cell1-public-svc" Dec 10 23:11:12 crc kubenswrapper[4791]: I1210 23:11:12.168958 4791 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-nova-novncproxy-cell1-vencrypt" Dec 10 23:11:12 crc kubenswrapper[4791]: I1210 23:11:12.169163 4791 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell1-novncproxy-config-data" Dec 10 23:11:12 crc kubenswrapper[4791]: I1210 23:11:12.184400 4791 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-novncproxy-0"] Dec 10 23:11:12 crc kubenswrapper[4791]: I1210 23:11:12.216162 4791 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"nova-novncproxy-tls-certs\" (UniqueName: \"kubernetes.io/secret/4c145576-0efc-440b-bc4d-5468995305b4-nova-novncproxy-tls-certs\") pod \"nova-cell1-novncproxy-0\" (UID: \"4c145576-0efc-440b-bc4d-5468995305b4\") " pod="openstack/nova-cell1-novncproxy-0" Dec 10 23:11:12 crc kubenswrapper[4791]: I1210 23:11:12.216227 4791 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"vencrypt-tls-certs\" (UniqueName: \"kubernetes.io/secret/4c145576-0efc-440b-bc4d-5468995305b4-vencrypt-tls-certs\") pod \"nova-cell1-novncproxy-0\" (UID: \"4c145576-0efc-440b-bc4d-5468995305b4\") " pod="openstack/nova-cell1-novncproxy-0" Dec 10 23:11:12 crc kubenswrapper[4791]: I1210 23:11:12.216283 4791 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/4c145576-0efc-440b-bc4d-5468995305b4-config-data\") pod \"nova-cell1-novncproxy-0\" (UID: \"4c145576-0efc-440b-bc4d-5468995305b4\") " pod="openstack/nova-cell1-novncproxy-0" Dec 10 23:11:12 crc kubenswrapper[4791]: I1210 23:11:12.216372 4791 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-mmvf2\" (UniqueName: \"kubernetes.io/projected/4c145576-0efc-440b-bc4d-5468995305b4-kube-api-access-mmvf2\") pod \"nova-cell1-novncproxy-0\" (UID: \"4c145576-0efc-440b-bc4d-5468995305b4\") " pod="openstack/nova-cell1-novncproxy-0" Dec 10 23:11:12 crc kubenswrapper[4791]: I1210 23:11:12.216406 4791 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/4c145576-0efc-440b-bc4d-5468995305b4-combined-ca-bundle\") pod \"nova-cell1-novncproxy-0\" (UID: \"4c145576-0efc-440b-bc4d-5468995305b4\") " pod="openstack/nova-cell1-novncproxy-0" Dec 10 23:11:12 crc kubenswrapper[4791]: I1210 23:11:12.290076 4791 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/nova-api-0" Dec 10 23:11:12 crc kubenswrapper[4791]: I1210 23:11:12.290634 4791 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-api-0" Dec 10 23:11:12 crc kubenswrapper[4791]: I1210 23:11:12.290742 4791 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/nova-api-0" Dec 10 23:11:12 crc kubenswrapper[4791]: I1210 23:11:12.293565 4791 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-api-0" Dec 10 23:11:12 crc kubenswrapper[4791]: I1210 23:11:12.317725 4791 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nova-novncproxy-tls-certs\" (UniqueName: \"kubernetes.io/secret/4c145576-0efc-440b-bc4d-5468995305b4-nova-novncproxy-tls-certs\") pod \"nova-cell1-novncproxy-0\" (UID: \"4c145576-0efc-440b-bc4d-5468995305b4\") " pod="openstack/nova-cell1-novncproxy-0" Dec 10 23:11:12 crc kubenswrapper[4791]: I1210 23:11:12.317786 4791 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"vencrypt-tls-certs\" (UniqueName: \"kubernetes.io/secret/4c145576-0efc-440b-bc4d-5468995305b4-vencrypt-tls-certs\") pod \"nova-cell1-novncproxy-0\" (UID: \"4c145576-0efc-440b-bc4d-5468995305b4\") " pod="openstack/nova-cell1-novncproxy-0" Dec 10 23:11:12 crc kubenswrapper[4791]: I1210 23:11:12.317840 4791 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/4c145576-0efc-440b-bc4d-5468995305b4-config-data\") pod \"nova-cell1-novncproxy-0\" (UID: \"4c145576-0efc-440b-bc4d-5468995305b4\") " pod="openstack/nova-cell1-novncproxy-0" Dec 10 23:11:12 crc kubenswrapper[4791]: I1210 23:11:12.317893 4791 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-mmvf2\" (UniqueName: \"kubernetes.io/projected/4c145576-0efc-440b-bc4d-5468995305b4-kube-api-access-mmvf2\") pod \"nova-cell1-novncproxy-0\" (UID: \"4c145576-0efc-440b-bc4d-5468995305b4\") " pod="openstack/nova-cell1-novncproxy-0" Dec 10 23:11:12 crc kubenswrapper[4791]: I1210 23:11:12.317925 4791 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/4c145576-0efc-440b-bc4d-5468995305b4-combined-ca-bundle\") pod \"nova-cell1-novncproxy-0\" (UID: \"4c145576-0efc-440b-bc4d-5468995305b4\") " pod="openstack/nova-cell1-novncproxy-0" Dec 10 23:11:12 crc kubenswrapper[4791]: I1210 23:11:12.323013 4791 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"nova-novncproxy-tls-certs\" (UniqueName: \"kubernetes.io/secret/4c145576-0efc-440b-bc4d-5468995305b4-nova-novncproxy-tls-certs\") pod \"nova-cell1-novncproxy-0\" (UID: \"4c145576-0efc-440b-bc4d-5468995305b4\") " pod="openstack/nova-cell1-novncproxy-0" Dec 10 23:11:12 crc kubenswrapper[4791]: I1210 23:11:12.323101 4791 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"vencrypt-tls-certs\" (UniqueName: \"kubernetes.io/secret/4c145576-0efc-440b-bc4d-5468995305b4-vencrypt-tls-certs\") pod \"nova-cell1-novncproxy-0\" (UID: \"4c145576-0efc-440b-bc4d-5468995305b4\") " pod="openstack/nova-cell1-novncproxy-0" Dec 10 23:11:12 crc kubenswrapper[4791]: I1210 23:11:12.335094 4791 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/4c145576-0efc-440b-bc4d-5468995305b4-combined-ca-bundle\") pod \"nova-cell1-novncproxy-0\" (UID: \"4c145576-0efc-440b-bc4d-5468995305b4\") " pod="openstack/nova-cell1-novncproxy-0" Dec 10 23:11:12 crc kubenswrapper[4791]: I1210 23:11:12.335297 4791 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/4c145576-0efc-440b-bc4d-5468995305b4-config-data\") pod \"nova-cell1-novncproxy-0\" (UID: \"4c145576-0efc-440b-bc4d-5468995305b4\") " pod="openstack/nova-cell1-novncproxy-0" Dec 10 23:11:12 crc kubenswrapper[4791]: I1210 23:11:12.343603 4791 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-mmvf2\" (UniqueName: \"kubernetes.io/projected/4c145576-0efc-440b-bc4d-5468995305b4-kube-api-access-mmvf2\") pod \"nova-cell1-novncproxy-0\" (UID: \"4c145576-0efc-440b-bc4d-5468995305b4\") " pod="openstack/nova-cell1-novncproxy-0" Dec 10 23:11:12 crc kubenswrapper[4791]: I1210 23:11:12.508163 4791 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-novncproxy-0" Dec 10 23:11:13 crc kubenswrapper[4791]: W1210 23:11:13.000105 4791 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod4c145576_0efc_440b_bc4d_5468995305b4.slice/crio-32090b3f6984b984bbf65bd2afcbc6e18c619c0b5ad92e5840ae3a7aecc08598 WatchSource:0}: Error finding container 32090b3f6984b984bbf65bd2afcbc6e18c619c0b5ad92e5840ae3a7aecc08598: Status 404 returned error can't find the container with id 32090b3f6984b984bbf65bd2afcbc6e18c619c0b5ad92e5840ae3a7aecc08598 Dec 10 23:11:13 crc kubenswrapper[4791]: I1210 23:11:13.003910 4791 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-novncproxy-0"] Dec 10 23:11:13 crc kubenswrapper[4791]: I1210 23:11:13.097855 4791 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-novncproxy-0" event={"ID":"4c145576-0efc-440b-bc4d-5468995305b4","Type":"ContainerStarted","Data":"32090b3f6984b984bbf65bd2afcbc6e18c619c0b5ad92e5840ae3a7aecc08598"} Dec 10 23:11:13 crc kubenswrapper[4791]: I1210 23:11:13.101856 4791 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"36c779d8-1b0f-4d7e-94ca-a05bda0b4ee8","Type":"ContainerStarted","Data":"4adc26eeae2e46577d7aee282669cf4154c110325b7c53f6a0f5089815dba157"} Dec 10 23:11:13 crc kubenswrapper[4791]: I1210 23:11:13.104558 4791 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-api-0" Dec 10 23:11:13 crc kubenswrapper[4791]: I1210 23:11:13.108237 4791 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-api-0" Dec 10 23:11:13 crc kubenswrapper[4791]: I1210 23:11:13.288923 4791 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-59cf4bdb65-p4l85"] Dec 10 23:11:13 crc kubenswrapper[4791]: I1210 23:11:13.306755 4791 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-59cf4bdb65-p4l85"] Dec 10 23:11:13 crc kubenswrapper[4791]: I1210 23:11:13.306900 4791 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-59cf4bdb65-p4l85" Dec 10 23:11:13 crc kubenswrapper[4791]: I1210 23:11:13.350160 4791 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/339ea8c9-715f-4394-ace8-6b66542bcc13-dns-svc\") pod \"dnsmasq-dns-59cf4bdb65-p4l85\" (UID: \"339ea8c9-715f-4394-ace8-6b66542bcc13\") " pod="openstack/dnsmasq-dns-59cf4bdb65-p4l85" Dec 10 23:11:13 crc kubenswrapper[4791]: I1210 23:11:13.350234 4791 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/339ea8c9-715f-4394-ace8-6b66542bcc13-ovsdbserver-sb\") pod \"dnsmasq-dns-59cf4bdb65-p4l85\" (UID: \"339ea8c9-715f-4394-ace8-6b66542bcc13\") " pod="openstack/dnsmasq-dns-59cf4bdb65-p4l85" Dec 10 23:11:13 crc kubenswrapper[4791]: I1210 23:11:13.350294 4791 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/339ea8c9-715f-4394-ace8-6b66542bcc13-ovsdbserver-nb\") pod \"dnsmasq-dns-59cf4bdb65-p4l85\" (UID: \"339ea8c9-715f-4394-ace8-6b66542bcc13\") " pod="openstack/dnsmasq-dns-59cf4bdb65-p4l85" Dec 10 23:11:13 crc kubenswrapper[4791]: I1210 23:11:13.350355 4791 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/339ea8c9-715f-4394-ace8-6b66542bcc13-config\") pod \"dnsmasq-dns-59cf4bdb65-p4l85\" (UID: \"339ea8c9-715f-4394-ace8-6b66542bcc13\") " pod="openstack/dnsmasq-dns-59cf4bdb65-p4l85" Dec 10 23:11:13 crc kubenswrapper[4791]: I1210 23:11:13.350392 4791 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/339ea8c9-715f-4394-ace8-6b66542bcc13-dns-swift-storage-0\") pod \"dnsmasq-dns-59cf4bdb65-p4l85\" (UID: \"339ea8c9-715f-4394-ace8-6b66542bcc13\") " pod="openstack/dnsmasq-dns-59cf4bdb65-p4l85" Dec 10 23:11:13 crc kubenswrapper[4791]: I1210 23:11:13.350519 4791 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-q6229\" (UniqueName: \"kubernetes.io/projected/339ea8c9-715f-4394-ace8-6b66542bcc13-kube-api-access-q6229\") pod \"dnsmasq-dns-59cf4bdb65-p4l85\" (UID: \"339ea8c9-715f-4394-ace8-6b66542bcc13\") " pod="openstack/dnsmasq-dns-59cf4bdb65-p4l85" Dec 10 23:11:13 crc kubenswrapper[4791]: I1210 23:11:13.452241 4791 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/339ea8c9-715f-4394-ace8-6b66542bcc13-ovsdbserver-nb\") pod \"dnsmasq-dns-59cf4bdb65-p4l85\" (UID: \"339ea8c9-715f-4394-ace8-6b66542bcc13\") " pod="openstack/dnsmasq-dns-59cf4bdb65-p4l85" Dec 10 23:11:13 crc kubenswrapper[4791]: I1210 23:11:13.452299 4791 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/339ea8c9-715f-4394-ace8-6b66542bcc13-config\") pod \"dnsmasq-dns-59cf4bdb65-p4l85\" (UID: \"339ea8c9-715f-4394-ace8-6b66542bcc13\") " pod="openstack/dnsmasq-dns-59cf4bdb65-p4l85" Dec 10 23:11:13 crc kubenswrapper[4791]: I1210 23:11:13.452324 4791 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/339ea8c9-715f-4394-ace8-6b66542bcc13-dns-swift-storage-0\") pod \"dnsmasq-dns-59cf4bdb65-p4l85\" (UID: \"339ea8c9-715f-4394-ace8-6b66542bcc13\") " pod="openstack/dnsmasq-dns-59cf4bdb65-p4l85" Dec 10 23:11:13 crc kubenswrapper[4791]: I1210 23:11:13.452421 4791 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-q6229\" (UniqueName: \"kubernetes.io/projected/339ea8c9-715f-4394-ace8-6b66542bcc13-kube-api-access-q6229\") pod \"dnsmasq-dns-59cf4bdb65-p4l85\" (UID: \"339ea8c9-715f-4394-ace8-6b66542bcc13\") " pod="openstack/dnsmasq-dns-59cf4bdb65-p4l85" Dec 10 23:11:13 crc kubenswrapper[4791]: I1210 23:11:13.452493 4791 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/339ea8c9-715f-4394-ace8-6b66542bcc13-dns-svc\") pod \"dnsmasq-dns-59cf4bdb65-p4l85\" (UID: \"339ea8c9-715f-4394-ace8-6b66542bcc13\") " pod="openstack/dnsmasq-dns-59cf4bdb65-p4l85" Dec 10 23:11:13 crc kubenswrapper[4791]: I1210 23:11:13.452517 4791 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/339ea8c9-715f-4394-ace8-6b66542bcc13-ovsdbserver-sb\") pod \"dnsmasq-dns-59cf4bdb65-p4l85\" (UID: \"339ea8c9-715f-4394-ace8-6b66542bcc13\") " pod="openstack/dnsmasq-dns-59cf4bdb65-p4l85" Dec 10 23:11:13 crc kubenswrapper[4791]: I1210 23:11:13.453714 4791 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/339ea8c9-715f-4394-ace8-6b66542bcc13-config\") pod \"dnsmasq-dns-59cf4bdb65-p4l85\" (UID: \"339ea8c9-715f-4394-ace8-6b66542bcc13\") " pod="openstack/dnsmasq-dns-59cf4bdb65-p4l85" Dec 10 23:11:13 crc kubenswrapper[4791]: I1210 23:11:13.453729 4791 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/339ea8c9-715f-4394-ace8-6b66542bcc13-ovsdbserver-sb\") pod \"dnsmasq-dns-59cf4bdb65-p4l85\" (UID: \"339ea8c9-715f-4394-ace8-6b66542bcc13\") " pod="openstack/dnsmasq-dns-59cf4bdb65-p4l85" Dec 10 23:11:13 crc kubenswrapper[4791]: I1210 23:11:13.453730 4791 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/339ea8c9-715f-4394-ace8-6b66542bcc13-dns-svc\") pod \"dnsmasq-dns-59cf4bdb65-p4l85\" (UID: \"339ea8c9-715f-4394-ace8-6b66542bcc13\") " pod="openstack/dnsmasq-dns-59cf4bdb65-p4l85" Dec 10 23:11:13 crc kubenswrapper[4791]: I1210 23:11:13.453733 4791 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/339ea8c9-715f-4394-ace8-6b66542bcc13-dns-swift-storage-0\") pod \"dnsmasq-dns-59cf4bdb65-p4l85\" (UID: \"339ea8c9-715f-4394-ace8-6b66542bcc13\") " pod="openstack/dnsmasq-dns-59cf4bdb65-p4l85" Dec 10 23:11:13 crc kubenswrapper[4791]: I1210 23:11:13.454936 4791 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/339ea8c9-715f-4394-ace8-6b66542bcc13-ovsdbserver-nb\") pod \"dnsmasq-dns-59cf4bdb65-p4l85\" (UID: \"339ea8c9-715f-4394-ace8-6b66542bcc13\") " pod="openstack/dnsmasq-dns-59cf4bdb65-p4l85" Dec 10 23:11:13 crc kubenswrapper[4791]: I1210 23:11:13.482067 4791 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-q6229\" (UniqueName: \"kubernetes.io/projected/339ea8c9-715f-4394-ace8-6b66542bcc13-kube-api-access-q6229\") pod \"dnsmasq-dns-59cf4bdb65-p4l85\" (UID: \"339ea8c9-715f-4394-ace8-6b66542bcc13\") " pod="openstack/dnsmasq-dns-59cf4bdb65-p4l85" Dec 10 23:11:13 crc kubenswrapper[4791]: I1210 23:11:13.626035 4791 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-59cf4bdb65-p4l85" Dec 10 23:11:13 crc kubenswrapper[4791]: I1210 23:11:13.899275 4791 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="e2d1527a-aac6-4ee0-9c41-60ef0784aa2a" path="/var/lib/kubelet/pods/e2d1527a-aac6-4ee0-9c41-60ef0784aa2a/volumes" Dec 10 23:11:14 crc kubenswrapper[4791]: I1210 23:11:14.098740 4791 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-59cf4bdb65-p4l85"] Dec 10 23:11:14 crc kubenswrapper[4791]: W1210 23:11:14.100951 4791 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod339ea8c9_715f_4394_ace8_6b66542bcc13.slice/crio-b6441e8d9725e479158a59e0b569853fe71039394deb10eae1e4118ea59fe783 WatchSource:0}: Error finding container b6441e8d9725e479158a59e0b569853fe71039394deb10eae1e4118ea59fe783: Status 404 returned error can't find the container with id b6441e8d9725e479158a59e0b569853fe71039394deb10eae1e4118ea59fe783 Dec 10 23:11:14 crc kubenswrapper[4791]: I1210 23:11:14.113531 4791 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-novncproxy-0" event={"ID":"4c145576-0efc-440b-bc4d-5468995305b4","Type":"ContainerStarted","Data":"110da40e132d6a4e7fbb3ddb4fe910a003670772f3c9bd7af22eea7b211650c2"} Dec 10 23:11:14 crc kubenswrapper[4791]: I1210 23:11:14.115357 4791 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-59cf4bdb65-p4l85" event={"ID":"339ea8c9-715f-4394-ace8-6b66542bcc13","Type":"ContainerStarted","Data":"b6441e8d9725e479158a59e0b569853fe71039394deb10eae1e4118ea59fe783"} Dec 10 23:11:14 crc kubenswrapper[4791]: I1210 23:11:14.142046 4791 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-cell1-novncproxy-0" podStartSLOduration=2.142023878 podStartE2EDuration="2.142023878s" podCreationTimestamp="2025-12-10 23:11:12 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-10 23:11:14.1348242 +0000 UTC m=+1308.564441813" watchObservedRunningTime="2025-12-10 23:11:14.142023878 +0000 UTC m=+1308.571641491" Dec 10 23:11:15 crc kubenswrapper[4791]: I1210 23:11:15.129593 4791 generic.go:334] "Generic (PLEG): container finished" podID="339ea8c9-715f-4394-ace8-6b66542bcc13" containerID="738564d5be05ed3564a50d6fb351796a769397fc4e53c48a772b76e7fde67bf7" exitCode=0 Dec 10 23:11:15 crc kubenswrapper[4791]: I1210 23:11:15.129705 4791 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-59cf4bdb65-p4l85" event={"ID":"339ea8c9-715f-4394-ace8-6b66542bcc13","Type":"ContainerDied","Data":"738564d5be05ed3564a50d6fb351796a769397fc4e53c48a772b76e7fde67bf7"} Dec 10 23:11:15 crc kubenswrapper[4791]: I1210 23:11:15.139413 4791 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"36c779d8-1b0f-4d7e-94ca-a05bda0b4ee8","Type":"ContainerStarted","Data":"73d5fbab8e8ad07fc3d37539521ec847f0d8ee31c1becd87c6730d66f610f4ac"} Dec 10 23:11:15 crc kubenswrapper[4791]: I1210 23:11:15.203042 4791 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ceilometer-0" podStartSLOduration=2.202808463 podStartE2EDuration="6.203014311s" podCreationTimestamp="2025-12-10 23:11:09 +0000 UTC" firstStartedPulling="2025-12-10 23:11:09.918867113 +0000 UTC m=+1304.348484726" lastFinishedPulling="2025-12-10 23:11:13.919072961 +0000 UTC m=+1308.348690574" observedRunningTime="2025-12-10 23:11:15.182834269 +0000 UTC m=+1309.612451892" watchObservedRunningTime="2025-12-10 23:11:15.203014311 +0000 UTC m=+1309.632631924" Dec 10 23:11:15 crc kubenswrapper[4791]: I1210 23:11:15.979971 4791 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-api-0"] Dec 10 23:11:16 crc kubenswrapper[4791]: I1210 23:11:16.152108 4791 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-59cf4bdb65-p4l85" event={"ID":"339ea8c9-715f-4394-ace8-6b66542bcc13","Type":"ContainerStarted","Data":"da15fdb47d176237ac9d5f6da99fb4dbbd6dd2db577a479864263339ab238add"} Dec 10 23:11:16 crc kubenswrapper[4791]: I1210 23:11:16.152266 4791 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-api-0" podUID="c435d469-c6b0-45a0-b61b-fd638540bf51" containerName="nova-api-log" containerID="cri-o://940bdefb6fe251888b9f0d4224ac2c43268c7a2c548cd5eec01678307a80317d" gracePeriod=30 Dec 10 23:11:16 crc kubenswrapper[4791]: I1210 23:11:16.152306 4791 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-api-0" podUID="c435d469-c6b0-45a0-b61b-fd638540bf51" containerName="nova-api-api" containerID="cri-o://c2bd07aa7c52e85c622f844ee2cd9b8761a2f8e9220b0e3214c3a421060ea90d" gracePeriod=30 Dec 10 23:11:16 crc kubenswrapper[4791]: I1210 23:11:16.152764 4791 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/ceilometer-0" Dec 10 23:11:16 crc kubenswrapper[4791]: I1210 23:11:16.179539 4791 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/dnsmasq-dns-59cf4bdb65-p4l85" podStartSLOduration=3.179513706 podStartE2EDuration="3.179513706s" podCreationTimestamp="2025-12-10 23:11:13 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-10 23:11:16.170616539 +0000 UTC m=+1310.600234162" watchObservedRunningTime="2025-12-10 23:11:16.179513706 +0000 UTC m=+1310.609131319" Dec 10 23:11:16 crc kubenswrapper[4791]: I1210 23:11:16.604468 4791 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ceilometer-0"] Dec 10 23:11:17 crc kubenswrapper[4791]: I1210 23:11:17.162941 4791 generic.go:334] "Generic (PLEG): container finished" podID="c435d469-c6b0-45a0-b61b-fd638540bf51" containerID="940bdefb6fe251888b9f0d4224ac2c43268c7a2c548cd5eec01678307a80317d" exitCode=143 Dec 10 23:11:17 crc kubenswrapper[4791]: I1210 23:11:17.163054 4791 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"c435d469-c6b0-45a0-b61b-fd638540bf51","Type":"ContainerDied","Data":"940bdefb6fe251888b9f0d4224ac2c43268c7a2c548cd5eec01678307a80317d"} Dec 10 23:11:17 crc kubenswrapper[4791]: I1210 23:11:17.163408 4791 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/dnsmasq-dns-59cf4bdb65-p4l85" Dec 10 23:11:17 crc kubenswrapper[4791]: I1210 23:11:17.509193 4791 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-cell1-novncproxy-0" Dec 10 23:11:18 crc kubenswrapper[4791]: I1210 23:11:18.171716 4791 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="36c779d8-1b0f-4d7e-94ca-a05bda0b4ee8" containerName="ceilometer-central-agent" containerID="cri-o://bd7013d70d1fb7fb35a7924c0c7006151aa72907faa8474540bc5912b7de09a2" gracePeriod=30 Dec 10 23:11:18 crc kubenswrapper[4791]: I1210 23:11:18.172402 4791 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="36c779d8-1b0f-4d7e-94ca-a05bda0b4ee8" containerName="sg-core" containerID="cri-o://4adc26eeae2e46577d7aee282669cf4154c110325b7c53f6a0f5089815dba157" gracePeriod=30 Dec 10 23:11:18 crc kubenswrapper[4791]: I1210 23:11:18.172454 4791 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="36c779d8-1b0f-4d7e-94ca-a05bda0b4ee8" containerName="proxy-httpd" containerID="cri-o://73d5fbab8e8ad07fc3d37539521ec847f0d8ee31c1becd87c6730d66f610f4ac" gracePeriod=30 Dec 10 23:11:18 crc kubenswrapper[4791]: I1210 23:11:18.172476 4791 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="36c779d8-1b0f-4d7e-94ca-a05bda0b4ee8" containerName="ceilometer-notification-agent" containerID="cri-o://517279520b8563ef48642d1a1f8a4fcc59abeb2393de710719a9f748d5ab9a24" gracePeriod=30 Dec 10 23:11:19 crc kubenswrapper[4791]: I1210 23:11:19.185827 4791 generic.go:334] "Generic (PLEG): container finished" podID="36c779d8-1b0f-4d7e-94ca-a05bda0b4ee8" containerID="73d5fbab8e8ad07fc3d37539521ec847f0d8ee31c1becd87c6730d66f610f4ac" exitCode=0 Dec 10 23:11:19 crc kubenswrapper[4791]: I1210 23:11:19.185867 4791 generic.go:334] "Generic (PLEG): container finished" podID="36c779d8-1b0f-4d7e-94ca-a05bda0b4ee8" containerID="4adc26eeae2e46577d7aee282669cf4154c110325b7c53f6a0f5089815dba157" exitCode=2 Dec 10 23:11:19 crc kubenswrapper[4791]: I1210 23:11:19.185879 4791 generic.go:334] "Generic (PLEG): container finished" podID="36c779d8-1b0f-4d7e-94ca-a05bda0b4ee8" containerID="517279520b8563ef48642d1a1f8a4fcc59abeb2393de710719a9f748d5ab9a24" exitCode=0 Dec 10 23:11:19 crc kubenswrapper[4791]: I1210 23:11:19.185878 4791 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"36c779d8-1b0f-4d7e-94ca-a05bda0b4ee8","Type":"ContainerDied","Data":"73d5fbab8e8ad07fc3d37539521ec847f0d8ee31c1becd87c6730d66f610f4ac"} Dec 10 23:11:19 crc kubenswrapper[4791]: I1210 23:11:19.186007 4791 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"36c779d8-1b0f-4d7e-94ca-a05bda0b4ee8","Type":"ContainerDied","Data":"4adc26eeae2e46577d7aee282669cf4154c110325b7c53f6a0f5089815dba157"} Dec 10 23:11:19 crc kubenswrapper[4791]: I1210 23:11:19.186018 4791 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"36c779d8-1b0f-4d7e-94ca-a05bda0b4ee8","Type":"ContainerDied","Data":"517279520b8563ef48642d1a1f8a4fcc59abeb2393de710719a9f748d5ab9a24"} Dec 10 23:11:19 crc kubenswrapper[4791]: I1210 23:11:19.692971 4791 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Dec 10 23:11:19 crc kubenswrapper[4791]: I1210 23:11:19.779236 4791 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/c435d469-c6b0-45a0-b61b-fd638540bf51-logs\") pod \"c435d469-c6b0-45a0-b61b-fd638540bf51\" (UID: \"c435d469-c6b0-45a0-b61b-fd638540bf51\") " Dec 10 23:11:19 crc kubenswrapper[4791]: I1210 23:11:19.779282 4791 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/c435d469-c6b0-45a0-b61b-fd638540bf51-config-data\") pod \"c435d469-c6b0-45a0-b61b-fd638540bf51\" (UID: \"c435d469-c6b0-45a0-b61b-fd638540bf51\") " Dec 10 23:11:19 crc kubenswrapper[4791]: I1210 23:11:19.779303 4791 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c435d469-c6b0-45a0-b61b-fd638540bf51-combined-ca-bundle\") pod \"c435d469-c6b0-45a0-b61b-fd638540bf51\" (UID: \"c435d469-c6b0-45a0-b61b-fd638540bf51\") " Dec 10 23:11:19 crc kubenswrapper[4791]: I1210 23:11:19.779493 4791 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-5ggxm\" (UniqueName: \"kubernetes.io/projected/c435d469-c6b0-45a0-b61b-fd638540bf51-kube-api-access-5ggxm\") pod \"c435d469-c6b0-45a0-b61b-fd638540bf51\" (UID: \"c435d469-c6b0-45a0-b61b-fd638540bf51\") " Dec 10 23:11:19 crc kubenswrapper[4791]: I1210 23:11:19.780530 4791 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/c435d469-c6b0-45a0-b61b-fd638540bf51-logs" (OuterVolumeSpecName: "logs") pod "c435d469-c6b0-45a0-b61b-fd638540bf51" (UID: "c435d469-c6b0-45a0-b61b-fd638540bf51"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 10 23:11:19 crc kubenswrapper[4791]: I1210 23:11:19.799633 4791 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/c435d469-c6b0-45a0-b61b-fd638540bf51-kube-api-access-5ggxm" (OuterVolumeSpecName: "kube-api-access-5ggxm") pod "c435d469-c6b0-45a0-b61b-fd638540bf51" (UID: "c435d469-c6b0-45a0-b61b-fd638540bf51"). InnerVolumeSpecName "kube-api-access-5ggxm". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 10 23:11:19 crc kubenswrapper[4791]: I1210 23:11:19.825117 4791 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/c435d469-c6b0-45a0-b61b-fd638540bf51-config-data" (OuterVolumeSpecName: "config-data") pod "c435d469-c6b0-45a0-b61b-fd638540bf51" (UID: "c435d469-c6b0-45a0-b61b-fd638540bf51"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 10 23:11:19 crc kubenswrapper[4791]: I1210 23:11:19.840793 4791 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/c435d469-c6b0-45a0-b61b-fd638540bf51-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "c435d469-c6b0-45a0-b61b-fd638540bf51" (UID: "c435d469-c6b0-45a0-b61b-fd638540bf51"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 10 23:11:19 crc kubenswrapper[4791]: I1210 23:11:19.882626 4791 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/c435d469-c6b0-45a0-b61b-fd638540bf51-logs\") on node \"crc\" DevicePath \"\"" Dec 10 23:11:19 crc kubenswrapper[4791]: I1210 23:11:19.882655 4791 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/c435d469-c6b0-45a0-b61b-fd638540bf51-config-data\") on node \"crc\" DevicePath \"\"" Dec 10 23:11:19 crc kubenswrapper[4791]: I1210 23:11:19.882665 4791 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c435d469-c6b0-45a0-b61b-fd638540bf51-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 10 23:11:19 crc kubenswrapper[4791]: I1210 23:11:19.882677 4791 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-5ggxm\" (UniqueName: \"kubernetes.io/projected/c435d469-c6b0-45a0-b61b-fd638540bf51-kube-api-access-5ggxm\") on node \"crc\" DevicePath \"\"" Dec 10 23:11:19 crc kubenswrapper[4791]: I1210 23:11:19.912552 4791 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Dec 10 23:11:19 crc kubenswrapper[4791]: I1210 23:11:19.984875 4791 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-rhtwc\" (UniqueName: \"kubernetes.io/projected/36c779d8-1b0f-4d7e-94ca-a05bda0b4ee8-kube-api-access-rhtwc\") pod \"36c779d8-1b0f-4d7e-94ca-a05bda0b4ee8\" (UID: \"36c779d8-1b0f-4d7e-94ca-a05bda0b4ee8\") " Dec 10 23:11:19 crc kubenswrapper[4791]: I1210 23:11:19.984964 4791 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/36c779d8-1b0f-4d7e-94ca-a05bda0b4ee8-log-httpd\") pod \"36c779d8-1b0f-4d7e-94ca-a05bda0b4ee8\" (UID: \"36c779d8-1b0f-4d7e-94ca-a05bda0b4ee8\") " Dec 10 23:11:19 crc kubenswrapper[4791]: I1210 23:11:19.985171 4791 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/36c779d8-1b0f-4d7e-94ca-a05bda0b4ee8-scripts\") pod \"36c779d8-1b0f-4d7e-94ca-a05bda0b4ee8\" (UID: \"36c779d8-1b0f-4d7e-94ca-a05bda0b4ee8\") " Dec 10 23:11:19 crc kubenswrapper[4791]: I1210 23:11:19.985302 4791 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/36c779d8-1b0f-4d7e-94ca-a05bda0b4ee8-combined-ca-bundle\") pod \"36c779d8-1b0f-4d7e-94ca-a05bda0b4ee8\" (UID: \"36c779d8-1b0f-4d7e-94ca-a05bda0b4ee8\") " Dec 10 23:11:19 crc kubenswrapper[4791]: I1210 23:11:19.985391 4791 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ceilometer-tls-certs\" (UniqueName: \"kubernetes.io/secret/36c779d8-1b0f-4d7e-94ca-a05bda0b4ee8-ceilometer-tls-certs\") pod \"36c779d8-1b0f-4d7e-94ca-a05bda0b4ee8\" (UID: \"36c779d8-1b0f-4d7e-94ca-a05bda0b4ee8\") " Dec 10 23:11:19 crc kubenswrapper[4791]: I1210 23:11:19.985495 4791 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/36c779d8-1b0f-4d7e-94ca-a05bda0b4ee8-sg-core-conf-yaml\") pod \"36c779d8-1b0f-4d7e-94ca-a05bda0b4ee8\" (UID: \"36c779d8-1b0f-4d7e-94ca-a05bda0b4ee8\") " Dec 10 23:11:19 crc kubenswrapper[4791]: I1210 23:11:19.985529 4791 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/36c779d8-1b0f-4d7e-94ca-a05bda0b4ee8-run-httpd\") pod \"36c779d8-1b0f-4d7e-94ca-a05bda0b4ee8\" (UID: \"36c779d8-1b0f-4d7e-94ca-a05bda0b4ee8\") " Dec 10 23:11:19 crc kubenswrapper[4791]: I1210 23:11:19.985627 4791 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/36c779d8-1b0f-4d7e-94ca-a05bda0b4ee8-config-data\") pod \"36c779d8-1b0f-4d7e-94ca-a05bda0b4ee8\" (UID: \"36c779d8-1b0f-4d7e-94ca-a05bda0b4ee8\") " Dec 10 23:11:19 crc kubenswrapper[4791]: I1210 23:11:19.985745 4791 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/36c779d8-1b0f-4d7e-94ca-a05bda0b4ee8-log-httpd" (OuterVolumeSpecName: "log-httpd") pod "36c779d8-1b0f-4d7e-94ca-a05bda0b4ee8" (UID: "36c779d8-1b0f-4d7e-94ca-a05bda0b4ee8"). InnerVolumeSpecName "log-httpd". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 10 23:11:19 crc kubenswrapper[4791]: I1210 23:11:19.986054 4791 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/36c779d8-1b0f-4d7e-94ca-a05bda0b4ee8-run-httpd" (OuterVolumeSpecName: "run-httpd") pod "36c779d8-1b0f-4d7e-94ca-a05bda0b4ee8" (UID: "36c779d8-1b0f-4d7e-94ca-a05bda0b4ee8"). InnerVolumeSpecName "run-httpd". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 10 23:11:19 crc kubenswrapper[4791]: I1210 23:11:19.986425 4791 reconciler_common.go:293] "Volume detached for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/36c779d8-1b0f-4d7e-94ca-a05bda0b4ee8-run-httpd\") on node \"crc\" DevicePath \"\"" Dec 10 23:11:19 crc kubenswrapper[4791]: I1210 23:11:19.986450 4791 reconciler_common.go:293] "Volume detached for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/36c779d8-1b0f-4d7e-94ca-a05bda0b4ee8-log-httpd\") on node \"crc\" DevicePath \"\"" Dec 10 23:11:19 crc kubenswrapper[4791]: I1210 23:11:19.990610 4791 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/36c779d8-1b0f-4d7e-94ca-a05bda0b4ee8-scripts" (OuterVolumeSpecName: "scripts") pod "36c779d8-1b0f-4d7e-94ca-a05bda0b4ee8" (UID: "36c779d8-1b0f-4d7e-94ca-a05bda0b4ee8"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 10 23:11:19 crc kubenswrapper[4791]: I1210 23:11:19.990643 4791 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/36c779d8-1b0f-4d7e-94ca-a05bda0b4ee8-kube-api-access-rhtwc" (OuterVolumeSpecName: "kube-api-access-rhtwc") pod "36c779d8-1b0f-4d7e-94ca-a05bda0b4ee8" (UID: "36c779d8-1b0f-4d7e-94ca-a05bda0b4ee8"). InnerVolumeSpecName "kube-api-access-rhtwc". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 10 23:11:20 crc kubenswrapper[4791]: I1210 23:11:20.010605 4791 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/36c779d8-1b0f-4d7e-94ca-a05bda0b4ee8-sg-core-conf-yaml" (OuterVolumeSpecName: "sg-core-conf-yaml") pod "36c779d8-1b0f-4d7e-94ca-a05bda0b4ee8" (UID: "36c779d8-1b0f-4d7e-94ca-a05bda0b4ee8"). InnerVolumeSpecName "sg-core-conf-yaml". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 10 23:11:20 crc kubenswrapper[4791]: I1210 23:11:20.038771 4791 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/36c779d8-1b0f-4d7e-94ca-a05bda0b4ee8-ceilometer-tls-certs" (OuterVolumeSpecName: "ceilometer-tls-certs") pod "36c779d8-1b0f-4d7e-94ca-a05bda0b4ee8" (UID: "36c779d8-1b0f-4d7e-94ca-a05bda0b4ee8"). InnerVolumeSpecName "ceilometer-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 10 23:11:20 crc kubenswrapper[4791]: I1210 23:11:20.064326 4791 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/36c779d8-1b0f-4d7e-94ca-a05bda0b4ee8-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "36c779d8-1b0f-4d7e-94ca-a05bda0b4ee8" (UID: "36c779d8-1b0f-4d7e-94ca-a05bda0b4ee8"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 10 23:11:20 crc kubenswrapper[4791]: I1210 23:11:20.088561 4791 reconciler_common.go:293] "Volume detached for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/36c779d8-1b0f-4d7e-94ca-a05bda0b4ee8-sg-core-conf-yaml\") on node \"crc\" DevicePath \"\"" Dec 10 23:11:20 crc kubenswrapper[4791]: I1210 23:11:20.088617 4791 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-rhtwc\" (UniqueName: \"kubernetes.io/projected/36c779d8-1b0f-4d7e-94ca-a05bda0b4ee8-kube-api-access-rhtwc\") on node \"crc\" DevicePath \"\"" Dec 10 23:11:20 crc kubenswrapper[4791]: I1210 23:11:20.088635 4791 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/36c779d8-1b0f-4d7e-94ca-a05bda0b4ee8-scripts\") on node \"crc\" DevicePath \"\"" Dec 10 23:11:20 crc kubenswrapper[4791]: I1210 23:11:20.088649 4791 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/36c779d8-1b0f-4d7e-94ca-a05bda0b4ee8-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 10 23:11:20 crc kubenswrapper[4791]: I1210 23:11:20.088663 4791 reconciler_common.go:293] "Volume detached for volume \"ceilometer-tls-certs\" (UniqueName: \"kubernetes.io/secret/36c779d8-1b0f-4d7e-94ca-a05bda0b4ee8-ceilometer-tls-certs\") on node \"crc\" DevicePath \"\"" Dec 10 23:11:20 crc kubenswrapper[4791]: I1210 23:11:20.090429 4791 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/36c779d8-1b0f-4d7e-94ca-a05bda0b4ee8-config-data" (OuterVolumeSpecName: "config-data") pod "36c779d8-1b0f-4d7e-94ca-a05bda0b4ee8" (UID: "36c779d8-1b0f-4d7e-94ca-a05bda0b4ee8"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 10 23:11:20 crc kubenswrapper[4791]: I1210 23:11:20.190739 4791 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/36c779d8-1b0f-4d7e-94ca-a05bda0b4ee8-config-data\") on node \"crc\" DevicePath \"\"" Dec 10 23:11:20 crc kubenswrapper[4791]: I1210 23:11:20.221623 4791 generic.go:334] "Generic (PLEG): container finished" podID="36c779d8-1b0f-4d7e-94ca-a05bda0b4ee8" containerID="bd7013d70d1fb7fb35a7924c0c7006151aa72907faa8474540bc5912b7de09a2" exitCode=0 Dec 10 23:11:20 crc kubenswrapper[4791]: I1210 23:11:20.221692 4791 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"36c779d8-1b0f-4d7e-94ca-a05bda0b4ee8","Type":"ContainerDied","Data":"bd7013d70d1fb7fb35a7924c0c7006151aa72907faa8474540bc5912b7de09a2"} Dec 10 23:11:20 crc kubenswrapper[4791]: I1210 23:11:20.222458 4791 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"36c779d8-1b0f-4d7e-94ca-a05bda0b4ee8","Type":"ContainerDied","Data":"5c9ac83bdac520a40246ebf6e5c87a217e53983f738e170e9314eebc7b756041"} Dec 10 23:11:20 crc kubenswrapper[4791]: I1210 23:11:20.221773 4791 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Dec 10 23:11:20 crc kubenswrapper[4791]: I1210 23:11:20.222540 4791 scope.go:117] "RemoveContainer" containerID="73d5fbab8e8ad07fc3d37539521ec847f0d8ee31c1becd87c6730d66f610f4ac" Dec 10 23:11:20 crc kubenswrapper[4791]: I1210 23:11:20.226602 4791 generic.go:334] "Generic (PLEG): container finished" podID="c435d469-c6b0-45a0-b61b-fd638540bf51" containerID="c2bd07aa7c52e85c622f844ee2cd9b8761a2f8e9220b0e3214c3a421060ea90d" exitCode=0 Dec 10 23:11:20 crc kubenswrapper[4791]: I1210 23:11:20.226642 4791 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"c435d469-c6b0-45a0-b61b-fd638540bf51","Type":"ContainerDied","Data":"c2bd07aa7c52e85c622f844ee2cd9b8761a2f8e9220b0e3214c3a421060ea90d"} Dec 10 23:11:20 crc kubenswrapper[4791]: I1210 23:11:20.226670 4791 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"c435d469-c6b0-45a0-b61b-fd638540bf51","Type":"ContainerDied","Data":"b3b5c917af99a1aaf2f7aac51b126d7ab683f8bfff1792bf9133276410a36beb"} Dec 10 23:11:20 crc kubenswrapper[4791]: I1210 23:11:20.226756 4791 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Dec 10 23:11:20 crc kubenswrapper[4791]: I1210 23:11:20.258307 4791 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-api-0"] Dec 10 23:11:20 crc kubenswrapper[4791]: I1210 23:11:20.260970 4791 scope.go:117] "RemoveContainer" containerID="4adc26eeae2e46577d7aee282669cf4154c110325b7c53f6a0f5089815dba157" Dec 10 23:11:20 crc kubenswrapper[4791]: I1210 23:11:20.277189 4791 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-api-0"] Dec 10 23:11:20 crc kubenswrapper[4791]: I1210 23:11:20.307905 4791 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ceilometer-0"] Dec 10 23:11:20 crc kubenswrapper[4791]: I1210 23:11:20.320507 4791 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/ceilometer-0"] Dec 10 23:11:20 crc kubenswrapper[4791]: I1210 23:11:20.323130 4791 scope.go:117] "RemoveContainer" containerID="517279520b8563ef48642d1a1f8a4fcc59abeb2393de710719a9f748d5ab9a24" Dec 10 23:11:20 crc kubenswrapper[4791]: I1210 23:11:20.327161 4791 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-api-0"] Dec 10 23:11:20 crc kubenswrapper[4791]: E1210 23:11:20.327572 4791 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="c435d469-c6b0-45a0-b61b-fd638540bf51" containerName="nova-api-api" Dec 10 23:11:20 crc kubenswrapper[4791]: I1210 23:11:20.327590 4791 state_mem.go:107] "Deleted CPUSet assignment" podUID="c435d469-c6b0-45a0-b61b-fd638540bf51" containerName="nova-api-api" Dec 10 23:11:20 crc kubenswrapper[4791]: E1210 23:11:20.327606 4791 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="36c779d8-1b0f-4d7e-94ca-a05bda0b4ee8" containerName="sg-core" Dec 10 23:11:20 crc kubenswrapper[4791]: I1210 23:11:20.327613 4791 state_mem.go:107] "Deleted CPUSet assignment" podUID="36c779d8-1b0f-4d7e-94ca-a05bda0b4ee8" containerName="sg-core" Dec 10 23:11:20 crc kubenswrapper[4791]: E1210 23:11:20.327630 4791 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="36c779d8-1b0f-4d7e-94ca-a05bda0b4ee8" containerName="proxy-httpd" Dec 10 23:11:20 crc kubenswrapper[4791]: I1210 23:11:20.327635 4791 state_mem.go:107] "Deleted CPUSet assignment" podUID="36c779d8-1b0f-4d7e-94ca-a05bda0b4ee8" containerName="proxy-httpd" Dec 10 23:11:20 crc kubenswrapper[4791]: E1210 23:11:20.327650 4791 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="c435d469-c6b0-45a0-b61b-fd638540bf51" containerName="nova-api-log" Dec 10 23:11:20 crc kubenswrapper[4791]: I1210 23:11:20.327655 4791 state_mem.go:107] "Deleted CPUSet assignment" podUID="c435d469-c6b0-45a0-b61b-fd638540bf51" containerName="nova-api-log" Dec 10 23:11:20 crc kubenswrapper[4791]: E1210 23:11:20.327663 4791 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="36c779d8-1b0f-4d7e-94ca-a05bda0b4ee8" containerName="ceilometer-central-agent" Dec 10 23:11:20 crc kubenswrapper[4791]: I1210 23:11:20.327669 4791 state_mem.go:107] "Deleted CPUSet assignment" podUID="36c779d8-1b0f-4d7e-94ca-a05bda0b4ee8" containerName="ceilometer-central-agent" Dec 10 23:11:20 crc kubenswrapper[4791]: E1210 23:11:20.327679 4791 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="36c779d8-1b0f-4d7e-94ca-a05bda0b4ee8" containerName="ceilometer-notification-agent" Dec 10 23:11:20 crc kubenswrapper[4791]: I1210 23:11:20.327684 4791 state_mem.go:107] "Deleted CPUSet assignment" podUID="36c779d8-1b0f-4d7e-94ca-a05bda0b4ee8" containerName="ceilometer-notification-agent" Dec 10 23:11:20 crc kubenswrapper[4791]: I1210 23:11:20.328166 4791 memory_manager.go:354] "RemoveStaleState removing state" podUID="c435d469-c6b0-45a0-b61b-fd638540bf51" containerName="nova-api-log" Dec 10 23:11:20 crc kubenswrapper[4791]: I1210 23:11:20.328190 4791 memory_manager.go:354] "RemoveStaleState removing state" podUID="36c779d8-1b0f-4d7e-94ca-a05bda0b4ee8" containerName="ceilometer-notification-agent" Dec 10 23:11:20 crc kubenswrapper[4791]: I1210 23:11:20.328217 4791 memory_manager.go:354] "RemoveStaleState removing state" podUID="36c779d8-1b0f-4d7e-94ca-a05bda0b4ee8" containerName="proxy-httpd" Dec 10 23:11:20 crc kubenswrapper[4791]: I1210 23:11:20.328226 4791 memory_manager.go:354] "RemoveStaleState removing state" podUID="36c779d8-1b0f-4d7e-94ca-a05bda0b4ee8" containerName="ceilometer-central-agent" Dec 10 23:11:20 crc kubenswrapper[4791]: I1210 23:11:20.328236 4791 memory_manager.go:354] "RemoveStaleState removing state" podUID="c435d469-c6b0-45a0-b61b-fd638540bf51" containerName="nova-api-api" Dec 10 23:11:20 crc kubenswrapper[4791]: I1210 23:11:20.328243 4791 memory_manager.go:354] "RemoveStaleState removing state" podUID="36c779d8-1b0f-4d7e-94ca-a05bda0b4ee8" containerName="sg-core" Dec 10 23:11:20 crc kubenswrapper[4791]: I1210 23:11:20.329227 4791 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Dec 10 23:11:20 crc kubenswrapper[4791]: I1210 23:11:20.331967 4791 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-api-config-data" Dec 10 23:11:20 crc kubenswrapper[4791]: I1210 23:11:20.332455 4791 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-nova-internal-svc" Dec 10 23:11:20 crc kubenswrapper[4791]: I1210 23:11:20.332512 4791 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-nova-public-svc" Dec 10 23:11:20 crc kubenswrapper[4791]: I1210 23:11:20.343092 4791 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-api-0"] Dec 10 23:11:20 crc kubenswrapper[4791]: I1210 23:11:20.350854 4791 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ceilometer-0"] Dec 10 23:11:20 crc kubenswrapper[4791]: I1210 23:11:20.357031 4791 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Dec 10 23:11:20 crc kubenswrapper[4791]: I1210 23:11:20.359737 4791 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceilometer-0"] Dec 10 23:11:20 crc kubenswrapper[4791]: I1210 23:11:20.362453 4791 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-config-data" Dec 10 23:11:20 crc kubenswrapper[4791]: I1210 23:11:20.362639 4791 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-ceilometer-internal-svc" Dec 10 23:11:20 crc kubenswrapper[4791]: I1210 23:11:20.363216 4791 scope.go:117] "RemoveContainer" containerID="bd7013d70d1fb7fb35a7924c0c7006151aa72907faa8474540bc5912b7de09a2" Dec 10 23:11:20 crc kubenswrapper[4791]: I1210 23:11:20.364740 4791 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-scripts" Dec 10 23:11:20 crc kubenswrapper[4791]: I1210 23:11:20.395544 4791 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ceilometer-tls-certs\" (UniqueName: \"kubernetes.io/secret/c85398ce-a5d4-414a-8811-775a4ce71193-ceilometer-tls-certs\") pod \"ceilometer-0\" (UID: \"c85398ce-a5d4-414a-8811-775a4ce71193\") " pod="openstack/ceilometer-0" Dec 10 23:11:20 crc kubenswrapper[4791]: I1210 23:11:20.395589 4791 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/c85398ce-a5d4-414a-8811-775a4ce71193-scripts\") pod \"ceilometer-0\" (UID: \"c85398ce-a5d4-414a-8811-775a4ce71193\") " pod="openstack/ceilometer-0" Dec 10 23:11:20 crc kubenswrapper[4791]: I1210 23:11:20.395624 4791 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/c85398ce-a5d4-414a-8811-775a4ce71193-log-httpd\") pod \"ceilometer-0\" (UID: \"c85398ce-a5d4-414a-8811-775a4ce71193\") " pod="openstack/ceilometer-0" Dec 10 23:11:20 crc kubenswrapper[4791]: I1210 23:11:20.395643 4791 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/c85398ce-a5d4-414a-8811-775a4ce71193-run-httpd\") pod \"ceilometer-0\" (UID: \"c85398ce-a5d4-414a-8811-775a4ce71193\") " pod="openstack/ceilometer-0" Dec 10 23:11:20 crc kubenswrapper[4791]: I1210 23:11:20.395698 4791 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/6460257f-3f13-4642-b5b5-fc05065df9d7-config-data\") pod \"nova-api-0\" (UID: \"6460257f-3f13-4642-b5b5-fc05065df9d7\") " pod="openstack/nova-api-0" Dec 10 23:11:20 crc kubenswrapper[4791]: I1210 23:11:20.395724 4791 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c85398ce-a5d4-414a-8811-775a4ce71193-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"c85398ce-a5d4-414a-8811-775a4ce71193\") " pod="openstack/ceilometer-0" Dec 10 23:11:20 crc kubenswrapper[4791]: I1210 23:11:20.395744 4791 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/6460257f-3f13-4642-b5b5-fc05065df9d7-logs\") pod \"nova-api-0\" (UID: \"6460257f-3f13-4642-b5b5-fc05065df9d7\") " pod="openstack/nova-api-0" Dec 10 23:11:20 crc kubenswrapper[4791]: I1210 23:11:20.395776 4791 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/6460257f-3f13-4642-b5b5-fc05065df9d7-public-tls-certs\") pod \"nova-api-0\" (UID: \"6460257f-3f13-4642-b5b5-fc05065df9d7\") " pod="openstack/nova-api-0" Dec 10 23:11:20 crc kubenswrapper[4791]: I1210 23:11:20.395803 4791 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-wmzk9\" (UniqueName: \"kubernetes.io/projected/c85398ce-a5d4-414a-8811-775a4ce71193-kube-api-access-wmzk9\") pod \"ceilometer-0\" (UID: \"c85398ce-a5d4-414a-8811-775a4ce71193\") " pod="openstack/ceilometer-0" Dec 10 23:11:20 crc kubenswrapper[4791]: I1210 23:11:20.395848 4791 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/6460257f-3f13-4642-b5b5-fc05065df9d7-internal-tls-certs\") pod \"nova-api-0\" (UID: \"6460257f-3f13-4642-b5b5-fc05065df9d7\") " pod="openstack/nova-api-0" Dec 10 23:11:20 crc kubenswrapper[4791]: I1210 23:11:20.395888 4791 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-6mrmt\" (UniqueName: \"kubernetes.io/projected/6460257f-3f13-4642-b5b5-fc05065df9d7-kube-api-access-6mrmt\") pod \"nova-api-0\" (UID: \"6460257f-3f13-4642-b5b5-fc05065df9d7\") " pod="openstack/nova-api-0" Dec 10 23:11:20 crc kubenswrapper[4791]: I1210 23:11:20.395911 4791 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/c85398ce-a5d4-414a-8811-775a4ce71193-config-data\") pod \"ceilometer-0\" (UID: \"c85398ce-a5d4-414a-8811-775a4ce71193\") " pod="openstack/ceilometer-0" Dec 10 23:11:20 crc kubenswrapper[4791]: I1210 23:11:20.395947 4791 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/6460257f-3f13-4642-b5b5-fc05065df9d7-combined-ca-bundle\") pod \"nova-api-0\" (UID: \"6460257f-3f13-4642-b5b5-fc05065df9d7\") " pod="openstack/nova-api-0" Dec 10 23:11:20 crc kubenswrapper[4791]: I1210 23:11:20.395993 4791 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/c85398ce-a5d4-414a-8811-775a4ce71193-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"c85398ce-a5d4-414a-8811-775a4ce71193\") " pod="openstack/ceilometer-0" Dec 10 23:11:20 crc kubenswrapper[4791]: I1210 23:11:20.414751 4791 scope.go:117] "RemoveContainer" containerID="73d5fbab8e8ad07fc3d37539521ec847f0d8ee31c1becd87c6730d66f610f4ac" Dec 10 23:11:20 crc kubenswrapper[4791]: E1210 23:11:20.415855 4791 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"73d5fbab8e8ad07fc3d37539521ec847f0d8ee31c1becd87c6730d66f610f4ac\": container with ID starting with 73d5fbab8e8ad07fc3d37539521ec847f0d8ee31c1becd87c6730d66f610f4ac not found: ID does not exist" containerID="73d5fbab8e8ad07fc3d37539521ec847f0d8ee31c1becd87c6730d66f610f4ac" Dec 10 23:11:20 crc kubenswrapper[4791]: I1210 23:11:20.415894 4791 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"73d5fbab8e8ad07fc3d37539521ec847f0d8ee31c1becd87c6730d66f610f4ac"} err="failed to get container status \"73d5fbab8e8ad07fc3d37539521ec847f0d8ee31c1becd87c6730d66f610f4ac\": rpc error: code = NotFound desc = could not find container \"73d5fbab8e8ad07fc3d37539521ec847f0d8ee31c1becd87c6730d66f610f4ac\": container with ID starting with 73d5fbab8e8ad07fc3d37539521ec847f0d8ee31c1becd87c6730d66f610f4ac not found: ID does not exist" Dec 10 23:11:20 crc kubenswrapper[4791]: I1210 23:11:20.415925 4791 scope.go:117] "RemoveContainer" containerID="4adc26eeae2e46577d7aee282669cf4154c110325b7c53f6a0f5089815dba157" Dec 10 23:11:20 crc kubenswrapper[4791]: E1210 23:11:20.419501 4791 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"4adc26eeae2e46577d7aee282669cf4154c110325b7c53f6a0f5089815dba157\": container with ID starting with 4adc26eeae2e46577d7aee282669cf4154c110325b7c53f6a0f5089815dba157 not found: ID does not exist" containerID="4adc26eeae2e46577d7aee282669cf4154c110325b7c53f6a0f5089815dba157" Dec 10 23:11:20 crc kubenswrapper[4791]: I1210 23:11:20.419540 4791 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"4adc26eeae2e46577d7aee282669cf4154c110325b7c53f6a0f5089815dba157"} err="failed to get container status \"4adc26eeae2e46577d7aee282669cf4154c110325b7c53f6a0f5089815dba157\": rpc error: code = NotFound desc = could not find container \"4adc26eeae2e46577d7aee282669cf4154c110325b7c53f6a0f5089815dba157\": container with ID starting with 4adc26eeae2e46577d7aee282669cf4154c110325b7c53f6a0f5089815dba157 not found: ID does not exist" Dec 10 23:11:20 crc kubenswrapper[4791]: I1210 23:11:20.419566 4791 scope.go:117] "RemoveContainer" containerID="517279520b8563ef48642d1a1f8a4fcc59abeb2393de710719a9f748d5ab9a24" Dec 10 23:11:20 crc kubenswrapper[4791]: E1210 23:11:20.419965 4791 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"517279520b8563ef48642d1a1f8a4fcc59abeb2393de710719a9f748d5ab9a24\": container with ID starting with 517279520b8563ef48642d1a1f8a4fcc59abeb2393de710719a9f748d5ab9a24 not found: ID does not exist" containerID="517279520b8563ef48642d1a1f8a4fcc59abeb2393de710719a9f748d5ab9a24" Dec 10 23:11:20 crc kubenswrapper[4791]: I1210 23:11:20.419986 4791 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"517279520b8563ef48642d1a1f8a4fcc59abeb2393de710719a9f748d5ab9a24"} err="failed to get container status \"517279520b8563ef48642d1a1f8a4fcc59abeb2393de710719a9f748d5ab9a24\": rpc error: code = NotFound desc = could not find container \"517279520b8563ef48642d1a1f8a4fcc59abeb2393de710719a9f748d5ab9a24\": container with ID starting with 517279520b8563ef48642d1a1f8a4fcc59abeb2393de710719a9f748d5ab9a24 not found: ID does not exist" Dec 10 23:11:20 crc kubenswrapper[4791]: I1210 23:11:20.420002 4791 scope.go:117] "RemoveContainer" containerID="bd7013d70d1fb7fb35a7924c0c7006151aa72907faa8474540bc5912b7de09a2" Dec 10 23:11:20 crc kubenswrapper[4791]: E1210 23:11:20.420815 4791 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"bd7013d70d1fb7fb35a7924c0c7006151aa72907faa8474540bc5912b7de09a2\": container with ID starting with bd7013d70d1fb7fb35a7924c0c7006151aa72907faa8474540bc5912b7de09a2 not found: ID does not exist" containerID="bd7013d70d1fb7fb35a7924c0c7006151aa72907faa8474540bc5912b7de09a2" Dec 10 23:11:20 crc kubenswrapper[4791]: I1210 23:11:20.420845 4791 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"bd7013d70d1fb7fb35a7924c0c7006151aa72907faa8474540bc5912b7de09a2"} err="failed to get container status \"bd7013d70d1fb7fb35a7924c0c7006151aa72907faa8474540bc5912b7de09a2\": rpc error: code = NotFound desc = could not find container \"bd7013d70d1fb7fb35a7924c0c7006151aa72907faa8474540bc5912b7de09a2\": container with ID starting with bd7013d70d1fb7fb35a7924c0c7006151aa72907faa8474540bc5912b7de09a2 not found: ID does not exist" Dec 10 23:11:20 crc kubenswrapper[4791]: I1210 23:11:20.420862 4791 scope.go:117] "RemoveContainer" containerID="c2bd07aa7c52e85c622f844ee2cd9b8761a2f8e9220b0e3214c3a421060ea90d" Dec 10 23:11:20 crc kubenswrapper[4791]: I1210 23:11:20.492526 4791 scope.go:117] "RemoveContainer" containerID="940bdefb6fe251888b9f0d4224ac2c43268c7a2c548cd5eec01678307a80317d" Dec 10 23:11:20 crc kubenswrapper[4791]: I1210 23:11:20.497466 4791 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/c85398ce-a5d4-414a-8811-775a4ce71193-log-httpd\") pod \"ceilometer-0\" (UID: \"c85398ce-a5d4-414a-8811-775a4ce71193\") " pod="openstack/ceilometer-0" Dec 10 23:11:20 crc kubenswrapper[4791]: I1210 23:11:20.497511 4791 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/c85398ce-a5d4-414a-8811-775a4ce71193-run-httpd\") pod \"ceilometer-0\" (UID: \"c85398ce-a5d4-414a-8811-775a4ce71193\") " pod="openstack/ceilometer-0" Dec 10 23:11:20 crc kubenswrapper[4791]: I1210 23:11:20.497596 4791 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/6460257f-3f13-4642-b5b5-fc05065df9d7-config-data\") pod \"nova-api-0\" (UID: \"6460257f-3f13-4642-b5b5-fc05065df9d7\") " pod="openstack/nova-api-0" Dec 10 23:11:20 crc kubenswrapper[4791]: I1210 23:11:20.497627 4791 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c85398ce-a5d4-414a-8811-775a4ce71193-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"c85398ce-a5d4-414a-8811-775a4ce71193\") " pod="openstack/ceilometer-0" Dec 10 23:11:20 crc kubenswrapper[4791]: I1210 23:11:20.497649 4791 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/6460257f-3f13-4642-b5b5-fc05065df9d7-logs\") pod \"nova-api-0\" (UID: \"6460257f-3f13-4642-b5b5-fc05065df9d7\") " pod="openstack/nova-api-0" Dec 10 23:11:20 crc kubenswrapper[4791]: I1210 23:11:20.497684 4791 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/6460257f-3f13-4642-b5b5-fc05065df9d7-public-tls-certs\") pod \"nova-api-0\" (UID: \"6460257f-3f13-4642-b5b5-fc05065df9d7\") " pod="openstack/nova-api-0" Dec 10 23:11:20 crc kubenswrapper[4791]: I1210 23:11:20.497710 4791 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-wmzk9\" (UniqueName: \"kubernetes.io/projected/c85398ce-a5d4-414a-8811-775a4ce71193-kube-api-access-wmzk9\") pod \"ceilometer-0\" (UID: \"c85398ce-a5d4-414a-8811-775a4ce71193\") " pod="openstack/ceilometer-0" Dec 10 23:11:20 crc kubenswrapper[4791]: I1210 23:11:20.497741 4791 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/6460257f-3f13-4642-b5b5-fc05065df9d7-internal-tls-certs\") pod \"nova-api-0\" (UID: \"6460257f-3f13-4642-b5b5-fc05065df9d7\") " pod="openstack/nova-api-0" Dec 10 23:11:20 crc kubenswrapper[4791]: I1210 23:11:20.497776 4791 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-6mrmt\" (UniqueName: \"kubernetes.io/projected/6460257f-3f13-4642-b5b5-fc05065df9d7-kube-api-access-6mrmt\") pod \"nova-api-0\" (UID: \"6460257f-3f13-4642-b5b5-fc05065df9d7\") " pod="openstack/nova-api-0" Dec 10 23:11:20 crc kubenswrapper[4791]: I1210 23:11:20.497797 4791 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/c85398ce-a5d4-414a-8811-775a4ce71193-config-data\") pod \"ceilometer-0\" (UID: \"c85398ce-a5d4-414a-8811-775a4ce71193\") " pod="openstack/ceilometer-0" Dec 10 23:11:20 crc kubenswrapper[4791]: I1210 23:11:20.497842 4791 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/6460257f-3f13-4642-b5b5-fc05065df9d7-combined-ca-bundle\") pod \"nova-api-0\" (UID: \"6460257f-3f13-4642-b5b5-fc05065df9d7\") " pod="openstack/nova-api-0" Dec 10 23:11:20 crc kubenswrapper[4791]: I1210 23:11:20.497868 4791 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/c85398ce-a5d4-414a-8811-775a4ce71193-log-httpd\") pod \"ceilometer-0\" (UID: \"c85398ce-a5d4-414a-8811-775a4ce71193\") " pod="openstack/ceilometer-0" Dec 10 23:11:20 crc kubenswrapper[4791]: I1210 23:11:20.497912 4791 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/c85398ce-a5d4-414a-8811-775a4ce71193-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"c85398ce-a5d4-414a-8811-775a4ce71193\") " pod="openstack/ceilometer-0" Dec 10 23:11:20 crc kubenswrapper[4791]: I1210 23:11:20.497933 4791 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ceilometer-tls-certs\" (UniqueName: \"kubernetes.io/secret/c85398ce-a5d4-414a-8811-775a4ce71193-ceilometer-tls-certs\") pod \"ceilometer-0\" (UID: \"c85398ce-a5d4-414a-8811-775a4ce71193\") " pod="openstack/ceilometer-0" Dec 10 23:11:20 crc kubenswrapper[4791]: I1210 23:11:20.497953 4791 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/c85398ce-a5d4-414a-8811-775a4ce71193-scripts\") pod \"ceilometer-0\" (UID: \"c85398ce-a5d4-414a-8811-775a4ce71193\") " pod="openstack/ceilometer-0" Dec 10 23:11:20 crc kubenswrapper[4791]: I1210 23:11:20.498216 4791 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/c85398ce-a5d4-414a-8811-775a4ce71193-run-httpd\") pod \"ceilometer-0\" (UID: \"c85398ce-a5d4-414a-8811-775a4ce71193\") " pod="openstack/ceilometer-0" Dec 10 23:11:20 crc kubenswrapper[4791]: I1210 23:11:20.498614 4791 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/6460257f-3f13-4642-b5b5-fc05065df9d7-logs\") pod \"nova-api-0\" (UID: \"6460257f-3f13-4642-b5b5-fc05065df9d7\") " pod="openstack/nova-api-0" Dec 10 23:11:20 crc kubenswrapper[4791]: I1210 23:11:20.504996 4791 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/6460257f-3f13-4642-b5b5-fc05065df9d7-config-data\") pod \"nova-api-0\" (UID: \"6460257f-3f13-4642-b5b5-fc05065df9d7\") " pod="openstack/nova-api-0" Dec 10 23:11:20 crc kubenswrapper[4791]: I1210 23:11:20.505036 4791 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/6460257f-3f13-4642-b5b5-fc05065df9d7-combined-ca-bundle\") pod \"nova-api-0\" (UID: \"6460257f-3f13-4642-b5b5-fc05065df9d7\") " pod="openstack/nova-api-0" Dec 10 23:11:20 crc kubenswrapper[4791]: I1210 23:11:20.505289 4791 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/6460257f-3f13-4642-b5b5-fc05065df9d7-internal-tls-certs\") pod \"nova-api-0\" (UID: \"6460257f-3f13-4642-b5b5-fc05065df9d7\") " pod="openstack/nova-api-0" Dec 10 23:11:20 crc kubenswrapper[4791]: I1210 23:11:20.508276 4791 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c85398ce-a5d4-414a-8811-775a4ce71193-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"c85398ce-a5d4-414a-8811-775a4ce71193\") " pod="openstack/ceilometer-0" Dec 10 23:11:20 crc kubenswrapper[4791]: I1210 23:11:20.508576 4791 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/c85398ce-a5d4-414a-8811-775a4ce71193-config-data\") pod \"ceilometer-0\" (UID: \"c85398ce-a5d4-414a-8811-775a4ce71193\") " pod="openstack/ceilometer-0" Dec 10 23:11:20 crc kubenswrapper[4791]: I1210 23:11:20.508960 4791 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ceilometer-tls-certs\" (UniqueName: \"kubernetes.io/secret/c85398ce-a5d4-414a-8811-775a4ce71193-ceilometer-tls-certs\") pod \"ceilometer-0\" (UID: \"c85398ce-a5d4-414a-8811-775a4ce71193\") " pod="openstack/ceilometer-0" Dec 10 23:11:20 crc kubenswrapper[4791]: I1210 23:11:20.509149 4791 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/c85398ce-a5d4-414a-8811-775a4ce71193-scripts\") pod \"ceilometer-0\" (UID: \"c85398ce-a5d4-414a-8811-775a4ce71193\") " pod="openstack/ceilometer-0" Dec 10 23:11:20 crc kubenswrapper[4791]: I1210 23:11:20.517568 4791 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/c85398ce-a5d4-414a-8811-775a4ce71193-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"c85398ce-a5d4-414a-8811-775a4ce71193\") " pod="openstack/ceilometer-0" Dec 10 23:11:20 crc kubenswrapper[4791]: I1210 23:11:20.519040 4791 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-6mrmt\" (UniqueName: \"kubernetes.io/projected/6460257f-3f13-4642-b5b5-fc05065df9d7-kube-api-access-6mrmt\") pod \"nova-api-0\" (UID: \"6460257f-3f13-4642-b5b5-fc05065df9d7\") " pod="openstack/nova-api-0" Dec 10 23:11:20 crc kubenswrapper[4791]: I1210 23:11:20.519197 4791 scope.go:117] "RemoveContainer" containerID="c2bd07aa7c52e85c622f844ee2cd9b8761a2f8e9220b0e3214c3a421060ea90d" Dec 10 23:11:20 crc kubenswrapper[4791]: I1210 23:11:20.519636 4791 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/6460257f-3f13-4642-b5b5-fc05065df9d7-public-tls-certs\") pod \"nova-api-0\" (UID: \"6460257f-3f13-4642-b5b5-fc05065df9d7\") " pod="openstack/nova-api-0" Dec 10 23:11:20 crc kubenswrapper[4791]: E1210 23:11:20.519742 4791 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"c2bd07aa7c52e85c622f844ee2cd9b8761a2f8e9220b0e3214c3a421060ea90d\": container with ID starting with c2bd07aa7c52e85c622f844ee2cd9b8761a2f8e9220b0e3214c3a421060ea90d not found: ID does not exist" containerID="c2bd07aa7c52e85c622f844ee2cd9b8761a2f8e9220b0e3214c3a421060ea90d" Dec 10 23:11:20 crc kubenswrapper[4791]: I1210 23:11:20.519779 4791 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"c2bd07aa7c52e85c622f844ee2cd9b8761a2f8e9220b0e3214c3a421060ea90d"} err="failed to get container status \"c2bd07aa7c52e85c622f844ee2cd9b8761a2f8e9220b0e3214c3a421060ea90d\": rpc error: code = NotFound desc = could not find container \"c2bd07aa7c52e85c622f844ee2cd9b8761a2f8e9220b0e3214c3a421060ea90d\": container with ID starting with c2bd07aa7c52e85c622f844ee2cd9b8761a2f8e9220b0e3214c3a421060ea90d not found: ID does not exist" Dec 10 23:11:20 crc kubenswrapper[4791]: I1210 23:11:20.519828 4791 scope.go:117] "RemoveContainer" containerID="940bdefb6fe251888b9f0d4224ac2c43268c7a2c548cd5eec01678307a80317d" Dec 10 23:11:20 crc kubenswrapper[4791]: E1210 23:11:20.520222 4791 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"940bdefb6fe251888b9f0d4224ac2c43268c7a2c548cd5eec01678307a80317d\": container with ID starting with 940bdefb6fe251888b9f0d4224ac2c43268c7a2c548cd5eec01678307a80317d not found: ID does not exist" containerID="940bdefb6fe251888b9f0d4224ac2c43268c7a2c548cd5eec01678307a80317d" Dec 10 23:11:20 crc kubenswrapper[4791]: I1210 23:11:20.520256 4791 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"940bdefb6fe251888b9f0d4224ac2c43268c7a2c548cd5eec01678307a80317d"} err="failed to get container status \"940bdefb6fe251888b9f0d4224ac2c43268c7a2c548cd5eec01678307a80317d\": rpc error: code = NotFound desc = could not find container \"940bdefb6fe251888b9f0d4224ac2c43268c7a2c548cd5eec01678307a80317d\": container with ID starting with 940bdefb6fe251888b9f0d4224ac2c43268c7a2c548cd5eec01678307a80317d not found: ID does not exist" Dec 10 23:11:20 crc kubenswrapper[4791]: I1210 23:11:20.526693 4791 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-wmzk9\" (UniqueName: \"kubernetes.io/projected/c85398ce-a5d4-414a-8811-775a4ce71193-kube-api-access-wmzk9\") pod \"ceilometer-0\" (UID: \"c85398ce-a5d4-414a-8811-775a4ce71193\") " pod="openstack/ceilometer-0" Dec 10 23:11:20 crc kubenswrapper[4791]: I1210 23:11:20.659037 4791 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Dec 10 23:11:20 crc kubenswrapper[4791]: I1210 23:11:20.676515 4791 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Dec 10 23:11:21 crc kubenswrapper[4791]: I1210 23:11:21.132270 4791 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-api-0"] Dec 10 23:11:21 crc kubenswrapper[4791]: W1210 23:11:21.136999 4791 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod6460257f_3f13_4642_b5b5_fc05065df9d7.slice/crio-478fbaaf7f1f01776b968365f05590a0059ae9631f821df7ffeb8f1e19f33612 WatchSource:0}: Error finding container 478fbaaf7f1f01776b968365f05590a0059ae9631f821df7ffeb8f1e19f33612: Status 404 returned error can't find the container with id 478fbaaf7f1f01776b968365f05590a0059ae9631f821df7ffeb8f1e19f33612 Dec 10 23:11:21 crc kubenswrapper[4791]: I1210 23:11:21.224111 4791 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceilometer-0"] Dec 10 23:11:21 crc kubenswrapper[4791]: I1210 23:11:21.243920 4791 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"6460257f-3f13-4642-b5b5-fc05065df9d7","Type":"ContainerStarted","Data":"478fbaaf7f1f01776b968365f05590a0059ae9631f821df7ffeb8f1e19f33612"} Dec 10 23:11:21 crc kubenswrapper[4791]: I1210 23:11:21.246891 4791 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"c85398ce-a5d4-414a-8811-775a4ce71193","Type":"ContainerStarted","Data":"959a0a8cecf6d8a4ca1011986f95d698204b567ede56bec388f806f33341b13a"} Dec 10 23:11:21 crc kubenswrapper[4791]: I1210 23:11:21.895981 4791 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="36c779d8-1b0f-4d7e-94ca-a05bda0b4ee8" path="/var/lib/kubelet/pods/36c779d8-1b0f-4d7e-94ca-a05bda0b4ee8/volumes" Dec 10 23:11:21 crc kubenswrapper[4791]: I1210 23:11:21.897460 4791 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="c435d469-c6b0-45a0-b61b-fd638540bf51" path="/var/lib/kubelet/pods/c435d469-c6b0-45a0-b61b-fd638540bf51/volumes" Dec 10 23:11:22 crc kubenswrapper[4791]: I1210 23:11:22.274586 4791 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"6460257f-3f13-4642-b5b5-fc05065df9d7","Type":"ContainerStarted","Data":"e6e79a11bb5efb0004d4c3a54c551264e04cb8c253bb15d795f2c9dfeb974154"} Dec 10 23:11:22 crc kubenswrapper[4791]: I1210 23:11:22.274649 4791 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"6460257f-3f13-4642-b5b5-fc05065df9d7","Type":"ContainerStarted","Data":"e17b5d105e85af3d0d1bfcd6dacac2d5d1b35811ddecc4140c5bfb94cb52c0bc"} Dec 10 23:11:22 crc kubenswrapper[4791]: I1210 23:11:22.312174 4791 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-api-0" podStartSLOduration=2.312153093 podStartE2EDuration="2.312153093s" podCreationTimestamp="2025-12-10 23:11:20 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-10 23:11:22.303608746 +0000 UTC m=+1316.733226379" watchObservedRunningTime="2025-12-10 23:11:22.312153093 +0000 UTC m=+1316.741770706" Dec 10 23:11:22 crc kubenswrapper[4791]: I1210 23:11:22.508829 4791 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-cell1-novncproxy-0" Dec 10 23:11:22 crc kubenswrapper[4791]: I1210 23:11:22.527602 4791 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/nova-cell1-novncproxy-0" Dec 10 23:11:23 crc kubenswrapper[4791]: I1210 23:11:23.307012 4791 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-cell1-novncproxy-0" Dec 10 23:11:23 crc kubenswrapper[4791]: I1210 23:11:23.545942 4791 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-cell1-cell-mapping-zcfpp"] Dec 10 23:11:23 crc kubenswrapper[4791]: I1210 23:11:23.548367 4791 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-cell-mapping-zcfpp" Dec 10 23:11:23 crc kubenswrapper[4791]: I1210 23:11:23.551546 4791 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell1-manage-scripts" Dec 10 23:11:23 crc kubenswrapper[4791]: I1210 23:11:23.551966 4791 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell1-manage-config-data" Dec 10 23:11:23 crc kubenswrapper[4791]: I1210 23:11:23.560795 4791 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-cell-mapping-zcfpp"] Dec 10 23:11:23 crc kubenswrapper[4791]: I1210 23:11:23.627722 4791 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/dnsmasq-dns-59cf4bdb65-p4l85" Dec 10 23:11:23 crc kubenswrapper[4791]: I1210 23:11:23.670253 4791 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/430975e4-f09a-4c3f-9c1d-0cab25b587b1-scripts\") pod \"nova-cell1-cell-mapping-zcfpp\" (UID: \"430975e4-f09a-4c3f-9c1d-0cab25b587b1\") " pod="openstack/nova-cell1-cell-mapping-zcfpp" Dec 10 23:11:23 crc kubenswrapper[4791]: I1210 23:11:23.670383 4791 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/430975e4-f09a-4c3f-9c1d-0cab25b587b1-combined-ca-bundle\") pod \"nova-cell1-cell-mapping-zcfpp\" (UID: \"430975e4-f09a-4c3f-9c1d-0cab25b587b1\") " pod="openstack/nova-cell1-cell-mapping-zcfpp" Dec 10 23:11:23 crc kubenswrapper[4791]: I1210 23:11:23.670444 4791 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/430975e4-f09a-4c3f-9c1d-0cab25b587b1-config-data\") pod \"nova-cell1-cell-mapping-zcfpp\" (UID: \"430975e4-f09a-4c3f-9c1d-0cab25b587b1\") " pod="openstack/nova-cell1-cell-mapping-zcfpp" Dec 10 23:11:23 crc kubenswrapper[4791]: I1210 23:11:23.670468 4791 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-pcfw2\" (UniqueName: \"kubernetes.io/projected/430975e4-f09a-4c3f-9c1d-0cab25b587b1-kube-api-access-pcfw2\") pod \"nova-cell1-cell-mapping-zcfpp\" (UID: \"430975e4-f09a-4c3f-9c1d-0cab25b587b1\") " pod="openstack/nova-cell1-cell-mapping-zcfpp" Dec 10 23:11:23 crc kubenswrapper[4791]: I1210 23:11:23.689308 4791 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-845d6d6f59-d6v8c"] Dec 10 23:11:23 crc kubenswrapper[4791]: I1210 23:11:23.690593 4791 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/dnsmasq-dns-845d6d6f59-d6v8c" podUID="6202cb66-3ca7-4988-a800-2682cba44d72" containerName="dnsmasq-dns" containerID="cri-o://23c8115e63afae1fa7a5d92c7142b2a8f9b026f2cd7fdba3e8bd355ae6b7935f" gracePeriod=10 Dec 10 23:11:23 crc kubenswrapper[4791]: I1210 23:11:23.772516 4791 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/430975e4-f09a-4c3f-9c1d-0cab25b587b1-config-data\") pod \"nova-cell1-cell-mapping-zcfpp\" (UID: \"430975e4-f09a-4c3f-9c1d-0cab25b587b1\") " pod="openstack/nova-cell1-cell-mapping-zcfpp" Dec 10 23:11:23 crc kubenswrapper[4791]: I1210 23:11:23.773376 4791 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-pcfw2\" (UniqueName: \"kubernetes.io/projected/430975e4-f09a-4c3f-9c1d-0cab25b587b1-kube-api-access-pcfw2\") pod \"nova-cell1-cell-mapping-zcfpp\" (UID: \"430975e4-f09a-4c3f-9c1d-0cab25b587b1\") " pod="openstack/nova-cell1-cell-mapping-zcfpp" Dec 10 23:11:23 crc kubenswrapper[4791]: I1210 23:11:23.773518 4791 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/430975e4-f09a-4c3f-9c1d-0cab25b587b1-scripts\") pod \"nova-cell1-cell-mapping-zcfpp\" (UID: \"430975e4-f09a-4c3f-9c1d-0cab25b587b1\") " pod="openstack/nova-cell1-cell-mapping-zcfpp" Dec 10 23:11:23 crc kubenswrapper[4791]: I1210 23:11:23.773645 4791 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/430975e4-f09a-4c3f-9c1d-0cab25b587b1-combined-ca-bundle\") pod \"nova-cell1-cell-mapping-zcfpp\" (UID: \"430975e4-f09a-4c3f-9c1d-0cab25b587b1\") " pod="openstack/nova-cell1-cell-mapping-zcfpp" Dec 10 23:11:23 crc kubenswrapper[4791]: I1210 23:11:23.782259 4791 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/430975e4-f09a-4c3f-9c1d-0cab25b587b1-config-data\") pod \"nova-cell1-cell-mapping-zcfpp\" (UID: \"430975e4-f09a-4c3f-9c1d-0cab25b587b1\") " pod="openstack/nova-cell1-cell-mapping-zcfpp" Dec 10 23:11:23 crc kubenswrapper[4791]: I1210 23:11:23.800000 4791 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/430975e4-f09a-4c3f-9c1d-0cab25b587b1-scripts\") pod \"nova-cell1-cell-mapping-zcfpp\" (UID: \"430975e4-f09a-4c3f-9c1d-0cab25b587b1\") " pod="openstack/nova-cell1-cell-mapping-zcfpp" Dec 10 23:11:23 crc kubenswrapper[4791]: I1210 23:11:23.801451 4791 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-pcfw2\" (UniqueName: \"kubernetes.io/projected/430975e4-f09a-4c3f-9c1d-0cab25b587b1-kube-api-access-pcfw2\") pod \"nova-cell1-cell-mapping-zcfpp\" (UID: \"430975e4-f09a-4c3f-9c1d-0cab25b587b1\") " pod="openstack/nova-cell1-cell-mapping-zcfpp" Dec 10 23:11:23 crc kubenswrapper[4791]: I1210 23:11:23.801784 4791 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/430975e4-f09a-4c3f-9c1d-0cab25b587b1-combined-ca-bundle\") pod \"nova-cell1-cell-mapping-zcfpp\" (UID: \"430975e4-f09a-4c3f-9c1d-0cab25b587b1\") " pod="openstack/nova-cell1-cell-mapping-zcfpp" Dec 10 23:11:23 crc kubenswrapper[4791]: I1210 23:11:23.876543 4791 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-cell-mapping-zcfpp" Dec 10 23:11:24 crc kubenswrapper[4791]: I1210 23:11:24.203384 4791 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-845d6d6f59-d6v8c" Dec 10 23:11:24 crc kubenswrapper[4791]: I1210 23:11:24.285400 4791 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-blchs\" (UniqueName: \"kubernetes.io/projected/6202cb66-3ca7-4988-a800-2682cba44d72-kube-api-access-blchs\") pod \"6202cb66-3ca7-4988-a800-2682cba44d72\" (UID: \"6202cb66-3ca7-4988-a800-2682cba44d72\") " Dec 10 23:11:24 crc kubenswrapper[4791]: I1210 23:11:24.285542 4791 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/6202cb66-3ca7-4988-a800-2682cba44d72-ovsdbserver-nb\") pod \"6202cb66-3ca7-4988-a800-2682cba44d72\" (UID: \"6202cb66-3ca7-4988-a800-2682cba44d72\") " Dec 10 23:11:24 crc kubenswrapper[4791]: I1210 23:11:24.285678 4791 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/6202cb66-3ca7-4988-a800-2682cba44d72-config\") pod \"6202cb66-3ca7-4988-a800-2682cba44d72\" (UID: \"6202cb66-3ca7-4988-a800-2682cba44d72\") " Dec 10 23:11:24 crc kubenswrapper[4791]: I1210 23:11:24.285825 4791 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/6202cb66-3ca7-4988-a800-2682cba44d72-dns-svc\") pod \"6202cb66-3ca7-4988-a800-2682cba44d72\" (UID: \"6202cb66-3ca7-4988-a800-2682cba44d72\") " Dec 10 23:11:24 crc kubenswrapper[4791]: I1210 23:11:24.286851 4791 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/6202cb66-3ca7-4988-a800-2682cba44d72-ovsdbserver-sb\") pod \"6202cb66-3ca7-4988-a800-2682cba44d72\" (UID: \"6202cb66-3ca7-4988-a800-2682cba44d72\") " Dec 10 23:11:24 crc kubenswrapper[4791]: I1210 23:11:24.286904 4791 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/6202cb66-3ca7-4988-a800-2682cba44d72-dns-swift-storage-0\") pod \"6202cb66-3ca7-4988-a800-2682cba44d72\" (UID: \"6202cb66-3ca7-4988-a800-2682cba44d72\") " Dec 10 23:11:24 crc kubenswrapper[4791]: I1210 23:11:24.307644 4791 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/6202cb66-3ca7-4988-a800-2682cba44d72-kube-api-access-blchs" (OuterVolumeSpecName: "kube-api-access-blchs") pod "6202cb66-3ca7-4988-a800-2682cba44d72" (UID: "6202cb66-3ca7-4988-a800-2682cba44d72"). InnerVolumeSpecName "kube-api-access-blchs". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 10 23:11:24 crc kubenswrapper[4791]: I1210 23:11:24.313394 4791 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"c85398ce-a5d4-414a-8811-775a4ce71193","Type":"ContainerStarted","Data":"1a2452d593e8960fda34e967cba0f1372d6f5949f06e11319089a92f4e35108d"} Dec 10 23:11:24 crc kubenswrapper[4791]: I1210 23:11:24.313454 4791 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"c85398ce-a5d4-414a-8811-775a4ce71193","Type":"ContainerStarted","Data":"ee1bcb8481a27990e9bb741e8d135361a18a7e494bc252e9cca0fdfd8dd29651"} Dec 10 23:11:24 crc kubenswrapper[4791]: I1210 23:11:24.316048 4791 generic.go:334] "Generic (PLEG): container finished" podID="6202cb66-3ca7-4988-a800-2682cba44d72" containerID="23c8115e63afae1fa7a5d92c7142b2a8f9b026f2cd7fdba3e8bd355ae6b7935f" exitCode=0 Dec 10 23:11:24 crc kubenswrapper[4791]: I1210 23:11:24.317206 4791 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-845d6d6f59-d6v8c" Dec 10 23:11:24 crc kubenswrapper[4791]: I1210 23:11:24.317651 4791 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-845d6d6f59-d6v8c" event={"ID":"6202cb66-3ca7-4988-a800-2682cba44d72","Type":"ContainerDied","Data":"23c8115e63afae1fa7a5d92c7142b2a8f9b026f2cd7fdba3e8bd355ae6b7935f"} Dec 10 23:11:24 crc kubenswrapper[4791]: I1210 23:11:24.317678 4791 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-845d6d6f59-d6v8c" event={"ID":"6202cb66-3ca7-4988-a800-2682cba44d72","Type":"ContainerDied","Data":"3190dbd7cddd5578a9e13db8ddfc19754d10d685bddb95bc7772fb7786e3334a"} Dec 10 23:11:24 crc kubenswrapper[4791]: I1210 23:11:24.317693 4791 scope.go:117] "RemoveContainer" containerID="23c8115e63afae1fa7a5d92c7142b2a8f9b026f2cd7fdba3e8bd355ae6b7935f" Dec 10 23:11:24 crc kubenswrapper[4791]: I1210 23:11:24.343794 4791 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/6202cb66-3ca7-4988-a800-2682cba44d72-ovsdbserver-sb" (OuterVolumeSpecName: "ovsdbserver-sb") pod "6202cb66-3ca7-4988-a800-2682cba44d72" (UID: "6202cb66-3ca7-4988-a800-2682cba44d72"). InnerVolumeSpecName "ovsdbserver-sb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 10 23:11:24 crc kubenswrapper[4791]: I1210 23:11:24.354785 4791 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/6202cb66-3ca7-4988-a800-2682cba44d72-ovsdbserver-nb" (OuterVolumeSpecName: "ovsdbserver-nb") pod "6202cb66-3ca7-4988-a800-2682cba44d72" (UID: "6202cb66-3ca7-4988-a800-2682cba44d72"). InnerVolumeSpecName "ovsdbserver-nb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 10 23:11:24 crc kubenswrapper[4791]: I1210 23:11:24.358024 4791 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/6202cb66-3ca7-4988-a800-2682cba44d72-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "6202cb66-3ca7-4988-a800-2682cba44d72" (UID: "6202cb66-3ca7-4988-a800-2682cba44d72"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 10 23:11:24 crc kubenswrapper[4791]: I1210 23:11:24.369029 4791 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/6202cb66-3ca7-4988-a800-2682cba44d72-config" (OuterVolumeSpecName: "config") pod "6202cb66-3ca7-4988-a800-2682cba44d72" (UID: "6202cb66-3ca7-4988-a800-2682cba44d72"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 10 23:11:24 crc kubenswrapper[4791]: I1210 23:11:24.372412 4791 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/6202cb66-3ca7-4988-a800-2682cba44d72-dns-swift-storage-0" (OuterVolumeSpecName: "dns-swift-storage-0") pod "6202cb66-3ca7-4988-a800-2682cba44d72" (UID: "6202cb66-3ca7-4988-a800-2682cba44d72"). InnerVolumeSpecName "dns-swift-storage-0". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 10 23:11:24 crc kubenswrapper[4791]: I1210 23:11:24.391061 4791 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/6202cb66-3ca7-4988-a800-2682cba44d72-dns-svc\") on node \"crc\" DevicePath \"\"" Dec 10 23:11:24 crc kubenswrapper[4791]: I1210 23:11:24.391095 4791 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/6202cb66-3ca7-4988-a800-2682cba44d72-ovsdbserver-sb\") on node \"crc\" DevicePath \"\"" Dec 10 23:11:24 crc kubenswrapper[4791]: I1210 23:11:24.391106 4791 reconciler_common.go:293] "Volume detached for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/6202cb66-3ca7-4988-a800-2682cba44d72-dns-swift-storage-0\") on node \"crc\" DevicePath \"\"" Dec 10 23:11:24 crc kubenswrapper[4791]: I1210 23:11:24.391116 4791 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-blchs\" (UniqueName: \"kubernetes.io/projected/6202cb66-3ca7-4988-a800-2682cba44d72-kube-api-access-blchs\") on node \"crc\" DevicePath \"\"" Dec 10 23:11:24 crc kubenswrapper[4791]: I1210 23:11:24.391124 4791 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/6202cb66-3ca7-4988-a800-2682cba44d72-ovsdbserver-nb\") on node \"crc\" DevicePath \"\"" Dec 10 23:11:24 crc kubenswrapper[4791]: I1210 23:11:24.391132 4791 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/6202cb66-3ca7-4988-a800-2682cba44d72-config\") on node \"crc\" DevicePath \"\"" Dec 10 23:11:24 crc kubenswrapper[4791]: W1210 23:11:24.470180 4791 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod430975e4_f09a_4c3f_9c1d_0cab25b587b1.slice/crio-57ce9ed1396199f8fb71bda9fb9b0e85a7ffc2659220789d969758c2f0d5c4b9 WatchSource:0}: Error finding container 57ce9ed1396199f8fb71bda9fb9b0e85a7ffc2659220789d969758c2f0d5c4b9: Status 404 returned error can't find the container with id 57ce9ed1396199f8fb71bda9fb9b0e85a7ffc2659220789d969758c2f0d5c4b9 Dec 10 23:11:24 crc kubenswrapper[4791]: I1210 23:11:24.488493 4791 scope.go:117] "RemoveContainer" containerID="6559890b469fbfa46c5812c5c8a37e419f0707fab3dc84b65e7d1f9a42fdd86e" Dec 10 23:11:24 crc kubenswrapper[4791]: I1210 23:11:24.488744 4791 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-cell-mapping-zcfpp"] Dec 10 23:11:24 crc kubenswrapper[4791]: I1210 23:11:24.536799 4791 scope.go:117] "RemoveContainer" containerID="23c8115e63afae1fa7a5d92c7142b2a8f9b026f2cd7fdba3e8bd355ae6b7935f" Dec 10 23:11:24 crc kubenswrapper[4791]: E1210 23:11:24.537425 4791 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"23c8115e63afae1fa7a5d92c7142b2a8f9b026f2cd7fdba3e8bd355ae6b7935f\": container with ID starting with 23c8115e63afae1fa7a5d92c7142b2a8f9b026f2cd7fdba3e8bd355ae6b7935f not found: ID does not exist" containerID="23c8115e63afae1fa7a5d92c7142b2a8f9b026f2cd7fdba3e8bd355ae6b7935f" Dec 10 23:11:24 crc kubenswrapper[4791]: I1210 23:11:24.537487 4791 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"23c8115e63afae1fa7a5d92c7142b2a8f9b026f2cd7fdba3e8bd355ae6b7935f"} err="failed to get container status \"23c8115e63afae1fa7a5d92c7142b2a8f9b026f2cd7fdba3e8bd355ae6b7935f\": rpc error: code = NotFound desc = could not find container \"23c8115e63afae1fa7a5d92c7142b2a8f9b026f2cd7fdba3e8bd355ae6b7935f\": container with ID starting with 23c8115e63afae1fa7a5d92c7142b2a8f9b026f2cd7fdba3e8bd355ae6b7935f not found: ID does not exist" Dec 10 23:11:24 crc kubenswrapper[4791]: I1210 23:11:24.537521 4791 scope.go:117] "RemoveContainer" containerID="6559890b469fbfa46c5812c5c8a37e419f0707fab3dc84b65e7d1f9a42fdd86e" Dec 10 23:11:24 crc kubenswrapper[4791]: E1210 23:11:24.538251 4791 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"6559890b469fbfa46c5812c5c8a37e419f0707fab3dc84b65e7d1f9a42fdd86e\": container with ID starting with 6559890b469fbfa46c5812c5c8a37e419f0707fab3dc84b65e7d1f9a42fdd86e not found: ID does not exist" containerID="6559890b469fbfa46c5812c5c8a37e419f0707fab3dc84b65e7d1f9a42fdd86e" Dec 10 23:11:24 crc kubenswrapper[4791]: I1210 23:11:24.538303 4791 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"6559890b469fbfa46c5812c5c8a37e419f0707fab3dc84b65e7d1f9a42fdd86e"} err="failed to get container status \"6559890b469fbfa46c5812c5c8a37e419f0707fab3dc84b65e7d1f9a42fdd86e\": rpc error: code = NotFound desc = could not find container \"6559890b469fbfa46c5812c5c8a37e419f0707fab3dc84b65e7d1f9a42fdd86e\": container with ID starting with 6559890b469fbfa46c5812c5c8a37e419f0707fab3dc84b65e7d1f9a42fdd86e not found: ID does not exist" Dec 10 23:11:24 crc kubenswrapper[4791]: I1210 23:11:24.663261 4791 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-845d6d6f59-d6v8c"] Dec 10 23:11:24 crc kubenswrapper[4791]: I1210 23:11:24.671185 4791 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-845d6d6f59-d6v8c"] Dec 10 23:11:25 crc kubenswrapper[4791]: I1210 23:11:25.038690 4791 patch_prober.go:28] interesting pod/machine-config-daemon-5rb5l container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 10 23:11:25 crc kubenswrapper[4791]: I1210 23:11:25.038766 4791 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-5rb5l" podUID="ba35653c-6e06-4cee-a4d6-137764090d18" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 10 23:11:25 crc kubenswrapper[4791]: I1210 23:11:25.340473 4791 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-cell-mapping-zcfpp" event={"ID":"430975e4-f09a-4c3f-9c1d-0cab25b587b1","Type":"ContainerStarted","Data":"8ad39df5144e867e9c45cb797ca31cbba3a462a23ea8808478880b36524e0c12"} Dec 10 23:11:25 crc kubenswrapper[4791]: I1210 23:11:25.340775 4791 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-cell-mapping-zcfpp" event={"ID":"430975e4-f09a-4c3f-9c1d-0cab25b587b1","Type":"ContainerStarted","Data":"57ce9ed1396199f8fb71bda9fb9b0e85a7ffc2659220789d969758c2f0d5c4b9"} Dec 10 23:11:25 crc kubenswrapper[4791]: I1210 23:11:25.357475 4791 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"c85398ce-a5d4-414a-8811-775a4ce71193","Type":"ContainerStarted","Data":"48c3dcfe33ad3cb676a58167d7d17175ed04de13ee4a823c894ca8d418429705"} Dec 10 23:11:25 crc kubenswrapper[4791]: I1210 23:11:25.897210 4791 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="6202cb66-3ca7-4988-a800-2682cba44d72" path="/var/lib/kubelet/pods/6202cb66-3ca7-4988-a800-2682cba44d72/volumes" Dec 10 23:11:25 crc kubenswrapper[4791]: I1210 23:11:25.915361 4791 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-cell1-cell-mapping-zcfpp" podStartSLOduration=2.915324977 podStartE2EDuration="2.915324977s" podCreationTimestamp="2025-12-10 23:11:23 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-10 23:11:25.368950452 +0000 UTC m=+1319.798568085" watchObservedRunningTime="2025-12-10 23:11:25.915324977 +0000 UTC m=+1320.344942590" Dec 10 23:11:27 crc kubenswrapper[4791]: I1210 23:11:27.375667 4791 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"c85398ce-a5d4-414a-8811-775a4ce71193","Type":"ContainerStarted","Data":"15d40e5efad14b926c99a8c728c2bd0b4fe093589bae2ddf8f275dc16e74475b"} Dec 10 23:11:27 crc kubenswrapper[4791]: I1210 23:11:27.376050 4791 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/ceilometer-0" Dec 10 23:11:27 crc kubenswrapper[4791]: I1210 23:11:27.402605 4791 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ceilometer-0" podStartSLOduration=2.371861328 podStartE2EDuration="7.402568269s" podCreationTimestamp="2025-12-10 23:11:20 +0000 UTC" firstStartedPulling="2025-12-10 23:11:21.229069942 +0000 UTC m=+1315.658687555" lastFinishedPulling="2025-12-10 23:11:26.259776893 +0000 UTC m=+1320.689394496" observedRunningTime="2025-12-10 23:11:27.398576563 +0000 UTC m=+1321.828194186" watchObservedRunningTime="2025-12-10 23:11:27.402568269 +0000 UTC m=+1321.832185882" Dec 10 23:11:30 crc kubenswrapper[4791]: I1210 23:11:30.419091 4791 generic.go:334] "Generic (PLEG): container finished" podID="430975e4-f09a-4c3f-9c1d-0cab25b587b1" containerID="8ad39df5144e867e9c45cb797ca31cbba3a462a23ea8808478880b36524e0c12" exitCode=0 Dec 10 23:11:30 crc kubenswrapper[4791]: I1210 23:11:30.419224 4791 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-cell-mapping-zcfpp" event={"ID":"430975e4-f09a-4c3f-9c1d-0cab25b587b1","Type":"ContainerDied","Data":"8ad39df5144e867e9c45cb797ca31cbba3a462a23ea8808478880b36524e0c12"} Dec 10 23:11:30 crc kubenswrapper[4791]: I1210 23:11:30.660732 4791 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-api-0" Dec 10 23:11:30 crc kubenswrapper[4791]: I1210 23:11:30.660983 4791 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-api-0" Dec 10 23:11:31 crc kubenswrapper[4791]: I1210 23:11:31.679545 4791 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/nova-api-0" podUID="6460257f-3f13-4642-b5b5-fc05065df9d7" containerName="nova-api-api" probeResult="failure" output="Get \"https://10.217.0.198:8774/\": net/http: request canceled (Client.Timeout exceeded while awaiting headers)" Dec 10 23:11:31 crc kubenswrapper[4791]: I1210 23:11:31.679684 4791 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/nova-api-0" podUID="6460257f-3f13-4642-b5b5-fc05065df9d7" containerName="nova-api-log" probeResult="failure" output="Get \"https://10.217.0.198:8774/\": net/http: request canceled (Client.Timeout exceeded while awaiting headers)" Dec 10 23:11:31 crc kubenswrapper[4791]: I1210 23:11:31.770913 4791 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-cell-mapping-zcfpp" Dec 10 23:11:31 crc kubenswrapper[4791]: I1210 23:11:31.943202 4791 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-pcfw2\" (UniqueName: \"kubernetes.io/projected/430975e4-f09a-4c3f-9c1d-0cab25b587b1-kube-api-access-pcfw2\") pod \"430975e4-f09a-4c3f-9c1d-0cab25b587b1\" (UID: \"430975e4-f09a-4c3f-9c1d-0cab25b587b1\") " Dec 10 23:11:31 crc kubenswrapper[4791]: I1210 23:11:31.943262 4791 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/430975e4-f09a-4c3f-9c1d-0cab25b587b1-combined-ca-bundle\") pod \"430975e4-f09a-4c3f-9c1d-0cab25b587b1\" (UID: \"430975e4-f09a-4c3f-9c1d-0cab25b587b1\") " Dec 10 23:11:31 crc kubenswrapper[4791]: I1210 23:11:31.943469 4791 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/430975e4-f09a-4c3f-9c1d-0cab25b587b1-config-data\") pod \"430975e4-f09a-4c3f-9c1d-0cab25b587b1\" (UID: \"430975e4-f09a-4c3f-9c1d-0cab25b587b1\") " Dec 10 23:11:31 crc kubenswrapper[4791]: I1210 23:11:31.943536 4791 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/430975e4-f09a-4c3f-9c1d-0cab25b587b1-scripts\") pod \"430975e4-f09a-4c3f-9c1d-0cab25b587b1\" (UID: \"430975e4-f09a-4c3f-9c1d-0cab25b587b1\") " Dec 10 23:11:31 crc kubenswrapper[4791]: I1210 23:11:31.954561 4791 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/430975e4-f09a-4c3f-9c1d-0cab25b587b1-scripts" (OuterVolumeSpecName: "scripts") pod "430975e4-f09a-4c3f-9c1d-0cab25b587b1" (UID: "430975e4-f09a-4c3f-9c1d-0cab25b587b1"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 10 23:11:31 crc kubenswrapper[4791]: I1210 23:11:31.960690 4791 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/430975e4-f09a-4c3f-9c1d-0cab25b587b1-kube-api-access-pcfw2" (OuterVolumeSpecName: "kube-api-access-pcfw2") pod "430975e4-f09a-4c3f-9c1d-0cab25b587b1" (UID: "430975e4-f09a-4c3f-9c1d-0cab25b587b1"). InnerVolumeSpecName "kube-api-access-pcfw2". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 10 23:11:31 crc kubenswrapper[4791]: I1210 23:11:31.974471 4791 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/430975e4-f09a-4c3f-9c1d-0cab25b587b1-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "430975e4-f09a-4c3f-9c1d-0cab25b587b1" (UID: "430975e4-f09a-4c3f-9c1d-0cab25b587b1"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 10 23:11:31 crc kubenswrapper[4791]: I1210 23:11:31.986447 4791 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/430975e4-f09a-4c3f-9c1d-0cab25b587b1-config-data" (OuterVolumeSpecName: "config-data") pod "430975e4-f09a-4c3f-9c1d-0cab25b587b1" (UID: "430975e4-f09a-4c3f-9c1d-0cab25b587b1"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 10 23:11:32 crc kubenswrapper[4791]: I1210 23:11:32.046067 4791 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/430975e4-f09a-4c3f-9c1d-0cab25b587b1-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 10 23:11:32 crc kubenswrapper[4791]: I1210 23:11:32.046098 4791 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/430975e4-f09a-4c3f-9c1d-0cab25b587b1-config-data\") on node \"crc\" DevicePath \"\"" Dec 10 23:11:32 crc kubenswrapper[4791]: I1210 23:11:32.046107 4791 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/430975e4-f09a-4c3f-9c1d-0cab25b587b1-scripts\") on node \"crc\" DevicePath \"\"" Dec 10 23:11:32 crc kubenswrapper[4791]: I1210 23:11:32.046116 4791 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-pcfw2\" (UniqueName: \"kubernetes.io/projected/430975e4-f09a-4c3f-9c1d-0cab25b587b1-kube-api-access-pcfw2\") on node \"crc\" DevicePath \"\"" Dec 10 23:11:32 crc kubenswrapper[4791]: I1210 23:11:32.441329 4791 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-cell-mapping-zcfpp" event={"ID":"430975e4-f09a-4c3f-9c1d-0cab25b587b1","Type":"ContainerDied","Data":"57ce9ed1396199f8fb71bda9fb9b0e85a7ffc2659220789d969758c2f0d5c4b9"} Dec 10 23:11:32 crc kubenswrapper[4791]: I1210 23:11:32.441392 4791 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="57ce9ed1396199f8fb71bda9fb9b0e85a7ffc2659220789d969758c2f0d5c4b9" Dec 10 23:11:32 crc kubenswrapper[4791]: I1210 23:11:32.441409 4791 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-cell-mapping-zcfpp" Dec 10 23:11:32 crc kubenswrapper[4791]: I1210 23:11:32.629627 4791 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-api-0"] Dec 10 23:11:32 crc kubenswrapper[4791]: I1210 23:11:32.630384 4791 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-api-0" podUID="6460257f-3f13-4642-b5b5-fc05065df9d7" containerName="nova-api-api" containerID="cri-o://e6e79a11bb5efb0004d4c3a54c551264e04cb8c253bb15d795f2c9dfeb974154" gracePeriod=30 Dec 10 23:11:32 crc kubenswrapper[4791]: I1210 23:11:32.629928 4791 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-api-0" podUID="6460257f-3f13-4642-b5b5-fc05065df9d7" containerName="nova-api-log" containerID="cri-o://e17b5d105e85af3d0d1bfcd6dacac2d5d1b35811ddecc4140c5bfb94cb52c0bc" gracePeriod=30 Dec 10 23:11:32 crc kubenswrapper[4791]: I1210 23:11:32.677587 4791 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-scheduler-0"] Dec 10 23:11:32 crc kubenswrapper[4791]: I1210 23:11:32.677930 4791 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-scheduler-0" podUID="f947977e-b1aa-43c5-b987-cea516e6a154" containerName="nova-scheduler-scheduler" containerID="cri-o://d5fbff44eb0dca42788f629ce6e1217af29146cb334e533f0e3c8327af2d22b4" gracePeriod=30 Dec 10 23:11:32 crc kubenswrapper[4791]: I1210 23:11:32.721013 4791 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-metadata-0"] Dec 10 23:11:32 crc kubenswrapper[4791]: I1210 23:11:32.721640 4791 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-metadata-0" podUID="1fab60e1-51e0-4101-8c37-c23d8f259cf2" containerName="nova-metadata-log" containerID="cri-o://1a13d9936e0cd13b45bac08f1af7f835c53e17adc738f8d92e00a0fd7f301300" gracePeriod=30 Dec 10 23:11:32 crc kubenswrapper[4791]: I1210 23:11:32.722239 4791 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-metadata-0" podUID="1fab60e1-51e0-4101-8c37-c23d8f259cf2" containerName="nova-metadata-metadata" containerID="cri-o://ea7f48b4a18e81bcdaa4b7d4980a08c16e8da208d46b06b9f6ae598a2931ea9d" gracePeriod=30 Dec 10 23:11:33 crc kubenswrapper[4791]: I1210 23:11:33.452221 4791 generic.go:334] "Generic (PLEG): container finished" podID="6460257f-3f13-4642-b5b5-fc05065df9d7" containerID="e17b5d105e85af3d0d1bfcd6dacac2d5d1b35811ddecc4140c5bfb94cb52c0bc" exitCode=143 Dec 10 23:11:33 crc kubenswrapper[4791]: I1210 23:11:33.452299 4791 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"6460257f-3f13-4642-b5b5-fc05065df9d7","Type":"ContainerDied","Data":"e17b5d105e85af3d0d1bfcd6dacac2d5d1b35811ddecc4140c5bfb94cb52c0bc"} Dec 10 23:11:33 crc kubenswrapper[4791]: I1210 23:11:33.457579 4791 generic.go:334] "Generic (PLEG): container finished" podID="1fab60e1-51e0-4101-8c37-c23d8f259cf2" containerID="1a13d9936e0cd13b45bac08f1af7f835c53e17adc738f8d92e00a0fd7f301300" exitCode=143 Dec 10 23:11:33 crc kubenswrapper[4791]: I1210 23:11:33.457700 4791 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"1fab60e1-51e0-4101-8c37-c23d8f259cf2","Type":"ContainerDied","Data":"1a13d9936e0cd13b45bac08f1af7f835c53e17adc738f8d92e00a0fd7f301300"} Dec 10 23:11:35 crc kubenswrapper[4791]: E1210 23:11:35.265814 4791 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = NotFound desc = container is not created or running: checking if PID of d5fbff44eb0dca42788f629ce6e1217af29146cb334e533f0e3c8327af2d22b4 is running failed: container process not found" containerID="d5fbff44eb0dca42788f629ce6e1217af29146cb334e533f0e3c8327af2d22b4" cmd=["/usr/bin/pgrep","-r","DRST","nova-scheduler"] Dec 10 23:11:35 crc kubenswrapper[4791]: E1210 23:11:35.266800 4791 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = NotFound desc = container is not created or running: checking if PID of d5fbff44eb0dca42788f629ce6e1217af29146cb334e533f0e3c8327af2d22b4 is running failed: container process not found" containerID="d5fbff44eb0dca42788f629ce6e1217af29146cb334e533f0e3c8327af2d22b4" cmd=["/usr/bin/pgrep","-r","DRST","nova-scheduler"] Dec 10 23:11:35 crc kubenswrapper[4791]: E1210 23:11:35.267316 4791 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = NotFound desc = container is not created or running: checking if PID of d5fbff44eb0dca42788f629ce6e1217af29146cb334e533f0e3c8327af2d22b4 is running failed: container process not found" containerID="d5fbff44eb0dca42788f629ce6e1217af29146cb334e533f0e3c8327af2d22b4" cmd=["/usr/bin/pgrep","-r","DRST","nova-scheduler"] Dec 10 23:11:35 crc kubenswrapper[4791]: E1210 23:11:35.267421 4791 prober.go:104] "Probe errored" err="rpc error: code = NotFound desc = container is not created or running: checking if PID of d5fbff44eb0dca42788f629ce6e1217af29146cb334e533f0e3c8327af2d22b4 is running failed: container process not found" probeType="Readiness" pod="openstack/nova-scheduler-0" podUID="f947977e-b1aa-43c5-b987-cea516e6a154" containerName="nova-scheduler-scheduler" Dec 10 23:11:35 crc kubenswrapper[4791]: I1210 23:11:35.475891 4791 generic.go:334] "Generic (PLEG): container finished" podID="f947977e-b1aa-43c5-b987-cea516e6a154" containerID="d5fbff44eb0dca42788f629ce6e1217af29146cb334e533f0e3c8327af2d22b4" exitCode=0 Dec 10 23:11:35 crc kubenswrapper[4791]: I1210 23:11:35.475975 4791 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-scheduler-0" event={"ID":"f947977e-b1aa-43c5-b987-cea516e6a154","Type":"ContainerDied","Data":"d5fbff44eb0dca42788f629ce6e1217af29146cb334e533f0e3c8327af2d22b4"} Dec 10 23:11:35 crc kubenswrapper[4791]: I1210 23:11:35.476196 4791 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-scheduler-0" event={"ID":"f947977e-b1aa-43c5-b987-cea516e6a154","Type":"ContainerDied","Data":"d41024783f0519ec8b3a38b1ab62cebe3f7c2d789f1d7ab6a7610be4985ae070"} Dec 10 23:11:35 crc kubenswrapper[4791]: I1210 23:11:35.476212 4791 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="d41024783f0519ec8b3a38b1ab62cebe3f7c2d789f1d7ab6a7610be4985ae070" Dec 10 23:11:35 crc kubenswrapper[4791]: I1210 23:11:35.519522 4791 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-scheduler-0" Dec 10 23:11:35 crc kubenswrapper[4791]: I1210 23:11:35.633174 4791 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f947977e-b1aa-43c5-b987-cea516e6a154-combined-ca-bundle\") pod \"f947977e-b1aa-43c5-b987-cea516e6a154\" (UID: \"f947977e-b1aa-43c5-b987-cea516e6a154\") " Dec 10 23:11:35 crc kubenswrapper[4791]: I1210 23:11:35.634466 4791 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/f947977e-b1aa-43c5-b987-cea516e6a154-config-data\") pod \"f947977e-b1aa-43c5-b987-cea516e6a154\" (UID: \"f947977e-b1aa-43c5-b987-cea516e6a154\") " Dec 10 23:11:35 crc kubenswrapper[4791]: I1210 23:11:35.634711 4791 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-vkc9h\" (UniqueName: \"kubernetes.io/projected/f947977e-b1aa-43c5-b987-cea516e6a154-kube-api-access-vkc9h\") pod \"f947977e-b1aa-43c5-b987-cea516e6a154\" (UID: \"f947977e-b1aa-43c5-b987-cea516e6a154\") " Dec 10 23:11:35 crc kubenswrapper[4791]: I1210 23:11:35.649109 4791 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/f947977e-b1aa-43c5-b987-cea516e6a154-kube-api-access-vkc9h" (OuterVolumeSpecName: "kube-api-access-vkc9h") pod "f947977e-b1aa-43c5-b987-cea516e6a154" (UID: "f947977e-b1aa-43c5-b987-cea516e6a154"). InnerVolumeSpecName "kube-api-access-vkc9h". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 10 23:11:35 crc kubenswrapper[4791]: I1210 23:11:35.674407 4791 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/f947977e-b1aa-43c5-b987-cea516e6a154-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "f947977e-b1aa-43c5-b987-cea516e6a154" (UID: "f947977e-b1aa-43c5-b987-cea516e6a154"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 10 23:11:35 crc kubenswrapper[4791]: I1210 23:11:35.676564 4791 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/f947977e-b1aa-43c5-b987-cea516e6a154-config-data" (OuterVolumeSpecName: "config-data") pod "f947977e-b1aa-43c5-b987-cea516e6a154" (UID: "f947977e-b1aa-43c5-b987-cea516e6a154"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 10 23:11:35 crc kubenswrapper[4791]: I1210 23:11:35.737197 4791 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/f947977e-b1aa-43c5-b987-cea516e6a154-config-data\") on node \"crc\" DevicePath \"\"" Dec 10 23:11:35 crc kubenswrapper[4791]: I1210 23:11:35.737480 4791 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-vkc9h\" (UniqueName: \"kubernetes.io/projected/f947977e-b1aa-43c5-b987-cea516e6a154-kube-api-access-vkc9h\") on node \"crc\" DevicePath \"\"" Dec 10 23:11:35 crc kubenswrapper[4791]: I1210 23:11:35.737558 4791 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f947977e-b1aa-43c5-b987-cea516e6a154-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 10 23:11:35 crc kubenswrapper[4791]: I1210 23:11:35.874270 4791 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/nova-metadata-0" podUID="1fab60e1-51e0-4101-8c37-c23d8f259cf2" containerName="nova-metadata-metadata" probeResult="failure" output="Get \"https://10.217.0.191:8775/\": read tcp 10.217.0.2:59774->10.217.0.191:8775: read: connection reset by peer" Dec 10 23:11:35 crc kubenswrapper[4791]: I1210 23:11:35.874308 4791 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/nova-metadata-0" podUID="1fab60e1-51e0-4101-8c37-c23d8f259cf2" containerName="nova-metadata-log" probeResult="failure" output="Get \"https://10.217.0.191:8775/\": read tcp 10.217.0.2:59766->10.217.0.191:8775: read: connection reset by peer" Dec 10 23:11:36 crc kubenswrapper[4791]: I1210 23:11:36.314880 4791 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-metadata-0" Dec 10 23:11:36 crc kubenswrapper[4791]: I1210 23:11:36.485581 4791 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"nova-metadata-tls-certs\" (UniqueName: \"kubernetes.io/secret/1fab60e1-51e0-4101-8c37-c23d8f259cf2-nova-metadata-tls-certs\") pod \"1fab60e1-51e0-4101-8c37-c23d8f259cf2\" (UID: \"1fab60e1-51e0-4101-8c37-c23d8f259cf2\") " Dec 10 23:11:36 crc kubenswrapper[4791]: I1210 23:11:36.485680 4791 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/1fab60e1-51e0-4101-8c37-c23d8f259cf2-logs\") pod \"1fab60e1-51e0-4101-8c37-c23d8f259cf2\" (UID: \"1fab60e1-51e0-4101-8c37-c23d8f259cf2\") " Dec 10 23:11:36 crc kubenswrapper[4791]: I1210 23:11:36.485704 4791 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/1fab60e1-51e0-4101-8c37-c23d8f259cf2-combined-ca-bundle\") pod \"1fab60e1-51e0-4101-8c37-c23d8f259cf2\" (UID: \"1fab60e1-51e0-4101-8c37-c23d8f259cf2\") " Dec 10 23:11:36 crc kubenswrapper[4791]: I1210 23:11:36.485725 4791 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/1fab60e1-51e0-4101-8c37-c23d8f259cf2-config-data\") pod \"1fab60e1-51e0-4101-8c37-c23d8f259cf2\" (UID: \"1fab60e1-51e0-4101-8c37-c23d8f259cf2\") " Dec 10 23:11:36 crc kubenswrapper[4791]: I1210 23:11:36.485864 4791 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-mrhgz\" (UniqueName: \"kubernetes.io/projected/1fab60e1-51e0-4101-8c37-c23d8f259cf2-kube-api-access-mrhgz\") pod \"1fab60e1-51e0-4101-8c37-c23d8f259cf2\" (UID: \"1fab60e1-51e0-4101-8c37-c23d8f259cf2\") " Dec 10 23:11:36 crc kubenswrapper[4791]: I1210 23:11:36.486202 4791 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/1fab60e1-51e0-4101-8c37-c23d8f259cf2-logs" (OuterVolumeSpecName: "logs") pod "1fab60e1-51e0-4101-8c37-c23d8f259cf2" (UID: "1fab60e1-51e0-4101-8c37-c23d8f259cf2"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 10 23:11:36 crc kubenswrapper[4791]: I1210 23:11:36.486517 4791 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/1fab60e1-51e0-4101-8c37-c23d8f259cf2-logs\") on node \"crc\" DevicePath \"\"" Dec 10 23:11:36 crc kubenswrapper[4791]: I1210 23:11:36.491821 4791 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/1fab60e1-51e0-4101-8c37-c23d8f259cf2-kube-api-access-mrhgz" (OuterVolumeSpecName: "kube-api-access-mrhgz") pod "1fab60e1-51e0-4101-8c37-c23d8f259cf2" (UID: "1fab60e1-51e0-4101-8c37-c23d8f259cf2"). InnerVolumeSpecName "kube-api-access-mrhgz". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 10 23:11:36 crc kubenswrapper[4791]: I1210 23:11:36.492947 4791 generic.go:334] "Generic (PLEG): container finished" podID="1fab60e1-51e0-4101-8c37-c23d8f259cf2" containerID="ea7f48b4a18e81bcdaa4b7d4980a08c16e8da208d46b06b9f6ae598a2931ea9d" exitCode=0 Dec 10 23:11:36 crc kubenswrapper[4791]: I1210 23:11:36.493039 4791 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"1fab60e1-51e0-4101-8c37-c23d8f259cf2","Type":"ContainerDied","Data":"ea7f48b4a18e81bcdaa4b7d4980a08c16e8da208d46b06b9f6ae598a2931ea9d"} Dec 10 23:11:36 crc kubenswrapper[4791]: I1210 23:11:36.493222 4791 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-metadata-0" Dec 10 23:11:36 crc kubenswrapper[4791]: I1210 23:11:36.493412 4791 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"1fab60e1-51e0-4101-8c37-c23d8f259cf2","Type":"ContainerDied","Data":"12a745ff744ef1b5f57d94629dd336396569e9fbc02be11d7305eb4edd2d4c83"} Dec 10 23:11:36 crc kubenswrapper[4791]: I1210 23:11:36.493462 4791 scope.go:117] "RemoveContainer" containerID="ea7f48b4a18e81bcdaa4b7d4980a08c16e8da208d46b06b9f6ae598a2931ea9d" Dec 10 23:11:36 crc kubenswrapper[4791]: I1210 23:11:36.493634 4791 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-scheduler-0" Dec 10 23:11:36 crc kubenswrapper[4791]: I1210 23:11:36.518439 4791 scope.go:117] "RemoveContainer" containerID="1a13d9936e0cd13b45bac08f1af7f835c53e17adc738f8d92e00a0fd7f301300" Dec 10 23:11:36 crc kubenswrapper[4791]: I1210 23:11:36.531028 4791 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/1fab60e1-51e0-4101-8c37-c23d8f259cf2-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "1fab60e1-51e0-4101-8c37-c23d8f259cf2" (UID: "1fab60e1-51e0-4101-8c37-c23d8f259cf2"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 10 23:11:36 crc kubenswrapper[4791]: I1210 23:11:36.541040 4791 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-scheduler-0"] Dec 10 23:11:36 crc kubenswrapper[4791]: I1210 23:11:36.555189 4791 scope.go:117] "RemoveContainer" containerID="ea7f48b4a18e81bcdaa4b7d4980a08c16e8da208d46b06b9f6ae598a2931ea9d" Dec 10 23:11:36 crc kubenswrapper[4791]: E1210 23:11:36.556094 4791 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"ea7f48b4a18e81bcdaa4b7d4980a08c16e8da208d46b06b9f6ae598a2931ea9d\": container with ID starting with ea7f48b4a18e81bcdaa4b7d4980a08c16e8da208d46b06b9f6ae598a2931ea9d not found: ID does not exist" containerID="ea7f48b4a18e81bcdaa4b7d4980a08c16e8da208d46b06b9f6ae598a2931ea9d" Dec 10 23:11:36 crc kubenswrapper[4791]: I1210 23:11:36.556184 4791 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"ea7f48b4a18e81bcdaa4b7d4980a08c16e8da208d46b06b9f6ae598a2931ea9d"} err="failed to get container status \"ea7f48b4a18e81bcdaa4b7d4980a08c16e8da208d46b06b9f6ae598a2931ea9d\": rpc error: code = NotFound desc = could not find container \"ea7f48b4a18e81bcdaa4b7d4980a08c16e8da208d46b06b9f6ae598a2931ea9d\": container with ID starting with ea7f48b4a18e81bcdaa4b7d4980a08c16e8da208d46b06b9f6ae598a2931ea9d not found: ID does not exist" Dec 10 23:11:36 crc kubenswrapper[4791]: I1210 23:11:36.556209 4791 scope.go:117] "RemoveContainer" containerID="1a13d9936e0cd13b45bac08f1af7f835c53e17adc738f8d92e00a0fd7f301300" Dec 10 23:11:36 crc kubenswrapper[4791]: E1210 23:11:36.556560 4791 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"1a13d9936e0cd13b45bac08f1af7f835c53e17adc738f8d92e00a0fd7f301300\": container with ID starting with 1a13d9936e0cd13b45bac08f1af7f835c53e17adc738f8d92e00a0fd7f301300 not found: ID does not exist" containerID="1a13d9936e0cd13b45bac08f1af7f835c53e17adc738f8d92e00a0fd7f301300" Dec 10 23:11:36 crc kubenswrapper[4791]: I1210 23:11:36.556628 4791 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"1a13d9936e0cd13b45bac08f1af7f835c53e17adc738f8d92e00a0fd7f301300"} err="failed to get container status \"1a13d9936e0cd13b45bac08f1af7f835c53e17adc738f8d92e00a0fd7f301300\": rpc error: code = NotFound desc = could not find container \"1a13d9936e0cd13b45bac08f1af7f835c53e17adc738f8d92e00a0fd7f301300\": container with ID starting with 1a13d9936e0cd13b45bac08f1af7f835c53e17adc738f8d92e00a0fd7f301300 not found: ID does not exist" Dec 10 23:11:36 crc kubenswrapper[4791]: I1210 23:11:36.558453 4791 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-scheduler-0"] Dec 10 23:11:36 crc kubenswrapper[4791]: I1210 23:11:36.563650 4791 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/1fab60e1-51e0-4101-8c37-c23d8f259cf2-config-data" (OuterVolumeSpecName: "config-data") pod "1fab60e1-51e0-4101-8c37-c23d8f259cf2" (UID: "1fab60e1-51e0-4101-8c37-c23d8f259cf2"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 10 23:11:36 crc kubenswrapper[4791]: I1210 23:11:36.573568 4791 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-scheduler-0"] Dec 10 23:11:36 crc kubenswrapper[4791]: E1210 23:11:36.574062 4791 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="6202cb66-3ca7-4988-a800-2682cba44d72" containerName="init" Dec 10 23:11:36 crc kubenswrapper[4791]: I1210 23:11:36.574078 4791 state_mem.go:107] "Deleted CPUSet assignment" podUID="6202cb66-3ca7-4988-a800-2682cba44d72" containerName="init" Dec 10 23:11:36 crc kubenswrapper[4791]: E1210 23:11:36.574099 4791 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="430975e4-f09a-4c3f-9c1d-0cab25b587b1" containerName="nova-manage" Dec 10 23:11:36 crc kubenswrapper[4791]: I1210 23:11:36.574105 4791 state_mem.go:107] "Deleted CPUSet assignment" podUID="430975e4-f09a-4c3f-9c1d-0cab25b587b1" containerName="nova-manage" Dec 10 23:11:36 crc kubenswrapper[4791]: E1210 23:11:36.574114 4791 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="1fab60e1-51e0-4101-8c37-c23d8f259cf2" containerName="nova-metadata-metadata" Dec 10 23:11:36 crc kubenswrapper[4791]: I1210 23:11:36.574120 4791 state_mem.go:107] "Deleted CPUSet assignment" podUID="1fab60e1-51e0-4101-8c37-c23d8f259cf2" containerName="nova-metadata-metadata" Dec 10 23:11:36 crc kubenswrapper[4791]: E1210 23:11:36.574136 4791 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="6202cb66-3ca7-4988-a800-2682cba44d72" containerName="dnsmasq-dns" Dec 10 23:11:36 crc kubenswrapper[4791]: I1210 23:11:36.574142 4791 state_mem.go:107] "Deleted CPUSet assignment" podUID="6202cb66-3ca7-4988-a800-2682cba44d72" containerName="dnsmasq-dns" Dec 10 23:11:36 crc kubenswrapper[4791]: E1210 23:11:36.574153 4791 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f947977e-b1aa-43c5-b987-cea516e6a154" containerName="nova-scheduler-scheduler" Dec 10 23:11:36 crc kubenswrapper[4791]: I1210 23:11:36.574162 4791 state_mem.go:107] "Deleted CPUSet assignment" podUID="f947977e-b1aa-43c5-b987-cea516e6a154" containerName="nova-scheduler-scheduler" Dec 10 23:11:36 crc kubenswrapper[4791]: E1210 23:11:36.574187 4791 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="1fab60e1-51e0-4101-8c37-c23d8f259cf2" containerName="nova-metadata-log" Dec 10 23:11:36 crc kubenswrapper[4791]: I1210 23:11:36.574193 4791 state_mem.go:107] "Deleted CPUSet assignment" podUID="1fab60e1-51e0-4101-8c37-c23d8f259cf2" containerName="nova-metadata-log" Dec 10 23:11:36 crc kubenswrapper[4791]: I1210 23:11:36.574468 4791 memory_manager.go:354] "RemoveStaleState removing state" podUID="1fab60e1-51e0-4101-8c37-c23d8f259cf2" containerName="nova-metadata-metadata" Dec 10 23:11:36 crc kubenswrapper[4791]: I1210 23:11:36.574491 4791 memory_manager.go:354] "RemoveStaleState removing state" podUID="1fab60e1-51e0-4101-8c37-c23d8f259cf2" containerName="nova-metadata-log" Dec 10 23:11:36 crc kubenswrapper[4791]: I1210 23:11:36.574511 4791 memory_manager.go:354] "RemoveStaleState removing state" podUID="430975e4-f09a-4c3f-9c1d-0cab25b587b1" containerName="nova-manage" Dec 10 23:11:36 crc kubenswrapper[4791]: I1210 23:11:36.574527 4791 memory_manager.go:354] "RemoveStaleState removing state" podUID="6202cb66-3ca7-4988-a800-2682cba44d72" containerName="dnsmasq-dns" Dec 10 23:11:36 crc kubenswrapper[4791]: I1210 23:11:36.574544 4791 memory_manager.go:354] "RemoveStaleState removing state" podUID="f947977e-b1aa-43c5-b987-cea516e6a154" containerName="nova-scheduler-scheduler" Dec 10 23:11:36 crc kubenswrapper[4791]: I1210 23:11:36.575671 4791 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-scheduler-0" Dec 10 23:11:36 crc kubenswrapper[4791]: I1210 23:11:36.583247 4791 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-scheduler-config-data" Dec 10 23:11:36 crc kubenswrapper[4791]: I1210 23:11:36.584614 4791 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-scheduler-0"] Dec 10 23:11:36 crc kubenswrapper[4791]: I1210 23:11:36.587685 4791 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/1fab60e1-51e0-4101-8c37-c23d8f259cf2-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 10 23:11:36 crc kubenswrapper[4791]: I1210 23:11:36.587724 4791 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/1fab60e1-51e0-4101-8c37-c23d8f259cf2-config-data\") on node \"crc\" DevicePath \"\"" Dec 10 23:11:36 crc kubenswrapper[4791]: I1210 23:11:36.587777 4791 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-mrhgz\" (UniqueName: \"kubernetes.io/projected/1fab60e1-51e0-4101-8c37-c23d8f259cf2-kube-api-access-mrhgz\") on node \"crc\" DevicePath \"\"" Dec 10 23:11:36 crc kubenswrapper[4791]: I1210 23:11:36.590644 4791 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/1fab60e1-51e0-4101-8c37-c23d8f259cf2-nova-metadata-tls-certs" (OuterVolumeSpecName: "nova-metadata-tls-certs") pod "1fab60e1-51e0-4101-8c37-c23d8f259cf2" (UID: "1fab60e1-51e0-4101-8c37-c23d8f259cf2"). InnerVolumeSpecName "nova-metadata-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 10 23:11:36 crc kubenswrapper[4791]: I1210 23:11:36.690950 4791 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-84vmm\" (UniqueName: \"kubernetes.io/projected/b1d1b38e-2897-4c52-a4e4-d994dfffe805-kube-api-access-84vmm\") pod \"nova-scheduler-0\" (UID: \"b1d1b38e-2897-4c52-a4e4-d994dfffe805\") " pod="openstack/nova-scheduler-0" Dec 10 23:11:36 crc kubenswrapper[4791]: I1210 23:11:36.691489 4791 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/b1d1b38e-2897-4c52-a4e4-d994dfffe805-combined-ca-bundle\") pod \"nova-scheduler-0\" (UID: \"b1d1b38e-2897-4c52-a4e4-d994dfffe805\") " pod="openstack/nova-scheduler-0" Dec 10 23:11:36 crc kubenswrapper[4791]: I1210 23:11:36.691641 4791 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/b1d1b38e-2897-4c52-a4e4-d994dfffe805-config-data\") pod \"nova-scheduler-0\" (UID: \"b1d1b38e-2897-4c52-a4e4-d994dfffe805\") " pod="openstack/nova-scheduler-0" Dec 10 23:11:36 crc kubenswrapper[4791]: I1210 23:11:36.691803 4791 reconciler_common.go:293] "Volume detached for volume \"nova-metadata-tls-certs\" (UniqueName: \"kubernetes.io/secret/1fab60e1-51e0-4101-8c37-c23d8f259cf2-nova-metadata-tls-certs\") on node \"crc\" DevicePath \"\"" Dec 10 23:11:36 crc kubenswrapper[4791]: I1210 23:11:36.792891 4791 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/b1d1b38e-2897-4c52-a4e4-d994dfffe805-combined-ca-bundle\") pod \"nova-scheduler-0\" (UID: \"b1d1b38e-2897-4c52-a4e4-d994dfffe805\") " pod="openstack/nova-scheduler-0" Dec 10 23:11:36 crc kubenswrapper[4791]: I1210 23:11:36.792964 4791 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/b1d1b38e-2897-4c52-a4e4-d994dfffe805-config-data\") pod \"nova-scheduler-0\" (UID: \"b1d1b38e-2897-4c52-a4e4-d994dfffe805\") " pod="openstack/nova-scheduler-0" Dec 10 23:11:36 crc kubenswrapper[4791]: I1210 23:11:36.793029 4791 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-84vmm\" (UniqueName: \"kubernetes.io/projected/b1d1b38e-2897-4c52-a4e4-d994dfffe805-kube-api-access-84vmm\") pod \"nova-scheduler-0\" (UID: \"b1d1b38e-2897-4c52-a4e4-d994dfffe805\") " pod="openstack/nova-scheduler-0" Dec 10 23:11:36 crc kubenswrapper[4791]: I1210 23:11:36.796575 4791 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/b1d1b38e-2897-4c52-a4e4-d994dfffe805-config-data\") pod \"nova-scheduler-0\" (UID: \"b1d1b38e-2897-4c52-a4e4-d994dfffe805\") " pod="openstack/nova-scheduler-0" Dec 10 23:11:36 crc kubenswrapper[4791]: I1210 23:11:36.804954 4791 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/b1d1b38e-2897-4c52-a4e4-d994dfffe805-combined-ca-bundle\") pod \"nova-scheduler-0\" (UID: \"b1d1b38e-2897-4c52-a4e4-d994dfffe805\") " pod="openstack/nova-scheduler-0" Dec 10 23:11:36 crc kubenswrapper[4791]: I1210 23:11:36.811526 4791 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-84vmm\" (UniqueName: \"kubernetes.io/projected/b1d1b38e-2897-4c52-a4e4-d994dfffe805-kube-api-access-84vmm\") pod \"nova-scheduler-0\" (UID: \"b1d1b38e-2897-4c52-a4e4-d994dfffe805\") " pod="openstack/nova-scheduler-0" Dec 10 23:11:36 crc kubenswrapper[4791]: I1210 23:11:36.898890 4791 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-metadata-0"] Dec 10 23:11:36 crc kubenswrapper[4791]: I1210 23:11:36.907725 4791 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-scheduler-0" Dec 10 23:11:36 crc kubenswrapper[4791]: I1210 23:11:36.911917 4791 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-metadata-0"] Dec 10 23:11:36 crc kubenswrapper[4791]: I1210 23:11:36.935136 4791 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-metadata-0"] Dec 10 23:11:36 crc kubenswrapper[4791]: I1210 23:11:36.937605 4791 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-metadata-0" Dec 10 23:11:36 crc kubenswrapper[4791]: I1210 23:11:36.940489 4791 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-nova-metadata-internal-svc" Dec 10 23:11:36 crc kubenswrapper[4791]: I1210 23:11:36.940709 4791 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-metadata-config-data" Dec 10 23:11:36 crc kubenswrapper[4791]: I1210 23:11:36.962011 4791 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-metadata-0"] Dec 10 23:11:37 crc kubenswrapper[4791]: I1210 23:11:37.002463 4791 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-w98j8\" (UniqueName: \"kubernetes.io/projected/8ee8d247-355c-4e17-965c-0b3c958a546c-kube-api-access-w98j8\") pod \"nova-metadata-0\" (UID: \"8ee8d247-355c-4e17-965c-0b3c958a546c\") " pod="openstack/nova-metadata-0" Dec 10 23:11:37 crc kubenswrapper[4791]: I1210 23:11:37.002557 4791 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/8ee8d247-355c-4e17-965c-0b3c958a546c-config-data\") pod \"nova-metadata-0\" (UID: \"8ee8d247-355c-4e17-965c-0b3c958a546c\") " pod="openstack/nova-metadata-0" Dec 10 23:11:37 crc kubenswrapper[4791]: I1210 23:11:37.002657 4791 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/8ee8d247-355c-4e17-965c-0b3c958a546c-combined-ca-bundle\") pod \"nova-metadata-0\" (UID: \"8ee8d247-355c-4e17-965c-0b3c958a546c\") " pod="openstack/nova-metadata-0" Dec 10 23:11:37 crc kubenswrapper[4791]: I1210 23:11:37.002679 4791 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"nova-metadata-tls-certs\" (UniqueName: \"kubernetes.io/secret/8ee8d247-355c-4e17-965c-0b3c958a546c-nova-metadata-tls-certs\") pod \"nova-metadata-0\" (UID: \"8ee8d247-355c-4e17-965c-0b3c958a546c\") " pod="openstack/nova-metadata-0" Dec 10 23:11:37 crc kubenswrapper[4791]: I1210 23:11:37.002736 4791 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/8ee8d247-355c-4e17-965c-0b3c958a546c-logs\") pod \"nova-metadata-0\" (UID: \"8ee8d247-355c-4e17-965c-0b3c958a546c\") " pod="openstack/nova-metadata-0" Dec 10 23:11:37 crc kubenswrapper[4791]: I1210 23:11:37.104197 4791 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-w98j8\" (UniqueName: \"kubernetes.io/projected/8ee8d247-355c-4e17-965c-0b3c958a546c-kube-api-access-w98j8\") pod \"nova-metadata-0\" (UID: \"8ee8d247-355c-4e17-965c-0b3c958a546c\") " pod="openstack/nova-metadata-0" Dec 10 23:11:37 crc kubenswrapper[4791]: I1210 23:11:37.104291 4791 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/8ee8d247-355c-4e17-965c-0b3c958a546c-config-data\") pod \"nova-metadata-0\" (UID: \"8ee8d247-355c-4e17-965c-0b3c958a546c\") " pod="openstack/nova-metadata-0" Dec 10 23:11:37 crc kubenswrapper[4791]: I1210 23:11:37.104357 4791 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/8ee8d247-355c-4e17-965c-0b3c958a546c-combined-ca-bundle\") pod \"nova-metadata-0\" (UID: \"8ee8d247-355c-4e17-965c-0b3c958a546c\") " pod="openstack/nova-metadata-0" Dec 10 23:11:37 crc kubenswrapper[4791]: I1210 23:11:37.104387 4791 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nova-metadata-tls-certs\" (UniqueName: \"kubernetes.io/secret/8ee8d247-355c-4e17-965c-0b3c958a546c-nova-metadata-tls-certs\") pod \"nova-metadata-0\" (UID: \"8ee8d247-355c-4e17-965c-0b3c958a546c\") " pod="openstack/nova-metadata-0" Dec 10 23:11:37 crc kubenswrapper[4791]: I1210 23:11:37.104442 4791 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/8ee8d247-355c-4e17-965c-0b3c958a546c-logs\") pod \"nova-metadata-0\" (UID: \"8ee8d247-355c-4e17-965c-0b3c958a546c\") " pod="openstack/nova-metadata-0" Dec 10 23:11:37 crc kubenswrapper[4791]: I1210 23:11:37.105852 4791 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/8ee8d247-355c-4e17-965c-0b3c958a546c-logs\") pod \"nova-metadata-0\" (UID: \"8ee8d247-355c-4e17-965c-0b3c958a546c\") " pod="openstack/nova-metadata-0" Dec 10 23:11:37 crc kubenswrapper[4791]: I1210 23:11:37.110295 4791 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"nova-metadata-tls-certs\" (UniqueName: \"kubernetes.io/secret/8ee8d247-355c-4e17-965c-0b3c958a546c-nova-metadata-tls-certs\") pod \"nova-metadata-0\" (UID: \"8ee8d247-355c-4e17-965c-0b3c958a546c\") " pod="openstack/nova-metadata-0" Dec 10 23:11:37 crc kubenswrapper[4791]: I1210 23:11:37.114453 4791 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/8ee8d247-355c-4e17-965c-0b3c958a546c-config-data\") pod \"nova-metadata-0\" (UID: \"8ee8d247-355c-4e17-965c-0b3c958a546c\") " pod="openstack/nova-metadata-0" Dec 10 23:11:37 crc kubenswrapper[4791]: I1210 23:11:37.123581 4791 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/8ee8d247-355c-4e17-965c-0b3c958a546c-combined-ca-bundle\") pod \"nova-metadata-0\" (UID: \"8ee8d247-355c-4e17-965c-0b3c958a546c\") " pod="openstack/nova-metadata-0" Dec 10 23:11:37 crc kubenswrapper[4791]: I1210 23:11:37.124360 4791 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-w98j8\" (UniqueName: \"kubernetes.io/projected/8ee8d247-355c-4e17-965c-0b3c958a546c-kube-api-access-w98j8\") pod \"nova-metadata-0\" (UID: \"8ee8d247-355c-4e17-965c-0b3c958a546c\") " pod="openstack/nova-metadata-0" Dec 10 23:11:37 crc kubenswrapper[4791]: I1210 23:11:37.357238 4791 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-metadata-0" Dec 10 23:11:37 crc kubenswrapper[4791]: I1210 23:11:37.421537 4791 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-scheduler-0"] Dec 10 23:11:37 crc kubenswrapper[4791]: W1210 23:11:37.424493 4791 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podb1d1b38e_2897_4c52_a4e4_d994dfffe805.slice/crio-8d1f53a0282e49859333085bd7947a5a7cec0cb18a0f118fd6e285fbca02c72f WatchSource:0}: Error finding container 8d1f53a0282e49859333085bd7947a5a7cec0cb18a0f118fd6e285fbca02c72f: Status 404 returned error can't find the container with id 8d1f53a0282e49859333085bd7947a5a7cec0cb18a0f118fd6e285fbca02c72f Dec 10 23:11:37 crc kubenswrapper[4791]: I1210 23:11:37.515378 4791 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-scheduler-0" event={"ID":"b1d1b38e-2897-4c52-a4e4-d994dfffe805","Type":"ContainerStarted","Data":"8d1f53a0282e49859333085bd7947a5a7cec0cb18a0f118fd6e285fbca02c72f"} Dec 10 23:11:37 crc kubenswrapper[4791]: I1210 23:11:37.519694 4791 generic.go:334] "Generic (PLEG): container finished" podID="6460257f-3f13-4642-b5b5-fc05065df9d7" containerID="e6e79a11bb5efb0004d4c3a54c551264e04cb8c253bb15d795f2c9dfeb974154" exitCode=0 Dec 10 23:11:37 crc kubenswrapper[4791]: I1210 23:11:37.519725 4791 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"6460257f-3f13-4642-b5b5-fc05065df9d7","Type":"ContainerDied","Data":"e6e79a11bb5efb0004d4c3a54c551264e04cb8c253bb15d795f2c9dfeb974154"} Dec 10 23:11:37 crc kubenswrapper[4791]: I1210 23:11:37.522696 4791 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Dec 10 23:11:37 crc kubenswrapper[4791]: I1210 23:11:37.616975 4791 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-6mrmt\" (UniqueName: \"kubernetes.io/projected/6460257f-3f13-4642-b5b5-fc05065df9d7-kube-api-access-6mrmt\") pod \"6460257f-3f13-4642-b5b5-fc05065df9d7\" (UID: \"6460257f-3f13-4642-b5b5-fc05065df9d7\") " Dec 10 23:11:37 crc kubenswrapper[4791]: I1210 23:11:37.617541 4791 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/6460257f-3f13-4642-b5b5-fc05065df9d7-internal-tls-certs\") pod \"6460257f-3f13-4642-b5b5-fc05065df9d7\" (UID: \"6460257f-3f13-4642-b5b5-fc05065df9d7\") " Dec 10 23:11:37 crc kubenswrapper[4791]: I1210 23:11:37.617563 4791 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/6460257f-3f13-4642-b5b5-fc05065df9d7-config-data\") pod \"6460257f-3f13-4642-b5b5-fc05065df9d7\" (UID: \"6460257f-3f13-4642-b5b5-fc05065df9d7\") " Dec 10 23:11:37 crc kubenswrapper[4791]: I1210 23:11:37.617614 4791 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/6460257f-3f13-4642-b5b5-fc05065df9d7-logs\") pod \"6460257f-3f13-4642-b5b5-fc05065df9d7\" (UID: \"6460257f-3f13-4642-b5b5-fc05065df9d7\") " Dec 10 23:11:37 crc kubenswrapper[4791]: I1210 23:11:37.617640 4791 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/6460257f-3f13-4642-b5b5-fc05065df9d7-public-tls-certs\") pod \"6460257f-3f13-4642-b5b5-fc05065df9d7\" (UID: \"6460257f-3f13-4642-b5b5-fc05065df9d7\") " Dec 10 23:11:37 crc kubenswrapper[4791]: I1210 23:11:37.617664 4791 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/6460257f-3f13-4642-b5b5-fc05065df9d7-combined-ca-bundle\") pod \"6460257f-3f13-4642-b5b5-fc05065df9d7\" (UID: \"6460257f-3f13-4642-b5b5-fc05065df9d7\") " Dec 10 23:11:37 crc kubenswrapper[4791]: I1210 23:11:37.619058 4791 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/6460257f-3f13-4642-b5b5-fc05065df9d7-logs" (OuterVolumeSpecName: "logs") pod "6460257f-3f13-4642-b5b5-fc05065df9d7" (UID: "6460257f-3f13-4642-b5b5-fc05065df9d7"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 10 23:11:37 crc kubenswrapper[4791]: I1210 23:11:37.623631 4791 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/6460257f-3f13-4642-b5b5-fc05065df9d7-kube-api-access-6mrmt" (OuterVolumeSpecName: "kube-api-access-6mrmt") pod "6460257f-3f13-4642-b5b5-fc05065df9d7" (UID: "6460257f-3f13-4642-b5b5-fc05065df9d7"). InnerVolumeSpecName "kube-api-access-6mrmt". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 10 23:11:37 crc kubenswrapper[4791]: I1210 23:11:37.652513 4791 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/6460257f-3f13-4642-b5b5-fc05065df9d7-config-data" (OuterVolumeSpecName: "config-data") pod "6460257f-3f13-4642-b5b5-fc05065df9d7" (UID: "6460257f-3f13-4642-b5b5-fc05065df9d7"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 10 23:11:37 crc kubenswrapper[4791]: I1210 23:11:37.673270 4791 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/6460257f-3f13-4642-b5b5-fc05065df9d7-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "6460257f-3f13-4642-b5b5-fc05065df9d7" (UID: "6460257f-3f13-4642-b5b5-fc05065df9d7"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 10 23:11:37 crc kubenswrapper[4791]: I1210 23:11:37.683835 4791 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/6460257f-3f13-4642-b5b5-fc05065df9d7-public-tls-certs" (OuterVolumeSpecName: "public-tls-certs") pod "6460257f-3f13-4642-b5b5-fc05065df9d7" (UID: "6460257f-3f13-4642-b5b5-fc05065df9d7"). InnerVolumeSpecName "public-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 10 23:11:37 crc kubenswrapper[4791]: I1210 23:11:37.691196 4791 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/6460257f-3f13-4642-b5b5-fc05065df9d7-internal-tls-certs" (OuterVolumeSpecName: "internal-tls-certs") pod "6460257f-3f13-4642-b5b5-fc05065df9d7" (UID: "6460257f-3f13-4642-b5b5-fc05065df9d7"). InnerVolumeSpecName "internal-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 10 23:11:37 crc kubenswrapper[4791]: I1210 23:11:37.719057 4791 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-6mrmt\" (UniqueName: \"kubernetes.io/projected/6460257f-3f13-4642-b5b5-fc05065df9d7-kube-api-access-6mrmt\") on node \"crc\" DevicePath \"\"" Dec 10 23:11:37 crc kubenswrapper[4791]: I1210 23:11:37.719101 4791 reconciler_common.go:293] "Volume detached for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/6460257f-3f13-4642-b5b5-fc05065df9d7-internal-tls-certs\") on node \"crc\" DevicePath \"\"" Dec 10 23:11:37 crc kubenswrapper[4791]: I1210 23:11:37.719114 4791 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/6460257f-3f13-4642-b5b5-fc05065df9d7-config-data\") on node \"crc\" DevicePath \"\"" Dec 10 23:11:37 crc kubenswrapper[4791]: I1210 23:11:37.719123 4791 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/6460257f-3f13-4642-b5b5-fc05065df9d7-logs\") on node \"crc\" DevicePath \"\"" Dec 10 23:11:37 crc kubenswrapper[4791]: I1210 23:11:37.719132 4791 reconciler_common.go:293] "Volume detached for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/6460257f-3f13-4642-b5b5-fc05065df9d7-public-tls-certs\") on node \"crc\" DevicePath \"\"" Dec 10 23:11:37 crc kubenswrapper[4791]: I1210 23:11:37.719142 4791 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/6460257f-3f13-4642-b5b5-fc05065df9d7-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 10 23:11:37 crc kubenswrapper[4791]: I1210 23:11:37.901537 4791 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="1fab60e1-51e0-4101-8c37-c23d8f259cf2" path="/var/lib/kubelet/pods/1fab60e1-51e0-4101-8c37-c23d8f259cf2/volumes" Dec 10 23:11:37 crc kubenswrapper[4791]: I1210 23:11:37.902252 4791 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="f947977e-b1aa-43c5-b987-cea516e6a154" path="/var/lib/kubelet/pods/f947977e-b1aa-43c5-b987-cea516e6a154/volumes" Dec 10 23:11:37 crc kubenswrapper[4791]: I1210 23:11:37.903485 4791 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-metadata-0"] Dec 10 23:11:38 crc kubenswrapper[4791]: I1210 23:11:38.534709 4791 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"8ee8d247-355c-4e17-965c-0b3c958a546c","Type":"ContainerStarted","Data":"91d72ebe9b7db28cb104fe5a66f6da391b62e4141c70e7f8107a4c26275d6549"} Dec 10 23:11:38 crc kubenswrapper[4791]: I1210 23:11:38.536140 4791 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"8ee8d247-355c-4e17-965c-0b3c958a546c","Type":"ContainerStarted","Data":"ab7e05308854073f8ef3eeead83e39afb9d6c54753541b152844dd4182b69cf3"} Dec 10 23:11:38 crc kubenswrapper[4791]: I1210 23:11:38.536213 4791 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"8ee8d247-355c-4e17-965c-0b3c958a546c","Type":"ContainerStarted","Data":"a747ad5d4a75012007c8be9d45f3680a3f911a9b19cf2e85a016d68767b3052f"} Dec 10 23:11:38 crc kubenswrapper[4791]: I1210 23:11:38.538034 4791 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"6460257f-3f13-4642-b5b5-fc05065df9d7","Type":"ContainerDied","Data":"478fbaaf7f1f01776b968365f05590a0059ae9631f821df7ffeb8f1e19f33612"} Dec 10 23:11:38 crc kubenswrapper[4791]: I1210 23:11:38.538079 4791 scope.go:117] "RemoveContainer" containerID="e6e79a11bb5efb0004d4c3a54c551264e04cb8c253bb15d795f2c9dfeb974154" Dec 10 23:11:38 crc kubenswrapper[4791]: I1210 23:11:38.538179 4791 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Dec 10 23:11:38 crc kubenswrapper[4791]: I1210 23:11:38.551625 4791 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-scheduler-0" event={"ID":"b1d1b38e-2897-4c52-a4e4-d994dfffe805","Type":"ContainerStarted","Data":"f5c7eac1f9d200398c8cae5bb32d2bcf845f0396a6a00774f39c1ae6ce42e3b3"} Dec 10 23:11:38 crc kubenswrapper[4791]: I1210 23:11:38.573566 4791 scope.go:117] "RemoveContainer" containerID="e17b5d105e85af3d0d1bfcd6dacac2d5d1b35811ddecc4140c5bfb94cb52c0bc" Dec 10 23:11:38 crc kubenswrapper[4791]: I1210 23:11:38.596568 4791 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-metadata-0" podStartSLOduration=2.596542122 podStartE2EDuration="2.596542122s" podCreationTimestamp="2025-12-10 23:11:36 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-10 23:11:38.562788687 +0000 UTC m=+1332.992406310" watchObservedRunningTime="2025-12-10 23:11:38.596542122 +0000 UTC m=+1333.026159735" Dec 10 23:11:38 crc kubenswrapper[4791]: I1210 23:11:38.598652 4791 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-api-0"] Dec 10 23:11:38 crc kubenswrapper[4791]: I1210 23:11:38.631130 4791 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-api-0"] Dec 10 23:11:38 crc kubenswrapper[4791]: I1210 23:11:38.636854 4791 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-scheduler-0" podStartSLOduration=2.636829815 podStartE2EDuration="2.636829815s" podCreationTimestamp="2025-12-10 23:11:36 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-10 23:11:38.613634005 +0000 UTC m=+1333.043251638" watchObservedRunningTime="2025-12-10 23:11:38.636829815 +0000 UTC m=+1333.066447428" Dec 10 23:11:38 crc kubenswrapper[4791]: I1210 23:11:38.664849 4791 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-api-0"] Dec 10 23:11:38 crc kubenswrapper[4791]: E1210 23:11:38.669077 4791 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="6460257f-3f13-4642-b5b5-fc05065df9d7" containerName="nova-api-log" Dec 10 23:11:38 crc kubenswrapper[4791]: I1210 23:11:38.669120 4791 state_mem.go:107] "Deleted CPUSet assignment" podUID="6460257f-3f13-4642-b5b5-fc05065df9d7" containerName="nova-api-log" Dec 10 23:11:38 crc kubenswrapper[4791]: E1210 23:11:38.669162 4791 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="6460257f-3f13-4642-b5b5-fc05065df9d7" containerName="nova-api-api" Dec 10 23:11:38 crc kubenswrapper[4791]: I1210 23:11:38.669172 4791 state_mem.go:107] "Deleted CPUSet assignment" podUID="6460257f-3f13-4642-b5b5-fc05065df9d7" containerName="nova-api-api" Dec 10 23:11:38 crc kubenswrapper[4791]: I1210 23:11:38.670056 4791 memory_manager.go:354] "RemoveStaleState removing state" podUID="6460257f-3f13-4642-b5b5-fc05065df9d7" containerName="nova-api-api" Dec 10 23:11:38 crc kubenswrapper[4791]: I1210 23:11:38.670088 4791 memory_manager.go:354] "RemoveStaleState removing state" podUID="6460257f-3f13-4642-b5b5-fc05065df9d7" containerName="nova-api-log" Dec 10 23:11:38 crc kubenswrapper[4791]: I1210 23:11:38.673036 4791 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Dec 10 23:11:38 crc kubenswrapper[4791]: I1210 23:11:38.677549 4791 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-nova-public-svc" Dec 10 23:11:38 crc kubenswrapper[4791]: I1210 23:11:38.677966 4791 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-api-config-data" Dec 10 23:11:38 crc kubenswrapper[4791]: I1210 23:11:38.678151 4791 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-nova-internal-svc" Dec 10 23:11:38 crc kubenswrapper[4791]: I1210 23:11:38.700992 4791 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-api-0"] Dec 10 23:11:38 crc kubenswrapper[4791]: I1210 23:11:38.843157 4791 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/70ddc653-9583-4f3c-a0f2-a2bfa952e7a4-public-tls-certs\") pod \"nova-api-0\" (UID: \"70ddc653-9583-4f3c-a0f2-a2bfa952e7a4\") " pod="openstack/nova-api-0" Dec 10 23:11:38 crc kubenswrapper[4791]: I1210 23:11:38.843231 4791 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/70ddc653-9583-4f3c-a0f2-a2bfa952e7a4-config-data\") pod \"nova-api-0\" (UID: \"70ddc653-9583-4f3c-a0f2-a2bfa952e7a4\") " pod="openstack/nova-api-0" Dec 10 23:11:38 crc kubenswrapper[4791]: I1210 23:11:38.843267 4791 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/70ddc653-9583-4f3c-a0f2-a2bfa952e7a4-internal-tls-certs\") pod \"nova-api-0\" (UID: \"70ddc653-9583-4f3c-a0f2-a2bfa952e7a4\") " pod="openstack/nova-api-0" Dec 10 23:11:38 crc kubenswrapper[4791]: I1210 23:11:38.843301 4791 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/70ddc653-9583-4f3c-a0f2-a2bfa952e7a4-logs\") pod \"nova-api-0\" (UID: \"70ddc653-9583-4f3c-a0f2-a2bfa952e7a4\") " pod="openstack/nova-api-0" Dec 10 23:11:38 crc kubenswrapper[4791]: I1210 23:11:38.843375 4791 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-kq577\" (UniqueName: \"kubernetes.io/projected/70ddc653-9583-4f3c-a0f2-a2bfa952e7a4-kube-api-access-kq577\") pod \"nova-api-0\" (UID: \"70ddc653-9583-4f3c-a0f2-a2bfa952e7a4\") " pod="openstack/nova-api-0" Dec 10 23:11:38 crc kubenswrapper[4791]: I1210 23:11:38.843415 4791 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/70ddc653-9583-4f3c-a0f2-a2bfa952e7a4-combined-ca-bundle\") pod \"nova-api-0\" (UID: \"70ddc653-9583-4f3c-a0f2-a2bfa952e7a4\") " pod="openstack/nova-api-0" Dec 10 23:11:38 crc kubenswrapper[4791]: I1210 23:11:38.945319 4791 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-kq577\" (UniqueName: \"kubernetes.io/projected/70ddc653-9583-4f3c-a0f2-a2bfa952e7a4-kube-api-access-kq577\") pod \"nova-api-0\" (UID: \"70ddc653-9583-4f3c-a0f2-a2bfa952e7a4\") " pod="openstack/nova-api-0" Dec 10 23:11:38 crc kubenswrapper[4791]: I1210 23:11:38.945421 4791 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/70ddc653-9583-4f3c-a0f2-a2bfa952e7a4-combined-ca-bundle\") pod \"nova-api-0\" (UID: \"70ddc653-9583-4f3c-a0f2-a2bfa952e7a4\") " pod="openstack/nova-api-0" Dec 10 23:11:38 crc kubenswrapper[4791]: I1210 23:11:38.945510 4791 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/70ddc653-9583-4f3c-a0f2-a2bfa952e7a4-public-tls-certs\") pod \"nova-api-0\" (UID: \"70ddc653-9583-4f3c-a0f2-a2bfa952e7a4\") " pod="openstack/nova-api-0" Dec 10 23:11:38 crc kubenswrapper[4791]: I1210 23:11:38.945561 4791 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/70ddc653-9583-4f3c-a0f2-a2bfa952e7a4-config-data\") pod \"nova-api-0\" (UID: \"70ddc653-9583-4f3c-a0f2-a2bfa952e7a4\") " pod="openstack/nova-api-0" Dec 10 23:11:38 crc kubenswrapper[4791]: I1210 23:11:38.945600 4791 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/70ddc653-9583-4f3c-a0f2-a2bfa952e7a4-internal-tls-certs\") pod \"nova-api-0\" (UID: \"70ddc653-9583-4f3c-a0f2-a2bfa952e7a4\") " pod="openstack/nova-api-0" Dec 10 23:11:38 crc kubenswrapper[4791]: I1210 23:11:38.945642 4791 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/70ddc653-9583-4f3c-a0f2-a2bfa952e7a4-logs\") pod \"nova-api-0\" (UID: \"70ddc653-9583-4f3c-a0f2-a2bfa952e7a4\") " pod="openstack/nova-api-0" Dec 10 23:11:38 crc kubenswrapper[4791]: I1210 23:11:38.946188 4791 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/70ddc653-9583-4f3c-a0f2-a2bfa952e7a4-logs\") pod \"nova-api-0\" (UID: \"70ddc653-9583-4f3c-a0f2-a2bfa952e7a4\") " pod="openstack/nova-api-0" Dec 10 23:11:38 crc kubenswrapper[4791]: I1210 23:11:38.949768 4791 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/70ddc653-9583-4f3c-a0f2-a2bfa952e7a4-public-tls-certs\") pod \"nova-api-0\" (UID: \"70ddc653-9583-4f3c-a0f2-a2bfa952e7a4\") " pod="openstack/nova-api-0" Dec 10 23:11:38 crc kubenswrapper[4791]: I1210 23:11:38.951183 4791 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/70ddc653-9583-4f3c-a0f2-a2bfa952e7a4-config-data\") pod \"nova-api-0\" (UID: \"70ddc653-9583-4f3c-a0f2-a2bfa952e7a4\") " pod="openstack/nova-api-0" Dec 10 23:11:38 crc kubenswrapper[4791]: I1210 23:11:38.952989 4791 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/70ddc653-9583-4f3c-a0f2-a2bfa952e7a4-combined-ca-bundle\") pod \"nova-api-0\" (UID: \"70ddc653-9583-4f3c-a0f2-a2bfa952e7a4\") " pod="openstack/nova-api-0" Dec 10 23:11:38 crc kubenswrapper[4791]: I1210 23:11:38.954046 4791 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/70ddc653-9583-4f3c-a0f2-a2bfa952e7a4-internal-tls-certs\") pod \"nova-api-0\" (UID: \"70ddc653-9583-4f3c-a0f2-a2bfa952e7a4\") " pod="openstack/nova-api-0" Dec 10 23:11:38 crc kubenswrapper[4791]: I1210 23:11:38.967207 4791 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-kq577\" (UniqueName: \"kubernetes.io/projected/70ddc653-9583-4f3c-a0f2-a2bfa952e7a4-kube-api-access-kq577\") pod \"nova-api-0\" (UID: \"70ddc653-9583-4f3c-a0f2-a2bfa952e7a4\") " pod="openstack/nova-api-0" Dec 10 23:11:39 crc kubenswrapper[4791]: I1210 23:11:39.022869 4791 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Dec 10 23:11:39 crc kubenswrapper[4791]: I1210 23:11:39.487575 4791 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-api-0"] Dec 10 23:11:39 crc kubenswrapper[4791]: I1210 23:11:39.567310 4791 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"70ddc653-9583-4f3c-a0f2-a2bfa952e7a4","Type":"ContainerStarted","Data":"9ebbdb45442687f52e11a95783cfe240f044127b3f65ea982c4d0d523580291c"} Dec 10 23:11:39 crc kubenswrapper[4791]: I1210 23:11:39.899086 4791 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="6460257f-3f13-4642-b5b5-fc05065df9d7" path="/var/lib/kubelet/pods/6460257f-3f13-4642-b5b5-fc05065df9d7/volumes" Dec 10 23:11:40 crc kubenswrapper[4791]: I1210 23:11:40.583796 4791 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"70ddc653-9583-4f3c-a0f2-a2bfa952e7a4","Type":"ContainerStarted","Data":"3681d49fc11a341210d84f8301046b638d9cdc30376450bc045caabc909762c2"} Dec 10 23:11:40 crc kubenswrapper[4791]: I1210 23:11:40.583850 4791 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"70ddc653-9583-4f3c-a0f2-a2bfa952e7a4","Type":"ContainerStarted","Data":"9d71f222cc2786796c23fce187c0d64b8d0c42e950990faa10f782fb7cede4bc"} Dec 10 23:11:40 crc kubenswrapper[4791]: I1210 23:11:40.614920 4791 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-api-0" podStartSLOduration=2.614896087 podStartE2EDuration="2.614896087s" podCreationTimestamp="2025-12-10 23:11:38 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-10 23:11:40.601700876 +0000 UTC m=+1335.031318499" watchObservedRunningTime="2025-12-10 23:11:40.614896087 +0000 UTC m=+1335.044513710" Dec 10 23:11:41 crc kubenswrapper[4791]: I1210 23:11:41.908841 4791 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-scheduler-0" Dec 10 23:11:42 crc kubenswrapper[4791]: I1210 23:11:42.358144 4791 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-metadata-0" Dec 10 23:11:42 crc kubenswrapper[4791]: I1210 23:11:42.358236 4791 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-metadata-0" Dec 10 23:11:46 crc kubenswrapper[4791]: I1210 23:11:46.907939 4791 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-scheduler-0" Dec 10 23:11:46 crc kubenswrapper[4791]: I1210 23:11:46.942927 4791 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/nova-scheduler-0" Dec 10 23:11:47 crc kubenswrapper[4791]: I1210 23:11:47.358244 4791 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-metadata-0" Dec 10 23:11:47 crc kubenswrapper[4791]: I1210 23:11:47.358305 4791 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-metadata-0" Dec 10 23:11:47 crc kubenswrapper[4791]: I1210 23:11:47.691362 4791 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-scheduler-0" Dec 10 23:11:48 crc kubenswrapper[4791]: I1210 23:11:48.371470 4791 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/nova-metadata-0" podUID="8ee8d247-355c-4e17-965c-0b3c958a546c" containerName="nova-metadata-metadata" probeResult="failure" output="Get \"https://10.217.0.202:8775/\": net/http: request canceled (Client.Timeout exceeded while awaiting headers)" Dec 10 23:11:48 crc kubenswrapper[4791]: I1210 23:11:48.371474 4791 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/nova-metadata-0" podUID="8ee8d247-355c-4e17-965c-0b3c958a546c" containerName="nova-metadata-log" probeResult="failure" output="Get \"https://10.217.0.202:8775/\": net/http: request canceled (Client.Timeout exceeded while awaiting headers)" Dec 10 23:11:49 crc kubenswrapper[4791]: I1210 23:11:49.023539 4791 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-api-0" Dec 10 23:11:49 crc kubenswrapper[4791]: I1210 23:11:49.025657 4791 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-api-0" Dec 10 23:11:50 crc kubenswrapper[4791]: I1210 23:11:50.033582 4791 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/nova-api-0" podUID="70ddc653-9583-4f3c-a0f2-a2bfa952e7a4" containerName="nova-api-log" probeResult="failure" output="Get \"https://10.217.0.203:8774/\": net/http: request canceled (Client.Timeout exceeded while awaiting headers)" Dec 10 23:11:50 crc kubenswrapper[4791]: I1210 23:11:50.033613 4791 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/nova-api-0" podUID="70ddc653-9583-4f3c-a0f2-a2bfa952e7a4" containerName="nova-api-api" probeResult="failure" output="Get \"https://10.217.0.203:8774/\": net/http: request canceled (Client.Timeout exceeded while awaiting headers)" Dec 10 23:11:50 crc kubenswrapper[4791]: I1210 23:11:50.689983 4791 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/ceilometer-0" Dec 10 23:11:55 crc kubenswrapper[4791]: I1210 23:11:55.038420 4791 patch_prober.go:28] interesting pod/machine-config-daemon-5rb5l container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 10 23:11:55 crc kubenswrapper[4791]: I1210 23:11:55.039534 4791 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-5rb5l" podUID="ba35653c-6e06-4cee-a4d6-137764090d18" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 10 23:11:57 crc kubenswrapper[4791]: I1210 23:11:57.364293 4791 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/nova-metadata-0" Dec 10 23:11:57 crc kubenswrapper[4791]: I1210 23:11:57.370819 4791 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/nova-metadata-0" Dec 10 23:11:57 crc kubenswrapper[4791]: I1210 23:11:57.372261 4791 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-metadata-0" Dec 10 23:11:57 crc kubenswrapper[4791]: I1210 23:11:57.847669 4791 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-metadata-0" Dec 10 23:11:59 crc kubenswrapper[4791]: I1210 23:11:59.219849 4791 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/nova-api-0" Dec 10 23:11:59 crc kubenswrapper[4791]: I1210 23:11:59.221998 4791 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-api-0" Dec 10 23:11:59 crc kubenswrapper[4791]: I1210 23:11:59.236857 4791 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/nova-api-0" Dec 10 23:11:59 crc kubenswrapper[4791]: I1210 23:11:59.328701 4791 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-api-0" Dec 10 23:12:00 crc kubenswrapper[4791]: I1210 23:12:00.112923 4791 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-api-0" Dec 10 23:12:00 crc kubenswrapper[4791]: I1210 23:12:00.131195 4791 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-api-0" Dec 10 23:12:08 crc kubenswrapper[4791]: I1210 23:12:08.742297 4791 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/rabbitmq-server-0"] Dec 10 23:12:09 crc kubenswrapper[4791]: I1210 23:12:09.674698 4791 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/rabbitmq-cell1-server-0"] Dec 10 23:12:14 crc kubenswrapper[4791]: I1210 23:12:14.186568 4791 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/rabbitmq-server-0" podUID="9ce8dbeb-dce8-47ac-98f1-fe92f74aa3de" containerName="rabbitmq" containerID="cri-o://f7a414fff9fd9d4a710f40666dd232a7704dad812d497846c5077aa27f08208d" gracePeriod=604795 Dec 10 23:12:14 crc kubenswrapper[4791]: I1210 23:12:14.911199 4791 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/rabbitmq-cell1-server-0" podUID="47eb9200-b0bc-41a6-abb9-f6167bd69c66" containerName="rabbitmq" containerID="cri-o://8dd3007abca59274d70261742380418ebbe002bfd3c95f64e21f07dcfab98397" gracePeriod=604795 Dec 10 23:12:14 crc kubenswrapper[4791]: I1210 23:12:14.991892 4791 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/rabbitmq-cell1-server-0" podUID="47eb9200-b0bc-41a6-abb9-f6167bd69c66" containerName="rabbitmq" probeResult="failure" output="dial tcp 10.217.0.99:5671: connect: connection refused" Dec 10 23:12:15 crc kubenswrapper[4791]: I1210 23:12:15.277222 4791 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/rabbitmq-server-0" podUID="9ce8dbeb-dce8-47ac-98f1-fe92f74aa3de" containerName="rabbitmq" probeResult="failure" output="dial tcp 10.217.0.98:5671: connect: connection refused" Dec 10 23:12:20 crc kubenswrapper[4791]: I1210 23:12:20.603301 4791 generic.go:334] "Generic (PLEG): container finished" podID="9ce8dbeb-dce8-47ac-98f1-fe92f74aa3de" containerID="f7a414fff9fd9d4a710f40666dd232a7704dad812d497846c5077aa27f08208d" exitCode=0 Dec 10 23:12:20 crc kubenswrapper[4791]: I1210 23:12:20.603402 4791 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-server-0" event={"ID":"9ce8dbeb-dce8-47ac-98f1-fe92f74aa3de","Type":"ContainerDied","Data":"f7a414fff9fd9d4a710f40666dd232a7704dad812d497846c5077aa27f08208d"} Dec 10 23:12:20 crc kubenswrapper[4791]: I1210 23:12:20.812055 4791 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/rabbitmq-server-0" Dec 10 23:12:21 crc kubenswrapper[4791]: I1210 23:12:21.183550 4791 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/9ce8dbeb-dce8-47ac-98f1-fe92f74aa3de-rabbitmq-erlang-cookie\") pod \"9ce8dbeb-dce8-47ac-98f1-fe92f74aa3de\" (UID: \"9ce8dbeb-dce8-47ac-98f1-fe92f74aa3de\") " Dec 10 23:12:21 crc kubenswrapper[4791]: I1210 23:12:21.184633 4791 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/9ce8dbeb-dce8-47ac-98f1-fe92f74aa3de-rabbitmq-confd\") pod \"9ce8dbeb-dce8-47ac-98f1-fe92f74aa3de\" (UID: \"9ce8dbeb-dce8-47ac-98f1-fe92f74aa3de\") " Dec 10 23:12:21 crc kubenswrapper[4791]: I1210 23:12:21.184551 4791 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/9ce8dbeb-dce8-47ac-98f1-fe92f74aa3de-rabbitmq-erlang-cookie" (OuterVolumeSpecName: "rabbitmq-erlang-cookie") pod "9ce8dbeb-dce8-47ac-98f1-fe92f74aa3de" (UID: "9ce8dbeb-dce8-47ac-98f1-fe92f74aa3de"). InnerVolumeSpecName "rabbitmq-erlang-cookie". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 10 23:12:21 crc kubenswrapper[4791]: I1210 23:12:21.184693 4791 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/9ce8dbeb-dce8-47ac-98f1-fe92f74aa3de-pod-info\") pod \"9ce8dbeb-dce8-47ac-98f1-fe92f74aa3de\" (UID: \"9ce8dbeb-dce8-47ac-98f1-fe92f74aa3de\") " Dec 10 23:12:21 crc kubenswrapper[4791]: I1210 23:12:21.184719 4791 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/9ce8dbeb-dce8-47ac-98f1-fe92f74aa3de-server-conf\") pod \"9ce8dbeb-dce8-47ac-98f1-fe92f74aa3de\" (UID: \"9ce8dbeb-dce8-47ac-98f1-fe92f74aa3de\") " Dec 10 23:12:21 crc kubenswrapper[4791]: I1210 23:12:21.185312 4791 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/9ce8dbeb-dce8-47ac-98f1-fe92f74aa3de-config-data\") pod \"9ce8dbeb-dce8-47ac-98f1-fe92f74aa3de\" (UID: \"9ce8dbeb-dce8-47ac-98f1-fe92f74aa3de\") " Dec 10 23:12:21 crc kubenswrapper[4791]: I1210 23:12:21.185359 4791 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"rabbitmq-tls\" (UniqueName: \"kubernetes.io/projected/9ce8dbeb-dce8-47ac-98f1-fe92f74aa3de-rabbitmq-tls\") pod \"9ce8dbeb-dce8-47ac-98f1-fe92f74aa3de\" (UID: \"9ce8dbeb-dce8-47ac-98f1-fe92f74aa3de\") " Dec 10 23:12:21 crc kubenswrapper[4791]: I1210 23:12:21.185379 4791 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/9ce8dbeb-dce8-47ac-98f1-fe92f74aa3de-rabbitmq-plugins\") pod \"9ce8dbeb-dce8-47ac-98f1-fe92f74aa3de\" (UID: \"9ce8dbeb-dce8-47ac-98f1-fe92f74aa3de\") " Dec 10 23:12:21 crc kubenswrapper[4791]: I1210 23:12:21.185411 4791 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"persistence\" (UniqueName: \"kubernetes.io/local-volume/local-storage07-crc\") pod \"9ce8dbeb-dce8-47ac-98f1-fe92f74aa3de\" (UID: \"9ce8dbeb-dce8-47ac-98f1-fe92f74aa3de\") " Dec 10 23:12:21 crc kubenswrapper[4791]: I1210 23:12:21.185455 4791 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-85rrv\" (UniqueName: \"kubernetes.io/projected/9ce8dbeb-dce8-47ac-98f1-fe92f74aa3de-kube-api-access-85rrv\") pod \"9ce8dbeb-dce8-47ac-98f1-fe92f74aa3de\" (UID: \"9ce8dbeb-dce8-47ac-98f1-fe92f74aa3de\") " Dec 10 23:12:21 crc kubenswrapper[4791]: I1210 23:12:21.185500 4791 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/9ce8dbeb-dce8-47ac-98f1-fe92f74aa3de-plugins-conf\") pod \"9ce8dbeb-dce8-47ac-98f1-fe92f74aa3de\" (UID: \"9ce8dbeb-dce8-47ac-98f1-fe92f74aa3de\") " Dec 10 23:12:21 crc kubenswrapper[4791]: I1210 23:12:21.185531 4791 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/9ce8dbeb-dce8-47ac-98f1-fe92f74aa3de-erlang-cookie-secret\") pod \"9ce8dbeb-dce8-47ac-98f1-fe92f74aa3de\" (UID: \"9ce8dbeb-dce8-47ac-98f1-fe92f74aa3de\") " Dec 10 23:12:21 crc kubenswrapper[4791]: I1210 23:12:21.186045 4791 reconciler_common.go:293] "Volume detached for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/9ce8dbeb-dce8-47ac-98f1-fe92f74aa3de-rabbitmq-erlang-cookie\") on node \"crc\" DevicePath \"\"" Dec 10 23:12:21 crc kubenswrapper[4791]: I1210 23:12:21.190176 4791 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/9ce8dbeb-dce8-47ac-98f1-fe92f74aa3de-rabbitmq-plugins" (OuterVolumeSpecName: "rabbitmq-plugins") pod "9ce8dbeb-dce8-47ac-98f1-fe92f74aa3de" (UID: "9ce8dbeb-dce8-47ac-98f1-fe92f74aa3de"). InnerVolumeSpecName "rabbitmq-plugins". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 10 23:12:21 crc kubenswrapper[4791]: I1210 23:12:21.195625 4791 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/9ce8dbeb-dce8-47ac-98f1-fe92f74aa3de-plugins-conf" (OuterVolumeSpecName: "plugins-conf") pod "9ce8dbeb-dce8-47ac-98f1-fe92f74aa3de" (UID: "9ce8dbeb-dce8-47ac-98f1-fe92f74aa3de"). InnerVolumeSpecName "plugins-conf". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 10 23:12:21 crc kubenswrapper[4791]: I1210 23:12:21.204403 4791 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/local-volume/local-storage07-crc" (OuterVolumeSpecName: "persistence") pod "9ce8dbeb-dce8-47ac-98f1-fe92f74aa3de" (UID: "9ce8dbeb-dce8-47ac-98f1-fe92f74aa3de"). InnerVolumeSpecName "local-storage07-crc". PluginName "kubernetes.io/local-volume", VolumeGidValue "" Dec 10 23:12:21 crc kubenswrapper[4791]: I1210 23:12:21.204588 4791 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/9ce8dbeb-dce8-47ac-98f1-fe92f74aa3de-rabbitmq-tls" (OuterVolumeSpecName: "rabbitmq-tls") pod "9ce8dbeb-dce8-47ac-98f1-fe92f74aa3de" (UID: "9ce8dbeb-dce8-47ac-98f1-fe92f74aa3de"). InnerVolumeSpecName "rabbitmq-tls". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 10 23:12:21 crc kubenswrapper[4791]: I1210 23:12:21.206952 4791 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/9ce8dbeb-dce8-47ac-98f1-fe92f74aa3de-erlang-cookie-secret" (OuterVolumeSpecName: "erlang-cookie-secret") pod "9ce8dbeb-dce8-47ac-98f1-fe92f74aa3de" (UID: "9ce8dbeb-dce8-47ac-98f1-fe92f74aa3de"). InnerVolumeSpecName "erlang-cookie-secret". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 10 23:12:21 crc kubenswrapper[4791]: I1210 23:12:21.211796 4791 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/downward-api/9ce8dbeb-dce8-47ac-98f1-fe92f74aa3de-pod-info" (OuterVolumeSpecName: "pod-info") pod "9ce8dbeb-dce8-47ac-98f1-fe92f74aa3de" (UID: "9ce8dbeb-dce8-47ac-98f1-fe92f74aa3de"). InnerVolumeSpecName "pod-info". PluginName "kubernetes.io/downward-api", VolumeGidValue "" Dec 10 23:12:21 crc kubenswrapper[4791]: I1210 23:12:21.215624 4791 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/9ce8dbeb-dce8-47ac-98f1-fe92f74aa3de-config-data" (OuterVolumeSpecName: "config-data") pod "9ce8dbeb-dce8-47ac-98f1-fe92f74aa3de" (UID: "9ce8dbeb-dce8-47ac-98f1-fe92f74aa3de"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 10 23:12:21 crc kubenswrapper[4791]: I1210 23:12:21.226422 4791 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/9ce8dbeb-dce8-47ac-98f1-fe92f74aa3de-kube-api-access-85rrv" (OuterVolumeSpecName: "kube-api-access-85rrv") pod "9ce8dbeb-dce8-47ac-98f1-fe92f74aa3de" (UID: "9ce8dbeb-dce8-47ac-98f1-fe92f74aa3de"). InnerVolumeSpecName "kube-api-access-85rrv". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 10 23:12:21 crc kubenswrapper[4791]: I1210 23:12:21.277193 4791 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/9ce8dbeb-dce8-47ac-98f1-fe92f74aa3de-server-conf" (OuterVolumeSpecName: "server-conf") pod "9ce8dbeb-dce8-47ac-98f1-fe92f74aa3de" (UID: "9ce8dbeb-dce8-47ac-98f1-fe92f74aa3de"). InnerVolumeSpecName "server-conf". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 10 23:12:21 crc kubenswrapper[4791]: I1210 23:12:21.292387 4791 reconciler_common.go:293] "Volume detached for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/9ce8dbeb-dce8-47ac-98f1-fe92f74aa3de-pod-info\") on node \"crc\" DevicePath \"\"" Dec 10 23:12:21 crc kubenswrapper[4791]: I1210 23:12:21.292644 4791 reconciler_common.go:293] "Volume detached for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/9ce8dbeb-dce8-47ac-98f1-fe92f74aa3de-server-conf\") on node \"crc\" DevicePath \"\"" Dec 10 23:12:21 crc kubenswrapper[4791]: I1210 23:12:21.292704 4791 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/9ce8dbeb-dce8-47ac-98f1-fe92f74aa3de-config-data\") on node \"crc\" DevicePath \"\"" Dec 10 23:12:21 crc kubenswrapper[4791]: I1210 23:12:21.292783 4791 reconciler_common.go:293] "Volume detached for volume \"rabbitmq-tls\" (UniqueName: \"kubernetes.io/projected/9ce8dbeb-dce8-47ac-98f1-fe92f74aa3de-rabbitmq-tls\") on node \"crc\" DevicePath \"\"" Dec 10 23:12:21 crc kubenswrapper[4791]: I1210 23:12:21.292854 4791 reconciler_common.go:293] "Volume detached for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/9ce8dbeb-dce8-47ac-98f1-fe92f74aa3de-rabbitmq-plugins\") on node \"crc\" DevicePath \"\"" Dec 10 23:12:21 crc kubenswrapper[4791]: I1210 23:12:21.292934 4791 reconciler_common.go:286] "operationExecutor.UnmountDevice started for volume \"local-storage07-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage07-crc\") on node \"crc\" " Dec 10 23:12:21 crc kubenswrapper[4791]: I1210 23:12:21.292991 4791 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-85rrv\" (UniqueName: \"kubernetes.io/projected/9ce8dbeb-dce8-47ac-98f1-fe92f74aa3de-kube-api-access-85rrv\") on node \"crc\" DevicePath \"\"" Dec 10 23:12:21 crc kubenswrapper[4791]: I1210 23:12:21.293057 4791 reconciler_common.go:293] "Volume detached for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/9ce8dbeb-dce8-47ac-98f1-fe92f74aa3de-plugins-conf\") on node \"crc\" DevicePath \"\"" Dec 10 23:12:21 crc kubenswrapper[4791]: I1210 23:12:21.293117 4791 reconciler_common.go:293] "Volume detached for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/9ce8dbeb-dce8-47ac-98f1-fe92f74aa3de-erlang-cookie-secret\") on node \"crc\" DevicePath \"\"" Dec 10 23:12:21 crc kubenswrapper[4791]: I1210 23:12:21.324171 4791 operation_generator.go:917] UnmountDevice succeeded for volume "local-storage07-crc" (UniqueName: "kubernetes.io/local-volume/local-storage07-crc") on node "crc" Dec 10 23:12:21 crc kubenswrapper[4791]: I1210 23:12:21.378450 4791 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/9ce8dbeb-dce8-47ac-98f1-fe92f74aa3de-rabbitmq-confd" (OuterVolumeSpecName: "rabbitmq-confd") pod "9ce8dbeb-dce8-47ac-98f1-fe92f74aa3de" (UID: "9ce8dbeb-dce8-47ac-98f1-fe92f74aa3de"). InnerVolumeSpecName "rabbitmq-confd". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 10 23:12:21 crc kubenswrapper[4791]: I1210 23:12:21.408525 4791 reconciler_common.go:293] "Volume detached for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/9ce8dbeb-dce8-47ac-98f1-fe92f74aa3de-rabbitmq-confd\") on node \"crc\" DevicePath \"\"" Dec 10 23:12:21 crc kubenswrapper[4791]: I1210 23:12:21.409013 4791 reconciler_common.go:293] "Volume detached for volume \"local-storage07-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage07-crc\") on node \"crc\" DevicePath \"\"" Dec 10 23:12:21 crc kubenswrapper[4791]: I1210 23:12:21.618008 4791 generic.go:334] "Generic (PLEG): container finished" podID="47eb9200-b0bc-41a6-abb9-f6167bd69c66" containerID="8dd3007abca59274d70261742380418ebbe002bfd3c95f64e21f07dcfab98397" exitCode=0 Dec 10 23:12:21 crc kubenswrapper[4791]: I1210 23:12:21.618094 4791 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-cell1-server-0" event={"ID":"47eb9200-b0bc-41a6-abb9-f6167bd69c66","Type":"ContainerDied","Data":"8dd3007abca59274d70261742380418ebbe002bfd3c95f64e21f07dcfab98397"} Dec 10 23:12:21 crc kubenswrapper[4791]: I1210 23:12:21.618155 4791 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-cell1-server-0" event={"ID":"47eb9200-b0bc-41a6-abb9-f6167bd69c66","Type":"ContainerDied","Data":"01e09f7fc253f735d96b6208f87608bbd0ff2e0759cea509621b849b58aad1f4"} Dec 10 23:12:21 crc kubenswrapper[4791]: I1210 23:12:21.618169 4791 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="01e09f7fc253f735d96b6208f87608bbd0ff2e0759cea509621b849b58aad1f4" Dec 10 23:12:21 crc kubenswrapper[4791]: I1210 23:12:21.620223 4791 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-server-0" event={"ID":"9ce8dbeb-dce8-47ac-98f1-fe92f74aa3de","Type":"ContainerDied","Data":"8b25683d606ab47b6eb65d1f8cd0a149b5a68f9bfbcb6f1b892270e6316094af"} Dec 10 23:12:21 crc kubenswrapper[4791]: I1210 23:12:21.620268 4791 scope.go:117] "RemoveContainer" containerID="f7a414fff9fd9d4a710f40666dd232a7704dad812d497846c5077aa27f08208d" Dec 10 23:12:21 crc kubenswrapper[4791]: I1210 23:12:21.620328 4791 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/rabbitmq-server-0" Dec 10 23:12:21 crc kubenswrapper[4791]: I1210 23:12:21.736709 4791 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/rabbitmq-cell1-server-0" Dec 10 23:12:21 crc kubenswrapper[4791]: I1210 23:12:21.746643 4791 scope.go:117] "RemoveContainer" containerID="d8a93d7e8bfe0416ef17df92a9875113cc702aca941f1f73df50121135822c94" Dec 10 23:12:21 crc kubenswrapper[4791]: I1210 23:12:21.759017 4791 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/rabbitmq-server-0"] Dec 10 23:12:21 crc kubenswrapper[4791]: I1210 23:12:21.785872 4791 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/rabbitmq-server-0"] Dec 10 23:12:21 crc kubenswrapper[4791]: I1210 23:12:21.812272 4791 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/rabbitmq-server-0"] Dec 10 23:12:21 crc kubenswrapper[4791]: E1210 23:12:21.813007 4791 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="9ce8dbeb-dce8-47ac-98f1-fe92f74aa3de" containerName="setup-container" Dec 10 23:12:21 crc kubenswrapper[4791]: I1210 23:12:21.813127 4791 state_mem.go:107] "Deleted CPUSet assignment" podUID="9ce8dbeb-dce8-47ac-98f1-fe92f74aa3de" containerName="setup-container" Dec 10 23:12:21 crc kubenswrapper[4791]: E1210 23:12:21.813216 4791 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="47eb9200-b0bc-41a6-abb9-f6167bd69c66" containerName="setup-container" Dec 10 23:12:21 crc kubenswrapper[4791]: I1210 23:12:21.813279 4791 state_mem.go:107] "Deleted CPUSet assignment" podUID="47eb9200-b0bc-41a6-abb9-f6167bd69c66" containerName="setup-container" Dec 10 23:12:21 crc kubenswrapper[4791]: E1210 23:12:21.813370 4791 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="47eb9200-b0bc-41a6-abb9-f6167bd69c66" containerName="rabbitmq" Dec 10 23:12:21 crc kubenswrapper[4791]: I1210 23:12:21.813440 4791 state_mem.go:107] "Deleted CPUSet assignment" podUID="47eb9200-b0bc-41a6-abb9-f6167bd69c66" containerName="rabbitmq" Dec 10 23:12:21 crc kubenswrapper[4791]: E1210 23:12:21.813606 4791 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="9ce8dbeb-dce8-47ac-98f1-fe92f74aa3de" containerName="rabbitmq" Dec 10 23:12:21 crc kubenswrapper[4791]: I1210 23:12:21.813680 4791 state_mem.go:107] "Deleted CPUSet assignment" podUID="9ce8dbeb-dce8-47ac-98f1-fe92f74aa3de" containerName="rabbitmq" Dec 10 23:12:21 crc kubenswrapper[4791]: I1210 23:12:21.813976 4791 memory_manager.go:354] "RemoveStaleState removing state" podUID="9ce8dbeb-dce8-47ac-98f1-fe92f74aa3de" containerName="rabbitmq" Dec 10 23:12:21 crc kubenswrapper[4791]: I1210 23:12:21.814077 4791 memory_manager.go:354] "RemoveStaleState removing state" podUID="47eb9200-b0bc-41a6-abb9-f6167bd69c66" containerName="rabbitmq" Dec 10 23:12:21 crc kubenswrapper[4791]: I1210 23:12:21.815811 4791 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/rabbitmq-server-0" Dec 10 23:12:21 crc kubenswrapper[4791]: I1210 23:12:21.823249 4791 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"rabbitmq-server-dockercfg-568t8" Dec 10 23:12:21 crc kubenswrapper[4791]: I1210 23:12:21.824837 4791 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"rabbitmq-erlang-cookie" Dec 10 23:12:21 crc kubenswrapper[4791]: I1210 23:12:21.827453 4791 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"rabbitmq-config-data" Dec 10 23:12:21 crc kubenswrapper[4791]: I1210 23:12:21.827669 4791 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"rabbitmq-plugins-conf" Dec 10 23:12:21 crc kubenswrapper[4791]: I1210 23:12:21.827841 4791 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-rabbitmq-svc" Dec 10 23:12:21 crc kubenswrapper[4791]: I1210 23:12:21.827964 4791 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"rabbitmq-default-user" Dec 10 23:12:21 crc kubenswrapper[4791]: I1210 23:12:21.828074 4791 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"rabbitmq-server-conf" Dec 10 23:12:21 crc kubenswrapper[4791]: I1210 23:12:21.839162 4791 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/rabbitmq-server-0"] Dec 10 23:12:21 crc kubenswrapper[4791]: I1210 23:12:21.901134 4791 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="9ce8dbeb-dce8-47ac-98f1-fe92f74aa3de" path="/var/lib/kubelet/pods/9ce8dbeb-dce8-47ac-98f1-fe92f74aa3de/volumes" Dec 10 23:12:21 crc kubenswrapper[4791]: I1210 23:12:21.925986 4791 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/47eb9200-b0bc-41a6-abb9-f6167bd69c66-pod-info\") pod \"47eb9200-b0bc-41a6-abb9-f6167bd69c66\" (UID: \"47eb9200-b0bc-41a6-abb9-f6167bd69c66\") " Dec 10 23:12:21 crc kubenswrapper[4791]: I1210 23:12:21.926052 4791 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-kmc8x\" (UniqueName: \"kubernetes.io/projected/47eb9200-b0bc-41a6-abb9-f6167bd69c66-kube-api-access-kmc8x\") pod \"47eb9200-b0bc-41a6-abb9-f6167bd69c66\" (UID: \"47eb9200-b0bc-41a6-abb9-f6167bd69c66\") " Dec 10 23:12:21 crc kubenswrapper[4791]: I1210 23:12:21.926116 4791 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/47eb9200-b0bc-41a6-abb9-f6167bd69c66-rabbitmq-confd\") pod \"47eb9200-b0bc-41a6-abb9-f6167bd69c66\" (UID: \"47eb9200-b0bc-41a6-abb9-f6167bd69c66\") " Dec 10 23:12:21 crc kubenswrapper[4791]: I1210 23:12:21.926154 4791 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"persistence\" (UniqueName: \"kubernetes.io/local-volume/local-storage08-crc\") pod \"47eb9200-b0bc-41a6-abb9-f6167bd69c66\" (UID: \"47eb9200-b0bc-41a6-abb9-f6167bd69c66\") " Dec 10 23:12:21 crc kubenswrapper[4791]: I1210 23:12:21.926182 4791 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/47eb9200-b0bc-41a6-abb9-f6167bd69c66-plugins-conf\") pod \"47eb9200-b0bc-41a6-abb9-f6167bd69c66\" (UID: \"47eb9200-b0bc-41a6-abb9-f6167bd69c66\") " Dec 10 23:12:21 crc kubenswrapper[4791]: I1210 23:12:21.926307 4791 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/47eb9200-b0bc-41a6-abb9-f6167bd69c66-config-data\") pod \"47eb9200-b0bc-41a6-abb9-f6167bd69c66\" (UID: \"47eb9200-b0bc-41a6-abb9-f6167bd69c66\") " Dec 10 23:12:21 crc kubenswrapper[4791]: I1210 23:12:21.926368 4791 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/47eb9200-b0bc-41a6-abb9-f6167bd69c66-erlang-cookie-secret\") pod \"47eb9200-b0bc-41a6-abb9-f6167bd69c66\" (UID: \"47eb9200-b0bc-41a6-abb9-f6167bd69c66\") " Dec 10 23:12:21 crc kubenswrapper[4791]: I1210 23:12:21.926431 4791 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/47eb9200-b0bc-41a6-abb9-f6167bd69c66-rabbitmq-plugins\") pod \"47eb9200-b0bc-41a6-abb9-f6167bd69c66\" (UID: \"47eb9200-b0bc-41a6-abb9-f6167bd69c66\") " Dec 10 23:12:21 crc kubenswrapper[4791]: I1210 23:12:21.926471 4791 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/47eb9200-b0bc-41a6-abb9-f6167bd69c66-rabbitmq-erlang-cookie\") pod \"47eb9200-b0bc-41a6-abb9-f6167bd69c66\" (UID: \"47eb9200-b0bc-41a6-abb9-f6167bd69c66\") " Dec 10 23:12:21 crc kubenswrapper[4791]: I1210 23:12:21.926514 4791 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/47eb9200-b0bc-41a6-abb9-f6167bd69c66-server-conf\") pod \"47eb9200-b0bc-41a6-abb9-f6167bd69c66\" (UID: \"47eb9200-b0bc-41a6-abb9-f6167bd69c66\") " Dec 10 23:12:21 crc kubenswrapper[4791]: I1210 23:12:21.926595 4791 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"rabbitmq-tls\" (UniqueName: \"kubernetes.io/projected/47eb9200-b0bc-41a6-abb9-f6167bd69c66-rabbitmq-tls\") pod \"47eb9200-b0bc-41a6-abb9-f6167bd69c66\" (UID: \"47eb9200-b0bc-41a6-abb9-f6167bd69c66\") " Dec 10 23:12:21 crc kubenswrapper[4791]: I1210 23:12:21.926914 4791 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage07-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage07-crc\") pod \"rabbitmq-server-0\" (UID: \"5f51c37d-2f6f-4ef0-a08c-77216e7db57b\") " pod="openstack/rabbitmq-server-0" Dec 10 23:12:21 crc kubenswrapper[4791]: I1210 23:12:21.926966 4791 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/5f51c37d-2f6f-4ef0-a08c-77216e7db57b-erlang-cookie-secret\") pod \"rabbitmq-server-0\" (UID: \"5f51c37d-2f6f-4ef0-a08c-77216e7db57b\") " pod="openstack/rabbitmq-server-0" Dec 10 23:12:21 crc kubenswrapper[4791]: I1210 23:12:21.927011 4791 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-hr8gb\" (UniqueName: \"kubernetes.io/projected/5f51c37d-2f6f-4ef0-a08c-77216e7db57b-kube-api-access-hr8gb\") pod \"rabbitmq-server-0\" (UID: \"5f51c37d-2f6f-4ef0-a08c-77216e7db57b\") " pod="openstack/rabbitmq-server-0" Dec 10 23:12:21 crc kubenswrapper[4791]: I1210 23:12:21.927047 4791 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/5f51c37d-2f6f-4ef0-a08c-77216e7db57b-config-data\") pod \"rabbitmq-server-0\" (UID: \"5f51c37d-2f6f-4ef0-a08c-77216e7db57b\") " pod="openstack/rabbitmq-server-0" Dec 10 23:12:21 crc kubenswrapper[4791]: I1210 23:12:21.927077 4791 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/5f51c37d-2f6f-4ef0-a08c-77216e7db57b-rabbitmq-erlang-cookie\") pod \"rabbitmq-server-0\" (UID: \"5f51c37d-2f6f-4ef0-a08c-77216e7db57b\") " pod="openstack/rabbitmq-server-0" Dec 10 23:12:21 crc kubenswrapper[4791]: I1210 23:12:21.927098 4791 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/5f51c37d-2f6f-4ef0-a08c-77216e7db57b-server-conf\") pod \"rabbitmq-server-0\" (UID: \"5f51c37d-2f6f-4ef0-a08c-77216e7db57b\") " pod="openstack/rabbitmq-server-0" Dec 10 23:12:21 crc kubenswrapper[4791]: I1210 23:12:21.927140 4791 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/5f51c37d-2f6f-4ef0-a08c-77216e7db57b-rabbitmq-confd\") pod \"rabbitmq-server-0\" (UID: \"5f51c37d-2f6f-4ef0-a08c-77216e7db57b\") " pod="openstack/rabbitmq-server-0" Dec 10 23:12:21 crc kubenswrapper[4791]: I1210 23:12:21.927203 4791 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-tls\" (UniqueName: \"kubernetes.io/projected/5f51c37d-2f6f-4ef0-a08c-77216e7db57b-rabbitmq-tls\") pod \"rabbitmq-server-0\" (UID: \"5f51c37d-2f6f-4ef0-a08c-77216e7db57b\") " pod="openstack/rabbitmq-server-0" Dec 10 23:12:21 crc kubenswrapper[4791]: I1210 23:12:21.927234 4791 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/5f51c37d-2f6f-4ef0-a08c-77216e7db57b-rabbitmq-plugins\") pod \"rabbitmq-server-0\" (UID: \"5f51c37d-2f6f-4ef0-a08c-77216e7db57b\") " pod="openstack/rabbitmq-server-0" Dec 10 23:12:21 crc kubenswrapper[4791]: I1210 23:12:21.927275 4791 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/5f51c37d-2f6f-4ef0-a08c-77216e7db57b-pod-info\") pod \"rabbitmq-server-0\" (UID: \"5f51c37d-2f6f-4ef0-a08c-77216e7db57b\") " pod="openstack/rabbitmq-server-0" Dec 10 23:12:21 crc kubenswrapper[4791]: I1210 23:12:21.927311 4791 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/5f51c37d-2f6f-4ef0-a08c-77216e7db57b-plugins-conf\") pod \"rabbitmq-server-0\" (UID: \"5f51c37d-2f6f-4ef0-a08c-77216e7db57b\") " pod="openstack/rabbitmq-server-0" Dec 10 23:12:21 crc kubenswrapper[4791]: I1210 23:12:21.927501 4791 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/47eb9200-b0bc-41a6-abb9-f6167bd69c66-rabbitmq-plugins" (OuterVolumeSpecName: "rabbitmq-plugins") pod "47eb9200-b0bc-41a6-abb9-f6167bd69c66" (UID: "47eb9200-b0bc-41a6-abb9-f6167bd69c66"). InnerVolumeSpecName "rabbitmq-plugins". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 10 23:12:21 crc kubenswrapper[4791]: I1210 23:12:21.928139 4791 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/47eb9200-b0bc-41a6-abb9-f6167bd69c66-rabbitmq-erlang-cookie" (OuterVolumeSpecName: "rabbitmq-erlang-cookie") pod "47eb9200-b0bc-41a6-abb9-f6167bd69c66" (UID: "47eb9200-b0bc-41a6-abb9-f6167bd69c66"). InnerVolumeSpecName "rabbitmq-erlang-cookie". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 10 23:12:21 crc kubenswrapper[4791]: I1210 23:12:21.928846 4791 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/47eb9200-b0bc-41a6-abb9-f6167bd69c66-plugins-conf" (OuterVolumeSpecName: "plugins-conf") pod "47eb9200-b0bc-41a6-abb9-f6167bd69c66" (UID: "47eb9200-b0bc-41a6-abb9-f6167bd69c66"). InnerVolumeSpecName "plugins-conf". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 10 23:12:21 crc kubenswrapper[4791]: I1210 23:12:21.932523 4791 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/47eb9200-b0bc-41a6-abb9-f6167bd69c66-erlang-cookie-secret" (OuterVolumeSpecName: "erlang-cookie-secret") pod "47eb9200-b0bc-41a6-abb9-f6167bd69c66" (UID: "47eb9200-b0bc-41a6-abb9-f6167bd69c66"). InnerVolumeSpecName "erlang-cookie-secret". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 10 23:12:21 crc kubenswrapper[4791]: I1210 23:12:21.933079 4791 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/downward-api/47eb9200-b0bc-41a6-abb9-f6167bd69c66-pod-info" (OuterVolumeSpecName: "pod-info") pod "47eb9200-b0bc-41a6-abb9-f6167bd69c66" (UID: "47eb9200-b0bc-41a6-abb9-f6167bd69c66"). InnerVolumeSpecName "pod-info". PluginName "kubernetes.io/downward-api", VolumeGidValue "" Dec 10 23:12:21 crc kubenswrapper[4791]: I1210 23:12:21.934226 4791 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/local-volume/local-storage08-crc" (OuterVolumeSpecName: "persistence") pod "47eb9200-b0bc-41a6-abb9-f6167bd69c66" (UID: "47eb9200-b0bc-41a6-abb9-f6167bd69c66"). InnerVolumeSpecName "local-storage08-crc". PluginName "kubernetes.io/local-volume", VolumeGidValue "" Dec 10 23:12:21 crc kubenswrapper[4791]: I1210 23:12:21.946393 4791 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/47eb9200-b0bc-41a6-abb9-f6167bd69c66-kube-api-access-kmc8x" (OuterVolumeSpecName: "kube-api-access-kmc8x") pod "47eb9200-b0bc-41a6-abb9-f6167bd69c66" (UID: "47eb9200-b0bc-41a6-abb9-f6167bd69c66"). InnerVolumeSpecName "kube-api-access-kmc8x". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 10 23:12:21 crc kubenswrapper[4791]: I1210 23:12:21.947563 4791 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/47eb9200-b0bc-41a6-abb9-f6167bd69c66-rabbitmq-tls" (OuterVolumeSpecName: "rabbitmq-tls") pod "47eb9200-b0bc-41a6-abb9-f6167bd69c66" (UID: "47eb9200-b0bc-41a6-abb9-f6167bd69c66"). InnerVolumeSpecName "rabbitmq-tls". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 10 23:12:22 crc kubenswrapper[4791]: I1210 23:12:22.010769 4791 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/47eb9200-b0bc-41a6-abb9-f6167bd69c66-config-data" (OuterVolumeSpecName: "config-data") pod "47eb9200-b0bc-41a6-abb9-f6167bd69c66" (UID: "47eb9200-b0bc-41a6-abb9-f6167bd69c66"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 10 23:12:22 crc kubenswrapper[4791]: I1210 23:12:22.015236 4791 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/47eb9200-b0bc-41a6-abb9-f6167bd69c66-server-conf" (OuterVolumeSpecName: "server-conf") pod "47eb9200-b0bc-41a6-abb9-f6167bd69c66" (UID: "47eb9200-b0bc-41a6-abb9-f6167bd69c66"). InnerVolumeSpecName "server-conf". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 10 23:12:22 crc kubenswrapper[4791]: I1210 23:12:22.028709 4791 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/5f51c37d-2f6f-4ef0-a08c-77216e7db57b-rabbitmq-confd\") pod \"rabbitmq-server-0\" (UID: \"5f51c37d-2f6f-4ef0-a08c-77216e7db57b\") " pod="openstack/rabbitmq-server-0" Dec 10 23:12:22 crc kubenswrapper[4791]: I1210 23:12:22.028793 4791 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-tls\" (UniqueName: \"kubernetes.io/projected/5f51c37d-2f6f-4ef0-a08c-77216e7db57b-rabbitmq-tls\") pod \"rabbitmq-server-0\" (UID: \"5f51c37d-2f6f-4ef0-a08c-77216e7db57b\") " pod="openstack/rabbitmq-server-0" Dec 10 23:12:22 crc kubenswrapper[4791]: I1210 23:12:22.028813 4791 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/5f51c37d-2f6f-4ef0-a08c-77216e7db57b-rabbitmq-plugins\") pod \"rabbitmq-server-0\" (UID: \"5f51c37d-2f6f-4ef0-a08c-77216e7db57b\") " pod="openstack/rabbitmq-server-0" Dec 10 23:12:22 crc kubenswrapper[4791]: I1210 23:12:22.029197 4791 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/5f51c37d-2f6f-4ef0-a08c-77216e7db57b-pod-info\") pod \"rabbitmq-server-0\" (UID: \"5f51c37d-2f6f-4ef0-a08c-77216e7db57b\") " pod="openstack/rabbitmq-server-0" Dec 10 23:12:22 crc kubenswrapper[4791]: I1210 23:12:22.029232 4791 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/5f51c37d-2f6f-4ef0-a08c-77216e7db57b-plugins-conf\") pod \"rabbitmq-server-0\" (UID: \"5f51c37d-2f6f-4ef0-a08c-77216e7db57b\") " pod="openstack/rabbitmq-server-0" Dec 10 23:12:22 crc kubenswrapper[4791]: I1210 23:12:22.029303 4791 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage07-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage07-crc\") pod \"rabbitmq-server-0\" (UID: \"5f51c37d-2f6f-4ef0-a08c-77216e7db57b\") " pod="openstack/rabbitmq-server-0" Dec 10 23:12:22 crc kubenswrapper[4791]: I1210 23:12:22.029332 4791 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/5f51c37d-2f6f-4ef0-a08c-77216e7db57b-erlang-cookie-secret\") pod \"rabbitmq-server-0\" (UID: \"5f51c37d-2f6f-4ef0-a08c-77216e7db57b\") " pod="openstack/rabbitmq-server-0" Dec 10 23:12:22 crc kubenswrapper[4791]: I1210 23:12:22.029393 4791 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-hr8gb\" (UniqueName: \"kubernetes.io/projected/5f51c37d-2f6f-4ef0-a08c-77216e7db57b-kube-api-access-hr8gb\") pod \"rabbitmq-server-0\" (UID: \"5f51c37d-2f6f-4ef0-a08c-77216e7db57b\") " pod="openstack/rabbitmq-server-0" Dec 10 23:12:22 crc kubenswrapper[4791]: I1210 23:12:22.029431 4791 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/5f51c37d-2f6f-4ef0-a08c-77216e7db57b-config-data\") pod \"rabbitmq-server-0\" (UID: \"5f51c37d-2f6f-4ef0-a08c-77216e7db57b\") " pod="openstack/rabbitmq-server-0" Dec 10 23:12:22 crc kubenswrapper[4791]: I1210 23:12:22.029463 4791 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/5f51c37d-2f6f-4ef0-a08c-77216e7db57b-rabbitmq-erlang-cookie\") pod \"rabbitmq-server-0\" (UID: \"5f51c37d-2f6f-4ef0-a08c-77216e7db57b\") " pod="openstack/rabbitmq-server-0" Dec 10 23:12:22 crc kubenswrapper[4791]: I1210 23:12:22.029480 4791 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/5f51c37d-2f6f-4ef0-a08c-77216e7db57b-server-conf\") pod \"rabbitmq-server-0\" (UID: \"5f51c37d-2f6f-4ef0-a08c-77216e7db57b\") " pod="openstack/rabbitmq-server-0" Dec 10 23:12:22 crc kubenswrapper[4791]: I1210 23:12:22.029529 4791 reconciler_common.go:293] "Volume detached for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/47eb9200-b0bc-41a6-abb9-f6167bd69c66-pod-info\") on node \"crc\" DevicePath \"\"" Dec 10 23:12:22 crc kubenswrapper[4791]: I1210 23:12:22.029547 4791 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-kmc8x\" (UniqueName: \"kubernetes.io/projected/47eb9200-b0bc-41a6-abb9-f6167bd69c66-kube-api-access-kmc8x\") on node \"crc\" DevicePath \"\"" Dec 10 23:12:22 crc kubenswrapper[4791]: I1210 23:12:22.029600 4791 reconciler_common.go:286] "operationExecutor.UnmountDevice started for volume \"local-storage08-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage08-crc\") on node \"crc\" " Dec 10 23:12:22 crc kubenswrapper[4791]: I1210 23:12:22.029610 4791 reconciler_common.go:293] "Volume detached for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/47eb9200-b0bc-41a6-abb9-f6167bd69c66-plugins-conf\") on node \"crc\" DevicePath \"\"" Dec 10 23:12:22 crc kubenswrapper[4791]: I1210 23:12:22.029619 4791 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/47eb9200-b0bc-41a6-abb9-f6167bd69c66-config-data\") on node \"crc\" DevicePath \"\"" Dec 10 23:12:22 crc kubenswrapper[4791]: I1210 23:12:22.029628 4791 reconciler_common.go:293] "Volume detached for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/47eb9200-b0bc-41a6-abb9-f6167bd69c66-erlang-cookie-secret\") on node \"crc\" DevicePath \"\"" Dec 10 23:12:22 crc kubenswrapper[4791]: I1210 23:12:22.029636 4791 reconciler_common.go:293] "Volume detached for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/47eb9200-b0bc-41a6-abb9-f6167bd69c66-rabbitmq-plugins\") on node \"crc\" DevicePath \"\"" Dec 10 23:12:22 crc kubenswrapper[4791]: I1210 23:12:22.029644 4791 reconciler_common.go:293] "Volume detached for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/47eb9200-b0bc-41a6-abb9-f6167bd69c66-rabbitmq-erlang-cookie\") on node \"crc\" DevicePath \"\"" Dec 10 23:12:22 crc kubenswrapper[4791]: I1210 23:12:22.029653 4791 reconciler_common.go:293] "Volume detached for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/47eb9200-b0bc-41a6-abb9-f6167bd69c66-server-conf\") on node \"crc\" DevicePath \"\"" Dec 10 23:12:22 crc kubenswrapper[4791]: I1210 23:12:22.029662 4791 reconciler_common.go:293] "Volume detached for volume \"rabbitmq-tls\" (UniqueName: \"kubernetes.io/projected/47eb9200-b0bc-41a6-abb9-f6167bd69c66-rabbitmq-tls\") on node \"crc\" DevicePath \"\"" Dec 10 23:12:22 crc kubenswrapper[4791]: I1210 23:12:22.033073 4791 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage07-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage07-crc\") pod \"rabbitmq-server-0\" (UID: \"5f51c37d-2f6f-4ef0-a08c-77216e7db57b\") device mount path \"/mnt/openstack/pv07\"" pod="openstack/rabbitmq-server-0" Dec 10 23:12:22 crc kubenswrapper[4791]: I1210 23:12:22.035875 4791 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/5f51c37d-2f6f-4ef0-a08c-77216e7db57b-server-conf\") pod \"rabbitmq-server-0\" (UID: \"5f51c37d-2f6f-4ef0-a08c-77216e7db57b\") " pod="openstack/rabbitmq-server-0" Dec 10 23:12:22 crc kubenswrapper[4791]: I1210 23:12:22.036029 4791 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-tls\" (UniqueName: \"kubernetes.io/projected/5f51c37d-2f6f-4ef0-a08c-77216e7db57b-rabbitmq-tls\") pod \"rabbitmq-server-0\" (UID: \"5f51c37d-2f6f-4ef0-a08c-77216e7db57b\") " pod="openstack/rabbitmq-server-0" Dec 10 23:12:22 crc kubenswrapper[4791]: I1210 23:12:22.037019 4791 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/5f51c37d-2f6f-4ef0-a08c-77216e7db57b-rabbitmq-confd\") pod \"rabbitmq-server-0\" (UID: \"5f51c37d-2f6f-4ef0-a08c-77216e7db57b\") " pod="openstack/rabbitmq-server-0" Dec 10 23:12:22 crc kubenswrapper[4791]: I1210 23:12:22.041658 4791 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/5f51c37d-2f6f-4ef0-a08c-77216e7db57b-config-data\") pod \"rabbitmq-server-0\" (UID: \"5f51c37d-2f6f-4ef0-a08c-77216e7db57b\") " pod="openstack/rabbitmq-server-0" Dec 10 23:12:22 crc kubenswrapper[4791]: I1210 23:12:22.043494 4791 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/5f51c37d-2f6f-4ef0-a08c-77216e7db57b-pod-info\") pod \"rabbitmq-server-0\" (UID: \"5f51c37d-2f6f-4ef0-a08c-77216e7db57b\") " pod="openstack/rabbitmq-server-0" Dec 10 23:12:22 crc kubenswrapper[4791]: I1210 23:12:22.045890 4791 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/5f51c37d-2f6f-4ef0-a08c-77216e7db57b-plugins-conf\") pod \"rabbitmq-server-0\" (UID: \"5f51c37d-2f6f-4ef0-a08c-77216e7db57b\") " pod="openstack/rabbitmq-server-0" Dec 10 23:12:22 crc kubenswrapper[4791]: I1210 23:12:22.047747 4791 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/5f51c37d-2f6f-4ef0-a08c-77216e7db57b-rabbitmq-plugins\") pod \"rabbitmq-server-0\" (UID: \"5f51c37d-2f6f-4ef0-a08c-77216e7db57b\") " pod="openstack/rabbitmq-server-0" Dec 10 23:12:22 crc kubenswrapper[4791]: I1210 23:12:22.047835 4791 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/5f51c37d-2f6f-4ef0-a08c-77216e7db57b-rabbitmq-erlang-cookie\") pod \"rabbitmq-server-0\" (UID: \"5f51c37d-2f6f-4ef0-a08c-77216e7db57b\") " pod="openstack/rabbitmq-server-0" Dec 10 23:12:22 crc kubenswrapper[4791]: I1210 23:12:22.051948 4791 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/5f51c37d-2f6f-4ef0-a08c-77216e7db57b-erlang-cookie-secret\") pod \"rabbitmq-server-0\" (UID: \"5f51c37d-2f6f-4ef0-a08c-77216e7db57b\") " pod="openstack/rabbitmq-server-0" Dec 10 23:12:22 crc kubenswrapper[4791]: I1210 23:12:22.052746 4791 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-hr8gb\" (UniqueName: \"kubernetes.io/projected/5f51c37d-2f6f-4ef0-a08c-77216e7db57b-kube-api-access-hr8gb\") pod \"rabbitmq-server-0\" (UID: \"5f51c37d-2f6f-4ef0-a08c-77216e7db57b\") " pod="openstack/rabbitmq-server-0" Dec 10 23:12:22 crc kubenswrapper[4791]: I1210 23:12:22.064482 4791 operation_generator.go:917] UnmountDevice succeeded for volume "local-storage08-crc" (UniqueName: "kubernetes.io/local-volume/local-storage08-crc") on node "crc" Dec 10 23:12:22 crc kubenswrapper[4791]: I1210 23:12:22.094542 4791 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/47eb9200-b0bc-41a6-abb9-f6167bd69c66-rabbitmq-confd" (OuterVolumeSpecName: "rabbitmq-confd") pod "47eb9200-b0bc-41a6-abb9-f6167bd69c66" (UID: "47eb9200-b0bc-41a6-abb9-f6167bd69c66"). InnerVolumeSpecName "rabbitmq-confd". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 10 23:12:22 crc kubenswrapper[4791]: I1210 23:12:22.112087 4791 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage07-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage07-crc\") pod \"rabbitmq-server-0\" (UID: \"5f51c37d-2f6f-4ef0-a08c-77216e7db57b\") " pod="openstack/rabbitmq-server-0" Dec 10 23:12:22 crc kubenswrapper[4791]: I1210 23:12:22.131654 4791 reconciler_common.go:293] "Volume detached for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/47eb9200-b0bc-41a6-abb9-f6167bd69c66-rabbitmq-confd\") on node \"crc\" DevicePath \"\"" Dec 10 23:12:22 crc kubenswrapper[4791]: I1210 23:12:22.131686 4791 reconciler_common.go:293] "Volume detached for volume \"local-storage08-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage08-crc\") on node \"crc\" DevicePath \"\"" Dec 10 23:12:22 crc kubenswrapper[4791]: I1210 23:12:22.154099 4791 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/rabbitmq-server-0" Dec 10 23:12:22 crc kubenswrapper[4791]: I1210 23:12:22.634444 4791 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/rabbitmq-cell1-server-0" Dec 10 23:12:22 crc kubenswrapper[4791]: I1210 23:12:22.637973 4791 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/rabbitmq-server-0"] Dec 10 23:12:22 crc kubenswrapper[4791]: I1210 23:12:22.975403 4791 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/rabbitmq-cell1-server-0"] Dec 10 23:12:22 crc kubenswrapper[4791]: I1210 23:12:22.987044 4791 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/rabbitmq-cell1-server-0"] Dec 10 23:12:23 crc kubenswrapper[4791]: I1210 23:12:23.001323 4791 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/rabbitmq-cell1-server-0"] Dec 10 23:12:23 crc kubenswrapper[4791]: I1210 23:12:23.003109 4791 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/rabbitmq-cell1-server-0" Dec 10 23:12:23 crc kubenswrapper[4791]: I1210 23:12:23.006848 4791 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"rabbitmq-cell1-erlang-cookie" Dec 10 23:12:23 crc kubenswrapper[4791]: I1210 23:12:23.007164 4791 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"rabbitmq-cell1-server-conf" Dec 10 23:12:23 crc kubenswrapper[4791]: I1210 23:12:23.007321 4791 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"rabbitmq-cell1-default-user" Dec 10 23:12:23 crc kubenswrapper[4791]: I1210 23:12:23.008352 4791 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"rabbitmq-cell1-server-dockercfg-scxq4" Dec 10 23:12:23 crc kubenswrapper[4791]: I1210 23:12:23.008633 4791 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"rabbitmq-cell1-config-data" Dec 10 23:12:23 crc kubenswrapper[4791]: I1210 23:12:23.008655 4791 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-rabbitmq-cell1-svc" Dec 10 23:12:23 crc kubenswrapper[4791]: I1210 23:12:23.008755 4791 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"rabbitmq-cell1-plugins-conf" Dec 10 23:12:23 crc kubenswrapper[4791]: I1210 23:12:23.019739 4791 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/rabbitmq-cell1-server-0"] Dec 10 23:12:23 crc kubenswrapper[4791]: I1210 23:12:23.156635 4791 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage08-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage08-crc\") pod \"rabbitmq-cell1-server-0\" (UID: \"3add67e0-b67a-438b-b83f-4c9f6733b5cb\") " pod="openstack/rabbitmq-cell1-server-0" Dec 10 23:12:23 crc kubenswrapper[4791]: I1210 23:12:23.156701 4791 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/3add67e0-b67a-438b-b83f-4c9f6733b5cb-rabbitmq-erlang-cookie\") pod \"rabbitmq-cell1-server-0\" (UID: \"3add67e0-b67a-438b-b83f-4c9f6733b5cb\") " pod="openstack/rabbitmq-cell1-server-0" Dec 10 23:12:23 crc kubenswrapper[4791]: I1210 23:12:23.156756 4791 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-kmsnr\" (UniqueName: \"kubernetes.io/projected/3add67e0-b67a-438b-b83f-4c9f6733b5cb-kube-api-access-kmsnr\") pod \"rabbitmq-cell1-server-0\" (UID: \"3add67e0-b67a-438b-b83f-4c9f6733b5cb\") " pod="openstack/rabbitmq-cell1-server-0" Dec 10 23:12:23 crc kubenswrapper[4791]: I1210 23:12:23.156841 4791 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-tls\" (UniqueName: \"kubernetes.io/projected/3add67e0-b67a-438b-b83f-4c9f6733b5cb-rabbitmq-tls\") pod \"rabbitmq-cell1-server-0\" (UID: \"3add67e0-b67a-438b-b83f-4c9f6733b5cb\") " pod="openstack/rabbitmq-cell1-server-0" Dec 10 23:12:23 crc kubenswrapper[4791]: I1210 23:12:23.156951 4791 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/3add67e0-b67a-438b-b83f-4c9f6733b5cb-rabbitmq-plugins\") pod \"rabbitmq-cell1-server-0\" (UID: \"3add67e0-b67a-438b-b83f-4c9f6733b5cb\") " pod="openstack/rabbitmq-cell1-server-0" Dec 10 23:12:23 crc kubenswrapper[4791]: I1210 23:12:23.157004 4791 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/3add67e0-b67a-438b-b83f-4c9f6733b5cb-erlang-cookie-secret\") pod \"rabbitmq-cell1-server-0\" (UID: \"3add67e0-b67a-438b-b83f-4c9f6733b5cb\") " pod="openstack/rabbitmq-cell1-server-0" Dec 10 23:12:23 crc kubenswrapper[4791]: I1210 23:12:23.157042 4791 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/3add67e0-b67a-438b-b83f-4c9f6733b5cb-rabbitmq-confd\") pod \"rabbitmq-cell1-server-0\" (UID: \"3add67e0-b67a-438b-b83f-4c9f6733b5cb\") " pod="openstack/rabbitmq-cell1-server-0" Dec 10 23:12:23 crc kubenswrapper[4791]: I1210 23:12:23.157095 4791 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/3add67e0-b67a-438b-b83f-4c9f6733b5cb-plugins-conf\") pod \"rabbitmq-cell1-server-0\" (UID: \"3add67e0-b67a-438b-b83f-4c9f6733b5cb\") " pod="openstack/rabbitmq-cell1-server-0" Dec 10 23:12:23 crc kubenswrapper[4791]: I1210 23:12:23.157120 4791 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/3add67e0-b67a-438b-b83f-4c9f6733b5cb-server-conf\") pod \"rabbitmq-cell1-server-0\" (UID: \"3add67e0-b67a-438b-b83f-4c9f6733b5cb\") " pod="openstack/rabbitmq-cell1-server-0" Dec 10 23:12:23 crc kubenswrapper[4791]: I1210 23:12:23.157145 4791 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/3add67e0-b67a-438b-b83f-4c9f6733b5cb-pod-info\") pod \"rabbitmq-cell1-server-0\" (UID: \"3add67e0-b67a-438b-b83f-4c9f6733b5cb\") " pod="openstack/rabbitmq-cell1-server-0" Dec 10 23:12:23 crc kubenswrapper[4791]: I1210 23:12:23.157195 4791 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/3add67e0-b67a-438b-b83f-4c9f6733b5cb-config-data\") pod \"rabbitmq-cell1-server-0\" (UID: \"3add67e0-b67a-438b-b83f-4c9f6733b5cb\") " pod="openstack/rabbitmq-cell1-server-0" Dec 10 23:12:23 crc kubenswrapper[4791]: I1210 23:12:23.259409 4791 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/3add67e0-b67a-438b-b83f-4c9f6733b5cb-rabbitmq-confd\") pod \"rabbitmq-cell1-server-0\" (UID: \"3add67e0-b67a-438b-b83f-4c9f6733b5cb\") " pod="openstack/rabbitmq-cell1-server-0" Dec 10 23:12:23 crc kubenswrapper[4791]: I1210 23:12:23.259479 4791 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/3add67e0-b67a-438b-b83f-4c9f6733b5cb-plugins-conf\") pod \"rabbitmq-cell1-server-0\" (UID: \"3add67e0-b67a-438b-b83f-4c9f6733b5cb\") " pod="openstack/rabbitmq-cell1-server-0" Dec 10 23:12:23 crc kubenswrapper[4791]: I1210 23:12:23.259497 4791 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/3add67e0-b67a-438b-b83f-4c9f6733b5cb-server-conf\") pod \"rabbitmq-cell1-server-0\" (UID: \"3add67e0-b67a-438b-b83f-4c9f6733b5cb\") " pod="openstack/rabbitmq-cell1-server-0" Dec 10 23:12:23 crc kubenswrapper[4791]: I1210 23:12:23.259535 4791 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/3add67e0-b67a-438b-b83f-4c9f6733b5cb-pod-info\") pod \"rabbitmq-cell1-server-0\" (UID: \"3add67e0-b67a-438b-b83f-4c9f6733b5cb\") " pod="openstack/rabbitmq-cell1-server-0" Dec 10 23:12:23 crc kubenswrapper[4791]: I1210 23:12:23.259570 4791 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/3add67e0-b67a-438b-b83f-4c9f6733b5cb-config-data\") pod \"rabbitmq-cell1-server-0\" (UID: \"3add67e0-b67a-438b-b83f-4c9f6733b5cb\") " pod="openstack/rabbitmq-cell1-server-0" Dec 10 23:12:23 crc kubenswrapper[4791]: I1210 23:12:23.259630 4791 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage08-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage08-crc\") pod \"rabbitmq-cell1-server-0\" (UID: \"3add67e0-b67a-438b-b83f-4c9f6733b5cb\") " pod="openstack/rabbitmq-cell1-server-0" Dec 10 23:12:23 crc kubenswrapper[4791]: I1210 23:12:23.259685 4791 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/3add67e0-b67a-438b-b83f-4c9f6733b5cb-rabbitmq-erlang-cookie\") pod \"rabbitmq-cell1-server-0\" (UID: \"3add67e0-b67a-438b-b83f-4c9f6733b5cb\") " pod="openstack/rabbitmq-cell1-server-0" Dec 10 23:12:23 crc kubenswrapper[4791]: I1210 23:12:23.259717 4791 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-kmsnr\" (UniqueName: \"kubernetes.io/projected/3add67e0-b67a-438b-b83f-4c9f6733b5cb-kube-api-access-kmsnr\") pod \"rabbitmq-cell1-server-0\" (UID: \"3add67e0-b67a-438b-b83f-4c9f6733b5cb\") " pod="openstack/rabbitmq-cell1-server-0" Dec 10 23:12:23 crc kubenswrapper[4791]: I1210 23:12:23.259763 4791 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-tls\" (UniqueName: \"kubernetes.io/projected/3add67e0-b67a-438b-b83f-4c9f6733b5cb-rabbitmq-tls\") pod \"rabbitmq-cell1-server-0\" (UID: \"3add67e0-b67a-438b-b83f-4c9f6733b5cb\") " pod="openstack/rabbitmq-cell1-server-0" Dec 10 23:12:23 crc kubenswrapper[4791]: I1210 23:12:23.259801 4791 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/3add67e0-b67a-438b-b83f-4c9f6733b5cb-rabbitmq-plugins\") pod \"rabbitmq-cell1-server-0\" (UID: \"3add67e0-b67a-438b-b83f-4c9f6733b5cb\") " pod="openstack/rabbitmq-cell1-server-0" Dec 10 23:12:23 crc kubenswrapper[4791]: I1210 23:12:23.259851 4791 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/3add67e0-b67a-438b-b83f-4c9f6733b5cb-erlang-cookie-secret\") pod \"rabbitmq-cell1-server-0\" (UID: \"3add67e0-b67a-438b-b83f-4c9f6733b5cb\") " pod="openstack/rabbitmq-cell1-server-0" Dec 10 23:12:23 crc kubenswrapper[4791]: I1210 23:12:23.260647 4791 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/3add67e0-b67a-438b-b83f-4c9f6733b5cb-rabbitmq-plugins\") pod \"rabbitmq-cell1-server-0\" (UID: \"3add67e0-b67a-438b-b83f-4c9f6733b5cb\") " pod="openstack/rabbitmq-cell1-server-0" Dec 10 23:12:23 crc kubenswrapper[4791]: I1210 23:12:23.260657 4791 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/3add67e0-b67a-438b-b83f-4c9f6733b5cb-rabbitmq-erlang-cookie\") pod \"rabbitmq-cell1-server-0\" (UID: \"3add67e0-b67a-438b-b83f-4c9f6733b5cb\") " pod="openstack/rabbitmq-cell1-server-0" Dec 10 23:12:23 crc kubenswrapper[4791]: I1210 23:12:23.260862 4791 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage08-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage08-crc\") pod \"rabbitmq-cell1-server-0\" (UID: \"3add67e0-b67a-438b-b83f-4c9f6733b5cb\") device mount path \"/mnt/openstack/pv08\"" pod="openstack/rabbitmq-cell1-server-0" Dec 10 23:12:23 crc kubenswrapper[4791]: I1210 23:12:23.261574 4791 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/3add67e0-b67a-438b-b83f-4c9f6733b5cb-config-data\") pod \"rabbitmq-cell1-server-0\" (UID: \"3add67e0-b67a-438b-b83f-4c9f6733b5cb\") " pod="openstack/rabbitmq-cell1-server-0" Dec 10 23:12:23 crc kubenswrapper[4791]: I1210 23:12:23.261634 4791 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/3add67e0-b67a-438b-b83f-4c9f6733b5cb-server-conf\") pod \"rabbitmq-cell1-server-0\" (UID: \"3add67e0-b67a-438b-b83f-4c9f6733b5cb\") " pod="openstack/rabbitmq-cell1-server-0" Dec 10 23:12:23 crc kubenswrapper[4791]: I1210 23:12:23.262185 4791 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/3add67e0-b67a-438b-b83f-4c9f6733b5cb-plugins-conf\") pod \"rabbitmq-cell1-server-0\" (UID: \"3add67e0-b67a-438b-b83f-4c9f6733b5cb\") " pod="openstack/rabbitmq-cell1-server-0" Dec 10 23:12:23 crc kubenswrapper[4791]: I1210 23:12:23.264825 4791 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/3add67e0-b67a-438b-b83f-4c9f6733b5cb-pod-info\") pod \"rabbitmq-cell1-server-0\" (UID: \"3add67e0-b67a-438b-b83f-4c9f6733b5cb\") " pod="openstack/rabbitmq-cell1-server-0" Dec 10 23:12:23 crc kubenswrapper[4791]: I1210 23:12:23.268304 4791 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/3add67e0-b67a-438b-b83f-4c9f6733b5cb-rabbitmq-confd\") pod \"rabbitmq-cell1-server-0\" (UID: \"3add67e0-b67a-438b-b83f-4c9f6733b5cb\") " pod="openstack/rabbitmq-cell1-server-0" Dec 10 23:12:23 crc kubenswrapper[4791]: I1210 23:12:23.268516 4791 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-tls\" (UniqueName: \"kubernetes.io/projected/3add67e0-b67a-438b-b83f-4c9f6733b5cb-rabbitmq-tls\") pod \"rabbitmq-cell1-server-0\" (UID: \"3add67e0-b67a-438b-b83f-4c9f6733b5cb\") " pod="openstack/rabbitmq-cell1-server-0" Dec 10 23:12:23 crc kubenswrapper[4791]: I1210 23:12:23.270733 4791 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/3add67e0-b67a-438b-b83f-4c9f6733b5cb-erlang-cookie-secret\") pod \"rabbitmq-cell1-server-0\" (UID: \"3add67e0-b67a-438b-b83f-4c9f6733b5cb\") " pod="openstack/rabbitmq-cell1-server-0" Dec 10 23:12:23 crc kubenswrapper[4791]: I1210 23:12:23.287204 4791 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-kmsnr\" (UniqueName: \"kubernetes.io/projected/3add67e0-b67a-438b-b83f-4c9f6733b5cb-kube-api-access-kmsnr\") pod \"rabbitmq-cell1-server-0\" (UID: \"3add67e0-b67a-438b-b83f-4c9f6733b5cb\") " pod="openstack/rabbitmq-cell1-server-0" Dec 10 23:12:23 crc kubenswrapper[4791]: I1210 23:12:23.300268 4791 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage08-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage08-crc\") pod \"rabbitmq-cell1-server-0\" (UID: \"3add67e0-b67a-438b-b83f-4c9f6733b5cb\") " pod="openstack/rabbitmq-cell1-server-0" Dec 10 23:12:23 crc kubenswrapper[4791]: I1210 23:12:23.325843 4791 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/rabbitmq-cell1-server-0" Dec 10 23:12:23 crc kubenswrapper[4791]: I1210 23:12:23.395029 4791 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-67b789f86c-cq8ml"] Dec 10 23:12:23 crc kubenswrapper[4791]: I1210 23:12:23.396653 4791 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-67b789f86c-cq8ml" Dec 10 23:12:23 crc kubenswrapper[4791]: I1210 23:12:23.402218 4791 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-edpm-ipam" Dec 10 23:12:23 crc kubenswrapper[4791]: I1210 23:12:23.408237 4791 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-67b789f86c-cq8ml"] Dec 10 23:12:23 crc kubenswrapper[4791]: I1210 23:12:23.588487 4791 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/53c95cd8-94b4-45f8-bc29-14fe2bb8049c-config\") pod \"dnsmasq-dns-67b789f86c-cq8ml\" (UID: \"53c95cd8-94b4-45f8-bc29-14fe2bb8049c\") " pod="openstack/dnsmasq-dns-67b789f86c-cq8ml" Dec 10 23:12:23 crc kubenswrapper[4791]: I1210 23:12:23.588752 4791 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/53c95cd8-94b4-45f8-bc29-14fe2bb8049c-ovsdbserver-nb\") pod \"dnsmasq-dns-67b789f86c-cq8ml\" (UID: \"53c95cd8-94b4-45f8-bc29-14fe2bb8049c\") " pod="openstack/dnsmasq-dns-67b789f86c-cq8ml" Dec 10 23:12:23 crc kubenswrapper[4791]: I1210 23:12:23.588790 4791 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/configmap/53c95cd8-94b4-45f8-bc29-14fe2bb8049c-openstack-edpm-ipam\") pod \"dnsmasq-dns-67b789f86c-cq8ml\" (UID: \"53c95cd8-94b4-45f8-bc29-14fe2bb8049c\") " pod="openstack/dnsmasq-dns-67b789f86c-cq8ml" Dec 10 23:12:23 crc kubenswrapper[4791]: I1210 23:12:23.588827 4791 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/53c95cd8-94b4-45f8-bc29-14fe2bb8049c-ovsdbserver-sb\") pod \"dnsmasq-dns-67b789f86c-cq8ml\" (UID: \"53c95cd8-94b4-45f8-bc29-14fe2bb8049c\") " pod="openstack/dnsmasq-dns-67b789f86c-cq8ml" Dec 10 23:12:23 crc kubenswrapper[4791]: I1210 23:12:23.595704 4791 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-zzrcq\" (UniqueName: \"kubernetes.io/projected/53c95cd8-94b4-45f8-bc29-14fe2bb8049c-kube-api-access-zzrcq\") pod \"dnsmasq-dns-67b789f86c-cq8ml\" (UID: \"53c95cd8-94b4-45f8-bc29-14fe2bb8049c\") " pod="openstack/dnsmasq-dns-67b789f86c-cq8ml" Dec 10 23:12:23 crc kubenswrapper[4791]: I1210 23:12:23.595792 4791 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/53c95cd8-94b4-45f8-bc29-14fe2bb8049c-dns-svc\") pod \"dnsmasq-dns-67b789f86c-cq8ml\" (UID: \"53c95cd8-94b4-45f8-bc29-14fe2bb8049c\") " pod="openstack/dnsmasq-dns-67b789f86c-cq8ml" Dec 10 23:12:23 crc kubenswrapper[4791]: I1210 23:12:23.595875 4791 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/53c95cd8-94b4-45f8-bc29-14fe2bb8049c-dns-swift-storage-0\") pod \"dnsmasq-dns-67b789f86c-cq8ml\" (UID: \"53c95cd8-94b4-45f8-bc29-14fe2bb8049c\") " pod="openstack/dnsmasq-dns-67b789f86c-cq8ml" Dec 10 23:12:23 crc kubenswrapper[4791]: I1210 23:12:23.650432 4791 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-server-0" event={"ID":"5f51c37d-2f6f-4ef0-a08c-77216e7db57b","Type":"ContainerStarted","Data":"9e8ad6dc9d161840d5da36e4d8d428f5477c829008f793924016c3ecbd4804d2"} Dec 10 23:12:23 crc kubenswrapper[4791]: I1210 23:12:23.696730 4791 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/53c95cd8-94b4-45f8-bc29-14fe2bb8049c-config\") pod \"dnsmasq-dns-67b789f86c-cq8ml\" (UID: \"53c95cd8-94b4-45f8-bc29-14fe2bb8049c\") " pod="openstack/dnsmasq-dns-67b789f86c-cq8ml" Dec 10 23:12:23 crc kubenswrapper[4791]: I1210 23:12:23.696774 4791 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/53c95cd8-94b4-45f8-bc29-14fe2bb8049c-ovsdbserver-nb\") pod \"dnsmasq-dns-67b789f86c-cq8ml\" (UID: \"53c95cd8-94b4-45f8-bc29-14fe2bb8049c\") " pod="openstack/dnsmasq-dns-67b789f86c-cq8ml" Dec 10 23:12:23 crc kubenswrapper[4791]: I1210 23:12:23.696797 4791 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/configmap/53c95cd8-94b4-45f8-bc29-14fe2bb8049c-openstack-edpm-ipam\") pod \"dnsmasq-dns-67b789f86c-cq8ml\" (UID: \"53c95cd8-94b4-45f8-bc29-14fe2bb8049c\") " pod="openstack/dnsmasq-dns-67b789f86c-cq8ml" Dec 10 23:12:23 crc kubenswrapper[4791]: I1210 23:12:23.696816 4791 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/53c95cd8-94b4-45f8-bc29-14fe2bb8049c-ovsdbserver-sb\") pod \"dnsmasq-dns-67b789f86c-cq8ml\" (UID: \"53c95cd8-94b4-45f8-bc29-14fe2bb8049c\") " pod="openstack/dnsmasq-dns-67b789f86c-cq8ml" Dec 10 23:12:23 crc kubenswrapper[4791]: I1210 23:12:23.696940 4791 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-zzrcq\" (UniqueName: \"kubernetes.io/projected/53c95cd8-94b4-45f8-bc29-14fe2bb8049c-kube-api-access-zzrcq\") pod \"dnsmasq-dns-67b789f86c-cq8ml\" (UID: \"53c95cd8-94b4-45f8-bc29-14fe2bb8049c\") " pod="openstack/dnsmasq-dns-67b789f86c-cq8ml" Dec 10 23:12:23 crc kubenswrapper[4791]: I1210 23:12:23.696969 4791 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/53c95cd8-94b4-45f8-bc29-14fe2bb8049c-dns-svc\") pod \"dnsmasq-dns-67b789f86c-cq8ml\" (UID: \"53c95cd8-94b4-45f8-bc29-14fe2bb8049c\") " pod="openstack/dnsmasq-dns-67b789f86c-cq8ml" Dec 10 23:12:23 crc kubenswrapper[4791]: I1210 23:12:23.696988 4791 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/53c95cd8-94b4-45f8-bc29-14fe2bb8049c-dns-swift-storage-0\") pod \"dnsmasq-dns-67b789f86c-cq8ml\" (UID: \"53c95cd8-94b4-45f8-bc29-14fe2bb8049c\") " pod="openstack/dnsmasq-dns-67b789f86c-cq8ml" Dec 10 23:12:23 crc kubenswrapper[4791]: I1210 23:12:23.697886 4791 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/53c95cd8-94b4-45f8-bc29-14fe2bb8049c-dns-swift-storage-0\") pod \"dnsmasq-dns-67b789f86c-cq8ml\" (UID: \"53c95cd8-94b4-45f8-bc29-14fe2bb8049c\") " pod="openstack/dnsmasq-dns-67b789f86c-cq8ml" Dec 10 23:12:23 crc kubenswrapper[4791]: I1210 23:12:23.698409 4791 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/53c95cd8-94b4-45f8-bc29-14fe2bb8049c-config\") pod \"dnsmasq-dns-67b789f86c-cq8ml\" (UID: \"53c95cd8-94b4-45f8-bc29-14fe2bb8049c\") " pod="openstack/dnsmasq-dns-67b789f86c-cq8ml" Dec 10 23:12:23 crc kubenswrapper[4791]: I1210 23:12:23.698954 4791 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/53c95cd8-94b4-45f8-bc29-14fe2bb8049c-ovsdbserver-nb\") pod \"dnsmasq-dns-67b789f86c-cq8ml\" (UID: \"53c95cd8-94b4-45f8-bc29-14fe2bb8049c\") " pod="openstack/dnsmasq-dns-67b789f86c-cq8ml" Dec 10 23:12:23 crc kubenswrapper[4791]: I1210 23:12:23.699480 4791 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/53c95cd8-94b4-45f8-bc29-14fe2bb8049c-ovsdbserver-sb\") pod \"dnsmasq-dns-67b789f86c-cq8ml\" (UID: \"53c95cd8-94b4-45f8-bc29-14fe2bb8049c\") " pod="openstack/dnsmasq-dns-67b789f86c-cq8ml" Dec 10 23:12:23 crc kubenswrapper[4791]: I1210 23:12:23.699632 4791 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/53c95cd8-94b4-45f8-bc29-14fe2bb8049c-dns-svc\") pod \"dnsmasq-dns-67b789f86c-cq8ml\" (UID: \"53c95cd8-94b4-45f8-bc29-14fe2bb8049c\") " pod="openstack/dnsmasq-dns-67b789f86c-cq8ml" Dec 10 23:12:23 crc kubenswrapper[4791]: I1210 23:12:23.699717 4791 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/configmap/53c95cd8-94b4-45f8-bc29-14fe2bb8049c-openstack-edpm-ipam\") pod \"dnsmasq-dns-67b789f86c-cq8ml\" (UID: \"53c95cd8-94b4-45f8-bc29-14fe2bb8049c\") " pod="openstack/dnsmasq-dns-67b789f86c-cq8ml" Dec 10 23:12:23 crc kubenswrapper[4791]: I1210 23:12:23.716817 4791 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-zzrcq\" (UniqueName: \"kubernetes.io/projected/53c95cd8-94b4-45f8-bc29-14fe2bb8049c-kube-api-access-zzrcq\") pod \"dnsmasq-dns-67b789f86c-cq8ml\" (UID: \"53c95cd8-94b4-45f8-bc29-14fe2bb8049c\") " pod="openstack/dnsmasq-dns-67b789f86c-cq8ml" Dec 10 23:12:23 crc kubenswrapper[4791]: I1210 23:12:23.723610 4791 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-67b789f86c-cq8ml" Dec 10 23:12:23 crc kubenswrapper[4791]: I1210 23:12:23.856960 4791 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/rabbitmq-cell1-server-0"] Dec 10 23:12:24 crc kubenswrapper[4791]: I1210 23:12:24.054011 4791 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="47eb9200-b0bc-41a6-abb9-f6167bd69c66" path="/var/lib/kubelet/pods/47eb9200-b0bc-41a6-abb9-f6167bd69c66/volumes" Dec 10 23:12:24 crc kubenswrapper[4791]: I1210 23:12:24.367922 4791 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-67b789f86c-cq8ml"] Dec 10 23:12:24 crc kubenswrapper[4791]: W1210 23:12:24.370458 4791 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod53c95cd8_94b4_45f8_bc29_14fe2bb8049c.slice/crio-614d6b0cc71eee92732118a925d187558411c90f10f774b92d27d182e3485097 WatchSource:0}: Error finding container 614d6b0cc71eee92732118a925d187558411c90f10f774b92d27d182e3485097: Status 404 returned error can't find the container with id 614d6b0cc71eee92732118a925d187558411c90f10f774b92d27d182e3485097 Dec 10 23:12:24 crc kubenswrapper[4791]: I1210 23:12:24.660384 4791 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-server-0" event={"ID":"5f51c37d-2f6f-4ef0-a08c-77216e7db57b","Type":"ContainerStarted","Data":"9ed929ea36f1a6f4cec971ac7e7d5a6838a4ce09161cfc8c65d2c50871abcbe2"} Dec 10 23:12:24 crc kubenswrapper[4791]: I1210 23:12:24.662575 4791 generic.go:334] "Generic (PLEG): container finished" podID="53c95cd8-94b4-45f8-bc29-14fe2bb8049c" containerID="abe2a817c9798b3353a352ad8918182e06184e89343fa9ef0850dbaaf6700678" exitCode=0 Dec 10 23:12:24 crc kubenswrapper[4791]: I1210 23:12:24.662644 4791 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-67b789f86c-cq8ml" event={"ID":"53c95cd8-94b4-45f8-bc29-14fe2bb8049c","Type":"ContainerDied","Data":"abe2a817c9798b3353a352ad8918182e06184e89343fa9ef0850dbaaf6700678"} Dec 10 23:12:24 crc kubenswrapper[4791]: I1210 23:12:24.662670 4791 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-67b789f86c-cq8ml" event={"ID":"53c95cd8-94b4-45f8-bc29-14fe2bb8049c","Type":"ContainerStarted","Data":"614d6b0cc71eee92732118a925d187558411c90f10f774b92d27d182e3485097"} Dec 10 23:12:24 crc kubenswrapper[4791]: I1210 23:12:24.664189 4791 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-cell1-server-0" event={"ID":"3add67e0-b67a-438b-b83f-4c9f6733b5cb","Type":"ContainerStarted","Data":"9688518b1063fee5c983181e2f0b2a0a0934a0bf175c472398f5582a6503dea5"} Dec 10 23:12:25 crc kubenswrapper[4791]: I1210 23:12:25.037709 4791 patch_prober.go:28] interesting pod/machine-config-daemon-5rb5l container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 10 23:12:25 crc kubenswrapper[4791]: I1210 23:12:25.037771 4791 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-5rb5l" podUID="ba35653c-6e06-4cee-a4d6-137764090d18" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 10 23:12:25 crc kubenswrapper[4791]: I1210 23:12:25.037824 4791 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-5rb5l" Dec 10 23:12:25 crc kubenswrapper[4791]: I1210 23:12:25.038745 4791 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"27ea6396cad6699de1e6134e18b2c18010b7367f4d673a0bdef71d4d1c354c04"} pod="openshift-machine-config-operator/machine-config-daemon-5rb5l" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Dec 10 23:12:25 crc kubenswrapper[4791]: I1210 23:12:25.038815 4791 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-5rb5l" podUID="ba35653c-6e06-4cee-a4d6-137764090d18" containerName="machine-config-daemon" containerID="cri-o://27ea6396cad6699de1e6134e18b2c18010b7367f4d673a0bdef71d4d1c354c04" gracePeriod=600 Dec 10 23:12:25 crc kubenswrapper[4791]: I1210 23:12:25.673505 4791 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-67b789f86c-cq8ml" event={"ID":"53c95cd8-94b4-45f8-bc29-14fe2bb8049c","Type":"ContainerStarted","Data":"a3dde81f9bff2a0f8bc2c74a006bc6eb02cd7703caae876aeedf170f58d474a6"} Dec 10 23:12:25 crc kubenswrapper[4791]: I1210 23:12:25.674021 4791 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/dnsmasq-dns-67b789f86c-cq8ml" Dec 10 23:12:25 crc kubenswrapper[4791]: I1210 23:12:25.675716 4791 generic.go:334] "Generic (PLEG): container finished" podID="ba35653c-6e06-4cee-a4d6-137764090d18" containerID="27ea6396cad6699de1e6134e18b2c18010b7367f4d673a0bdef71d4d1c354c04" exitCode=0 Dec 10 23:12:25 crc kubenswrapper[4791]: I1210 23:12:25.676298 4791 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-5rb5l" event={"ID":"ba35653c-6e06-4cee-a4d6-137764090d18","Type":"ContainerDied","Data":"27ea6396cad6699de1e6134e18b2c18010b7367f4d673a0bdef71d4d1c354c04"} Dec 10 23:12:25 crc kubenswrapper[4791]: I1210 23:12:25.676331 4791 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-5rb5l" event={"ID":"ba35653c-6e06-4cee-a4d6-137764090d18","Type":"ContainerStarted","Data":"f12779a795c9f6f63a93027d53c67c89d0179d87b9df554c25bfc607f06caff1"} Dec 10 23:12:25 crc kubenswrapper[4791]: I1210 23:12:25.676363 4791 scope.go:117] "RemoveContainer" containerID="2b6bbc8a626f5244441ef712b0db42022f2a15236c1846adb0b1f8c097905a73" Dec 10 23:12:25 crc kubenswrapper[4791]: I1210 23:12:25.695882 4791 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/dnsmasq-dns-67b789f86c-cq8ml" podStartSLOduration=2.695863166 podStartE2EDuration="2.695863166s" podCreationTimestamp="2025-12-10 23:12:23 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-10 23:12:25.691964124 +0000 UTC m=+1380.121581747" watchObservedRunningTime="2025-12-10 23:12:25.695863166 +0000 UTC m=+1380.125480779" Dec 10 23:12:26 crc kubenswrapper[4791]: I1210 23:12:26.688975 4791 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-cell1-server-0" event={"ID":"3add67e0-b67a-438b-b83f-4c9f6733b5cb","Type":"ContainerStarted","Data":"2bc6b221ebeead7a88db3ab622e885d16946fac93e077d0559fe60a6208fcaf6"} Dec 10 23:12:33 crc kubenswrapper[4791]: I1210 23:12:33.725596 4791 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/dnsmasq-dns-67b789f86c-cq8ml" Dec 10 23:12:33 crc kubenswrapper[4791]: I1210 23:12:33.836069 4791 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-59cf4bdb65-p4l85"] Dec 10 23:12:33 crc kubenswrapper[4791]: I1210 23:12:33.836406 4791 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/dnsmasq-dns-59cf4bdb65-p4l85" podUID="339ea8c9-715f-4394-ace8-6b66542bcc13" containerName="dnsmasq-dns" containerID="cri-o://da15fdb47d176237ac9d5f6da99fb4dbbd6dd2db577a479864263339ab238add" gracePeriod=10 Dec 10 23:12:34 crc kubenswrapper[4791]: I1210 23:12:34.016962 4791 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-cb6ffcf87-wcxkk"] Dec 10 23:12:34 crc kubenswrapper[4791]: I1210 23:12:34.018412 4791 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-cb6ffcf87-wcxkk" Dec 10 23:12:34 crc kubenswrapper[4791]: I1210 23:12:34.053467 4791 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-cb6ffcf87-wcxkk"] Dec 10 23:12:34 crc kubenswrapper[4791]: I1210 23:12:34.140604 4791 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/07473608-7cf2-4bcb-b7bb-046165c73afe-ovsdbserver-sb\") pod \"dnsmasq-dns-cb6ffcf87-wcxkk\" (UID: \"07473608-7cf2-4bcb-b7bb-046165c73afe\") " pod="openstack/dnsmasq-dns-cb6ffcf87-wcxkk" Dec 10 23:12:34 crc kubenswrapper[4791]: I1210 23:12:34.140700 4791 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/07473608-7cf2-4bcb-b7bb-046165c73afe-ovsdbserver-nb\") pod \"dnsmasq-dns-cb6ffcf87-wcxkk\" (UID: \"07473608-7cf2-4bcb-b7bb-046165c73afe\") " pod="openstack/dnsmasq-dns-cb6ffcf87-wcxkk" Dec 10 23:12:34 crc kubenswrapper[4791]: I1210 23:12:34.140742 4791 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/07473608-7cf2-4bcb-b7bb-046165c73afe-dns-swift-storage-0\") pod \"dnsmasq-dns-cb6ffcf87-wcxkk\" (UID: \"07473608-7cf2-4bcb-b7bb-046165c73afe\") " pod="openstack/dnsmasq-dns-cb6ffcf87-wcxkk" Dec 10 23:12:34 crc kubenswrapper[4791]: I1210 23:12:34.140759 4791 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/07473608-7cf2-4bcb-b7bb-046165c73afe-dns-svc\") pod \"dnsmasq-dns-cb6ffcf87-wcxkk\" (UID: \"07473608-7cf2-4bcb-b7bb-046165c73afe\") " pod="openstack/dnsmasq-dns-cb6ffcf87-wcxkk" Dec 10 23:12:34 crc kubenswrapper[4791]: I1210 23:12:34.140787 4791 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-ghhl2\" (UniqueName: \"kubernetes.io/projected/07473608-7cf2-4bcb-b7bb-046165c73afe-kube-api-access-ghhl2\") pod \"dnsmasq-dns-cb6ffcf87-wcxkk\" (UID: \"07473608-7cf2-4bcb-b7bb-046165c73afe\") " pod="openstack/dnsmasq-dns-cb6ffcf87-wcxkk" Dec 10 23:12:34 crc kubenswrapper[4791]: I1210 23:12:34.140838 4791 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/configmap/07473608-7cf2-4bcb-b7bb-046165c73afe-openstack-edpm-ipam\") pod \"dnsmasq-dns-cb6ffcf87-wcxkk\" (UID: \"07473608-7cf2-4bcb-b7bb-046165c73afe\") " pod="openstack/dnsmasq-dns-cb6ffcf87-wcxkk" Dec 10 23:12:34 crc kubenswrapper[4791]: I1210 23:12:34.140856 4791 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/07473608-7cf2-4bcb-b7bb-046165c73afe-config\") pod \"dnsmasq-dns-cb6ffcf87-wcxkk\" (UID: \"07473608-7cf2-4bcb-b7bb-046165c73afe\") " pod="openstack/dnsmasq-dns-cb6ffcf87-wcxkk" Dec 10 23:12:34 crc kubenswrapper[4791]: I1210 23:12:34.242994 4791 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/07473608-7cf2-4bcb-b7bb-046165c73afe-ovsdbserver-sb\") pod \"dnsmasq-dns-cb6ffcf87-wcxkk\" (UID: \"07473608-7cf2-4bcb-b7bb-046165c73afe\") " pod="openstack/dnsmasq-dns-cb6ffcf87-wcxkk" Dec 10 23:12:34 crc kubenswrapper[4791]: I1210 23:12:34.243110 4791 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/07473608-7cf2-4bcb-b7bb-046165c73afe-ovsdbserver-nb\") pod \"dnsmasq-dns-cb6ffcf87-wcxkk\" (UID: \"07473608-7cf2-4bcb-b7bb-046165c73afe\") " pod="openstack/dnsmasq-dns-cb6ffcf87-wcxkk" Dec 10 23:12:34 crc kubenswrapper[4791]: I1210 23:12:34.243155 4791 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/07473608-7cf2-4bcb-b7bb-046165c73afe-dns-swift-storage-0\") pod \"dnsmasq-dns-cb6ffcf87-wcxkk\" (UID: \"07473608-7cf2-4bcb-b7bb-046165c73afe\") " pod="openstack/dnsmasq-dns-cb6ffcf87-wcxkk" Dec 10 23:12:34 crc kubenswrapper[4791]: I1210 23:12:34.243228 4791 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/07473608-7cf2-4bcb-b7bb-046165c73afe-dns-svc\") pod \"dnsmasq-dns-cb6ffcf87-wcxkk\" (UID: \"07473608-7cf2-4bcb-b7bb-046165c73afe\") " pod="openstack/dnsmasq-dns-cb6ffcf87-wcxkk" Dec 10 23:12:34 crc kubenswrapper[4791]: I1210 23:12:34.243289 4791 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-ghhl2\" (UniqueName: \"kubernetes.io/projected/07473608-7cf2-4bcb-b7bb-046165c73afe-kube-api-access-ghhl2\") pod \"dnsmasq-dns-cb6ffcf87-wcxkk\" (UID: \"07473608-7cf2-4bcb-b7bb-046165c73afe\") " pod="openstack/dnsmasq-dns-cb6ffcf87-wcxkk" Dec 10 23:12:34 crc kubenswrapper[4791]: I1210 23:12:34.243382 4791 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/configmap/07473608-7cf2-4bcb-b7bb-046165c73afe-openstack-edpm-ipam\") pod \"dnsmasq-dns-cb6ffcf87-wcxkk\" (UID: \"07473608-7cf2-4bcb-b7bb-046165c73afe\") " pod="openstack/dnsmasq-dns-cb6ffcf87-wcxkk" Dec 10 23:12:34 crc kubenswrapper[4791]: I1210 23:12:34.243422 4791 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/07473608-7cf2-4bcb-b7bb-046165c73afe-config\") pod \"dnsmasq-dns-cb6ffcf87-wcxkk\" (UID: \"07473608-7cf2-4bcb-b7bb-046165c73afe\") " pod="openstack/dnsmasq-dns-cb6ffcf87-wcxkk" Dec 10 23:12:34 crc kubenswrapper[4791]: I1210 23:12:34.244362 4791 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/07473608-7cf2-4bcb-b7bb-046165c73afe-config\") pod \"dnsmasq-dns-cb6ffcf87-wcxkk\" (UID: \"07473608-7cf2-4bcb-b7bb-046165c73afe\") " pod="openstack/dnsmasq-dns-cb6ffcf87-wcxkk" Dec 10 23:12:34 crc kubenswrapper[4791]: I1210 23:12:34.245134 4791 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/07473608-7cf2-4bcb-b7bb-046165c73afe-dns-swift-storage-0\") pod \"dnsmasq-dns-cb6ffcf87-wcxkk\" (UID: \"07473608-7cf2-4bcb-b7bb-046165c73afe\") " pod="openstack/dnsmasq-dns-cb6ffcf87-wcxkk" Dec 10 23:12:34 crc kubenswrapper[4791]: I1210 23:12:34.258127 4791 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/07473608-7cf2-4bcb-b7bb-046165c73afe-ovsdbserver-nb\") pod \"dnsmasq-dns-cb6ffcf87-wcxkk\" (UID: \"07473608-7cf2-4bcb-b7bb-046165c73afe\") " pod="openstack/dnsmasq-dns-cb6ffcf87-wcxkk" Dec 10 23:12:34 crc kubenswrapper[4791]: I1210 23:12:34.259140 4791 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/07473608-7cf2-4bcb-b7bb-046165c73afe-dns-svc\") pod \"dnsmasq-dns-cb6ffcf87-wcxkk\" (UID: \"07473608-7cf2-4bcb-b7bb-046165c73afe\") " pod="openstack/dnsmasq-dns-cb6ffcf87-wcxkk" Dec 10 23:12:34 crc kubenswrapper[4791]: I1210 23:12:34.261088 4791 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/07473608-7cf2-4bcb-b7bb-046165c73afe-ovsdbserver-sb\") pod \"dnsmasq-dns-cb6ffcf87-wcxkk\" (UID: \"07473608-7cf2-4bcb-b7bb-046165c73afe\") " pod="openstack/dnsmasq-dns-cb6ffcf87-wcxkk" Dec 10 23:12:34 crc kubenswrapper[4791]: I1210 23:12:34.261759 4791 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/configmap/07473608-7cf2-4bcb-b7bb-046165c73afe-openstack-edpm-ipam\") pod \"dnsmasq-dns-cb6ffcf87-wcxkk\" (UID: \"07473608-7cf2-4bcb-b7bb-046165c73afe\") " pod="openstack/dnsmasq-dns-cb6ffcf87-wcxkk" Dec 10 23:12:34 crc kubenswrapper[4791]: I1210 23:12:34.273603 4791 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-ghhl2\" (UniqueName: \"kubernetes.io/projected/07473608-7cf2-4bcb-b7bb-046165c73afe-kube-api-access-ghhl2\") pod \"dnsmasq-dns-cb6ffcf87-wcxkk\" (UID: \"07473608-7cf2-4bcb-b7bb-046165c73afe\") " pod="openstack/dnsmasq-dns-cb6ffcf87-wcxkk" Dec 10 23:12:34 crc kubenswrapper[4791]: I1210 23:12:34.347437 4791 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-cb6ffcf87-wcxkk" Dec 10 23:12:34 crc kubenswrapper[4791]: I1210 23:12:34.351485 4791 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-59cf4bdb65-p4l85" Dec 10 23:12:34 crc kubenswrapper[4791]: I1210 23:12:34.446008 4791 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/339ea8c9-715f-4394-ace8-6b66542bcc13-dns-svc\") pod \"339ea8c9-715f-4394-ace8-6b66542bcc13\" (UID: \"339ea8c9-715f-4394-ace8-6b66542bcc13\") " Dec 10 23:12:34 crc kubenswrapper[4791]: I1210 23:12:34.446117 4791 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/339ea8c9-715f-4394-ace8-6b66542bcc13-dns-swift-storage-0\") pod \"339ea8c9-715f-4394-ace8-6b66542bcc13\" (UID: \"339ea8c9-715f-4394-ace8-6b66542bcc13\") " Dec 10 23:12:34 crc kubenswrapper[4791]: I1210 23:12:34.446234 4791 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/339ea8c9-715f-4394-ace8-6b66542bcc13-ovsdbserver-nb\") pod \"339ea8c9-715f-4394-ace8-6b66542bcc13\" (UID: \"339ea8c9-715f-4394-ace8-6b66542bcc13\") " Dec 10 23:12:34 crc kubenswrapper[4791]: I1210 23:12:34.446299 4791 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/339ea8c9-715f-4394-ace8-6b66542bcc13-config\") pod \"339ea8c9-715f-4394-ace8-6b66542bcc13\" (UID: \"339ea8c9-715f-4394-ace8-6b66542bcc13\") " Dec 10 23:12:34 crc kubenswrapper[4791]: I1210 23:12:34.446322 4791 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-q6229\" (UniqueName: \"kubernetes.io/projected/339ea8c9-715f-4394-ace8-6b66542bcc13-kube-api-access-q6229\") pod \"339ea8c9-715f-4394-ace8-6b66542bcc13\" (UID: \"339ea8c9-715f-4394-ace8-6b66542bcc13\") " Dec 10 23:12:34 crc kubenswrapper[4791]: I1210 23:12:34.446369 4791 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/339ea8c9-715f-4394-ace8-6b66542bcc13-ovsdbserver-sb\") pod \"339ea8c9-715f-4394-ace8-6b66542bcc13\" (UID: \"339ea8c9-715f-4394-ace8-6b66542bcc13\") " Dec 10 23:12:34 crc kubenswrapper[4791]: I1210 23:12:34.466713 4791 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/339ea8c9-715f-4394-ace8-6b66542bcc13-kube-api-access-q6229" (OuterVolumeSpecName: "kube-api-access-q6229") pod "339ea8c9-715f-4394-ace8-6b66542bcc13" (UID: "339ea8c9-715f-4394-ace8-6b66542bcc13"). InnerVolumeSpecName "kube-api-access-q6229". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 10 23:12:34 crc kubenswrapper[4791]: I1210 23:12:34.508168 4791 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/339ea8c9-715f-4394-ace8-6b66542bcc13-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "339ea8c9-715f-4394-ace8-6b66542bcc13" (UID: "339ea8c9-715f-4394-ace8-6b66542bcc13"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 10 23:12:34 crc kubenswrapper[4791]: I1210 23:12:34.526884 4791 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/339ea8c9-715f-4394-ace8-6b66542bcc13-dns-swift-storage-0" (OuterVolumeSpecName: "dns-swift-storage-0") pod "339ea8c9-715f-4394-ace8-6b66542bcc13" (UID: "339ea8c9-715f-4394-ace8-6b66542bcc13"). InnerVolumeSpecName "dns-swift-storage-0". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 10 23:12:34 crc kubenswrapper[4791]: I1210 23:12:34.539023 4791 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/339ea8c9-715f-4394-ace8-6b66542bcc13-ovsdbserver-sb" (OuterVolumeSpecName: "ovsdbserver-sb") pod "339ea8c9-715f-4394-ace8-6b66542bcc13" (UID: "339ea8c9-715f-4394-ace8-6b66542bcc13"). InnerVolumeSpecName "ovsdbserver-sb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 10 23:12:34 crc kubenswrapper[4791]: I1210 23:12:34.550119 4791 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/339ea8c9-715f-4394-ace8-6b66542bcc13-ovsdbserver-nb" (OuterVolumeSpecName: "ovsdbserver-nb") pod "339ea8c9-715f-4394-ace8-6b66542bcc13" (UID: "339ea8c9-715f-4394-ace8-6b66542bcc13"). InnerVolumeSpecName "ovsdbserver-nb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 10 23:12:34 crc kubenswrapper[4791]: I1210 23:12:34.550396 4791 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/339ea8c9-715f-4394-ace8-6b66542bcc13-ovsdbserver-nb\") on node \"crc\" DevicePath \"\"" Dec 10 23:12:34 crc kubenswrapper[4791]: I1210 23:12:34.550435 4791 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-q6229\" (UniqueName: \"kubernetes.io/projected/339ea8c9-715f-4394-ace8-6b66542bcc13-kube-api-access-q6229\") on node \"crc\" DevicePath \"\"" Dec 10 23:12:34 crc kubenswrapper[4791]: I1210 23:12:34.550452 4791 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/339ea8c9-715f-4394-ace8-6b66542bcc13-ovsdbserver-sb\") on node \"crc\" DevicePath \"\"" Dec 10 23:12:34 crc kubenswrapper[4791]: I1210 23:12:34.550463 4791 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/339ea8c9-715f-4394-ace8-6b66542bcc13-dns-svc\") on node \"crc\" DevicePath \"\"" Dec 10 23:12:34 crc kubenswrapper[4791]: I1210 23:12:34.550474 4791 reconciler_common.go:293] "Volume detached for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/339ea8c9-715f-4394-ace8-6b66542bcc13-dns-swift-storage-0\") on node \"crc\" DevicePath \"\"" Dec 10 23:12:34 crc kubenswrapper[4791]: I1210 23:12:34.551845 4791 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/339ea8c9-715f-4394-ace8-6b66542bcc13-config" (OuterVolumeSpecName: "config") pod "339ea8c9-715f-4394-ace8-6b66542bcc13" (UID: "339ea8c9-715f-4394-ace8-6b66542bcc13"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 10 23:12:34 crc kubenswrapper[4791]: I1210 23:12:34.652534 4791 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/339ea8c9-715f-4394-ace8-6b66542bcc13-config\") on node \"crc\" DevicePath \"\"" Dec 10 23:12:34 crc kubenswrapper[4791]: I1210 23:12:34.794683 4791 generic.go:334] "Generic (PLEG): container finished" podID="339ea8c9-715f-4394-ace8-6b66542bcc13" containerID="da15fdb47d176237ac9d5f6da99fb4dbbd6dd2db577a479864263339ab238add" exitCode=0 Dec 10 23:12:34 crc kubenswrapper[4791]: I1210 23:12:34.794740 4791 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-59cf4bdb65-p4l85" event={"ID":"339ea8c9-715f-4394-ace8-6b66542bcc13","Type":"ContainerDied","Data":"da15fdb47d176237ac9d5f6da99fb4dbbd6dd2db577a479864263339ab238add"} Dec 10 23:12:34 crc kubenswrapper[4791]: I1210 23:12:34.794797 4791 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-59cf4bdb65-p4l85" Dec 10 23:12:34 crc kubenswrapper[4791]: I1210 23:12:34.794895 4791 scope.go:117] "RemoveContainer" containerID="da15fdb47d176237ac9d5f6da99fb4dbbd6dd2db577a479864263339ab238add" Dec 10 23:12:34 crc kubenswrapper[4791]: I1210 23:12:34.795009 4791 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-59cf4bdb65-p4l85" event={"ID":"339ea8c9-715f-4394-ace8-6b66542bcc13","Type":"ContainerDied","Data":"b6441e8d9725e479158a59e0b569853fe71039394deb10eae1e4118ea59fe783"} Dec 10 23:12:34 crc kubenswrapper[4791]: I1210 23:12:34.828842 4791 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-cb6ffcf87-wcxkk"] Dec 10 23:12:34 crc kubenswrapper[4791]: I1210 23:12:34.844524 4791 scope.go:117] "RemoveContainer" containerID="738564d5be05ed3564a50d6fb351796a769397fc4e53c48a772b76e7fde67bf7" Dec 10 23:12:34 crc kubenswrapper[4791]: I1210 23:12:34.857387 4791 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-59cf4bdb65-p4l85"] Dec 10 23:12:34 crc kubenswrapper[4791]: I1210 23:12:34.866308 4791 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-59cf4bdb65-p4l85"] Dec 10 23:12:34 crc kubenswrapper[4791]: I1210 23:12:34.869838 4791 scope.go:117] "RemoveContainer" containerID="da15fdb47d176237ac9d5f6da99fb4dbbd6dd2db577a479864263339ab238add" Dec 10 23:12:34 crc kubenswrapper[4791]: E1210 23:12:34.870442 4791 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"da15fdb47d176237ac9d5f6da99fb4dbbd6dd2db577a479864263339ab238add\": container with ID starting with da15fdb47d176237ac9d5f6da99fb4dbbd6dd2db577a479864263339ab238add not found: ID does not exist" containerID="da15fdb47d176237ac9d5f6da99fb4dbbd6dd2db577a479864263339ab238add" Dec 10 23:12:34 crc kubenswrapper[4791]: I1210 23:12:34.870481 4791 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"da15fdb47d176237ac9d5f6da99fb4dbbd6dd2db577a479864263339ab238add"} err="failed to get container status \"da15fdb47d176237ac9d5f6da99fb4dbbd6dd2db577a479864263339ab238add\": rpc error: code = NotFound desc = could not find container \"da15fdb47d176237ac9d5f6da99fb4dbbd6dd2db577a479864263339ab238add\": container with ID starting with da15fdb47d176237ac9d5f6da99fb4dbbd6dd2db577a479864263339ab238add not found: ID does not exist" Dec 10 23:12:34 crc kubenswrapper[4791]: I1210 23:12:34.870537 4791 scope.go:117] "RemoveContainer" containerID="738564d5be05ed3564a50d6fb351796a769397fc4e53c48a772b76e7fde67bf7" Dec 10 23:12:34 crc kubenswrapper[4791]: E1210 23:12:34.870852 4791 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"738564d5be05ed3564a50d6fb351796a769397fc4e53c48a772b76e7fde67bf7\": container with ID starting with 738564d5be05ed3564a50d6fb351796a769397fc4e53c48a772b76e7fde67bf7 not found: ID does not exist" containerID="738564d5be05ed3564a50d6fb351796a769397fc4e53c48a772b76e7fde67bf7" Dec 10 23:12:34 crc kubenswrapper[4791]: I1210 23:12:34.870911 4791 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"738564d5be05ed3564a50d6fb351796a769397fc4e53c48a772b76e7fde67bf7"} err="failed to get container status \"738564d5be05ed3564a50d6fb351796a769397fc4e53c48a772b76e7fde67bf7\": rpc error: code = NotFound desc = could not find container \"738564d5be05ed3564a50d6fb351796a769397fc4e53c48a772b76e7fde67bf7\": container with ID starting with 738564d5be05ed3564a50d6fb351796a769397fc4e53c48a772b76e7fde67bf7 not found: ID does not exist" Dec 10 23:12:35 crc kubenswrapper[4791]: E1210 23:12:35.348420 4791 cadvisor_stats_provider.go:516] "Partial failure issuing cadvisor.ContainerInfoV2" err="partial failures: [\"/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod07473608_7cf2_4bcb_b7bb_046165c73afe.slice/crio-conmon-1a683d0c0eca414fba4599aec35b0bcd4e788ea5875c1fb26c8dfffa712bcaf1.scope\": RecentStats: unable to find data in memory cache]" Dec 10 23:12:35 crc kubenswrapper[4791]: I1210 23:12:35.809646 4791 generic.go:334] "Generic (PLEG): container finished" podID="07473608-7cf2-4bcb-b7bb-046165c73afe" containerID="1a683d0c0eca414fba4599aec35b0bcd4e788ea5875c1fb26c8dfffa712bcaf1" exitCode=0 Dec 10 23:12:35 crc kubenswrapper[4791]: I1210 23:12:35.809754 4791 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-cb6ffcf87-wcxkk" event={"ID":"07473608-7cf2-4bcb-b7bb-046165c73afe","Type":"ContainerDied","Data":"1a683d0c0eca414fba4599aec35b0bcd4e788ea5875c1fb26c8dfffa712bcaf1"} Dec 10 23:12:35 crc kubenswrapper[4791]: I1210 23:12:35.810014 4791 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-cb6ffcf87-wcxkk" event={"ID":"07473608-7cf2-4bcb-b7bb-046165c73afe","Type":"ContainerStarted","Data":"0f4514d9dcab50bb86d8b66d730d0027887fe650f088f0929de3de9d7130630c"} Dec 10 23:12:35 crc kubenswrapper[4791]: I1210 23:12:35.896328 4791 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="339ea8c9-715f-4394-ace8-6b66542bcc13" path="/var/lib/kubelet/pods/339ea8c9-715f-4394-ace8-6b66542bcc13/volumes" Dec 10 23:12:36 crc kubenswrapper[4791]: I1210 23:12:36.823469 4791 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-cb6ffcf87-wcxkk" event={"ID":"07473608-7cf2-4bcb-b7bb-046165c73afe","Type":"ContainerStarted","Data":"b801208c711fc4d475212f5f6e9064ba0ca6e7a0e00b49932563c953fb6753f5"} Dec 10 23:12:36 crc kubenswrapper[4791]: I1210 23:12:36.824044 4791 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/dnsmasq-dns-cb6ffcf87-wcxkk" Dec 10 23:12:36 crc kubenswrapper[4791]: I1210 23:12:36.848687 4791 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/dnsmasq-dns-cb6ffcf87-wcxkk" podStartSLOduration=3.84866684 podStartE2EDuration="3.84866684s" podCreationTimestamp="2025-12-10 23:12:33 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-10 23:12:36.840167535 +0000 UTC m=+1391.269785148" watchObservedRunningTime="2025-12-10 23:12:36.84866684 +0000 UTC m=+1391.278284453" Dec 10 23:12:44 crc kubenswrapper[4791]: I1210 23:12:44.349474 4791 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/dnsmasq-dns-cb6ffcf87-wcxkk" Dec 10 23:12:44 crc kubenswrapper[4791]: I1210 23:12:44.437363 4791 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-67b789f86c-cq8ml"] Dec 10 23:12:44 crc kubenswrapper[4791]: I1210 23:12:44.437620 4791 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/dnsmasq-dns-67b789f86c-cq8ml" podUID="53c95cd8-94b4-45f8-bc29-14fe2bb8049c" containerName="dnsmasq-dns" containerID="cri-o://a3dde81f9bff2a0f8bc2c74a006bc6eb02cd7703caae876aeedf170f58d474a6" gracePeriod=10 Dec 10 23:12:44 crc kubenswrapper[4791]: I1210 23:12:44.930088 4791 generic.go:334] "Generic (PLEG): container finished" podID="53c95cd8-94b4-45f8-bc29-14fe2bb8049c" containerID="a3dde81f9bff2a0f8bc2c74a006bc6eb02cd7703caae876aeedf170f58d474a6" exitCode=0 Dec 10 23:12:44 crc kubenswrapper[4791]: I1210 23:12:44.930184 4791 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-67b789f86c-cq8ml" event={"ID":"53c95cd8-94b4-45f8-bc29-14fe2bb8049c","Type":"ContainerDied","Data":"a3dde81f9bff2a0f8bc2c74a006bc6eb02cd7703caae876aeedf170f58d474a6"} Dec 10 23:12:44 crc kubenswrapper[4791]: I1210 23:12:44.930755 4791 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-67b789f86c-cq8ml" event={"ID":"53c95cd8-94b4-45f8-bc29-14fe2bb8049c","Type":"ContainerDied","Data":"614d6b0cc71eee92732118a925d187558411c90f10f774b92d27d182e3485097"} Dec 10 23:12:44 crc kubenswrapper[4791]: I1210 23:12:44.930773 4791 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="614d6b0cc71eee92732118a925d187558411c90f10f774b92d27d182e3485097" Dec 10 23:12:44 crc kubenswrapper[4791]: I1210 23:12:44.941687 4791 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-67b789f86c-cq8ml" Dec 10 23:12:45 crc kubenswrapper[4791]: I1210 23:12:45.082781 4791 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/configmap/53c95cd8-94b4-45f8-bc29-14fe2bb8049c-openstack-edpm-ipam\") pod \"53c95cd8-94b4-45f8-bc29-14fe2bb8049c\" (UID: \"53c95cd8-94b4-45f8-bc29-14fe2bb8049c\") " Dec 10 23:12:45 crc kubenswrapper[4791]: I1210 23:12:45.082917 4791 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-zzrcq\" (UniqueName: \"kubernetes.io/projected/53c95cd8-94b4-45f8-bc29-14fe2bb8049c-kube-api-access-zzrcq\") pod \"53c95cd8-94b4-45f8-bc29-14fe2bb8049c\" (UID: \"53c95cd8-94b4-45f8-bc29-14fe2bb8049c\") " Dec 10 23:12:45 crc kubenswrapper[4791]: I1210 23:12:45.082955 4791 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/53c95cd8-94b4-45f8-bc29-14fe2bb8049c-dns-svc\") pod \"53c95cd8-94b4-45f8-bc29-14fe2bb8049c\" (UID: \"53c95cd8-94b4-45f8-bc29-14fe2bb8049c\") " Dec 10 23:12:45 crc kubenswrapper[4791]: I1210 23:12:45.082992 4791 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/53c95cd8-94b4-45f8-bc29-14fe2bb8049c-dns-swift-storage-0\") pod \"53c95cd8-94b4-45f8-bc29-14fe2bb8049c\" (UID: \"53c95cd8-94b4-45f8-bc29-14fe2bb8049c\") " Dec 10 23:12:45 crc kubenswrapper[4791]: I1210 23:12:45.083039 4791 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/53c95cd8-94b4-45f8-bc29-14fe2bb8049c-ovsdbserver-nb\") pod \"53c95cd8-94b4-45f8-bc29-14fe2bb8049c\" (UID: \"53c95cd8-94b4-45f8-bc29-14fe2bb8049c\") " Dec 10 23:12:45 crc kubenswrapper[4791]: I1210 23:12:45.083076 4791 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/53c95cd8-94b4-45f8-bc29-14fe2bb8049c-ovsdbserver-sb\") pod \"53c95cd8-94b4-45f8-bc29-14fe2bb8049c\" (UID: \"53c95cd8-94b4-45f8-bc29-14fe2bb8049c\") " Dec 10 23:12:45 crc kubenswrapper[4791]: I1210 23:12:45.083164 4791 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/53c95cd8-94b4-45f8-bc29-14fe2bb8049c-config\") pod \"53c95cd8-94b4-45f8-bc29-14fe2bb8049c\" (UID: \"53c95cd8-94b4-45f8-bc29-14fe2bb8049c\") " Dec 10 23:12:45 crc kubenswrapper[4791]: I1210 23:12:45.092695 4791 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/53c95cd8-94b4-45f8-bc29-14fe2bb8049c-kube-api-access-zzrcq" (OuterVolumeSpecName: "kube-api-access-zzrcq") pod "53c95cd8-94b4-45f8-bc29-14fe2bb8049c" (UID: "53c95cd8-94b4-45f8-bc29-14fe2bb8049c"). InnerVolumeSpecName "kube-api-access-zzrcq". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 10 23:12:45 crc kubenswrapper[4791]: I1210 23:12:45.145799 4791 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/53c95cd8-94b4-45f8-bc29-14fe2bb8049c-openstack-edpm-ipam" (OuterVolumeSpecName: "openstack-edpm-ipam") pod "53c95cd8-94b4-45f8-bc29-14fe2bb8049c" (UID: "53c95cd8-94b4-45f8-bc29-14fe2bb8049c"). InnerVolumeSpecName "openstack-edpm-ipam". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 10 23:12:45 crc kubenswrapper[4791]: I1210 23:12:45.146412 4791 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/53c95cd8-94b4-45f8-bc29-14fe2bb8049c-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "53c95cd8-94b4-45f8-bc29-14fe2bb8049c" (UID: "53c95cd8-94b4-45f8-bc29-14fe2bb8049c"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 10 23:12:45 crc kubenswrapper[4791]: I1210 23:12:45.146769 4791 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/53c95cd8-94b4-45f8-bc29-14fe2bb8049c-dns-swift-storage-0" (OuterVolumeSpecName: "dns-swift-storage-0") pod "53c95cd8-94b4-45f8-bc29-14fe2bb8049c" (UID: "53c95cd8-94b4-45f8-bc29-14fe2bb8049c"). InnerVolumeSpecName "dns-swift-storage-0". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 10 23:12:45 crc kubenswrapper[4791]: I1210 23:12:45.154245 4791 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/53c95cd8-94b4-45f8-bc29-14fe2bb8049c-ovsdbserver-nb" (OuterVolumeSpecName: "ovsdbserver-nb") pod "53c95cd8-94b4-45f8-bc29-14fe2bb8049c" (UID: "53c95cd8-94b4-45f8-bc29-14fe2bb8049c"). InnerVolumeSpecName "ovsdbserver-nb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 10 23:12:45 crc kubenswrapper[4791]: I1210 23:12:45.156009 4791 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/53c95cd8-94b4-45f8-bc29-14fe2bb8049c-config" (OuterVolumeSpecName: "config") pod "53c95cd8-94b4-45f8-bc29-14fe2bb8049c" (UID: "53c95cd8-94b4-45f8-bc29-14fe2bb8049c"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 10 23:12:45 crc kubenswrapper[4791]: I1210 23:12:45.168688 4791 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/53c95cd8-94b4-45f8-bc29-14fe2bb8049c-ovsdbserver-sb" (OuterVolumeSpecName: "ovsdbserver-sb") pod "53c95cd8-94b4-45f8-bc29-14fe2bb8049c" (UID: "53c95cd8-94b4-45f8-bc29-14fe2bb8049c"). InnerVolumeSpecName "ovsdbserver-sb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 10 23:12:45 crc kubenswrapper[4791]: I1210 23:12:45.184877 4791 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/53c95cd8-94b4-45f8-bc29-14fe2bb8049c-ovsdbserver-nb\") on node \"crc\" DevicePath \"\"" Dec 10 23:12:45 crc kubenswrapper[4791]: I1210 23:12:45.184914 4791 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/53c95cd8-94b4-45f8-bc29-14fe2bb8049c-ovsdbserver-sb\") on node \"crc\" DevicePath \"\"" Dec 10 23:12:45 crc kubenswrapper[4791]: I1210 23:12:45.184925 4791 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/53c95cd8-94b4-45f8-bc29-14fe2bb8049c-config\") on node \"crc\" DevicePath \"\"" Dec 10 23:12:45 crc kubenswrapper[4791]: I1210 23:12:45.184936 4791 reconciler_common.go:293] "Volume detached for volume \"openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/configmap/53c95cd8-94b4-45f8-bc29-14fe2bb8049c-openstack-edpm-ipam\") on node \"crc\" DevicePath \"\"" Dec 10 23:12:45 crc kubenswrapper[4791]: I1210 23:12:45.184945 4791 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-zzrcq\" (UniqueName: \"kubernetes.io/projected/53c95cd8-94b4-45f8-bc29-14fe2bb8049c-kube-api-access-zzrcq\") on node \"crc\" DevicePath \"\"" Dec 10 23:12:45 crc kubenswrapper[4791]: I1210 23:12:45.184956 4791 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/53c95cd8-94b4-45f8-bc29-14fe2bb8049c-dns-svc\") on node \"crc\" DevicePath \"\"" Dec 10 23:12:45 crc kubenswrapper[4791]: I1210 23:12:45.184965 4791 reconciler_common.go:293] "Volume detached for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/53c95cd8-94b4-45f8-bc29-14fe2bb8049c-dns-swift-storage-0\") on node \"crc\" DevicePath \"\"" Dec 10 23:12:45 crc kubenswrapper[4791]: I1210 23:12:45.942602 4791 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-67b789f86c-cq8ml" Dec 10 23:12:45 crc kubenswrapper[4791]: I1210 23:12:45.972597 4791 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-67b789f86c-cq8ml"] Dec 10 23:12:45 crc kubenswrapper[4791]: I1210 23:12:45.983887 4791 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-67b789f86c-cq8ml"] Dec 10 23:12:47 crc kubenswrapper[4791]: I1210 23:12:47.981042 4791 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="53c95cd8-94b4-45f8-bc29-14fe2bb8049c" path="/var/lib/kubelet/pods/53c95cd8-94b4-45f8-bc29-14fe2bb8049c/volumes" Dec 10 23:12:56 crc kubenswrapper[4791]: I1210 23:12:56.511983 4791 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-xb2ph"] Dec 10 23:12:56 crc kubenswrapper[4791]: E1210 23:12:56.513032 4791 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="53c95cd8-94b4-45f8-bc29-14fe2bb8049c" containerName="init" Dec 10 23:12:56 crc kubenswrapper[4791]: I1210 23:12:56.513048 4791 state_mem.go:107] "Deleted CPUSet assignment" podUID="53c95cd8-94b4-45f8-bc29-14fe2bb8049c" containerName="init" Dec 10 23:12:56 crc kubenswrapper[4791]: E1210 23:12:56.513083 4791 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="339ea8c9-715f-4394-ace8-6b66542bcc13" containerName="dnsmasq-dns" Dec 10 23:12:56 crc kubenswrapper[4791]: I1210 23:12:56.513089 4791 state_mem.go:107] "Deleted CPUSet assignment" podUID="339ea8c9-715f-4394-ace8-6b66542bcc13" containerName="dnsmasq-dns" Dec 10 23:12:56 crc kubenswrapper[4791]: E1210 23:12:56.513102 4791 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="339ea8c9-715f-4394-ace8-6b66542bcc13" containerName="init" Dec 10 23:12:56 crc kubenswrapper[4791]: I1210 23:12:56.513109 4791 state_mem.go:107] "Deleted CPUSet assignment" podUID="339ea8c9-715f-4394-ace8-6b66542bcc13" containerName="init" Dec 10 23:12:56 crc kubenswrapper[4791]: E1210 23:12:56.513119 4791 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="53c95cd8-94b4-45f8-bc29-14fe2bb8049c" containerName="dnsmasq-dns" Dec 10 23:12:56 crc kubenswrapper[4791]: I1210 23:12:56.513125 4791 state_mem.go:107] "Deleted CPUSet assignment" podUID="53c95cd8-94b4-45f8-bc29-14fe2bb8049c" containerName="dnsmasq-dns" Dec 10 23:12:56 crc kubenswrapper[4791]: I1210 23:12:56.513303 4791 memory_manager.go:354] "RemoveStaleState removing state" podUID="339ea8c9-715f-4394-ace8-6b66542bcc13" containerName="dnsmasq-dns" Dec 10 23:12:56 crc kubenswrapper[4791]: I1210 23:12:56.513330 4791 memory_manager.go:354] "RemoveStaleState removing state" podUID="53c95cd8-94b4-45f8-bc29-14fe2bb8049c" containerName="dnsmasq-dns" Dec 10 23:12:56 crc kubenswrapper[4791]: I1210 23:12:56.514020 4791 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-xb2ph" Dec 10 23:12:56 crc kubenswrapper[4791]: I1210 23:12:56.516420 4791 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-ansible-ssh-private-key-secret" Dec 10 23:12:56 crc kubenswrapper[4791]: I1210 23:12:56.516949 4791 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-dockercfg-hf8pr" Dec 10 23:12:56 crc kubenswrapper[4791]: I1210 23:12:56.517617 4791 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Dec 10 23:12:56 crc kubenswrapper[4791]: I1210 23:12:56.519445 4791 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-edpm-ipam" Dec 10 23:12:56 crc kubenswrapper[4791]: I1210 23:12:56.527863 4791 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-xb2ph"] Dec 10 23:12:56 crc kubenswrapper[4791]: I1210 23:12:56.532548 4791 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/eff77dbf-4e46-4119-9db0-eb32a1d9d11c-inventory\") pod \"repo-setup-edpm-deployment-openstack-edpm-ipam-xb2ph\" (UID: \"eff77dbf-4e46-4119-9db0-eb32a1d9d11c\") " pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-xb2ph" Dec 10 23:12:56 crc kubenswrapper[4791]: I1210 23:12:56.532725 4791 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-z87xr\" (UniqueName: \"kubernetes.io/projected/eff77dbf-4e46-4119-9db0-eb32a1d9d11c-kube-api-access-z87xr\") pod \"repo-setup-edpm-deployment-openstack-edpm-ipam-xb2ph\" (UID: \"eff77dbf-4e46-4119-9db0-eb32a1d9d11c\") " pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-xb2ph" Dec 10 23:12:56 crc kubenswrapper[4791]: I1210 23:12:56.532917 4791 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/eff77dbf-4e46-4119-9db0-eb32a1d9d11c-ssh-key\") pod \"repo-setup-edpm-deployment-openstack-edpm-ipam-xb2ph\" (UID: \"eff77dbf-4e46-4119-9db0-eb32a1d9d11c\") " pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-xb2ph" Dec 10 23:12:56 crc kubenswrapper[4791]: I1210 23:12:56.532992 4791 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"repo-setup-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/eff77dbf-4e46-4119-9db0-eb32a1d9d11c-repo-setup-combined-ca-bundle\") pod \"repo-setup-edpm-deployment-openstack-edpm-ipam-xb2ph\" (UID: \"eff77dbf-4e46-4119-9db0-eb32a1d9d11c\") " pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-xb2ph" Dec 10 23:12:56 crc kubenswrapper[4791]: I1210 23:12:56.638109 4791 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/eff77dbf-4e46-4119-9db0-eb32a1d9d11c-inventory\") pod \"repo-setup-edpm-deployment-openstack-edpm-ipam-xb2ph\" (UID: \"eff77dbf-4e46-4119-9db0-eb32a1d9d11c\") " pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-xb2ph" Dec 10 23:12:56 crc kubenswrapper[4791]: I1210 23:12:56.639216 4791 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-z87xr\" (UniqueName: \"kubernetes.io/projected/eff77dbf-4e46-4119-9db0-eb32a1d9d11c-kube-api-access-z87xr\") pod \"repo-setup-edpm-deployment-openstack-edpm-ipam-xb2ph\" (UID: \"eff77dbf-4e46-4119-9db0-eb32a1d9d11c\") " pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-xb2ph" Dec 10 23:12:56 crc kubenswrapper[4791]: I1210 23:12:56.639907 4791 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/eff77dbf-4e46-4119-9db0-eb32a1d9d11c-ssh-key\") pod \"repo-setup-edpm-deployment-openstack-edpm-ipam-xb2ph\" (UID: \"eff77dbf-4e46-4119-9db0-eb32a1d9d11c\") " pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-xb2ph" Dec 10 23:12:56 crc kubenswrapper[4791]: I1210 23:12:56.639954 4791 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"repo-setup-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/eff77dbf-4e46-4119-9db0-eb32a1d9d11c-repo-setup-combined-ca-bundle\") pod \"repo-setup-edpm-deployment-openstack-edpm-ipam-xb2ph\" (UID: \"eff77dbf-4e46-4119-9db0-eb32a1d9d11c\") " pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-xb2ph" Dec 10 23:12:56 crc kubenswrapper[4791]: I1210 23:12:56.648541 4791 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/eff77dbf-4e46-4119-9db0-eb32a1d9d11c-ssh-key\") pod \"repo-setup-edpm-deployment-openstack-edpm-ipam-xb2ph\" (UID: \"eff77dbf-4e46-4119-9db0-eb32a1d9d11c\") " pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-xb2ph" Dec 10 23:12:56 crc kubenswrapper[4791]: I1210 23:12:56.649404 4791 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"repo-setup-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/eff77dbf-4e46-4119-9db0-eb32a1d9d11c-repo-setup-combined-ca-bundle\") pod \"repo-setup-edpm-deployment-openstack-edpm-ipam-xb2ph\" (UID: \"eff77dbf-4e46-4119-9db0-eb32a1d9d11c\") " pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-xb2ph" Dec 10 23:12:56 crc kubenswrapper[4791]: I1210 23:12:56.650032 4791 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/eff77dbf-4e46-4119-9db0-eb32a1d9d11c-inventory\") pod \"repo-setup-edpm-deployment-openstack-edpm-ipam-xb2ph\" (UID: \"eff77dbf-4e46-4119-9db0-eb32a1d9d11c\") " pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-xb2ph" Dec 10 23:12:56 crc kubenswrapper[4791]: I1210 23:12:56.660289 4791 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-z87xr\" (UniqueName: \"kubernetes.io/projected/eff77dbf-4e46-4119-9db0-eb32a1d9d11c-kube-api-access-z87xr\") pod \"repo-setup-edpm-deployment-openstack-edpm-ipam-xb2ph\" (UID: \"eff77dbf-4e46-4119-9db0-eb32a1d9d11c\") " pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-xb2ph" Dec 10 23:12:56 crc kubenswrapper[4791]: I1210 23:12:56.849566 4791 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-xb2ph" Dec 10 23:12:57 crc kubenswrapper[4791]: I1210 23:12:57.128254 4791 generic.go:334] "Generic (PLEG): container finished" podID="5f51c37d-2f6f-4ef0-a08c-77216e7db57b" containerID="9ed929ea36f1a6f4cec971ac7e7d5a6838a4ce09161cfc8c65d2c50871abcbe2" exitCode=0 Dec 10 23:12:57 crc kubenswrapper[4791]: I1210 23:12:57.128554 4791 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-server-0" event={"ID":"5f51c37d-2f6f-4ef0-a08c-77216e7db57b","Type":"ContainerDied","Data":"9ed929ea36f1a6f4cec971ac7e7d5a6838a4ce09161cfc8c65d2c50871abcbe2"} Dec 10 23:12:57 crc kubenswrapper[4791]: I1210 23:12:57.416638 4791 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-xb2ph"] Dec 10 23:12:58 crc kubenswrapper[4791]: I1210 23:12:58.139053 4791 generic.go:334] "Generic (PLEG): container finished" podID="3add67e0-b67a-438b-b83f-4c9f6733b5cb" containerID="2bc6b221ebeead7a88db3ab622e885d16946fac93e077d0559fe60a6208fcaf6" exitCode=0 Dec 10 23:12:58 crc kubenswrapper[4791]: I1210 23:12:58.139182 4791 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-cell1-server-0" event={"ID":"3add67e0-b67a-438b-b83f-4c9f6733b5cb","Type":"ContainerDied","Data":"2bc6b221ebeead7a88db3ab622e885d16946fac93e077d0559fe60a6208fcaf6"} Dec 10 23:12:58 crc kubenswrapper[4791]: I1210 23:12:58.145549 4791 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-server-0" event={"ID":"5f51c37d-2f6f-4ef0-a08c-77216e7db57b","Type":"ContainerStarted","Data":"661550375f128c1b794d200b7897680bc4e8dd177ea4cf771a491962696f37d8"} Dec 10 23:12:58 crc kubenswrapper[4791]: I1210 23:12:58.145750 4791 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/rabbitmq-server-0" Dec 10 23:12:58 crc kubenswrapper[4791]: I1210 23:12:58.147023 4791 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-xb2ph" event={"ID":"eff77dbf-4e46-4119-9db0-eb32a1d9d11c","Type":"ContainerStarted","Data":"7729e0b350752d053accee9aeaa2f3e2da900b1c65e7c25df9a20acc3734de6f"} Dec 10 23:12:58 crc kubenswrapper[4791]: I1210 23:12:58.231090 4791 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/rabbitmq-server-0" podStartSLOduration=37.231066974 podStartE2EDuration="37.231066974s" podCreationTimestamp="2025-12-10 23:12:21 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-10 23:12:58.20669749 +0000 UTC m=+1412.636315113" watchObservedRunningTime="2025-12-10 23:12:58.231066974 +0000 UTC m=+1412.660684587" Dec 10 23:12:59 crc kubenswrapper[4791]: I1210 23:12:59.156871 4791 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-cell1-server-0" event={"ID":"3add67e0-b67a-438b-b83f-4c9f6733b5cb","Type":"ContainerStarted","Data":"a460bca4736a4cb19df962e97412f0df21d214a7cccb3da0844fe68cf72c2e94"} Dec 10 23:12:59 crc kubenswrapper[4791]: I1210 23:12:59.158148 4791 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/rabbitmq-cell1-server-0" Dec 10 23:12:59 crc kubenswrapper[4791]: I1210 23:12:59.188899 4791 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/rabbitmq-cell1-server-0" podStartSLOduration=37.188883949 podStartE2EDuration="37.188883949s" podCreationTimestamp="2025-12-10 23:12:22 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-10 23:12:59.187564251 +0000 UTC m=+1413.617181864" watchObservedRunningTime="2025-12-10 23:12:59.188883949 +0000 UTC m=+1413.618501552" Dec 10 23:13:11 crc kubenswrapper[4791]: I1210 23:13:11.314795 4791 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-xb2ph" event={"ID":"eff77dbf-4e46-4119-9db0-eb32a1d9d11c","Type":"ContainerStarted","Data":"03c68f2e4c3989cc000e8d75b1d05f3e54b9ebc0801a268fbf95b167a3ea2af0"} Dec 10 23:13:11 crc kubenswrapper[4791]: I1210 23:13:11.343573 4791 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-xb2ph" podStartSLOduration=2.477419374 podStartE2EDuration="15.34355351s" podCreationTimestamp="2025-12-10 23:12:56 +0000 UTC" firstStartedPulling="2025-12-10 23:12:57.423840776 +0000 UTC m=+1411.853458379" lastFinishedPulling="2025-12-10 23:13:10.289974902 +0000 UTC m=+1424.719592515" observedRunningTime="2025-12-10 23:13:11.33327113 +0000 UTC m=+1425.762888773" watchObservedRunningTime="2025-12-10 23:13:11.34355351 +0000 UTC m=+1425.773171123" Dec 10 23:13:12 crc kubenswrapper[4791]: I1210 23:13:12.158948 4791 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/rabbitmq-server-0" Dec 10 23:13:13 crc kubenswrapper[4791]: I1210 23:13:13.330568 4791 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/rabbitmq-cell1-server-0" Dec 10 23:13:22 crc kubenswrapper[4791]: I1210 23:13:22.439163 4791 generic.go:334] "Generic (PLEG): container finished" podID="eff77dbf-4e46-4119-9db0-eb32a1d9d11c" containerID="03c68f2e4c3989cc000e8d75b1d05f3e54b9ebc0801a268fbf95b167a3ea2af0" exitCode=0 Dec 10 23:13:22 crc kubenswrapper[4791]: I1210 23:13:22.439249 4791 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-xb2ph" event={"ID":"eff77dbf-4e46-4119-9db0-eb32a1d9d11c","Type":"ContainerDied","Data":"03c68f2e4c3989cc000e8d75b1d05f3e54b9ebc0801a268fbf95b167a3ea2af0"} Dec 10 23:13:23 crc kubenswrapper[4791]: I1210 23:13:23.867680 4791 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-xb2ph" Dec 10 23:13:24 crc kubenswrapper[4791]: I1210 23:13:24.014649 4791 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"repo-setup-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/eff77dbf-4e46-4119-9db0-eb32a1d9d11c-repo-setup-combined-ca-bundle\") pod \"eff77dbf-4e46-4119-9db0-eb32a1d9d11c\" (UID: \"eff77dbf-4e46-4119-9db0-eb32a1d9d11c\") " Dec 10 23:13:24 crc kubenswrapper[4791]: I1210 23:13:24.014931 4791 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/eff77dbf-4e46-4119-9db0-eb32a1d9d11c-inventory\") pod \"eff77dbf-4e46-4119-9db0-eb32a1d9d11c\" (UID: \"eff77dbf-4e46-4119-9db0-eb32a1d9d11c\") " Dec 10 23:13:24 crc kubenswrapper[4791]: I1210 23:13:24.016120 4791 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-z87xr\" (UniqueName: \"kubernetes.io/projected/eff77dbf-4e46-4119-9db0-eb32a1d9d11c-kube-api-access-z87xr\") pod \"eff77dbf-4e46-4119-9db0-eb32a1d9d11c\" (UID: \"eff77dbf-4e46-4119-9db0-eb32a1d9d11c\") " Dec 10 23:13:24 crc kubenswrapper[4791]: I1210 23:13:24.016295 4791 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/eff77dbf-4e46-4119-9db0-eb32a1d9d11c-ssh-key\") pod \"eff77dbf-4e46-4119-9db0-eb32a1d9d11c\" (UID: \"eff77dbf-4e46-4119-9db0-eb32a1d9d11c\") " Dec 10 23:13:24 crc kubenswrapper[4791]: I1210 23:13:24.031282 4791 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/eff77dbf-4e46-4119-9db0-eb32a1d9d11c-repo-setup-combined-ca-bundle" (OuterVolumeSpecName: "repo-setup-combined-ca-bundle") pod "eff77dbf-4e46-4119-9db0-eb32a1d9d11c" (UID: "eff77dbf-4e46-4119-9db0-eb32a1d9d11c"). InnerVolumeSpecName "repo-setup-combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 10 23:13:24 crc kubenswrapper[4791]: I1210 23:13:24.034174 4791 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/eff77dbf-4e46-4119-9db0-eb32a1d9d11c-kube-api-access-z87xr" (OuterVolumeSpecName: "kube-api-access-z87xr") pod "eff77dbf-4e46-4119-9db0-eb32a1d9d11c" (UID: "eff77dbf-4e46-4119-9db0-eb32a1d9d11c"). InnerVolumeSpecName "kube-api-access-z87xr". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 10 23:13:24 crc kubenswrapper[4791]: I1210 23:13:24.057985 4791 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/eff77dbf-4e46-4119-9db0-eb32a1d9d11c-inventory" (OuterVolumeSpecName: "inventory") pod "eff77dbf-4e46-4119-9db0-eb32a1d9d11c" (UID: "eff77dbf-4e46-4119-9db0-eb32a1d9d11c"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 10 23:13:24 crc kubenswrapper[4791]: I1210 23:13:24.063062 4791 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/eff77dbf-4e46-4119-9db0-eb32a1d9d11c-ssh-key" (OuterVolumeSpecName: "ssh-key") pod "eff77dbf-4e46-4119-9db0-eb32a1d9d11c" (UID: "eff77dbf-4e46-4119-9db0-eb32a1d9d11c"). InnerVolumeSpecName "ssh-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 10 23:13:24 crc kubenswrapper[4791]: I1210 23:13:24.118495 4791 reconciler_common.go:293] "Volume detached for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/eff77dbf-4e46-4119-9db0-eb32a1d9d11c-ssh-key\") on node \"crc\" DevicePath \"\"" Dec 10 23:13:24 crc kubenswrapper[4791]: I1210 23:13:24.118533 4791 reconciler_common.go:293] "Volume detached for volume \"repo-setup-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/eff77dbf-4e46-4119-9db0-eb32a1d9d11c-repo-setup-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 10 23:13:24 crc kubenswrapper[4791]: I1210 23:13:24.118552 4791 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/eff77dbf-4e46-4119-9db0-eb32a1d9d11c-inventory\") on node \"crc\" DevicePath \"\"" Dec 10 23:13:24 crc kubenswrapper[4791]: I1210 23:13:24.118565 4791 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-z87xr\" (UniqueName: \"kubernetes.io/projected/eff77dbf-4e46-4119-9db0-eb32a1d9d11c-kube-api-access-z87xr\") on node \"crc\" DevicePath \"\"" Dec 10 23:13:24 crc kubenswrapper[4791]: I1210 23:13:24.464479 4791 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-xb2ph" event={"ID":"eff77dbf-4e46-4119-9db0-eb32a1d9d11c","Type":"ContainerDied","Data":"7729e0b350752d053accee9aeaa2f3e2da900b1c65e7c25df9a20acc3734de6f"} Dec 10 23:13:24 crc kubenswrapper[4791]: I1210 23:13:24.464943 4791 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="7729e0b350752d053accee9aeaa2f3e2da900b1c65e7c25df9a20acc3734de6f" Dec 10 23:13:24 crc kubenswrapper[4791]: I1210 23:13:24.464524 4791 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-xb2ph" Dec 10 23:13:24 crc kubenswrapper[4791]: I1210 23:13:24.578756 4791 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/redhat-edpm-deployment-openstack-edpm-ipam-2zj6q"] Dec 10 23:13:24 crc kubenswrapper[4791]: E1210 23:13:24.579180 4791 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="eff77dbf-4e46-4119-9db0-eb32a1d9d11c" containerName="repo-setup-edpm-deployment-openstack-edpm-ipam" Dec 10 23:13:24 crc kubenswrapper[4791]: I1210 23:13:24.579202 4791 state_mem.go:107] "Deleted CPUSet assignment" podUID="eff77dbf-4e46-4119-9db0-eb32a1d9d11c" containerName="repo-setup-edpm-deployment-openstack-edpm-ipam" Dec 10 23:13:24 crc kubenswrapper[4791]: I1210 23:13:24.579392 4791 memory_manager.go:354] "RemoveStaleState removing state" podUID="eff77dbf-4e46-4119-9db0-eb32a1d9d11c" containerName="repo-setup-edpm-deployment-openstack-edpm-ipam" Dec 10 23:13:24 crc kubenswrapper[4791]: I1210 23:13:24.580003 4791 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/redhat-edpm-deployment-openstack-edpm-ipam-2zj6q" Dec 10 23:13:24 crc kubenswrapper[4791]: I1210 23:13:24.584790 4791 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-edpm-ipam" Dec 10 23:13:24 crc kubenswrapper[4791]: I1210 23:13:24.584833 4791 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-ansible-ssh-private-key-secret" Dec 10 23:13:24 crc kubenswrapper[4791]: I1210 23:13:24.585575 4791 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-dockercfg-hf8pr" Dec 10 23:13:24 crc kubenswrapper[4791]: I1210 23:13:24.589428 4791 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Dec 10 23:13:24 crc kubenswrapper[4791]: I1210 23:13:24.596488 4791 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/redhat-edpm-deployment-openstack-edpm-ipam-2zj6q"] Dec 10 23:13:24 crc kubenswrapper[4791]: I1210 23:13:24.731687 4791 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/e77c2ca5-1bf1-4739-b1c3-b0913fcc9d99-ssh-key\") pod \"redhat-edpm-deployment-openstack-edpm-ipam-2zj6q\" (UID: \"e77c2ca5-1bf1-4739-b1c3-b0913fcc9d99\") " pod="openstack/redhat-edpm-deployment-openstack-edpm-ipam-2zj6q" Dec 10 23:13:24 crc kubenswrapper[4791]: I1210 23:13:24.731760 4791 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-vtmb5\" (UniqueName: \"kubernetes.io/projected/e77c2ca5-1bf1-4739-b1c3-b0913fcc9d99-kube-api-access-vtmb5\") pod \"redhat-edpm-deployment-openstack-edpm-ipam-2zj6q\" (UID: \"e77c2ca5-1bf1-4739-b1c3-b0913fcc9d99\") " pod="openstack/redhat-edpm-deployment-openstack-edpm-ipam-2zj6q" Dec 10 23:13:24 crc kubenswrapper[4791]: I1210 23:13:24.732914 4791 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/e77c2ca5-1bf1-4739-b1c3-b0913fcc9d99-inventory\") pod \"redhat-edpm-deployment-openstack-edpm-ipam-2zj6q\" (UID: \"e77c2ca5-1bf1-4739-b1c3-b0913fcc9d99\") " pod="openstack/redhat-edpm-deployment-openstack-edpm-ipam-2zj6q" Dec 10 23:13:24 crc kubenswrapper[4791]: I1210 23:13:24.834959 4791 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/e77c2ca5-1bf1-4739-b1c3-b0913fcc9d99-inventory\") pod \"redhat-edpm-deployment-openstack-edpm-ipam-2zj6q\" (UID: \"e77c2ca5-1bf1-4739-b1c3-b0913fcc9d99\") " pod="openstack/redhat-edpm-deployment-openstack-edpm-ipam-2zj6q" Dec 10 23:13:24 crc kubenswrapper[4791]: I1210 23:13:24.835093 4791 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/e77c2ca5-1bf1-4739-b1c3-b0913fcc9d99-ssh-key\") pod \"redhat-edpm-deployment-openstack-edpm-ipam-2zj6q\" (UID: \"e77c2ca5-1bf1-4739-b1c3-b0913fcc9d99\") " pod="openstack/redhat-edpm-deployment-openstack-edpm-ipam-2zj6q" Dec 10 23:13:24 crc kubenswrapper[4791]: I1210 23:13:24.835131 4791 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-vtmb5\" (UniqueName: \"kubernetes.io/projected/e77c2ca5-1bf1-4739-b1c3-b0913fcc9d99-kube-api-access-vtmb5\") pod \"redhat-edpm-deployment-openstack-edpm-ipam-2zj6q\" (UID: \"e77c2ca5-1bf1-4739-b1c3-b0913fcc9d99\") " pod="openstack/redhat-edpm-deployment-openstack-edpm-ipam-2zj6q" Dec 10 23:13:24 crc kubenswrapper[4791]: I1210 23:13:24.839915 4791 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/e77c2ca5-1bf1-4739-b1c3-b0913fcc9d99-ssh-key\") pod \"redhat-edpm-deployment-openstack-edpm-ipam-2zj6q\" (UID: \"e77c2ca5-1bf1-4739-b1c3-b0913fcc9d99\") " pod="openstack/redhat-edpm-deployment-openstack-edpm-ipam-2zj6q" Dec 10 23:13:24 crc kubenswrapper[4791]: I1210 23:13:24.851814 4791 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/e77c2ca5-1bf1-4739-b1c3-b0913fcc9d99-inventory\") pod \"redhat-edpm-deployment-openstack-edpm-ipam-2zj6q\" (UID: \"e77c2ca5-1bf1-4739-b1c3-b0913fcc9d99\") " pod="openstack/redhat-edpm-deployment-openstack-edpm-ipam-2zj6q" Dec 10 23:13:24 crc kubenswrapper[4791]: I1210 23:13:24.855079 4791 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-vtmb5\" (UniqueName: \"kubernetes.io/projected/e77c2ca5-1bf1-4739-b1c3-b0913fcc9d99-kube-api-access-vtmb5\") pod \"redhat-edpm-deployment-openstack-edpm-ipam-2zj6q\" (UID: \"e77c2ca5-1bf1-4739-b1c3-b0913fcc9d99\") " pod="openstack/redhat-edpm-deployment-openstack-edpm-ipam-2zj6q" Dec 10 23:13:24 crc kubenswrapper[4791]: I1210 23:13:24.929965 4791 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/redhat-edpm-deployment-openstack-edpm-ipam-2zj6q" Dec 10 23:13:25 crc kubenswrapper[4791]: I1210 23:13:25.486728 4791 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/redhat-edpm-deployment-openstack-edpm-ipam-2zj6q"] Dec 10 23:13:25 crc kubenswrapper[4791]: W1210 23:13:25.491615 4791 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pode77c2ca5_1bf1_4739_b1c3_b0913fcc9d99.slice/crio-ef7a7366af565becd6bd55f192f6e8ce41c59bfe51bc9f5d418fa4fa07601808 WatchSource:0}: Error finding container ef7a7366af565becd6bd55f192f6e8ce41c59bfe51bc9f5d418fa4fa07601808: Status 404 returned error can't find the container with id ef7a7366af565becd6bd55f192f6e8ce41c59bfe51bc9f5d418fa4fa07601808 Dec 10 23:13:26 crc kubenswrapper[4791]: I1210 23:13:26.092318 4791 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Dec 10 23:13:26 crc kubenswrapper[4791]: I1210 23:13:26.487210 4791 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/redhat-edpm-deployment-openstack-edpm-ipam-2zj6q" event={"ID":"e77c2ca5-1bf1-4739-b1c3-b0913fcc9d99","Type":"ContainerStarted","Data":"a7b7b12db198ccc6445673ed60216ff0beed0d0cb25b77cc687ed5b502ccbfee"} Dec 10 23:13:26 crc kubenswrapper[4791]: I1210 23:13:26.487270 4791 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/redhat-edpm-deployment-openstack-edpm-ipam-2zj6q" event={"ID":"e77c2ca5-1bf1-4739-b1c3-b0913fcc9d99","Type":"ContainerStarted","Data":"ef7a7366af565becd6bd55f192f6e8ce41c59bfe51bc9f5d418fa4fa07601808"} Dec 10 23:13:26 crc kubenswrapper[4791]: I1210 23:13:26.512552 4791 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/redhat-edpm-deployment-openstack-edpm-ipam-2zj6q" podStartSLOduration=1.916589747 podStartE2EDuration="2.512531895s" podCreationTimestamp="2025-12-10 23:13:24 +0000 UTC" firstStartedPulling="2025-12-10 23:13:25.493872511 +0000 UTC m=+1439.923490124" lastFinishedPulling="2025-12-10 23:13:26.089814659 +0000 UTC m=+1440.519432272" observedRunningTime="2025-12-10 23:13:26.501209746 +0000 UTC m=+1440.930827359" watchObservedRunningTime="2025-12-10 23:13:26.512531895 +0000 UTC m=+1440.942149508" Dec 10 23:13:29 crc kubenswrapper[4791]: I1210 23:13:29.529631 4791 generic.go:334] "Generic (PLEG): container finished" podID="e77c2ca5-1bf1-4739-b1c3-b0913fcc9d99" containerID="a7b7b12db198ccc6445673ed60216ff0beed0d0cb25b77cc687ed5b502ccbfee" exitCode=0 Dec 10 23:13:29 crc kubenswrapper[4791]: I1210 23:13:29.529735 4791 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/redhat-edpm-deployment-openstack-edpm-ipam-2zj6q" event={"ID":"e77c2ca5-1bf1-4739-b1c3-b0913fcc9d99","Type":"ContainerDied","Data":"a7b7b12db198ccc6445673ed60216ff0beed0d0cb25b77cc687ed5b502ccbfee"} Dec 10 23:13:30 crc kubenswrapper[4791]: I1210 23:13:30.965497 4791 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/redhat-edpm-deployment-openstack-edpm-ipam-2zj6q" Dec 10 23:13:31 crc kubenswrapper[4791]: I1210 23:13:31.119068 4791 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-vtmb5\" (UniqueName: \"kubernetes.io/projected/e77c2ca5-1bf1-4739-b1c3-b0913fcc9d99-kube-api-access-vtmb5\") pod \"e77c2ca5-1bf1-4739-b1c3-b0913fcc9d99\" (UID: \"e77c2ca5-1bf1-4739-b1c3-b0913fcc9d99\") " Dec 10 23:13:31 crc kubenswrapper[4791]: I1210 23:13:31.119321 4791 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/e77c2ca5-1bf1-4739-b1c3-b0913fcc9d99-inventory\") pod \"e77c2ca5-1bf1-4739-b1c3-b0913fcc9d99\" (UID: \"e77c2ca5-1bf1-4739-b1c3-b0913fcc9d99\") " Dec 10 23:13:31 crc kubenswrapper[4791]: I1210 23:13:31.119447 4791 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/e77c2ca5-1bf1-4739-b1c3-b0913fcc9d99-ssh-key\") pod \"e77c2ca5-1bf1-4739-b1c3-b0913fcc9d99\" (UID: \"e77c2ca5-1bf1-4739-b1c3-b0913fcc9d99\") " Dec 10 23:13:31 crc kubenswrapper[4791]: I1210 23:13:31.125254 4791 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/e77c2ca5-1bf1-4739-b1c3-b0913fcc9d99-kube-api-access-vtmb5" (OuterVolumeSpecName: "kube-api-access-vtmb5") pod "e77c2ca5-1bf1-4739-b1c3-b0913fcc9d99" (UID: "e77c2ca5-1bf1-4739-b1c3-b0913fcc9d99"). InnerVolumeSpecName "kube-api-access-vtmb5". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 10 23:13:31 crc kubenswrapper[4791]: I1210 23:13:31.150273 4791 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/e77c2ca5-1bf1-4739-b1c3-b0913fcc9d99-inventory" (OuterVolumeSpecName: "inventory") pod "e77c2ca5-1bf1-4739-b1c3-b0913fcc9d99" (UID: "e77c2ca5-1bf1-4739-b1c3-b0913fcc9d99"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 10 23:13:31 crc kubenswrapper[4791]: I1210 23:13:31.172017 4791 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/e77c2ca5-1bf1-4739-b1c3-b0913fcc9d99-ssh-key" (OuterVolumeSpecName: "ssh-key") pod "e77c2ca5-1bf1-4739-b1c3-b0913fcc9d99" (UID: "e77c2ca5-1bf1-4739-b1c3-b0913fcc9d99"). InnerVolumeSpecName "ssh-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 10 23:13:31 crc kubenswrapper[4791]: I1210 23:13:31.222213 4791 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/e77c2ca5-1bf1-4739-b1c3-b0913fcc9d99-inventory\") on node \"crc\" DevicePath \"\"" Dec 10 23:13:31 crc kubenswrapper[4791]: I1210 23:13:31.222248 4791 reconciler_common.go:293] "Volume detached for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/e77c2ca5-1bf1-4739-b1c3-b0913fcc9d99-ssh-key\") on node \"crc\" DevicePath \"\"" Dec 10 23:13:31 crc kubenswrapper[4791]: I1210 23:13:31.222260 4791 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-vtmb5\" (UniqueName: \"kubernetes.io/projected/e77c2ca5-1bf1-4739-b1c3-b0913fcc9d99-kube-api-access-vtmb5\") on node \"crc\" DevicePath \"\"" Dec 10 23:13:31 crc kubenswrapper[4791]: I1210 23:13:31.556016 4791 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/redhat-edpm-deployment-openstack-edpm-ipam-2zj6q" event={"ID":"e77c2ca5-1bf1-4739-b1c3-b0913fcc9d99","Type":"ContainerDied","Data":"ef7a7366af565becd6bd55f192f6e8ce41c59bfe51bc9f5d418fa4fa07601808"} Dec 10 23:13:31 crc kubenswrapper[4791]: I1210 23:13:31.556057 4791 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="ef7a7366af565becd6bd55f192f6e8ce41c59bfe51bc9f5d418fa4fa07601808" Dec 10 23:13:31 crc kubenswrapper[4791]: I1210 23:13:31.556111 4791 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/redhat-edpm-deployment-openstack-edpm-ipam-2zj6q" Dec 10 23:13:31 crc kubenswrapper[4791]: I1210 23:13:31.630263 4791 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-fpng5"] Dec 10 23:13:31 crc kubenswrapper[4791]: E1210 23:13:31.630869 4791 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="e77c2ca5-1bf1-4739-b1c3-b0913fcc9d99" containerName="redhat-edpm-deployment-openstack-edpm-ipam" Dec 10 23:13:31 crc kubenswrapper[4791]: I1210 23:13:31.630893 4791 state_mem.go:107] "Deleted CPUSet assignment" podUID="e77c2ca5-1bf1-4739-b1c3-b0913fcc9d99" containerName="redhat-edpm-deployment-openstack-edpm-ipam" Dec 10 23:13:31 crc kubenswrapper[4791]: I1210 23:13:31.631096 4791 memory_manager.go:354] "RemoveStaleState removing state" podUID="e77c2ca5-1bf1-4739-b1c3-b0913fcc9d99" containerName="redhat-edpm-deployment-openstack-edpm-ipam" Dec 10 23:13:31 crc kubenswrapper[4791]: I1210 23:13:31.631744 4791 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-fpng5" Dec 10 23:13:31 crc kubenswrapper[4791]: I1210 23:13:31.635021 4791 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-ansible-ssh-private-key-secret" Dec 10 23:13:31 crc kubenswrapper[4791]: I1210 23:13:31.635248 4791 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-dockercfg-hf8pr" Dec 10 23:13:31 crc kubenswrapper[4791]: I1210 23:13:31.635706 4791 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Dec 10 23:13:31 crc kubenswrapper[4791]: I1210 23:13:31.635761 4791 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-edpm-ipam" Dec 10 23:13:31 crc kubenswrapper[4791]: I1210 23:13:31.647778 4791 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-fpng5"] Dec 10 23:13:31 crc kubenswrapper[4791]: I1210 23:13:31.832517 4791 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/07a0187d-1677-4da9-9a3b-8da49e8f6819-inventory\") pod \"bootstrap-edpm-deployment-openstack-edpm-ipam-fpng5\" (UID: \"07a0187d-1677-4da9-9a3b-8da49e8f6819\") " pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-fpng5" Dec 10 23:13:31 crc kubenswrapper[4791]: I1210 23:13:31.832856 4791 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-dc9ck\" (UniqueName: \"kubernetes.io/projected/07a0187d-1677-4da9-9a3b-8da49e8f6819-kube-api-access-dc9ck\") pod \"bootstrap-edpm-deployment-openstack-edpm-ipam-fpng5\" (UID: \"07a0187d-1677-4da9-9a3b-8da49e8f6819\") " pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-fpng5" Dec 10 23:13:31 crc kubenswrapper[4791]: I1210 23:13:31.832892 4791 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/07a0187d-1677-4da9-9a3b-8da49e8f6819-ssh-key\") pod \"bootstrap-edpm-deployment-openstack-edpm-ipam-fpng5\" (UID: \"07a0187d-1677-4da9-9a3b-8da49e8f6819\") " pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-fpng5" Dec 10 23:13:31 crc kubenswrapper[4791]: I1210 23:13:31.832917 4791 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"bootstrap-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/07a0187d-1677-4da9-9a3b-8da49e8f6819-bootstrap-combined-ca-bundle\") pod \"bootstrap-edpm-deployment-openstack-edpm-ipam-fpng5\" (UID: \"07a0187d-1677-4da9-9a3b-8da49e8f6819\") " pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-fpng5" Dec 10 23:13:31 crc kubenswrapper[4791]: I1210 23:13:31.935251 4791 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/07a0187d-1677-4da9-9a3b-8da49e8f6819-inventory\") pod \"bootstrap-edpm-deployment-openstack-edpm-ipam-fpng5\" (UID: \"07a0187d-1677-4da9-9a3b-8da49e8f6819\") " pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-fpng5" Dec 10 23:13:31 crc kubenswrapper[4791]: I1210 23:13:31.935328 4791 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-dc9ck\" (UniqueName: \"kubernetes.io/projected/07a0187d-1677-4da9-9a3b-8da49e8f6819-kube-api-access-dc9ck\") pod \"bootstrap-edpm-deployment-openstack-edpm-ipam-fpng5\" (UID: \"07a0187d-1677-4da9-9a3b-8da49e8f6819\") " pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-fpng5" Dec 10 23:13:31 crc kubenswrapper[4791]: I1210 23:13:31.935402 4791 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/07a0187d-1677-4da9-9a3b-8da49e8f6819-ssh-key\") pod \"bootstrap-edpm-deployment-openstack-edpm-ipam-fpng5\" (UID: \"07a0187d-1677-4da9-9a3b-8da49e8f6819\") " pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-fpng5" Dec 10 23:13:31 crc kubenswrapper[4791]: I1210 23:13:31.935452 4791 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"bootstrap-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/07a0187d-1677-4da9-9a3b-8da49e8f6819-bootstrap-combined-ca-bundle\") pod \"bootstrap-edpm-deployment-openstack-edpm-ipam-fpng5\" (UID: \"07a0187d-1677-4da9-9a3b-8da49e8f6819\") " pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-fpng5" Dec 10 23:13:31 crc kubenswrapper[4791]: I1210 23:13:31.943689 4791 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/07a0187d-1677-4da9-9a3b-8da49e8f6819-ssh-key\") pod \"bootstrap-edpm-deployment-openstack-edpm-ipam-fpng5\" (UID: \"07a0187d-1677-4da9-9a3b-8da49e8f6819\") " pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-fpng5" Dec 10 23:13:31 crc kubenswrapper[4791]: I1210 23:13:31.944169 4791 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"bootstrap-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/07a0187d-1677-4da9-9a3b-8da49e8f6819-bootstrap-combined-ca-bundle\") pod \"bootstrap-edpm-deployment-openstack-edpm-ipam-fpng5\" (UID: \"07a0187d-1677-4da9-9a3b-8da49e8f6819\") " pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-fpng5" Dec 10 23:13:31 crc kubenswrapper[4791]: I1210 23:13:31.945306 4791 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/07a0187d-1677-4da9-9a3b-8da49e8f6819-inventory\") pod \"bootstrap-edpm-deployment-openstack-edpm-ipam-fpng5\" (UID: \"07a0187d-1677-4da9-9a3b-8da49e8f6819\") " pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-fpng5" Dec 10 23:13:31 crc kubenswrapper[4791]: I1210 23:13:31.970218 4791 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-dc9ck\" (UniqueName: \"kubernetes.io/projected/07a0187d-1677-4da9-9a3b-8da49e8f6819-kube-api-access-dc9ck\") pod \"bootstrap-edpm-deployment-openstack-edpm-ipam-fpng5\" (UID: \"07a0187d-1677-4da9-9a3b-8da49e8f6819\") " pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-fpng5" Dec 10 23:13:32 crc kubenswrapper[4791]: I1210 23:13:32.258656 4791 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-fpng5" Dec 10 23:13:32 crc kubenswrapper[4791]: I1210 23:13:32.789309 4791 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-fpng5"] Dec 10 23:13:33 crc kubenswrapper[4791]: I1210 23:13:33.584286 4791 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-fpng5" event={"ID":"07a0187d-1677-4da9-9a3b-8da49e8f6819","Type":"ContainerStarted","Data":"f100c8400fb1baf14371050d2bedf09b75fa03b225878c6f0168e865e1d1ecef"} Dec 10 23:13:34 crc kubenswrapper[4791]: I1210 23:13:34.594675 4791 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-fpng5" event={"ID":"07a0187d-1677-4da9-9a3b-8da49e8f6819","Type":"ContainerStarted","Data":"5eb7ff160fc3e10b947e2e0545d1afd197556543489b13d321b4d64ac233973c"} Dec 10 23:13:34 crc kubenswrapper[4791]: I1210 23:13:34.633764 4791 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-fpng5" podStartSLOduration=2.969101886 podStartE2EDuration="3.63373078s" podCreationTimestamp="2025-12-10 23:13:31 +0000 UTC" firstStartedPulling="2025-12-10 23:13:32.795209317 +0000 UTC m=+1447.224826930" lastFinishedPulling="2025-12-10 23:13:33.459838191 +0000 UTC m=+1447.889455824" observedRunningTime="2025-12-10 23:13:34.61883869 +0000 UTC m=+1449.048456343" watchObservedRunningTime="2025-12-10 23:13:34.63373078 +0000 UTC m=+1449.063348423" Dec 10 23:13:51 crc kubenswrapper[4791]: I1210 23:13:51.199831 4791 scope.go:117] "RemoveContainer" containerID="f0d087440ba482240de855bdc96648601e34f74962b3bf2123c4205414310fef" Dec 10 23:13:51 crc kubenswrapper[4791]: I1210 23:13:51.235475 4791 scope.go:117] "RemoveContainer" containerID="f4aaa5aed3505e556bbf35f76215681ffa6ce51d231f26498e8185599e152873" Dec 10 23:13:51 crc kubenswrapper[4791]: I1210 23:13:51.279848 4791 scope.go:117] "RemoveContainer" containerID="8dd3007abca59274d70261742380418ebbe002bfd3c95f64e21f07dcfab98397" Dec 10 23:14:25 crc kubenswrapper[4791]: I1210 23:14:25.038161 4791 patch_prober.go:28] interesting pod/machine-config-daemon-5rb5l container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 10 23:14:25 crc kubenswrapper[4791]: I1210 23:14:25.039000 4791 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-5rb5l" podUID="ba35653c-6e06-4cee-a4d6-137764090d18" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 10 23:14:31 crc kubenswrapper[4791]: I1210 23:14:31.447558 4791 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-marketplace-tfshq"] Dec 10 23:14:31 crc kubenswrapper[4791]: I1210 23:14:31.450498 4791 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-tfshq" Dec 10 23:14:31 crc kubenswrapper[4791]: I1210 23:14:31.468874 4791 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-tfshq"] Dec 10 23:14:31 crc kubenswrapper[4791]: I1210 23:14:31.554280 4791 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-2pkzw\" (UniqueName: \"kubernetes.io/projected/4cb92e7d-83db-48a9-ba3b-937bf14c900f-kube-api-access-2pkzw\") pod \"redhat-marketplace-tfshq\" (UID: \"4cb92e7d-83db-48a9-ba3b-937bf14c900f\") " pod="openshift-marketplace/redhat-marketplace-tfshq" Dec 10 23:14:31 crc kubenswrapper[4791]: I1210 23:14:31.554449 4791 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/4cb92e7d-83db-48a9-ba3b-937bf14c900f-utilities\") pod \"redhat-marketplace-tfshq\" (UID: \"4cb92e7d-83db-48a9-ba3b-937bf14c900f\") " pod="openshift-marketplace/redhat-marketplace-tfshq" Dec 10 23:14:31 crc kubenswrapper[4791]: I1210 23:14:31.554526 4791 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/4cb92e7d-83db-48a9-ba3b-937bf14c900f-catalog-content\") pod \"redhat-marketplace-tfshq\" (UID: \"4cb92e7d-83db-48a9-ba3b-937bf14c900f\") " pod="openshift-marketplace/redhat-marketplace-tfshq" Dec 10 23:14:31 crc kubenswrapper[4791]: I1210 23:14:31.656120 4791 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-2pkzw\" (UniqueName: \"kubernetes.io/projected/4cb92e7d-83db-48a9-ba3b-937bf14c900f-kube-api-access-2pkzw\") pod \"redhat-marketplace-tfshq\" (UID: \"4cb92e7d-83db-48a9-ba3b-937bf14c900f\") " pod="openshift-marketplace/redhat-marketplace-tfshq" Dec 10 23:14:31 crc kubenswrapper[4791]: I1210 23:14:31.656449 4791 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/4cb92e7d-83db-48a9-ba3b-937bf14c900f-utilities\") pod \"redhat-marketplace-tfshq\" (UID: \"4cb92e7d-83db-48a9-ba3b-937bf14c900f\") " pod="openshift-marketplace/redhat-marketplace-tfshq" Dec 10 23:14:31 crc kubenswrapper[4791]: I1210 23:14:31.656572 4791 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/4cb92e7d-83db-48a9-ba3b-937bf14c900f-catalog-content\") pod \"redhat-marketplace-tfshq\" (UID: \"4cb92e7d-83db-48a9-ba3b-937bf14c900f\") " pod="openshift-marketplace/redhat-marketplace-tfshq" Dec 10 23:14:31 crc kubenswrapper[4791]: I1210 23:14:31.657088 4791 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/4cb92e7d-83db-48a9-ba3b-937bf14c900f-utilities\") pod \"redhat-marketplace-tfshq\" (UID: \"4cb92e7d-83db-48a9-ba3b-937bf14c900f\") " pod="openshift-marketplace/redhat-marketplace-tfshq" Dec 10 23:14:31 crc kubenswrapper[4791]: I1210 23:14:31.657143 4791 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/4cb92e7d-83db-48a9-ba3b-937bf14c900f-catalog-content\") pod \"redhat-marketplace-tfshq\" (UID: \"4cb92e7d-83db-48a9-ba3b-937bf14c900f\") " pod="openshift-marketplace/redhat-marketplace-tfshq" Dec 10 23:14:31 crc kubenswrapper[4791]: I1210 23:14:31.690715 4791 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-2pkzw\" (UniqueName: \"kubernetes.io/projected/4cb92e7d-83db-48a9-ba3b-937bf14c900f-kube-api-access-2pkzw\") pod \"redhat-marketplace-tfshq\" (UID: \"4cb92e7d-83db-48a9-ba3b-937bf14c900f\") " pod="openshift-marketplace/redhat-marketplace-tfshq" Dec 10 23:14:31 crc kubenswrapper[4791]: I1210 23:14:31.773191 4791 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-tfshq" Dec 10 23:14:32 crc kubenswrapper[4791]: I1210 23:14:32.289937 4791 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-tfshq"] Dec 10 23:14:32 crc kubenswrapper[4791]: I1210 23:14:32.350297 4791 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-tfshq" event={"ID":"4cb92e7d-83db-48a9-ba3b-937bf14c900f","Type":"ContainerStarted","Data":"5f5c0091297d4b2536a5d2d4cdb50f498ea15b7c3e7d22872643c214ac7fecb5"} Dec 10 23:14:33 crc kubenswrapper[4791]: I1210 23:14:33.363178 4791 generic.go:334] "Generic (PLEG): container finished" podID="4cb92e7d-83db-48a9-ba3b-937bf14c900f" containerID="e4fe7f3efbaa80ccc63cb354f0d936b574267d523bbf454ecf2550e7d9054fca" exitCode=0 Dec 10 23:14:33 crc kubenswrapper[4791]: I1210 23:14:33.363230 4791 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-tfshq" event={"ID":"4cb92e7d-83db-48a9-ba3b-937bf14c900f","Type":"ContainerDied","Data":"e4fe7f3efbaa80ccc63cb354f0d936b574267d523bbf454ecf2550e7d9054fca"} Dec 10 23:14:34 crc kubenswrapper[4791]: I1210 23:14:34.377666 4791 generic.go:334] "Generic (PLEG): container finished" podID="4cb92e7d-83db-48a9-ba3b-937bf14c900f" containerID="98d19cda6d8a497a0ce8fe14a859acec95863931a66e6c23a7c2a3cfe854128b" exitCode=0 Dec 10 23:14:34 crc kubenswrapper[4791]: I1210 23:14:34.377745 4791 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-tfshq" event={"ID":"4cb92e7d-83db-48a9-ba3b-937bf14c900f","Type":"ContainerDied","Data":"98d19cda6d8a497a0ce8fe14a859acec95863931a66e6c23a7c2a3cfe854128b"} Dec 10 23:14:35 crc kubenswrapper[4791]: I1210 23:14:35.390147 4791 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-tfshq" event={"ID":"4cb92e7d-83db-48a9-ba3b-937bf14c900f","Type":"ContainerStarted","Data":"1e75acaf17025dae2e878ab738fd7aa7f2e7703f4b8abfa87f3ae1e043fdb8f1"} Dec 10 23:14:41 crc kubenswrapper[4791]: I1210 23:14:41.773933 4791 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-marketplace-tfshq" Dec 10 23:14:41 crc kubenswrapper[4791]: I1210 23:14:41.775201 4791 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-marketplace-tfshq" Dec 10 23:14:41 crc kubenswrapper[4791]: I1210 23:14:41.825598 4791 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-marketplace-tfshq" Dec 10 23:14:41 crc kubenswrapper[4791]: I1210 23:14:41.845240 4791 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-marketplace-tfshq" podStartSLOduration=9.372093856 podStartE2EDuration="10.845224209s" podCreationTimestamp="2025-12-10 23:14:31 +0000 UTC" firstStartedPulling="2025-12-10 23:14:33.365436305 +0000 UTC m=+1507.795053918" lastFinishedPulling="2025-12-10 23:14:34.838566648 +0000 UTC m=+1509.268184271" observedRunningTime="2025-12-10 23:14:35.412038703 +0000 UTC m=+1509.841656356" watchObservedRunningTime="2025-12-10 23:14:41.845224209 +0000 UTC m=+1516.274841822" Dec 10 23:14:42 crc kubenswrapper[4791]: I1210 23:14:42.526609 4791 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-marketplace-tfshq" Dec 10 23:14:42 crc kubenswrapper[4791]: I1210 23:14:42.584804 4791 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-tfshq"] Dec 10 23:14:44 crc kubenswrapper[4791]: I1210 23:14:44.483437 4791 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-marketplace-tfshq" podUID="4cb92e7d-83db-48a9-ba3b-937bf14c900f" containerName="registry-server" containerID="cri-o://1e75acaf17025dae2e878ab738fd7aa7f2e7703f4b8abfa87f3ae1e043fdb8f1" gracePeriod=2 Dec 10 23:14:45 crc kubenswrapper[4791]: I1210 23:14:45.512425 4791 generic.go:334] "Generic (PLEG): container finished" podID="4cb92e7d-83db-48a9-ba3b-937bf14c900f" containerID="1e75acaf17025dae2e878ab738fd7aa7f2e7703f4b8abfa87f3ae1e043fdb8f1" exitCode=0 Dec 10 23:14:45 crc kubenswrapper[4791]: I1210 23:14:45.512472 4791 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-tfshq" event={"ID":"4cb92e7d-83db-48a9-ba3b-937bf14c900f","Type":"ContainerDied","Data":"1e75acaf17025dae2e878ab738fd7aa7f2e7703f4b8abfa87f3ae1e043fdb8f1"} Dec 10 23:14:46 crc kubenswrapper[4791]: I1210 23:14:46.167523 4791 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-tfshq" Dec 10 23:14:46 crc kubenswrapper[4791]: I1210 23:14:46.360972 4791 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-2pkzw\" (UniqueName: \"kubernetes.io/projected/4cb92e7d-83db-48a9-ba3b-937bf14c900f-kube-api-access-2pkzw\") pod \"4cb92e7d-83db-48a9-ba3b-937bf14c900f\" (UID: \"4cb92e7d-83db-48a9-ba3b-937bf14c900f\") " Dec 10 23:14:46 crc kubenswrapper[4791]: I1210 23:14:46.361205 4791 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/4cb92e7d-83db-48a9-ba3b-937bf14c900f-utilities\") pod \"4cb92e7d-83db-48a9-ba3b-937bf14c900f\" (UID: \"4cb92e7d-83db-48a9-ba3b-937bf14c900f\") " Dec 10 23:14:46 crc kubenswrapper[4791]: I1210 23:14:46.361247 4791 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/4cb92e7d-83db-48a9-ba3b-937bf14c900f-catalog-content\") pod \"4cb92e7d-83db-48a9-ba3b-937bf14c900f\" (UID: \"4cb92e7d-83db-48a9-ba3b-937bf14c900f\") " Dec 10 23:14:46 crc kubenswrapper[4791]: I1210 23:14:46.363301 4791 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/4cb92e7d-83db-48a9-ba3b-937bf14c900f-utilities" (OuterVolumeSpecName: "utilities") pod "4cb92e7d-83db-48a9-ba3b-937bf14c900f" (UID: "4cb92e7d-83db-48a9-ba3b-937bf14c900f"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 10 23:14:46 crc kubenswrapper[4791]: I1210 23:14:46.368702 4791 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/4cb92e7d-83db-48a9-ba3b-937bf14c900f-kube-api-access-2pkzw" (OuterVolumeSpecName: "kube-api-access-2pkzw") pod "4cb92e7d-83db-48a9-ba3b-937bf14c900f" (UID: "4cb92e7d-83db-48a9-ba3b-937bf14c900f"). InnerVolumeSpecName "kube-api-access-2pkzw". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 10 23:14:46 crc kubenswrapper[4791]: I1210 23:14:46.391377 4791 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/4cb92e7d-83db-48a9-ba3b-937bf14c900f-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "4cb92e7d-83db-48a9-ba3b-937bf14c900f" (UID: "4cb92e7d-83db-48a9-ba3b-937bf14c900f"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 10 23:14:46 crc kubenswrapper[4791]: I1210 23:14:46.464357 4791 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/4cb92e7d-83db-48a9-ba3b-937bf14c900f-utilities\") on node \"crc\" DevicePath \"\"" Dec 10 23:14:46 crc kubenswrapper[4791]: I1210 23:14:46.464396 4791 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/4cb92e7d-83db-48a9-ba3b-937bf14c900f-catalog-content\") on node \"crc\" DevicePath \"\"" Dec 10 23:14:46 crc kubenswrapper[4791]: I1210 23:14:46.464411 4791 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-2pkzw\" (UniqueName: \"kubernetes.io/projected/4cb92e7d-83db-48a9-ba3b-937bf14c900f-kube-api-access-2pkzw\") on node \"crc\" DevicePath \"\"" Dec 10 23:14:46 crc kubenswrapper[4791]: I1210 23:14:46.528138 4791 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-tfshq" event={"ID":"4cb92e7d-83db-48a9-ba3b-937bf14c900f","Type":"ContainerDied","Data":"5f5c0091297d4b2536a5d2d4cdb50f498ea15b7c3e7d22872643c214ac7fecb5"} Dec 10 23:14:46 crc kubenswrapper[4791]: I1210 23:14:46.528226 4791 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-tfshq" Dec 10 23:14:46 crc kubenswrapper[4791]: I1210 23:14:46.528242 4791 scope.go:117] "RemoveContainer" containerID="1e75acaf17025dae2e878ab738fd7aa7f2e7703f4b8abfa87f3ae1e043fdb8f1" Dec 10 23:14:46 crc kubenswrapper[4791]: I1210 23:14:46.557464 4791 scope.go:117] "RemoveContainer" containerID="98d19cda6d8a497a0ce8fe14a859acec95863931a66e6c23a7c2a3cfe854128b" Dec 10 23:14:46 crc kubenswrapper[4791]: I1210 23:14:46.596253 4791 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-tfshq"] Dec 10 23:14:46 crc kubenswrapper[4791]: I1210 23:14:46.597058 4791 scope.go:117] "RemoveContainer" containerID="e4fe7f3efbaa80ccc63cb354f0d936b574267d523bbf454ecf2550e7d9054fca" Dec 10 23:14:46 crc kubenswrapper[4791]: I1210 23:14:46.611404 4791 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-marketplace-tfshq"] Dec 10 23:14:47 crc kubenswrapper[4791]: I1210 23:14:47.899718 4791 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="4cb92e7d-83db-48a9-ba3b-937bf14c900f" path="/var/lib/kubelet/pods/4cb92e7d-83db-48a9-ba3b-937bf14c900f/volumes" Dec 10 23:14:51 crc kubenswrapper[4791]: I1210 23:14:51.380663 4791 scope.go:117] "RemoveContainer" containerID="07fb0486c1e3fbfccc4819aa21c7ffea14c6bbcfde3f6c560ded3b5e60ee75a1" Dec 10 23:14:55 crc kubenswrapper[4791]: I1210 23:14:55.038302 4791 patch_prober.go:28] interesting pod/machine-config-daemon-5rb5l container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 10 23:14:55 crc kubenswrapper[4791]: I1210 23:14:55.039293 4791 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-5rb5l" podUID="ba35653c-6e06-4cee-a4d6-137764090d18" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 10 23:15:00 crc kubenswrapper[4791]: I1210 23:15:00.167745 4791 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29423475-l2jwz"] Dec 10 23:15:00 crc kubenswrapper[4791]: E1210 23:15:00.168901 4791 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="4cb92e7d-83db-48a9-ba3b-937bf14c900f" containerName="extract-content" Dec 10 23:15:00 crc kubenswrapper[4791]: I1210 23:15:00.168921 4791 state_mem.go:107] "Deleted CPUSet assignment" podUID="4cb92e7d-83db-48a9-ba3b-937bf14c900f" containerName="extract-content" Dec 10 23:15:00 crc kubenswrapper[4791]: E1210 23:15:00.168950 4791 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="4cb92e7d-83db-48a9-ba3b-937bf14c900f" containerName="extract-utilities" Dec 10 23:15:00 crc kubenswrapper[4791]: I1210 23:15:00.168958 4791 state_mem.go:107] "Deleted CPUSet assignment" podUID="4cb92e7d-83db-48a9-ba3b-937bf14c900f" containerName="extract-utilities" Dec 10 23:15:00 crc kubenswrapper[4791]: E1210 23:15:00.168982 4791 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="4cb92e7d-83db-48a9-ba3b-937bf14c900f" containerName="registry-server" Dec 10 23:15:00 crc kubenswrapper[4791]: I1210 23:15:00.168989 4791 state_mem.go:107] "Deleted CPUSet assignment" podUID="4cb92e7d-83db-48a9-ba3b-937bf14c900f" containerName="registry-server" Dec 10 23:15:00 crc kubenswrapper[4791]: I1210 23:15:00.169224 4791 memory_manager.go:354] "RemoveStaleState removing state" podUID="4cb92e7d-83db-48a9-ba3b-937bf14c900f" containerName="registry-server" Dec 10 23:15:00 crc kubenswrapper[4791]: I1210 23:15:00.170047 4791 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29423475-l2jwz" Dec 10 23:15:00 crc kubenswrapper[4791]: I1210 23:15:00.172190 4791 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-operator-lifecycle-manager"/"collect-profiles-config" Dec 10 23:15:00 crc kubenswrapper[4791]: I1210 23:15:00.177985 4791 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29423475-l2jwz"] Dec 10 23:15:00 crc kubenswrapper[4791]: I1210 23:15:00.179282 4791 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"collect-profiles-dockercfg-kzf4t" Dec 10 23:15:00 crc kubenswrapper[4791]: I1210 23:15:00.237437 4791 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/ce778cac-da93-4e3e-89ed-769568cb4b52-secret-volume\") pod \"collect-profiles-29423475-l2jwz\" (UID: \"ce778cac-da93-4e3e-89ed-769568cb4b52\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29423475-l2jwz" Dec 10 23:15:00 crc kubenswrapper[4791]: I1210 23:15:00.237846 4791 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/ce778cac-da93-4e3e-89ed-769568cb4b52-config-volume\") pod \"collect-profiles-29423475-l2jwz\" (UID: \"ce778cac-da93-4e3e-89ed-769568cb4b52\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29423475-l2jwz" Dec 10 23:15:00 crc kubenswrapper[4791]: I1210 23:15:00.237958 4791 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-l9pfv\" (UniqueName: \"kubernetes.io/projected/ce778cac-da93-4e3e-89ed-769568cb4b52-kube-api-access-l9pfv\") pod \"collect-profiles-29423475-l2jwz\" (UID: \"ce778cac-da93-4e3e-89ed-769568cb4b52\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29423475-l2jwz" Dec 10 23:15:00 crc kubenswrapper[4791]: I1210 23:15:00.340215 4791 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-l9pfv\" (UniqueName: \"kubernetes.io/projected/ce778cac-da93-4e3e-89ed-769568cb4b52-kube-api-access-l9pfv\") pod \"collect-profiles-29423475-l2jwz\" (UID: \"ce778cac-da93-4e3e-89ed-769568cb4b52\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29423475-l2jwz" Dec 10 23:15:00 crc kubenswrapper[4791]: I1210 23:15:00.340354 4791 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/ce778cac-da93-4e3e-89ed-769568cb4b52-secret-volume\") pod \"collect-profiles-29423475-l2jwz\" (UID: \"ce778cac-da93-4e3e-89ed-769568cb4b52\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29423475-l2jwz" Dec 10 23:15:00 crc kubenswrapper[4791]: I1210 23:15:00.340382 4791 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/ce778cac-da93-4e3e-89ed-769568cb4b52-config-volume\") pod \"collect-profiles-29423475-l2jwz\" (UID: \"ce778cac-da93-4e3e-89ed-769568cb4b52\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29423475-l2jwz" Dec 10 23:15:00 crc kubenswrapper[4791]: I1210 23:15:00.341238 4791 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/ce778cac-da93-4e3e-89ed-769568cb4b52-config-volume\") pod \"collect-profiles-29423475-l2jwz\" (UID: \"ce778cac-da93-4e3e-89ed-769568cb4b52\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29423475-l2jwz" Dec 10 23:15:00 crc kubenswrapper[4791]: I1210 23:15:00.351348 4791 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/ce778cac-da93-4e3e-89ed-769568cb4b52-secret-volume\") pod \"collect-profiles-29423475-l2jwz\" (UID: \"ce778cac-da93-4e3e-89ed-769568cb4b52\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29423475-l2jwz" Dec 10 23:15:00 crc kubenswrapper[4791]: I1210 23:15:00.358025 4791 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-l9pfv\" (UniqueName: \"kubernetes.io/projected/ce778cac-da93-4e3e-89ed-769568cb4b52-kube-api-access-l9pfv\") pod \"collect-profiles-29423475-l2jwz\" (UID: \"ce778cac-da93-4e3e-89ed-769568cb4b52\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29423475-l2jwz" Dec 10 23:15:00 crc kubenswrapper[4791]: I1210 23:15:00.500682 4791 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29423475-l2jwz" Dec 10 23:15:00 crc kubenswrapper[4791]: I1210 23:15:00.981158 4791 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29423475-l2jwz"] Dec 10 23:15:01 crc kubenswrapper[4791]: I1210 23:15:01.697565 4791 generic.go:334] "Generic (PLEG): container finished" podID="ce778cac-da93-4e3e-89ed-769568cb4b52" containerID="4f8e348e6239dcaf799551ffbdf8449ca88dd21d073c94cf6252f32b69d9ae1f" exitCode=0 Dec 10 23:15:01 crc kubenswrapper[4791]: I1210 23:15:01.697658 4791 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29423475-l2jwz" event={"ID":"ce778cac-da93-4e3e-89ed-769568cb4b52","Type":"ContainerDied","Data":"4f8e348e6239dcaf799551ffbdf8449ca88dd21d073c94cf6252f32b69d9ae1f"} Dec 10 23:15:01 crc kubenswrapper[4791]: I1210 23:15:01.699613 4791 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29423475-l2jwz" event={"ID":"ce778cac-da93-4e3e-89ed-769568cb4b52","Type":"ContainerStarted","Data":"2c14c5a069084eacd1f2d02854ecf8243531da6814cc8d7a78b63392a147c3ec"} Dec 10 23:15:03 crc kubenswrapper[4791]: I1210 23:15:03.087794 4791 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29423475-l2jwz" Dec 10 23:15:03 crc kubenswrapper[4791]: I1210 23:15:03.192039 4791 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/ce778cac-da93-4e3e-89ed-769568cb4b52-config-volume\") pod \"ce778cac-da93-4e3e-89ed-769568cb4b52\" (UID: \"ce778cac-da93-4e3e-89ed-769568cb4b52\") " Dec 10 23:15:03 crc kubenswrapper[4791]: I1210 23:15:03.192318 4791 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-l9pfv\" (UniqueName: \"kubernetes.io/projected/ce778cac-da93-4e3e-89ed-769568cb4b52-kube-api-access-l9pfv\") pod \"ce778cac-da93-4e3e-89ed-769568cb4b52\" (UID: \"ce778cac-da93-4e3e-89ed-769568cb4b52\") " Dec 10 23:15:03 crc kubenswrapper[4791]: I1210 23:15:03.192419 4791 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/ce778cac-da93-4e3e-89ed-769568cb4b52-secret-volume\") pod \"ce778cac-da93-4e3e-89ed-769568cb4b52\" (UID: \"ce778cac-da93-4e3e-89ed-769568cb4b52\") " Dec 10 23:15:03 crc kubenswrapper[4791]: I1210 23:15:03.193285 4791 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/ce778cac-da93-4e3e-89ed-769568cb4b52-config-volume" (OuterVolumeSpecName: "config-volume") pod "ce778cac-da93-4e3e-89ed-769568cb4b52" (UID: "ce778cac-da93-4e3e-89ed-769568cb4b52"). InnerVolumeSpecName "config-volume". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 10 23:15:03 crc kubenswrapper[4791]: I1210 23:15:03.195024 4791 reconciler_common.go:293] "Volume detached for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/ce778cac-da93-4e3e-89ed-769568cb4b52-config-volume\") on node \"crc\" DevicePath \"\"" Dec 10 23:15:03 crc kubenswrapper[4791]: I1210 23:15:03.215347 4791 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/ce778cac-da93-4e3e-89ed-769568cb4b52-secret-volume" (OuterVolumeSpecName: "secret-volume") pod "ce778cac-da93-4e3e-89ed-769568cb4b52" (UID: "ce778cac-da93-4e3e-89ed-769568cb4b52"). InnerVolumeSpecName "secret-volume". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 10 23:15:03 crc kubenswrapper[4791]: I1210 23:15:03.215419 4791 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/ce778cac-da93-4e3e-89ed-769568cb4b52-kube-api-access-l9pfv" (OuterVolumeSpecName: "kube-api-access-l9pfv") pod "ce778cac-da93-4e3e-89ed-769568cb4b52" (UID: "ce778cac-da93-4e3e-89ed-769568cb4b52"). InnerVolumeSpecName "kube-api-access-l9pfv". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 10 23:15:03 crc kubenswrapper[4791]: I1210 23:15:03.296974 4791 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-l9pfv\" (UniqueName: \"kubernetes.io/projected/ce778cac-da93-4e3e-89ed-769568cb4b52-kube-api-access-l9pfv\") on node \"crc\" DevicePath \"\"" Dec 10 23:15:03 crc kubenswrapper[4791]: I1210 23:15:03.297009 4791 reconciler_common.go:293] "Volume detached for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/ce778cac-da93-4e3e-89ed-769568cb4b52-secret-volume\") on node \"crc\" DevicePath \"\"" Dec 10 23:15:03 crc kubenswrapper[4791]: I1210 23:15:03.724362 4791 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29423475-l2jwz" event={"ID":"ce778cac-da93-4e3e-89ed-769568cb4b52","Type":"ContainerDied","Data":"2c14c5a069084eacd1f2d02854ecf8243531da6814cc8d7a78b63392a147c3ec"} Dec 10 23:15:03 crc kubenswrapper[4791]: I1210 23:15:03.724420 4791 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="2c14c5a069084eacd1f2d02854ecf8243531da6814cc8d7a78b63392a147c3ec" Dec 10 23:15:03 crc kubenswrapper[4791]: I1210 23:15:03.724457 4791 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29423475-l2jwz" Dec 10 23:15:08 crc kubenswrapper[4791]: I1210 23:15:08.619068 4791 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/certified-operators-658k2"] Dec 10 23:15:08 crc kubenswrapper[4791]: E1210 23:15:08.621608 4791 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="ce778cac-da93-4e3e-89ed-769568cb4b52" containerName="collect-profiles" Dec 10 23:15:08 crc kubenswrapper[4791]: I1210 23:15:08.621738 4791 state_mem.go:107] "Deleted CPUSet assignment" podUID="ce778cac-da93-4e3e-89ed-769568cb4b52" containerName="collect-profiles" Dec 10 23:15:08 crc kubenswrapper[4791]: I1210 23:15:08.622333 4791 memory_manager.go:354] "RemoveStaleState removing state" podUID="ce778cac-da93-4e3e-89ed-769568cb4b52" containerName="collect-profiles" Dec 10 23:15:08 crc kubenswrapper[4791]: I1210 23:15:08.624906 4791 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-658k2" Dec 10 23:15:08 crc kubenswrapper[4791]: I1210 23:15:08.631544 4791 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-658k2"] Dec 10 23:15:08 crc kubenswrapper[4791]: I1210 23:15:08.699672 4791 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-6rlqj\" (UniqueName: \"kubernetes.io/projected/3c0bf306-a118-4131-9888-17cb46b832ac-kube-api-access-6rlqj\") pod \"certified-operators-658k2\" (UID: \"3c0bf306-a118-4131-9888-17cb46b832ac\") " pod="openshift-marketplace/certified-operators-658k2" Dec 10 23:15:08 crc kubenswrapper[4791]: I1210 23:15:08.699931 4791 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/3c0bf306-a118-4131-9888-17cb46b832ac-catalog-content\") pod \"certified-operators-658k2\" (UID: \"3c0bf306-a118-4131-9888-17cb46b832ac\") " pod="openshift-marketplace/certified-operators-658k2" Dec 10 23:15:08 crc kubenswrapper[4791]: I1210 23:15:08.700034 4791 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/3c0bf306-a118-4131-9888-17cb46b832ac-utilities\") pod \"certified-operators-658k2\" (UID: \"3c0bf306-a118-4131-9888-17cb46b832ac\") " pod="openshift-marketplace/certified-operators-658k2" Dec 10 23:15:08 crc kubenswrapper[4791]: I1210 23:15:08.801889 4791 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/3c0bf306-a118-4131-9888-17cb46b832ac-utilities\") pod \"certified-operators-658k2\" (UID: \"3c0bf306-a118-4131-9888-17cb46b832ac\") " pod="openshift-marketplace/certified-operators-658k2" Dec 10 23:15:08 crc kubenswrapper[4791]: I1210 23:15:08.802099 4791 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-6rlqj\" (UniqueName: \"kubernetes.io/projected/3c0bf306-a118-4131-9888-17cb46b832ac-kube-api-access-6rlqj\") pod \"certified-operators-658k2\" (UID: \"3c0bf306-a118-4131-9888-17cb46b832ac\") " pod="openshift-marketplace/certified-operators-658k2" Dec 10 23:15:08 crc kubenswrapper[4791]: I1210 23:15:08.802123 4791 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/3c0bf306-a118-4131-9888-17cb46b832ac-catalog-content\") pod \"certified-operators-658k2\" (UID: \"3c0bf306-a118-4131-9888-17cb46b832ac\") " pod="openshift-marketplace/certified-operators-658k2" Dec 10 23:15:08 crc kubenswrapper[4791]: I1210 23:15:08.802732 4791 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/3c0bf306-a118-4131-9888-17cb46b832ac-catalog-content\") pod \"certified-operators-658k2\" (UID: \"3c0bf306-a118-4131-9888-17cb46b832ac\") " pod="openshift-marketplace/certified-operators-658k2" Dec 10 23:15:08 crc kubenswrapper[4791]: I1210 23:15:08.802945 4791 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/3c0bf306-a118-4131-9888-17cb46b832ac-utilities\") pod \"certified-operators-658k2\" (UID: \"3c0bf306-a118-4131-9888-17cb46b832ac\") " pod="openshift-marketplace/certified-operators-658k2" Dec 10 23:15:08 crc kubenswrapper[4791]: I1210 23:15:08.821744 4791 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-6rlqj\" (UniqueName: \"kubernetes.io/projected/3c0bf306-a118-4131-9888-17cb46b832ac-kube-api-access-6rlqj\") pod \"certified-operators-658k2\" (UID: \"3c0bf306-a118-4131-9888-17cb46b832ac\") " pod="openshift-marketplace/certified-operators-658k2" Dec 10 23:15:08 crc kubenswrapper[4791]: I1210 23:15:08.957728 4791 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-658k2" Dec 10 23:15:09 crc kubenswrapper[4791]: I1210 23:15:09.443630 4791 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-658k2"] Dec 10 23:15:09 crc kubenswrapper[4791]: I1210 23:15:09.786269 4791 generic.go:334] "Generic (PLEG): container finished" podID="3c0bf306-a118-4131-9888-17cb46b832ac" containerID="d69cecbe0cf9ac8005b5fa04c7ec40cdfdc06981ffa2e667ffb1e36e6954161a" exitCode=0 Dec 10 23:15:09 crc kubenswrapper[4791]: I1210 23:15:09.786478 4791 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-658k2" event={"ID":"3c0bf306-a118-4131-9888-17cb46b832ac","Type":"ContainerDied","Data":"d69cecbe0cf9ac8005b5fa04c7ec40cdfdc06981ffa2e667ffb1e36e6954161a"} Dec 10 23:15:09 crc kubenswrapper[4791]: I1210 23:15:09.786668 4791 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-658k2" event={"ID":"3c0bf306-a118-4131-9888-17cb46b832ac","Type":"ContainerStarted","Data":"e29ecbe1e6a8f8b33f8c56e642091c475c866ea90221ba1783aa6774e56e4a3f"} Dec 10 23:15:17 crc kubenswrapper[4791]: I1210 23:15:17.872782 4791 generic.go:334] "Generic (PLEG): container finished" podID="3c0bf306-a118-4131-9888-17cb46b832ac" containerID="c0a0a94f1f423d22c3b351631f5690046b276000a4a13392f21a648fb53856a2" exitCode=0 Dec 10 23:15:17 crc kubenswrapper[4791]: I1210 23:15:17.872866 4791 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-658k2" event={"ID":"3c0bf306-a118-4131-9888-17cb46b832ac","Type":"ContainerDied","Data":"c0a0a94f1f423d22c3b351631f5690046b276000a4a13392f21a648fb53856a2"} Dec 10 23:15:18 crc kubenswrapper[4791]: I1210 23:15:18.888946 4791 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-658k2" event={"ID":"3c0bf306-a118-4131-9888-17cb46b832ac","Type":"ContainerStarted","Data":"e125b0c8405a37578f72d03da423f5dd68e703bcae56bf123ed904bf0a200b0c"} Dec 10 23:15:18 crc kubenswrapper[4791]: I1210 23:15:18.930453 4791 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/certified-operators-658k2" podStartSLOduration=2.379858346 podStartE2EDuration="10.930411903s" podCreationTimestamp="2025-12-10 23:15:08 +0000 UTC" firstStartedPulling="2025-12-10 23:15:09.790695599 +0000 UTC m=+1544.220313212" lastFinishedPulling="2025-12-10 23:15:18.341249136 +0000 UTC m=+1552.770866769" observedRunningTime="2025-12-10 23:15:18.923965302 +0000 UTC m=+1553.353582915" watchObservedRunningTime="2025-12-10 23:15:18.930411903 +0000 UTC m=+1553.360029566" Dec 10 23:15:18 crc kubenswrapper[4791]: I1210 23:15:18.958153 4791 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/certified-operators-658k2" Dec 10 23:15:18 crc kubenswrapper[4791]: I1210 23:15:18.958312 4791 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/certified-operators-658k2" Dec 10 23:15:20 crc kubenswrapper[4791]: I1210 23:15:20.013076 4791 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-marketplace/certified-operators-658k2" podUID="3c0bf306-a118-4131-9888-17cb46b832ac" containerName="registry-server" probeResult="failure" output=< Dec 10 23:15:20 crc kubenswrapper[4791]: timeout: failed to connect service ":50051" within 1s Dec 10 23:15:20 crc kubenswrapper[4791]: > Dec 10 23:15:25 crc kubenswrapper[4791]: I1210 23:15:25.038045 4791 patch_prober.go:28] interesting pod/machine-config-daemon-5rb5l container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 10 23:15:25 crc kubenswrapper[4791]: I1210 23:15:25.039593 4791 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-5rb5l" podUID="ba35653c-6e06-4cee-a4d6-137764090d18" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 10 23:15:25 crc kubenswrapper[4791]: I1210 23:15:25.039716 4791 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-5rb5l" Dec 10 23:15:25 crc kubenswrapper[4791]: I1210 23:15:25.040528 4791 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"f12779a795c9f6f63a93027d53c67c89d0179d87b9df554c25bfc607f06caff1"} pod="openshift-machine-config-operator/machine-config-daemon-5rb5l" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Dec 10 23:15:25 crc kubenswrapper[4791]: I1210 23:15:25.040699 4791 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-5rb5l" podUID="ba35653c-6e06-4cee-a4d6-137764090d18" containerName="machine-config-daemon" containerID="cri-o://f12779a795c9f6f63a93027d53c67c89d0179d87b9df554c25bfc607f06caff1" gracePeriod=600 Dec 10 23:15:25 crc kubenswrapper[4791]: E1210 23:15:25.167812 4791 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-5rb5l_openshift-machine-config-operator(ba35653c-6e06-4cee-a4d6-137764090d18)\"" pod="openshift-machine-config-operator/machine-config-daemon-5rb5l" podUID="ba35653c-6e06-4cee-a4d6-137764090d18" Dec 10 23:15:25 crc kubenswrapper[4791]: I1210 23:15:25.969595 4791 generic.go:334] "Generic (PLEG): container finished" podID="ba35653c-6e06-4cee-a4d6-137764090d18" containerID="f12779a795c9f6f63a93027d53c67c89d0179d87b9df554c25bfc607f06caff1" exitCode=0 Dec 10 23:15:25 crc kubenswrapper[4791]: I1210 23:15:25.969648 4791 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-5rb5l" event={"ID":"ba35653c-6e06-4cee-a4d6-137764090d18","Type":"ContainerDied","Data":"f12779a795c9f6f63a93027d53c67c89d0179d87b9df554c25bfc607f06caff1"} Dec 10 23:15:25 crc kubenswrapper[4791]: I1210 23:15:25.969688 4791 scope.go:117] "RemoveContainer" containerID="27ea6396cad6699de1e6134e18b2c18010b7367f4d673a0bdef71d4d1c354c04" Dec 10 23:15:25 crc kubenswrapper[4791]: I1210 23:15:25.973767 4791 scope.go:117] "RemoveContainer" containerID="f12779a795c9f6f63a93027d53c67c89d0179d87b9df554c25bfc607f06caff1" Dec 10 23:15:25 crc kubenswrapper[4791]: E1210 23:15:25.974135 4791 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-5rb5l_openshift-machine-config-operator(ba35653c-6e06-4cee-a4d6-137764090d18)\"" pod="openshift-machine-config-operator/machine-config-daemon-5rb5l" podUID="ba35653c-6e06-4cee-a4d6-137764090d18" Dec 10 23:15:29 crc kubenswrapper[4791]: I1210 23:15:29.037951 4791 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/certified-operators-658k2" Dec 10 23:15:29 crc kubenswrapper[4791]: I1210 23:15:29.082766 4791 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/certified-operators-658k2" Dec 10 23:15:29 crc kubenswrapper[4791]: I1210 23:15:29.156590 4791 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-658k2"] Dec 10 23:15:29 crc kubenswrapper[4791]: I1210 23:15:29.286617 4791 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-ttrzg"] Dec 10 23:15:29 crc kubenswrapper[4791]: I1210 23:15:29.286904 4791 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/certified-operators-ttrzg" podUID="f00159a6-00e0-4975-87a0-9d8b2b67b157" containerName="registry-server" containerID="cri-o://a510617cedd04f0ed52a6e86f460ffc4061bdd0b9f5d38be528eaebf33162a4e" gracePeriod=2 Dec 10 23:15:29 crc kubenswrapper[4791]: E1210 23:15:29.847404 4791 cadvisor_stats_provider.go:516] "Partial failure issuing cadvisor.ContainerInfoV2" err="partial failures: [\"/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podf00159a6_00e0_4975_87a0_9d8b2b67b157.slice/crio-conmon-a510617cedd04f0ed52a6e86f460ffc4061bdd0b9f5d38be528eaebf33162a4e.scope\": RecentStats: unable to find data in memory cache]" Dec 10 23:15:30 crc kubenswrapper[4791]: I1210 23:15:30.010740 4791 generic.go:334] "Generic (PLEG): container finished" podID="f00159a6-00e0-4975-87a0-9d8b2b67b157" containerID="a510617cedd04f0ed52a6e86f460ffc4061bdd0b9f5d38be528eaebf33162a4e" exitCode=0 Dec 10 23:15:30 crc kubenswrapper[4791]: I1210 23:15:30.011007 4791 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-ttrzg" event={"ID":"f00159a6-00e0-4975-87a0-9d8b2b67b157","Type":"ContainerDied","Data":"a510617cedd04f0ed52a6e86f460ffc4061bdd0b9f5d38be528eaebf33162a4e"} Dec 10 23:15:30 crc kubenswrapper[4791]: I1210 23:15:30.258651 4791 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-ttrzg" Dec 10 23:15:30 crc kubenswrapper[4791]: I1210 23:15:30.384049 4791 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/f00159a6-00e0-4975-87a0-9d8b2b67b157-utilities\") pod \"f00159a6-00e0-4975-87a0-9d8b2b67b157\" (UID: \"f00159a6-00e0-4975-87a0-9d8b2b67b157\") " Dec 10 23:15:30 crc kubenswrapper[4791]: I1210 23:15:30.384175 4791 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/f00159a6-00e0-4975-87a0-9d8b2b67b157-catalog-content\") pod \"f00159a6-00e0-4975-87a0-9d8b2b67b157\" (UID: \"f00159a6-00e0-4975-87a0-9d8b2b67b157\") " Dec 10 23:15:30 crc kubenswrapper[4791]: I1210 23:15:30.384281 4791 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-v9bhf\" (UniqueName: \"kubernetes.io/projected/f00159a6-00e0-4975-87a0-9d8b2b67b157-kube-api-access-v9bhf\") pod \"f00159a6-00e0-4975-87a0-9d8b2b67b157\" (UID: \"f00159a6-00e0-4975-87a0-9d8b2b67b157\") " Dec 10 23:15:30 crc kubenswrapper[4791]: I1210 23:15:30.384692 4791 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/f00159a6-00e0-4975-87a0-9d8b2b67b157-utilities" (OuterVolumeSpecName: "utilities") pod "f00159a6-00e0-4975-87a0-9d8b2b67b157" (UID: "f00159a6-00e0-4975-87a0-9d8b2b67b157"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 10 23:15:30 crc kubenswrapper[4791]: I1210 23:15:30.390638 4791 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/f00159a6-00e0-4975-87a0-9d8b2b67b157-kube-api-access-v9bhf" (OuterVolumeSpecName: "kube-api-access-v9bhf") pod "f00159a6-00e0-4975-87a0-9d8b2b67b157" (UID: "f00159a6-00e0-4975-87a0-9d8b2b67b157"). InnerVolumeSpecName "kube-api-access-v9bhf". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 10 23:15:30 crc kubenswrapper[4791]: I1210 23:15:30.437899 4791 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/f00159a6-00e0-4975-87a0-9d8b2b67b157-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "f00159a6-00e0-4975-87a0-9d8b2b67b157" (UID: "f00159a6-00e0-4975-87a0-9d8b2b67b157"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 10 23:15:30 crc kubenswrapper[4791]: I1210 23:15:30.486911 4791 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/f00159a6-00e0-4975-87a0-9d8b2b67b157-utilities\") on node \"crc\" DevicePath \"\"" Dec 10 23:15:30 crc kubenswrapper[4791]: I1210 23:15:30.486941 4791 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/f00159a6-00e0-4975-87a0-9d8b2b67b157-catalog-content\") on node \"crc\" DevicePath \"\"" Dec 10 23:15:30 crc kubenswrapper[4791]: I1210 23:15:30.486953 4791 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-v9bhf\" (UniqueName: \"kubernetes.io/projected/f00159a6-00e0-4975-87a0-9d8b2b67b157-kube-api-access-v9bhf\") on node \"crc\" DevicePath \"\"" Dec 10 23:15:31 crc kubenswrapper[4791]: I1210 23:15:31.023563 4791 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-ttrzg" event={"ID":"f00159a6-00e0-4975-87a0-9d8b2b67b157","Type":"ContainerDied","Data":"a144abeebcd2dbd886fa194cfaad11b723fbdd129a7b14b88655666921aa98c9"} Dec 10 23:15:31 crc kubenswrapper[4791]: I1210 23:15:31.023900 4791 scope.go:117] "RemoveContainer" containerID="a510617cedd04f0ed52a6e86f460ffc4061bdd0b9f5d38be528eaebf33162a4e" Dec 10 23:15:31 crc kubenswrapper[4791]: I1210 23:15:31.023605 4791 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-ttrzg" Dec 10 23:15:31 crc kubenswrapper[4791]: I1210 23:15:31.067744 4791 scope.go:117] "RemoveContainer" containerID="c364bedbb065523a05d7f6e6823a440f9098a3dd7c9baff34c2b16acb2d6458d" Dec 10 23:15:31 crc kubenswrapper[4791]: I1210 23:15:31.067888 4791 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-ttrzg"] Dec 10 23:15:31 crc kubenswrapper[4791]: I1210 23:15:31.076653 4791 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/certified-operators-ttrzg"] Dec 10 23:15:31 crc kubenswrapper[4791]: I1210 23:15:31.099811 4791 scope.go:117] "RemoveContainer" containerID="03d58c5c8454efc83b466213d9227a8c922b9de1db055bcd3da961cae612f83a" Dec 10 23:15:31 crc kubenswrapper[4791]: I1210 23:15:31.899100 4791 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="f00159a6-00e0-4975-87a0-9d8b2b67b157" path="/var/lib/kubelet/pods/f00159a6-00e0-4975-87a0-9d8b2b67b157/volumes" Dec 10 23:15:38 crc kubenswrapper[4791]: I1210 23:15:38.884794 4791 scope.go:117] "RemoveContainer" containerID="f12779a795c9f6f63a93027d53c67c89d0179d87b9df554c25bfc607f06caff1" Dec 10 23:15:38 crc kubenswrapper[4791]: E1210 23:15:38.885575 4791 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-5rb5l_openshift-machine-config-operator(ba35653c-6e06-4cee-a4d6-137764090d18)\"" pod="openshift-machine-config-operator/machine-config-daemon-5rb5l" podUID="ba35653c-6e06-4cee-a4d6-137764090d18" Dec 10 23:15:51 crc kubenswrapper[4791]: I1210 23:15:51.499839 4791 scope.go:117] "RemoveContainer" containerID="3a843f3dd0a7f3fa2fe6719f2d7fcb8bb9ae41ef430830e5a912a792e2fe3c4a" Dec 10 23:15:51 crc kubenswrapper[4791]: I1210 23:15:51.531701 4791 scope.go:117] "RemoveContainer" containerID="b4f9dd4411c75b64e8ac6561f42e13b64709dd0bfd4f266b0d540210a7768a27" Dec 10 23:15:51 crc kubenswrapper[4791]: I1210 23:15:51.560464 4791 scope.go:117] "RemoveContainer" containerID="b9fbe7eb299965c84c9556998893e0f9a21817ebbe4e48d8cd0d072a837e119a" Dec 10 23:15:51 crc kubenswrapper[4791]: I1210 23:15:51.592935 4791 scope.go:117] "RemoveContainer" containerID="1cd4657d399538b1cc65c711e3e11791d84c93aef6bed6e7009f412fb56b3453" Dec 10 23:15:51 crc kubenswrapper[4791]: I1210 23:15:51.620867 4791 scope.go:117] "RemoveContainer" containerID="521f2aafeb6650480862ba7b822c1a7b87190962da901d6b233d82ced08d46f6" Dec 10 23:15:51 crc kubenswrapper[4791]: I1210 23:15:51.643271 4791 scope.go:117] "RemoveContainer" containerID="d42186dda997e7eea2ad8e2fe49e4065a0a7fa16abde74c6eb02667a3aad9b30" Dec 10 23:15:51 crc kubenswrapper[4791]: I1210 23:15:51.885235 4791 scope.go:117] "RemoveContainer" containerID="f12779a795c9f6f63a93027d53c67c89d0179d87b9df554c25bfc607f06caff1" Dec 10 23:15:51 crc kubenswrapper[4791]: E1210 23:15:51.885960 4791 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-5rb5l_openshift-machine-config-operator(ba35653c-6e06-4cee-a4d6-137764090d18)\"" pod="openshift-machine-config-operator/machine-config-daemon-5rb5l" podUID="ba35653c-6e06-4cee-a4d6-137764090d18" Dec 10 23:16:06 crc kubenswrapper[4791]: I1210 23:16:06.884948 4791 scope.go:117] "RemoveContainer" containerID="f12779a795c9f6f63a93027d53c67c89d0179d87b9df554c25bfc607f06caff1" Dec 10 23:16:06 crc kubenswrapper[4791]: E1210 23:16:06.885720 4791 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-5rb5l_openshift-machine-config-operator(ba35653c-6e06-4cee-a4d6-137764090d18)\"" pod="openshift-machine-config-operator/machine-config-daemon-5rb5l" podUID="ba35653c-6e06-4cee-a4d6-137764090d18" Dec 10 23:16:18 crc kubenswrapper[4791]: I1210 23:16:18.885086 4791 scope.go:117] "RemoveContainer" containerID="f12779a795c9f6f63a93027d53c67c89d0179d87b9df554c25bfc607f06caff1" Dec 10 23:16:18 crc kubenswrapper[4791]: E1210 23:16:18.886202 4791 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-5rb5l_openshift-machine-config-operator(ba35653c-6e06-4cee-a4d6-137764090d18)\"" pod="openshift-machine-config-operator/machine-config-daemon-5rb5l" podUID="ba35653c-6e06-4cee-a4d6-137764090d18" Dec 10 23:16:29 crc kubenswrapper[4791]: I1210 23:16:29.884672 4791 scope.go:117] "RemoveContainer" containerID="f12779a795c9f6f63a93027d53c67c89d0179d87b9df554c25bfc607f06caff1" Dec 10 23:16:29 crc kubenswrapper[4791]: E1210 23:16:29.885431 4791 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-5rb5l_openshift-machine-config-operator(ba35653c-6e06-4cee-a4d6-137764090d18)\"" pod="openshift-machine-config-operator/machine-config-daemon-5rb5l" podUID="ba35653c-6e06-4cee-a4d6-137764090d18" Dec 10 23:16:40 crc kubenswrapper[4791]: I1210 23:16:40.884577 4791 scope.go:117] "RemoveContainer" containerID="f12779a795c9f6f63a93027d53c67c89d0179d87b9df554c25bfc607f06caff1" Dec 10 23:16:40 crc kubenswrapper[4791]: E1210 23:16:40.885379 4791 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-5rb5l_openshift-machine-config-operator(ba35653c-6e06-4cee-a4d6-137764090d18)\"" pod="openshift-machine-config-operator/machine-config-daemon-5rb5l" podUID="ba35653c-6e06-4cee-a4d6-137764090d18" Dec 10 23:16:51 crc kubenswrapper[4791]: I1210 23:16:51.725037 4791 scope.go:117] "RemoveContainer" containerID="d5fbff44eb0dca42788f629ce6e1217af29146cb334e533f0e3c8327af2d22b4" Dec 10 23:16:53 crc kubenswrapper[4791]: I1210 23:16:53.935201 4791 generic.go:334] "Generic (PLEG): container finished" podID="07a0187d-1677-4da9-9a3b-8da49e8f6819" containerID="5eb7ff160fc3e10b947e2e0545d1afd197556543489b13d321b4d64ac233973c" exitCode=0 Dec 10 23:16:53 crc kubenswrapper[4791]: I1210 23:16:53.935265 4791 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-fpng5" event={"ID":"07a0187d-1677-4da9-9a3b-8da49e8f6819","Type":"ContainerDied","Data":"5eb7ff160fc3e10b947e2e0545d1afd197556543489b13d321b4d64ac233973c"} Dec 10 23:16:55 crc kubenswrapper[4791]: I1210 23:16:55.480057 4791 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-fpng5" Dec 10 23:16:55 crc kubenswrapper[4791]: I1210 23:16:55.565899 4791 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/07a0187d-1677-4da9-9a3b-8da49e8f6819-inventory\") pod \"07a0187d-1677-4da9-9a3b-8da49e8f6819\" (UID: \"07a0187d-1677-4da9-9a3b-8da49e8f6819\") " Dec 10 23:16:55 crc kubenswrapper[4791]: I1210 23:16:55.566067 4791 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/07a0187d-1677-4da9-9a3b-8da49e8f6819-ssh-key\") pod \"07a0187d-1677-4da9-9a3b-8da49e8f6819\" (UID: \"07a0187d-1677-4da9-9a3b-8da49e8f6819\") " Dec 10 23:16:55 crc kubenswrapper[4791]: I1210 23:16:55.566101 4791 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"bootstrap-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/07a0187d-1677-4da9-9a3b-8da49e8f6819-bootstrap-combined-ca-bundle\") pod \"07a0187d-1677-4da9-9a3b-8da49e8f6819\" (UID: \"07a0187d-1677-4da9-9a3b-8da49e8f6819\") " Dec 10 23:16:55 crc kubenswrapper[4791]: I1210 23:16:55.566168 4791 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-dc9ck\" (UniqueName: \"kubernetes.io/projected/07a0187d-1677-4da9-9a3b-8da49e8f6819-kube-api-access-dc9ck\") pod \"07a0187d-1677-4da9-9a3b-8da49e8f6819\" (UID: \"07a0187d-1677-4da9-9a3b-8da49e8f6819\") " Dec 10 23:16:55 crc kubenswrapper[4791]: I1210 23:16:55.572118 4791 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/07a0187d-1677-4da9-9a3b-8da49e8f6819-bootstrap-combined-ca-bundle" (OuterVolumeSpecName: "bootstrap-combined-ca-bundle") pod "07a0187d-1677-4da9-9a3b-8da49e8f6819" (UID: "07a0187d-1677-4da9-9a3b-8da49e8f6819"). InnerVolumeSpecName "bootstrap-combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 10 23:16:55 crc kubenswrapper[4791]: I1210 23:16:55.572189 4791 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/07a0187d-1677-4da9-9a3b-8da49e8f6819-kube-api-access-dc9ck" (OuterVolumeSpecName: "kube-api-access-dc9ck") pod "07a0187d-1677-4da9-9a3b-8da49e8f6819" (UID: "07a0187d-1677-4da9-9a3b-8da49e8f6819"). InnerVolumeSpecName "kube-api-access-dc9ck". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 10 23:16:55 crc kubenswrapper[4791]: I1210 23:16:55.592857 4791 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/07a0187d-1677-4da9-9a3b-8da49e8f6819-inventory" (OuterVolumeSpecName: "inventory") pod "07a0187d-1677-4da9-9a3b-8da49e8f6819" (UID: "07a0187d-1677-4da9-9a3b-8da49e8f6819"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 10 23:16:55 crc kubenswrapper[4791]: I1210 23:16:55.604693 4791 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/07a0187d-1677-4da9-9a3b-8da49e8f6819-ssh-key" (OuterVolumeSpecName: "ssh-key") pod "07a0187d-1677-4da9-9a3b-8da49e8f6819" (UID: "07a0187d-1677-4da9-9a3b-8da49e8f6819"). InnerVolumeSpecName "ssh-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 10 23:16:55 crc kubenswrapper[4791]: I1210 23:16:55.668494 4791 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/07a0187d-1677-4da9-9a3b-8da49e8f6819-inventory\") on node \"crc\" DevicePath \"\"" Dec 10 23:16:55 crc kubenswrapper[4791]: I1210 23:16:55.668667 4791 reconciler_common.go:293] "Volume detached for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/07a0187d-1677-4da9-9a3b-8da49e8f6819-ssh-key\") on node \"crc\" DevicePath \"\"" Dec 10 23:16:55 crc kubenswrapper[4791]: I1210 23:16:55.668686 4791 reconciler_common.go:293] "Volume detached for volume \"bootstrap-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/07a0187d-1677-4da9-9a3b-8da49e8f6819-bootstrap-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 10 23:16:55 crc kubenswrapper[4791]: I1210 23:16:55.668702 4791 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-dc9ck\" (UniqueName: \"kubernetes.io/projected/07a0187d-1677-4da9-9a3b-8da49e8f6819-kube-api-access-dc9ck\") on node \"crc\" DevicePath \"\"" Dec 10 23:16:55 crc kubenswrapper[4791]: I1210 23:16:55.907115 4791 scope.go:117] "RemoveContainer" containerID="f12779a795c9f6f63a93027d53c67c89d0179d87b9df554c25bfc607f06caff1" Dec 10 23:16:55 crc kubenswrapper[4791]: E1210 23:16:55.908260 4791 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-5rb5l_openshift-machine-config-operator(ba35653c-6e06-4cee-a4d6-137764090d18)\"" pod="openshift-machine-config-operator/machine-config-daemon-5rb5l" podUID="ba35653c-6e06-4cee-a4d6-137764090d18" Dec 10 23:16:55 crc kubenswrapper[4791]: I1210 23:16:55.962563 4791 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-fpng5" event={"ID":"07a0187d-1677-4da9-9a3b-8da49e8f6819","Type":"ContainerDied","Data":"f100c8400fb1baf14371050d2bedf09b75fa03b225878c6f0168e865e1d1ecef"} Dec 10 23:16:55 crc kubenswrapper[4791]: I1210 23:16:55.962609 4791 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="f100c8400fb1baf14371050d2bedf09b75fa03b225878c6f0168e865e1d1ecef" Dec 10 23:16:55 crc kubenswrapper[4791]: I1210 23:16:55.962927 4791 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-fpng5" Dec 10 23:16:56 crc kubenswrapper[4791]: I1210 23:16:56.041163 4791 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/download-cache-edpm-deployment-openstack-edpm-ipam-nx9lt"] Dec 10 23:16:56 crc kubenswrapper[4791]: E1210 23:16:56.046424 4791 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f00159a6-00e0-4975-87a0-9d8b2b67b157" containerName="extract-utilities" Dec 10 23:16:56 crc kubenswrapper[4791]: I1210 23:16:56.046462 4791 state_mem.go:107] "Deleted CPUSet assignment" podUID="f00159a6-00e0-4975-87a0-9d8b2b67b157" containerName="extract-utilities" Dec 10 23:16:56 crc kubenswrapper[4791]: E1210 23:16:56.046478 4791 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f00159a6-00e0-4975-87a0-9d8b2b67b157" containerName="extract-content" Dec 10 23:16:56 crc kubenswrapper[4791]: I1210 23:16:56.046489 4791 state_mem.go:107] "Deleted CPUSet assignment" podUID="f00159a6-00e0-4975-87a0-9d8b2b67b157" containerName="extract-content" Dec 10 23:16:56 crc kubenswrapper[4791]: E1210 23:16:56.046527 4791 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="07a0187d-1677-4da9-9a3b-8da49e8f6819" containerName="bootstrap-edpm-deployment-openstack-edpm-ipam" Dec 10 23:16:56 crc kubenswrapper[4791]: I1210 23:16:56.046536 4791 state_mem.go:107] "Deleted CPUSet assignment" podUID="07a0187d-1677-4da9-9a3b-8da49e8f6819" containerName="bootstrap-edpm-deployment-openstack-edpm-ipam" Dec 10 23:16:56 crc kubenswrapper[4791]: E1210 23:16:56.046556 4791 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f00159a6-00e0-4975-87a0-9d8b2b67b157" containerName="registry-server" Dec 10 23:16:56 crc kubenswrapper[4791]: I1210 23:16:56.046564 4791 state_mem.go:107] "Deleted CPUSet assignment" podUID="f00159a6-00e0-4975-87a0-9d8b2b67b157" containerName="registry-server" Dec 10 23:16:56 crc kubenswrapper[4791]: I1210 23:16:56.046757 4791 memory_manager.go:354] "RemoveStaleState removing state" podUID="07a0187d-1677-4da9-9a3b-8da49e8f6819" containerName="bootstrap-edpm-deployment-openstack-edpm-ipam" Dec 10 23:16:56 crc kubenswrapper[4791]: I1210 23:16:56.046776 4791 memory_manager.go:354] "RemoveStaleState removing state" podUID="f00159a6-00e0-4975-87a0-9d8b2b67b157" containerName="registry-server" Dec 10 23:16:56 crc kubenswrapper[4791]: I1210 23:16:56.047405 4791 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/download-cache-edpm-deployment-openstack-edpm-ipam-nx9lt" Dec 10 23:16:56 crc kubenswrapper[4791]: I1210 23:16:56.049538 4791 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Dec 10 23:16:56 crc kubenswrapper[4791]: I1210 23:16:56.049739 4791 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-ansible-ssh-private-key-secret" Dec 10 23:16:56 crc kubenswrapper[4791]: I1210 23:16:56.049712 4791 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-dockercfg-hf8pr" Dec 10 23:16:56 crc kubenswrapper[4791]: I1210 23:16:56.049791 4791 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-edpm-ipam" Dec 10 23:16:56 crc kubenswrapper[4791]: I1210 23:16:56.054916 4791 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/download-cache-edpm-deployment-openstack-edpm-ipam-nx9lt"] Dec 10 23:16:56 crc kubenswrapper[4791]: I1210 23:16:56.077600 4791 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/36b35496-ca5d-49ac-bb9f-cb697fa27af6-inventory\") pod \"download-cache-edpm-deployment-openstack-edpm-ipam-nx9lt\" (UID: \"36b35496-ca5d-49ac-bb9f-cb697fa27af6\") " pod="openstack/download-cache-edpm-deployment-openstack-edpm-ipam-nx9lt" Dec 10 23:16:56 crc kubenswrapper[4791]: I1210 23:16:56.077683 4791 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/36b35496-ca5d-49ac-bb9f-cb697fa27af6-ssh-key\") pod \"download-cache-edpm-deployment-openstack-edpm-ipam-nx9lt\" (UID: \"36b35496-ca5d-49ac-bb9f-cb697fa27af6\") " pod="openstack/download-cache-edpm-deployment-openstack-edpm-ipam-nx9lt" Dec 10 23:16:56 crc kubenswrapper[4791]: I1210 23:16:56.078039 4791 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-pfrfn\" (UniqueName: \"kubernetes.io/projected/36b35496-ca5d-49ac-bb9f-cb697fa27af6-kube-api-access-pfrfn\") pod \"download-cache-edpm-deployment-openstack-edpm-ipam-nx9lt\" (UID: \"36b35496-ca5d-49ac-bb9f-cb697fa27af6\") " pod="openstack/download-cache-edpm-deployment-openstack-edpm-ipam-nx9lt" Dec 10 23:16:56 crc kubenswrapper[4791]: I1210 23:16:56.180537 4791 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/36b35496-ca5d-49ac-bb9f-cb697fa27af6-inventory\") pod \"download-cache-edpm-deployment-openstack-edpm-ipam-nx9lt\" (UID: \"36b35496-ca5d-49ac-bb9f-cb697fa27af6\") " pod="openstack/download-cache-edpm-deployment-openstack-edpm-ipam-nx9lt" Dec 10 23:16:56 crc kubenswrapper[4791]: I1210 23:16:56.180846 4791 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/36b35496-ca5d-49ac-bb9f-cb697fa27af6-ssh-key\") pod \"download-cache-edpm-deployment-openstack-edpm-ipam-nx9lt\" (UID: \"36b35496-ca5d-49ac-bb9f-cb697fa27af6\") " pod="openstack/download-cache-edpm-deployment-openstack-edpm-ipam-nx9lt" Dec 10 23:16:56 crc kubenswrapper[4791]: I1210 23:16:56.181010 4791 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-pfrfn\" (UniqueName: \"kubernetes.io/projected/36b35496-ca5d-49ac-bb9f-cb697fa27af6-kube-api-access-pfrfn\") pod \"download-cache-edpm-deployment-openstack-edpm-ipam-nx9lt\" (UID: \"36b35496-ca5d-49ac-bb9f-cb697fa27af6\") " pod="openstack/download-cache-edpm-deployment-openstack-edpm-ipam-nx9lt" Dec 10 23:16:56 crc kubenswrapper[4791]: I1210 23:16:56.184054 4791 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/36b35496-ca5d-49ac-bb9f-cb697fa27af6-inventory\") pod \"download-cache-edpm-deployment-openstack-edpm-ipam-nx9lt\" (UID: \"36b35496-ca5d-49ac-bb9f-cb697fa27af6\") " pod="openstack/download-cache-edpm-deployment-openstack-edpm-ipam-nx9lt" Dec 10 23:16:56 crc kubenswrapper[4791]: I1210 23:16:56.184482 4791 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/36b35496-ca5d-49ac-bb9f-cb697fa27af6-ssh-key\") pod \"download-cache-edpm-deployment-openstack-edpm-ipam-nx9lt\" (UID: \"36b35496-ca5d-49ac-bb9f-cb697fa27af6\") " pod="openstack/download-cache-edpm-deployment-openstack-edpm-ipam-nx9lt" Dec 10 23:16:56 crc kubenswrapper[4791]: I1210 23:16:56.203691 4791 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-pfrfn\" (UniqueName: \"kubernetes.io/projected/36b35496-ca5d-49ac-bb9f-cb697fa27af6-kube-api-access-pfrfn\") pod \"download-cache-edpm-deployment-openstack-edpm-ipam-nx9lt\" (UID: \"36b35496-ca5d-49ac-bb9f-cb697fa27af6\") " pod="openstack/download-cache-edpm-deployment-openstack-edpm-ipam-nx9lt" Dec 10 23:16:56 crc kubenswrapper[4791]: I1210 23:16:56.368673 4791 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/download-cache-edpm-deployment-openstack-edpm-ipam-nx9lt" Dec 10 23:16:56 crc kubenswrapper[4791]: I1210 23:16:56.922242 4791 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/download-cache-edpm-deployment-openstack-edpm-ipam-nx9lt"] Dec 10 23:16:56 crc kubenswrapper[4791]: I1210 23:16:56.931294 4791 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Dec 10 23:16:56 crc kubenswrapper[4791]: I1210 23:16:56.972133 4791 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/download-cache-edpm-deployment-openstack-edpm-ipam-nx9lt" event={"ID":"36b35496-ca5d-49ac-bb9f-cb697fa27af6","Type":"ContainerStarted","Data":"536aa86d6767a578de114c7b0f46ff87e0685da61a4461bb69e5f649ae2fc88d"} Dec 10 23:16:58 crc kubenswrapper[4791]: I1210 23:16:58.995084 4791 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/download-cache-edpm-deployment-openstack-edpm-ipam-nx9lt" event={"ID":"36b35496-ca5d-49ac-bb9f-cb697fa27af6","Type":"ContainerStarted","Data":"283d087ba8890bc3d594f839e99b3a2210adfbcb7f3aa3f82437e1d0e0a55328"} Dec 10 23:16:59 crc kubenswrapper[4791]: I1210 23:16:59.020571 4791 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/download-cache-edpm-deployment-openstack-edpm-ipam-nx9lt" podStartSLOduration=1.8414615429999999 podStartE2EDuration="3.020551337s" podCreationTimestamp="2025-12-10 23:16:56 +0000 UTC" firstStartedPulling="2025-12-10 23:16:56.931026109 +0000 UTC m=+1651.360643722" lastFinishedPulling="2025-12-10 23:16:58.110115893 +0000 UTC m=+1652.539733516" observedRunningTime="2025-12-10 23:16:59.018998482 +0000 UTC m=+1653.448616095" watchObservedRunningTime="2025-12-10 23:16:59.020551337 +0000 UTC m=+1653.450168950" Dec 10 23:17:07 crc kubenswrapper[4791]: I1210 23:17:07.885003 4791 scope.go:117] "RemoveContainer" containerID="f12779a795c9f6f63a93027d53c67c89d0179d87b9df554c25bfc607f06caff1" Dec 10 23:17:07 crc kubenswrapper[4791]: E1210 23:17:07.885781 4791 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-5rb5l_openshift-machine-config-operator(ba35653c-6e06-4cee-a4d6-137764090d18)\"" pod="openshift-machine-config-operator/machine-config-daemon-5rb5l" podUID="ba35653c-6e06-4cee-a4d6-137764090d18" Dec 10 23:17:19 crc kubenswrapper[4791]: I1210 23:17:19.885372 4791 scope.go:117] "RemoveContainer" containerID="f12779a795c9f6f63a93027d53c67c89d0179d87b9df554c25bfc607f06caff1" Dec 10 23:17:19 crc kubenswrapper[4791]: E1210 23:17:19.885903 4791 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-5rb5l_openshift-machine-config-operator(ba35653c-6e06-4cee-a4d6-137764090d18)\"" pod="openshift-machine-config-operator/machine-config-daemon-5rb5l" podUID="ba35653c-6e06-4cee-a4d6-137764090d18" Dec 10 23:17:30 crc kubenswrapper[4791]: I1210 23:17:30.884544 4791 scope.go:117] "RemoveContainer" containerID="f12779a795c9f6f63a93027d53c67c89d0179d87b9df554c25bfc607f06caff1" Dec 10 23:17:30 crc kubenswrapper[4791]: E1210 23:17:30.885285 4791 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-5rb5l_openshift-machine-config-operator(ba35653c-6e06-4cee-a4d6-137764090d18)\"" pod="openshift-machine-config-operator/machine-config-daemon-5rb5l" podUID="ba35653c-6e06-4cee-a4d6-137764090d18" Dec 10 23:17:32 crc kubenswrapper[4791]: I1210 23:17:32.041418 4791 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/keystone-db-create-qfvp4"] Dec 10 23:17:32 crc kubenswrapper[4791]: I1210 23:17:32.053533 4791 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/placement-492e-account-create-update-xfsh4"] Dec 10 23:17:32 crc kubenswrapper[4791]: I1210 23:17:32.064410 4791 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/keystone-db-create-qfvp4"] Dec 10 23:17:32 crc kubenswrapper[4791]: I1210 23:17:32.076435 4791 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/placement-db-create-g4swg"] Dec 10 23:17:32 crc kubenswrapper[4791]: I1210 23:17:32.086403 4791 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/placement-492e-account-create-update-xfsh4"] Dec 10 23:17:32 crc kubenswrapper[4791]: I1210 23:17:32.092859 4791 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/placement-db-create-g4swg"] Dec 10 23:17:33 crc kubenswrapper[4791]: I1210 23:17:33.036420 4791 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/glance-db-create-czcrm"] Dec 10 23:17:33 crc kubenswrapper[4791]: I1210 23:17:33.048940 4791 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/keystone-85e7-account-create-update-25zns"] Dec 10 23:17:33 crc kubenswrapper[4791]: I1210 23:17:33.057709 4791 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/keystone-85e7-account-create-update-25zns"] Dec 10 23:17:33 crc kubenswrapper[4791]: I1210 23:17:33.066662 4791 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/glance-db-create-czcrm"] Dec 10 23:17:33 crc kubenswrapper[4791]: I1210 23:17:33.078126 4791 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/glance-073e-account-create-update-hwgzh"] Dec 10 23:17:33 crc kubenswrapper[4791]: I1210 23:17:33.085077 4791 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/glance-073e-account-create-update-hwgzh"] Dec 10 23:17:33 crc kubenswrapper[4791]: I1210 23:17:33.896719 4791 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="2a8736f4-b93d-4bde-a270-0acb0c730ab0" path="/var/lib/kubelet/pods/2a8736f4-b93d-4bde-a270-0acb0c730ab0/volumes" Dec 10 23:17:33 crc kubenswrapper[4791]: I1210 23:17:33.897649 4791 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="2c828334-1c90-4871-b203-71f82ac6fb8b" path="/var/lib/kubelet/pods/2c828334-1c90-4871-b203-71f82ac6fb8b/volumes" Dec 10 23:17:33 crc kubenswrapper[4791]: I1210 23:17:33.898584 4791 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="4d16f575-dc6d-4a78-8303-10a1c8202c0e" path="/var/lib/kubelet/pods/4d16f575-dc6d-4a78-8303-10a1c8202c0e/volumes" Dec 10 23:17:33 crc kubenswrapper[4791]: I1210 23:17:33.899720 4791 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="831d2f93-31c5-469f-b4bf-1dd85fa21cbf" path="/var/lib/kubelet/pods/831d2f93-31c5-469f-b4bf-1dd85fa21cbf/volumes" Dec 10 23:17:33 crc kubenswrapper[4791]: I1210 23:17:33.901294 4791 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="af2a4d0a-289c-4070-8371-2eca62ff0ca2" path="/var/lib/kubelet/pods/af2a4d0a-289c-4070-8371-2eca62ff0ca2/volumes" Dec 10 23:17:33 crc kubenswrapper[4791]: I1210 23:17:33.902109 4791 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="afcdc302-63c5-45c5-bf5b-46ef1caf953d" path="/var/lib/kubelet/pods/afcdc302-63c5-45c5-bf5b-46ef1caf953d/volumes" Dec 10 23:17:42 crc kubenswrapper[4791]: I1210 23:17:42.884944 4791 scope.go:117] "RemoveContainer" containerID="f12779a795c9f6f63a93027d53c67c89d0179d87b9df554c25bfc607f06caff1" Dec 10 23:17:42 crc kubenswrapper[4791]: E1210 23:17:42.886035 4791 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-5rb5l_openshift-machine-config-operator(ba35653c-6e06-4cee-a4d6-137764090d18)\"" pod="openshift-machine-config-operator/machine-config-daemon-5rb5l" podUID="ba35653c-6e06-4cee-a4d6-137764090d18" Dec 10 23:17:51 crc kubenswrapper[4791]: I1210 23:17:51.782662 4791 scope.go:117] "RemoveContainer" containerID="8203de309099603138b658f2b92d9f6938a0d58392a05ced22b78b6a5b6d1bcf" Dec 10 23:17:51 crc kubenswrapper[4791]: I1210 23:17:51.835396 4791 scope.go:117] "RemoveContainer" containerID="b4d9076c09753643536c9ad794e20a5b314f2e20f13af8e7e2cc4329ea297990" Dec 10 23:17:51 crc kubenswrapper[4791]: I1210 23:17:51.873524 4791 scope.go:117] "RemoveContainer" containerID="eb263c1b91e1c38df8b56867ab31368ec0c95f2353b67c05eb17173caf26ea53" Dec 10 23:17:51 crc kubenswrapper[4791]: I1210 23:17:51.923498 4791 scope.go:117] "RemoveContainer" containerID="8030250db2495a83cbbd9faee36c9964e867a4e460c9bdb7b6cd4661b2928250" Dec 10 23:17:51 crc kubenswrapper[4791]: I1210 23:17:51.980510 4791 scope.go:117] "RemoveContainer" containerID="54e3ad356567141309d0c6aa74e1bb0becd794cdad3c11cd66345605af663fb3" Dec 10 23:17:52 crc kubenswrapper[4791]: I1210 23:17:52.025066 4791 scope.go:117] "RemoveContainer" containerID="74c66018017c1765bad98de000d0470b38bafa4c766ec9da4d05f8312538ab4b" Dec 10 23:17:56 crc kubenswrapper[4791]: I1210 23:17:56.885379 4791 scope.go:117] "RemoveContainer" containerID="f12779a795c9f6f63a93027d53c67c89d0179d87b9df554c25bfc607f06caff1" Dec 10 23:17:56 crc kubenswrapper[4791]: E1210 23:17:56.886146 4791 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-5rb5l_openshift-machine-config-operator(ba35653c-6e06-4cee-a4d6-137764090d18)\"" pod="openshift-machine-config-operator/machine-config-daemon-5rb5l" podUID="ba35653c-6e06-4cee-a4d6-137764090d18" Dec 10 23:18:10 crc kubenswrapper[4791]: I1210 23:18:10.072303 4791 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/barbican-db-create-w8q8q"] Dec 10 23:18:10 crc kubenswrapper[4791]: I1210 23:18:10.082767 4791 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/neutron-db-create-l5xkj"] Dec 10 23:18:10 crc kubenswrapper[4791]: I1210 23:18:10.094672 4791 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/barbican-a84b-account-create-update-7ct6w"] Dec 10 23:18:10 crc kubenswrapper[4791]: I1210 23:18:10.102901 4791 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/cinder-8c6c-account-create-update-s9h2s"] Dec 10 23:18:10 crc kubenswrapper[4791]: I1210 23:18:10.109748 4791 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/cinder-db-create-58mvg"] Dec 10 23:18:10 crc kubenswrapper[4791]: I1210 23:18:10.116569 4791 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/cinder-8c6c-account-create-update-s9h2s"] Dec 10 23:18:10 crc kubenswrapper[4791]: I1210 23:18:10.123112 4791 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/barbican-db-create-w8q8q"] Dec 10 23:18:10 crc kubenswrapper[4791]: I1210 23:18:10.129568 4791 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/neutron-db-create-l5xkj"] Dec 10 23:18:10 crc kubenswrapper[4791]: I1210 23:18:10.136740 4791 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/barbican-a84b-account-create-update-7ct6w"] Dec 10 23:18:10 crc kubenswrapper[4791]: I1210 23:18:10.143685 4791 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/cinder-db-create-58mvg"] Dec 10 23:18:11 crc kubenswrapper[4791]: I1210 23:18:11.884782 4791 scope.go:117] "RemoveContainer" containerID="f12779a795c9f6f63a93027d53c67c89d0179d87b9df554c25bfc607f06caff1" Dec 10 23:18:11 crc kubenswrapper[4791]: E1210 23:18:11.885319 4791 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-5rb5l_openshift-machine-config-operator(ba35653c-6e06-4cee-a4d6-137764090d18)\"" pod="openshift-machine-config-operator/machine-config-daemon-5rb5l" podUID="ba35653c-6e06-4cee-a4d6-137764090d18" Dec 10 23:18:11 crc kubenswrapper[4791]: I1210 23:18:11.895887 4791 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="25d98ac3-17a9-4885-a64a-2ae593acf177" path="/var/lib/kubelet/pods/25d98ac3-17a9-4885-a64a-2ae593acf177/volumes" Dec 10 23:18:11 crc kubenswrapper[4791]: I1210 23:18:11.896600 4791 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="77886642-7b06-476f-acb0-7e4df70190a8" path="/var/lib/kubelet/pods/77886642-7b06-476f-acb0-7e4df70190a8/volumes" Dec 10 23:18:11 crc kubenswrapper[4791]: I1210 23:18:11.897271 4791 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="c289a9e6-9d03-499b-82f6-f0c898d94f04" path="/var/lib/kubelet/pods/c289a9e6-9d03-499b-82f6-f0c898d94f04/volumes" Dec 10 23:18:11 crc kubenswrapper[4791]: I1210 23:18:11.897979 4791 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="d2eecb39-349b-48d4-8819-aa392095d5d0" path="/var/lib/kubelet/pods/d2eecb39-349b-48d4-8819-aa392095d5d0/volumes" Dec 10 23:18:11 crc kubenswrapper[4791]: I1210 23:18:11.899170 4791 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="ebf742dc-fd16-49ae-a30a-b0f61e4d15fd" path="/var/lib/kubelet/pods/ebf742dc-fd16-49ae-a30a-b0f61e4d15fd/volumes" Dec 10 23:18:15 crc kubenswrapper[4791]: I1210 23:18:15.039895 4791 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/neutron-e982-account-create-update-7fdhc"] Dec 10 23:18:15 crc kubenswrapper[4791]: I1210 23:18:15.051666 4791 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/neutron-e982-account-create-update-7fdhc"] Dec 10 23:18:15 crc kubenswrapper[4791]: I1210 23:18:15.895403 4791 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="e7649319-d8a5-48d8-b9f8-f236602c72a0" path="/var/lib/kubelet/pods/e7649319-d8a5-48d8-b9f8-f236602c72a0/volumes" Dec 10 23:18:19 crc kubenswrapper[4791]: I1210 23:18:19.055977 4791 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/keystone-db-sync-kt8jj"] Dec 10 23:18:19 crc kubenswrapper[4791]: I1210 23:18:19.069098 4791 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/keystone-db-sync-kt8jj"] Dec 10 23:18:19 crc kubenswrapper[4791]: I1210 23:18:19.899714 4791 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="5ad26083-9c93-4ec4-9159-ef991f116f80" path="/var/lib/kubelet/pods/5ad26083-9c93-4ec4-9159-ef991f116f80/volumes" Dec 10 23:18:25 crc kubenswrapper[4791]: I1210 23:18:25.904947 4791 scope.go:117] "RemoveContainer" containerID="f12779a795c9f6f63a93027d53c67c89d0179d87b9df554c25bfc607f06caff1" Dec 10 23:18:25 crc kubenswrapper[4791]: E1210 23:18:25.906151 4791 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-5rb5l_openshift-machine-config-operator(ba35653c-6e06-4cee-a4d6-137764090d18)\"" pod="openshift-machine-config-operator/machine-config-daemon-5rb5l" podUID="ba35653c-6e06-4cee-a4d6-137764090d18" Dec 10 23:18:36 crc kubenswrapper[4791]: I1210 23:18:36.100666 4791 generic.go:334] "Generic (PLEG): container finished" podID="36b35496-ca5d-49ac-bb9f-cb697fa27af6" containerID="283d087ba8890bc3d594f839e99b3a2210adfbcb7f3aa3f82437e1d0e0a55328" exitCode=0 Dec 10 23:18:36 crc kubenswrapper[4791]: I1210 23:18:36.100792 4791 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/download-cache-edpm-deployment-openstack-edpm-ipam-nx9lt" event={"ID":"36b35496-ca5d-49ac-bb9f-cb697fa27af6","Type":"ContainerDied","Data":"283d087ba8890bc3d594f839e99b3a2210adfbcb7f3aa3f82437e1d0e0a55328"} Dec 10 23:18:37 crc kubenswrapper[4791]: I1210 23:18:37.073035 4791 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/glance-db-sync-ccl4n"] Dec 10 23:18:37 crc kubenswrapper[4791]: I1210 23:18:37.084659 4791 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/glance-db-sync-ccl4n"] Dec 10 23:18:37 crc kubenswrapper[4791]: I1210 23:18:37.556371 4791 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/download-cache-edpm-deployment-openstack-edpm-ipam-nx9lt" Dec 10 23:18:37 crc kubenswrapper[4791]: I1210 23:18:37.578964 4791 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/36b35496-ca5d-49ac-bb9f-cb697fa27af6-inventory\") pod \"36b35496-ca5d-49ac-bb9f-cb697fa27af6\" (UID: \"36b35496-ca5d-49ac-bb9f-cb697fa27af6\") " Dec 10 23:18:37 crc kubenswrapper[4791]: I1210 23:18:37.579217 4791 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/36b35496-ca5d-49ac-bb9f-cb697fa27af6-ssh-key\") pod \"36b35496-ca5d-49ac-bb9f-cb697fa27af6\" (UID: \"36b35496-ca5d-49ac-bb9f-cb697fa27af6\") " Dec 10 23:18:37 crc kubenswrapper[4791]: I1210 23:18:37.579242 4791 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-pfrfn\" (UniqueName: \"kubernetes.io/projected/36b35496-ca5d-49ac-bb9f-cb697fa27af6-kube-api-access-pfrfn\") pod \"36b35496-ca5d-49ac-bb9f-cb697fa27af6\" (UID: \"36b35496-ca5d-49ac-bb9f-cb697fa27af6\") " Dec 10 23:18:37 crc kubenswrapper[4791]: I1210 23:18:37.585492 4791 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/36b35496-ca5d-49ac-bb9f-cb697fa27af6-kube-api-access-pfrfn" (OuterVolumeSpecName: "kube-api-access-pfrfn") pod "36b35496-ca5d-49ac-bb9f-cb697fa27af6" (UID: "36b35496-ca5d-49ac-bb9f-cb697fa27af6"). InnerVolumeSpecName "kube-api-access-pfrfn". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 10 23:18:37 crc kubenswrapper[4791]: I1210 23:18:37.610481 4791 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/36b35496-ca5d-49ac-bb9f-cb697fa27af6-ssh-key" (OuterVolumeSpecName: "ssh-key") pod "36b35496-ca5d-49ac-bb9f-cb697fa27af6" (UID: "36b35496-ca5d-49ac-bb9f-cb697fa27af6"). InnerVolumeSpecName "ssh-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 10 23:18:37 crc kubenswrapper[4791]: I1210 23:18:37.613528 4791 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/36b35496-ca5d-49ac-bb9f-cb697fa27af6-inventory" (OuterVolumeSpecName: "inventory") pod "36b35496-ca5d-49ac-bb9f-cb697fa27af6" (UID: "36b35496-ca5d-49ac-bb9f-cb697fa27af6"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 10 23:18:37 crc kubenswrapper[4791]: I1210 23:18:37.681663 4791 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-pfrfn\" (UniqueName: \"kubernetes.io/projected/36b35496-ca5d-49ac-bb9f-cb697fa27af6-kube-api-access-pfrfn\") on node \"crc\" DevicePath \"\"" Dec 10 23:18:37 crc kubenswrapper[4791]: I1210 23:18:37.682006 4791 reconciler_common.go:293] "Volume detached for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/36b35496-ca5d-49ac-bb9f-cb697fa27af6-ssh-key\") on node \"crc\" DevicePath \"\"" Dec 10 23:18:37 crc kubenswrapper[4791]: I1210 23:18:37.682028 4791 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/36b35496-ca5d-49ac-bb9f-cb697fa27af6-inventory\") on node \"crc\" DevicePath \"\"" Dec 10 23:18:37 crc kubenswrapper[4791]: I1210 23:18:37.901856 4791 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="a2c44c96-70c2-414a-9890-cb95b66f10b2" path="/var/lib/kubelet/pods/a2c44c96-70c2-414a-9890-cb95b66f10b2/volumes" Dec 10 23:18:38 crc kubenswrapper[4791]: I1210 23:18:38.119891 4791 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/download-cache-edpm-deployment-openstack-edpm-ipam-nx9lt" event={"ID":"36b35496-ca5d-49ac-bb9f-cb697fa27af6","Type":"ContainerDied","Data":"536aa86d6767a578de114c7b0f46ff87e0685da61a4461bb69e5f649ae2fc88d"} Dec 10 23:18:38 crc kubenswrapper[4791]: I1210 23:18:38.119922 4791 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/download-cache-edpm-deployment-openstack-edpm-ipam-nx9lt" Dec 10 23:18:38 crc kubenswrapper[4791]: I1210 23:18:38.119927 4791 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="536aa86d6767a578de114c7b0f46ff87e0685da61a4461bb69e5f649ae2fc88d" Dec 10 23:18:38 crc kubenswrapper[4791]: I1210 23:18:38.203627 4791 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/configure-network-edpm-deployment-openstack-edpm-ipam-wf5sq"] Dec 10 23:18:38 crc kubenswrapper[4791]: E1210 23:18:38.204620 4791 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="36b35496-ca5d-49ac-bb9f-cb697fa27af6" containerName="download-cache-edpm-deployment-openstack-edpm-ipam" Dec 10 23:18:38 crc kubenswrapper[4791]: I1210 23:18:38.204646 4791 state_mem.go:107] "Deleted CPUSet assignment" podUID="36b35496-ca5d-49ac-bb9f-cb697fa27af6" containerName="download-cache-edpm-deployment-openstack-edpm-ipam" Dec 10 23:18:38 crc kubenswrapper[4791]: I1210 23:18:38.205233 4791 memory_manager.go:354] "RemoveStaleState removing state" podUID="36b35496-ca5d-49ac-bb9f-cb697fa27af6" containerName="download-cache-edpm-deployment-openstack-edpm-ipam" Dec 10 23:18:38 crc kubenswrapper[4791]: I1210 23:18:38.206158 4791 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/configure-network-edpm-deployment-openstack-edpm-ipam-wf5sq" Dec 10 23:18:38 crc kubenswrapper[4791]: I1210 23:18:38.217748 4791 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-ansible-ssh-private-key-secret" Dec 10 23:18:38 crc kubenswrapper[4791]: I1210 23:18:38.217955 4791 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-edpm-ipam" Dec 10 23:18:38 crc kubenswrapper[4791]: I1210 23:18:38.218210 4791 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Dec 10 23:18:38 crc kubenswrapper[4791]: I1210 23:18:38.218446 4791 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-dockercfg-hf8pr" Dec 10 23:18:38 crc kubenswrapper[4791]: I1210 23:18:38.239634 4791 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/configure-network-edpm-deployment-openstack-edpm-ipam-wf5sq"] Dec 10 23:18:38 crc kubenswrapper[4791]: I1210 23:18:38.290932 4791 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-52twz\" (UniqueName: \"kubernetes.io/projected/852ca005-3b2f-407e-ae86-9b1c5a5fe182-kube-api-access-52twz\") pod \"configure-network-edpm-deployment-openstack-edpm-ipam-wf5sq\" (UID: \"852ca005-3b2f-407e-ae86-9b1c5a5fe182\") " pod="openstack/configure-network-edpm-deployment-openstack-edpm-ipam-wf5sq" Dec 10 23:18:38 crc kubenswrapper[4791]: I1210 23:18:38.290992 4791 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/852ca005-3b2f-407e-ae86-9b1c5a5fe182-ssh-key\") pod \"configure-network-edpm-deployment-openstack-edpm-ipam-wf5sq\" (UID: \"852ca005-3b2f-407e-ae86-9b1c5a5fe182\") " pod="openstack/configure-network-edpm-deployment-openstack-edpm-ipam-wf5sq" Dec 10 23:18:38 crc kubenswrapper[4791]: I1210 23:18:38.291056 4791 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/852ca005-3b2f-407e-ae86-9b1c5a5fe182-inventory\") pod \"configure-network-edpm-deployment-openstack-edpm-ipam-wf5sq\" (UID: \"852ca005-3b2f-407e-ae86-9b1c5a5fe182\") " pod="openstack/configure-network-edpm-deployment-openstack-edpm-ipam-wf5sq" Dec 10 23:18:38 crc kubenswrapper[4791]: I1210 23:18:38.393388 4791 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/852ca005-3b2f-407e-ae86-9b1c5a5fe182-inventory\") pod \"configure-network-edpm-deployment-openstack-edpm-ipam-wf5sq\" (UID: \"852ca005-3b2f-407e-ae86-9b1c5a5fe182\") " pod="openstack/configure-network-edpm-deployment-openstack-edpm-ipam-wf5sq" Dec 10 23:18:38 crc kubenswrapper[4791]: I1210 23:18:38.393526 4791 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-52twz\" (UniqueName: \"kubernetes.io/projected/852ca005-3b2f-407e-ae86-9b1c5a5fe182-kube-api-access-52twz\") pod \"configure-network-edpm-deployment-openstack-edpm-ipam-wf5sq\" (UID: \"852ca005-3b2f-407e-ae86-9b1c5a5fe182\") " pod="openstack/configure-network-edpm-deployment-openstack-edpm-ipam-wf5sq" Dec 10 23:18:38 crc kubenswrapper[4791]: I1210 23:18:38.393561 4791 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/852ca005-3b2f-407e-ae86-9b1c5a5fe182-ssh-key\") pod \"configure-network-edpm-deployment-openstack-edpm-ipam-wf5sq\" (UID: \"852ca005-3b2f-407e-ae86-9b1c5a5fe182\") " pod="openstack/configure-network-edpm-deployment-openstack-edpm-ipam-wf5sq" Dec 10 23:18:38 crc kubenswrapper[4791]: I1210 23:18:38.397975 4791 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/852ca005-3b2f-407e-ae86-9b1c5a5fe182-ssh-key\") pod \"configure-network-edpm-deployment-openstack-edpm-ipam-wf5sq\" (UID: \"852ca005-3b2f-407e-ae86-9b1c5a5fe182\") " pod="openstack/configure-network-edpm-deployment-openstack-edpm-ipam-wf5sq" Dec 10 23:18:38 crc kubenswrapper[4791]: I1210 23:18:38.399377 4791 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/852ca005-3b2f-407e-ae86-9b1c5a5fe182-inventory\") pod \"configure-network-edpm-deployment-openstack-edpm-ipam-wf5sq\" (UID: \"852ca005-3b2f-407e-ae86-9b1c5a5fe182\") " pod="openstack/configure-network-edpm-deployment-openstack-edpm-ipam-wf5sq" Dec 10 23:18:38 crc kubenswrapper[4791]: I1210 23:18:38.410371 4791 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-52twz\" (UniqueName: \"kubernetes.io/projected/852ca005-3b2f-407e-ae86-9b1c5a5fe182-kube-api-access-52twz\") pod \"configure-network-edpm-deployment-openstack-edpm-ipam-wf5sq\" (UID: \"852ca005-3b2f-407e-ae86-9b1c5a5fe182\") " pod="openstack/configure-network-edpm-deployment-openstack-edpm-ipam-wf5sq" Dec 10 23:18:38 crc kubenswrapper[4791]: I1210 23:18:38.536692 4791 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/configure-network-edpm-deployment-openstack-edpm-ipam-wf5sq" Dec 10 23:18:39 crc kubenswrapper[4791]: I1210 23:18:39.138422 4791 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/configure-network-edpm-deployment-openstack-edpm-ipam-wf5sq"] Dec 10 23:18:39 crc kubenswrapper[4791]: W1210 23:18:39.140306 4791 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod852ca005_3b2f_407e_ae86_9b1c5a5fe182.slice/crio-6c0802937126fd1839647890c5e56c806b8a8870202d71af6d0c883fd2bec85a WatchSource:0}: Error finding container 6c0802937126fd1839647890c5e56c806b8a8870202d71af6d0c883fd2bec85a: Status 404 returned error can't find the container with id 6c0802937126fd1839647890c5e56c806b8a8870202d71af6d0c883fd2bec85a Dec 10 23:18:40 crc kubenswrapper[4791]: I1210 23:18:40.140920 4791 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/configure-network-edpm-deployment-openstack-edpm-ipam-wf5sq" event={"ID":"852ca005-3b2f-407e-ae86-9b1c5a5fe182","Type":"ContainerStarted","Data":"7a1d98890d7381e4ea9261c3cf3af4edcef1cff3f649a44f2ed84a6ec6d2c0c6"} Dec 10 23:18:40 crc kubenswrapper[4791]: I1210 23:18:40.141267 4791 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/configure-network-edpm-deployment-openstack-edpm-ipam-wf5sq" event={"ID":"852ca005-3b2f-407e-ae86-9b1c5a5fe182","Type":"ContainerStarted","Data":"6c0802937126fd1839647890c5e56c806b8a8870202d71af6d0c883fd2bec85a"} Dec 10 23:18:40 crc kubenswrapper[4791]: I1210 23:18:40.166512 4791 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/configure-network-edpm-deployment-openstack-edpm-ipam-wf5sq" podStartSLOduration=1.542768974 podStartE2EDuration="2.166498559s" podCreationTimestamp="2025-12-10 23:18:38 +0000 UTC" firstStartedPulling="2025-12-10 23:18:39.143122107 +0000 UTC m=+1753.572739730" lastFinishedPulling="2025-12-10 23:18:39.766851692 +0000 UTC m=+1754.196469315" observedRunningTime="2025-12-10 23:18:40.162079662 +0000 UTC m=+1754.591697275" watchObservedRunningTime="2025-12-10 23:18:40.166498559 +0000 UTC m=+1754.596116172" Dec 10 23:18:40 crc kubenswrapper[4791]: I1210 23:18:40.885064 4791 scope.go:117] "RemoveContainer" containerID="f12779a795c9f6f63a93027d53c67c89d0179d87b9df554c25bfc607f06caff1" Dec 10 23:18:40 crc kubenswrapper[4791]: E1210 23:18:40.885812 4791 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-5rb5l_openshift-machine-config-operator(ba35653c-6e06-4cee-a4d6-137764090d18)\"" pod="openshift-machine-config-operator/machine-config-daemon-5rb5l" podUID="ba35653c-6e06-4cee-a4d6-137764090d18" Dec 10 23:18:51 crc kubenswrapper[4791]: I1210 23:18:51.052999 4791 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/neutron-db-sync-kg4vv"] Dec 10 23:18:51 crc kubenswrapper[4791]: I1210 23:18:51.066197 4791 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/neutron-db-sync-kg4vv"] Dec 10 23:18:51 crc kubenswrapper[4791]: I1210 23:18:51.900090 4791 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="6076a930-1583-41b7-b5ae-bc6c61086e48" path="/var/lib/kubelet/pods/6076a930-1583-41b7-b5ae-bc6c61086e48/volumes" Dec 10 23:18:52 crc kubenswrapper[4791]: I1210 23:18:52.293261 4791 scope.go:117] "RemoveContainer" containerID="e728059b66b1ca7c2831961436238782b27e22a524a6ac12d7c043a19d44d59e" Dec 10 23:18:52 crc kubenswrapper[4791]: I1210 23:18:52.352035 4791 scope.go:117] "RemoveContainer" containerID="abe2a817c9798b3353a352ad8918182e06184e89343fa9ef0850dbaaf6700678" Dec 10 23:18:52 crc kubenswrapper[4791]: I1210 23:18:52.406245 4791 scope.go:117] "RemoveContainer" containerID="1246dddd2dc57e9b79396e6ae705e086d5153488fee1cd18e4dfa7a916a08eed" Dec 10 23:18:52 crc kubenswrapper[4791]: I1210 23:18:52.483364 4791 scope.go:117] "RemoveContainer" containerID="da6698cdf0ff45e48a9d6b08ecbf12721c5b95ee247f5490e3f0ee1861d0c94a" Dec 10 23:18:52 crc kubenswrapper[4791]: I1210 23:18:52.526270 4791 scope.go:117] "RemoveContainer" containerID="88d56927bc9f6ed7c032ad8701f19d0e57298caa95444f144d523b4c8b5ef2bc" Dec 10 23:18:52 crc kubenswrapper[4791]: I1210 23:18:52.558561 4791 scope.go:117] "RemoveContainer" containerID="a3dde81f9bff2a0f8bc2c74a006bc6eb02cd7703caae876aeedf170f58d474a6" Dec 10 23:18:52 crc kubenswrapper[4791]: I1210 23:18:52.594188 4791 scope.go:117] "RemoveContainer" containerID="b409fe2d3a34ee8ca490698ce05841ae136b5e63e5b04879ee4a783bcca8809e" Dec 10 23:18:52 crc kubenswrapper[4791]: I1210 23:18:52.642696 4791 scope.go:117] "RemoveContainer" containerID="eeb768e1a992b95d7fd4932dcc0d91575705cb14db754bba405b594192a5334d" Dec 10 23:18:52 crc kubenswrapper[4791]: I1210 23:18:52.674617 4791 scope.go:117] "RemoveContainer" containerID="ef2c1e1726ac3fe2b5200e0a1b6289e8846bb2fc6d6fd1dbaa755fddce4936b5" Dec 10 23:18:52 crc kubenswrapper[4791]: I1210 23:18:52.703461 4791 scope.go:117] "RemoveContainer" containerID="a3eeafd6ea8ce8bb3f96e912604f8a520c9ceb4ac7d92a96d2414c6cfb5c3ec8" Dec 10 23:18:52 crc kubenswrapper[4791]: I1210 23:18:52.728359 4791 scope.go:117] "RemoveContainer" containerID="309e845d753d39f9e77682b06f34c9acbc81eb099f4a3042921ed02e899bf5e8" Dec 10 23:18:55 crc kubenswrapper[4791]: I1210 23:18:55.891487 4791 scope.go:117] "RemoveContainer" containerID="f12779a795c9f6f63a93027d53c67c89d0179d87b9df554c25bfc607f06caff1" Dec 10 23:18:55 crc kubenswrapper[4791]: E1210 23:18:55.892245 4791 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-5rb5l_openshift-machine-config-operator(ba35653c-6e06-4cee-a4d6-137764090d18)\"" pod="openshift-machine-config-operator/machine-config-daemon-5rb5l" podUID="ba35653c-6e06-4cee-a4d6-137764090d18" Dec 10 23:18:59 crc kubenswrapper[4791]: I1210 23:18:59.063300 4791 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/keystone-bootstrap-mwqkg"] Dec 10 23:18:59 crc kubenswrapper[4791]: I1210 23:18:59.081317 4791 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/keystone-bootstrap-mwqkg"] Dec 10 23:18:59 crc kubenswrapper[4791]: I1210 23:18:59.899504 4791 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="7c65a05d-c2cc-45c4-b469-e6a271e01856" path="/var/lib/kubelet/pods/7c65a05d-c2cc-45c4-b469-e6a271e01856/volumes" Dec 10 23:19:00 crc kubenswrapper[4791]: I1210 23:19:00.050610 4791 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/placement-db-sync-svgbr"] Dec 10 23:19:00 crc kubenswrapper[4791]: I1210 23:19:00.073681 4791 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/placement-db-sync-svgbr"] Dec 10 23:19:01 crc kubenswrapper[4791]: I1210 23:19:01.905326 4791 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="b9efeadd-d54b-4f78-a0ad-85cc62321b84" path="/var/lib/kubelet/pods/b9efeadd-d54b-4f78-a0ad-85cc62321b84/volumes" Dec 10 23:19:07 crc kubenswrapper[4791]: I1210 23:19:07.886159 4791 scope.go:117] "RemoveContainer" containerID="f12779a795c9f6f63a93027d53c67c89d0179d87b9df554c25bfc607f06caff1" Dec 10 23:19:07 crc kubenswrapper[4791]: E1210 23:19:07.888824 4791 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-5rb5l_openshift-machine-config-operator(ba35653c-6e06-4cee-a4d6-137764090d18)\"" pod="openshift-machine-config-operator/machine-config-daemon-5rb5l" podUID="ba35653c-6e06-4cee-a4d6-137764090d18" Dec 10 23:19:10 crc kubenswrapper[4791]: I1210 23:19:10.051477 4791 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/cinder-db-sync-g4s5p"] Dec 10 23:19:10 crc kubenswrapper[4791]: I1210 23:19:10.065095 4791 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/cinder-db-sync-g4s5p"] Dec 10 23:19:11 crc kubenswrapper[4791]: I1210 23:19:11.898956 4791 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="a85df2ed-be84-41b3-8993-c3d963daef08" path="/var/lib/kubelet/pods/a85df2ed-be84-41b3-8993-c3d963daef08/volumes" Dec 10 23:19:14 crc kubenswrapper[4791]: I1210 23:19:14.036505 4791 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/barbican-db-sync-zllvm"] Dec 10 23:19:14 crc kubenswrapper[4791]: I1210 23:19:14.047014 4791 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/barbican-db-sync-zllvm"] Dec 10 23:19:15 crc kubenswrapper[4791]: I1210 23:19:15.899794 4791 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="851556d7-fa38-4eec-ae74-de989d66d27b" path="/var/lib/kubelet/pods/851556d7-fa38-4eec-ae74-de989d66d27b/volumes" Dec 10 23:19:20 crc kubenswrapper[4791]: I1210 23:19:20.884798 4791 scope.go:117] "RemoveContainer" containerID="f12779a795c9f6f63a93027d53c67c89d0179d87b9df554c25bfc607f06caff1" Dec 10 23:19:20 crc kubenswrapper[4791]: E1210 23:19:20.885323 4791 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-5rb5l_openshift-machine-config-operator(ba35653c-6e06-4cee-a4d6-137764090d18)\"" pod="openshift-machine-config-operator/machine-config-daemon-5rb5l" podUID="ba35653c-6e06-4cee-a4d6-137764090d18" Dec 10 23:19:23 crc kubenswrapper[4791]: I1210 23:19:23.128539 4791 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/community-operators-qz8bj"] Dec 10 23:19:23 crc kubenswrapper[4791]: I1210 23:19:23.131134 4791 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-qz8bj" Dec 10 23:19:23 crc kubenswrapper[4791]: I1210 23:19:23.153072 4791 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-qz8bj"] Dec 10 23:19:23 crc kubenswrapper[4791]: I1210 23:19:23.210295 4791 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/ee228122-ccd4-4d15-9ba8-a8b75ad66fa3-utilities\") pod \"community-operators-qz8bj\" (UID: \"ee228122-ccd4-4d15-9ba8-a8b75ad66fa3\") " pod="openshift-marketplace/community-operators-qz8bj" Dec 10 23:19:23 crc kubenswrapper[4791]: I1210 23:19:23.210466 4791 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-6h9vw\" (UniqueName: \"kubernetes.io/projected/ee228122-ccd4-4d15-9ba8-a8b75ad66fa3-kube-api-access-6h9vw\") pod \"community-operators-qz8bj\" (UID: \"ee228122-ccd4-4d15-9ba8-a8b75ad66fa3\") " pod="openshift-marketplace/community-operators-qz8bj" Dec 10 23:19:23 crc kubenswrapper[4791]: I1210 23:19:23.210586 4791 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/ee228122-ccd4-4d15-9ba8-a8b75ad66fa3-catalog-content\") pod \"community-operators-qz8bj\" (UID: \"ee228122-ccd4-4d15-9ba8-a8b75ad66fa3\") " pod="openshift-marketplace/community-operators-qz8bj" Dec 10 23:19:23 crc kubenswrapper[4791]: I1210 23:19:23.312311 4791 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/ee228122-ccd4-4d15-9ba8-a8b75ad66fa3-utilities\") pod \"community-operators-qz8bj\" (UID: \"ee228122-ccd4-4d15-9ba8-a8b75ad66fa3\") " pod="openshift-marketplace/community-operators-qz8bj" Dec 10 23:19:23 crc kubenswrapper[4791]: I1210 23:19:23.312466 4791 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-6h9vw\" (UniqueName: \"kubernetes.io/projected/ee228122-ccd4-4d15-9ba8-a8b75ad66fa3-kube-api-access-6h9vw\") pod \"community-operators-qz8bj\" (UID: \"ee228122-ccd4-4d15-9ba8-a8b75ad66fa3\") " pod="openshift-marketplace/community-operators-qz8bj" Dec 10 23:19:23 crc kubenswrapper[4791]: I1210 23:19:23.312561 4791 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/ee228122-ccd4-4d15-9ba8-a8b75ad66fa3-catalog-content\") pod \"community-operators-qz8bj\" (UID: \"ee228122-ccd4-4d15-9ba8-a8b75ad66fa3\") " pod="openshift-marketplace/community-operators-qz8bj" Dec 10 23:19:23 crc kubenswrapper[4791]: I1210 23:19:23.312961 4791 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/ee228122-ccd4-4d15-9ba8-a8b75ad66fa3-utilities\") pod \"community-operators-qz8bj\" (UID: \"ee228122-ccd4-4d15-9ba8-a8b75ad66fa3\") " pod="openshift-marketplace/community-operators-qz8bj" Dec 10 23:19:23 crc kubenswrapper[4791]: I1210 23:19:23.313028 4791 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/ee228122-ccd4-4d15-9ba8-a8b75ad66fa3-catalog-content\") pod \"community-operators-qz8bj\" (UID: \"ee228122-ccd4-4d15-9ba8-a8b75ad66fa3\") " pod="openshift-marketplace/community-operators-qz8bj" Dec 10 23:19:23 crc kubenswrapper[4791]: I1210 23:19:23.334039 4791 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-6h9vw\" (UniqueName: \"kubernetes.io/projected/ee228122-ccd4-4d15-9ba8-a8b75ad66fa3-kube-api-access-6h9vw\") pod \"community-operators-qz8bj\" (UID: \"ee228122-ccd4-4d15-9ba8-a8b75ad66fa3\") " pod="openshift-marketplace/community-operators-qz8bj" Dec 10 23:19:23 crc kubenswrapper[4791]: I1210 23:19:23.474388 4791 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-qz8bj" Dec 10 23:19:23 crc kubenswrapper[4791]: I1210 23:19:23.970958 4791 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-qz8bj"] Dec 10 23:19:24 crc kubenswrapper[4791]: I1210 23:19:24.658898 4791 generic.go:334] "Generic (PLEG): container finished" podID="ee228122-ccd4-4d15-9ba8-a8b75ad66fa3" containerID="178371a38a9a8ff14eb3a60f7dfd8af3af6a7bb75b3c70b04412521d48453387" exitCode=0 Dec 10 23:19:24 crc kubenswrapper[4791]: I1210 23:19:24.658949 4791 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-qz8bj" event={"ID":"ee228122-ccd4-4d15-9ba8-a8b75ad66fa3","Type":"ContainerDied","Data":"178371a38a9a8ff14eb3a60f7dfd8af3af6a7bb75b3c70b04412521d48453387"} Dec 10 23:19:24 crc kubenswrapper[4791]: I1210 23:19:24.658980 4791 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-qz8bj" event={"ID":"ee228122-ccd4-4d15-9ba8-a8b75ad66fa3","Type":"ContainerStarted","Data":"63584d75a8fe1b4d2b3c54068243512a37acff57a1317417984c17b75c64bd7d"} Dec 10 23:19:25 crc kubenswrapper[4791]: I1210 23:19:25.668247 4791 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-qz8bj" event={"ID":"ee228122-ccd4-4d15-9ba8-a8b75ad66fa3","Type":"ContainerStarted","Data":"1f601b495c523b51fe65ba26775c6406db8f2e7dce6d9a7cb184f03fa76f2d69"} Dec 10 23:19:26 crc kubenswrapper[4791]: I1210 23:19:26.127216 4791 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-operators-rccn4"] Dec 10 23:19:26 crc kubenswrapper[4791]: I1210 23:19:26.129585 4791 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-rccn4" Dec 10 23:19:26 crc kubenswrapper[4791]: I1210 23:19:26.135316 4791 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-rccn4"] Dec 10 23:19:26 crc kubenswrapper[4791]: I1210 23:19:26.262909 4791 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-bqkkb\" (UniqueName: \"kubernetes.io/projected/05961153-0a72-4c04-bea9-f622c3b54460-kube-api-access-bqkkb\") pod \"redhat-operators-rccn4\" (UID: \"05961153-0a72-4c04-bea9-f622c3b54460\") " pod="openshift-marketplace/redhat-operators-rccn4" Dec 10 23:19:26 crc kubenswrapper[4791]: I1210 23:19:26.263077 4791 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/05961153-0a72-4c04-bea9-f622c3b54460-catalog-content\") pod \"redhat-operators-rccn4\" (UID: \"05961153-0a72-4c04-bea9-f622c3b54460\") " pod="openshift-marketplace/redhat-operators-rccn4" Dec 10 23:19:26 crc kubenswrapper[4791]: I1210 23:19:26.263260 4791 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/05961153-0a72-4c04-bea9-f622c3b54460-utilities\") pod \"redhat-operators-rccn4\" (UID: \"05961153-0a72-4c04-bea9-f622c3b54460\") " pod="openshift-marketplace/redhat-operators-rccn4" Dec 10 23:19:26 crc kubenswrapper[4791]: I1210 23:19:26.364973 4791 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-bqkkb\" (UniqueName: \"kubernetes.io/projected/05961153-0a72-4c04-bea9-f622c3b54460-kube-api-access-bqkkb\") pod \"redhat-operators-rccn4\" (UID: \"05961153-0a72-4c04-bea9-f622c3b54460\") " pod="openshift-marketplace/redhat-operators-rccn4" Dec 10 23:19:26 crc kubenswrapper[4791]: I1210 23:19:26.365033 4791 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/05961153-0a72-4c04-bea9-f622c3b54460-catalog-content\") pod \"redhat-operators-rccn4\" (UID: \"05961153-0a72-4c04-bea9-f622c3b54460\") " pod="openshift-marketplace/redhat-operators-rccn4" Dec 10 23:19:26 crc kubenswrapper[4791]: I1210 23:19:26.365067 4791 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/05961153-0a72-4c04-bea9-f622c3b54460-utilities\") pod \"redhat-operators-rccn4\" (UID: \"05961153-0a72-4c04-bea9-f622c3b54460\") " pod="openshift-marketplace/redhat-operators-rccn4" Dec 10 23:19:26 crc kubenswrapper[4791]: I1210 23:19:26.365673 4791 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/05961153-0a72-4c04-bea9-f622c3b54460-utilities\") pod \"redhat-operators-rccn4\" (UID: \"05961153-0a72-4c04-bea9-f622c3b54460\") " pod="openshift-marketplace/redhat-operators-rccn4" Dec 10 23:19:26 crc kubenswrapper[4791]: I1210 23:19:26.365675 4791 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/05961153-0a72-4c04-bea9-f622c3b54460-catalog-content\") pod \"redhat-operators-rccn4\" (UID: \"05961153-0a72-4c04-bea9-f622c3b54460\") " pod="openshift-marketplace/redhat-operators-rccn4" Dec 10 23:19:26 crc kubenswrapper[4791]: I1210 23:19:26.385890 4791 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-bqkkb\" (UniqueName: \"kubernetes.io/projected/05961153-0a72-4c04-bea9-f622c3b54460-kube-api-access-bqkkb\") pod \"redhat-operators-rccn4\" (UID: \"05961153-0a72-4c04-bea9-f622c3b54460\") " pod="openshift-marketplace/redhat-operators-rccn4" Dec 10 23:19:26 crc kubenswrapper[4791]: I1210 23:19:26.467359 4791 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-rccn4" Dec 10 23:19:26 crc kubenswrapper[4791]: I1210 23:19:26.679069 4791 generic.go:334] "Generic (PLEG): container finished" podID="ee228122-ccd4-4d15-9ba8-a8b75ad66fa3" containerID="1f601b495c523b51fe65ba26775c6406db8f2e7dce6d9a7cb184f03fa76f2d69" exitCode=0 Dec 10 23:19:26 crc kubenswrapper[4791]: I1210 23:19:26.679107 4791 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-qz8bj" event={"ID":"ee228122-ccd4-4d15-9ba8-a8b75ad66fa3","Type":"ContainerDied","Data":"1f601b495c523b51fe65ba26775c6406db8f2e7dce6d9a7cb184f03fa76f2d69"} Dec 10 23:19:26 crc kubenswrapper[4791]: I1210 23:19:26.943831 4791 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-rccn4"] Dec 10 23:19:26 crc kubenswrapper[4791]: W1210 23:19:26.947542 4791 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod05961153_0a72_4c04_bea9_f622c3b54460.slice/crio-cb8927435bb2dcfc29b2dba2d7aad6a7c27c0dbae9577df995407e6df54520b7 WatchSource:0}: Error finding container cb8927435bb2dcfc29b2dba2d7aad6a7c27c0dbae9577df995407e6df54520b7: Status 404 returned error can't find the container with id cb8927435bb2dcfc29b2dba2d7aad6a7c27c0dbae9577df995407e6df54520b7 Dec 10 23:19:27 crc kubenswrapper[4791]: I1210 23:19:27.688643 4791 generic.go:334] "Generic (PLEG): container finished" podID="05961153-0a72-4c04-bea9-f622c3b54460" containerID="ad3d7800b15aeb482e1f77cd764aeed417388112e418a2145627373ca625b9c3" exitCode=0 Dec 10 23:19:27 crc kubenswrapper[4791]: I1210 23:19:27.688969 4791 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-rccn4" event={"ID":"05961153-0a72-4c04-bea9-f622c3b54460","Type":"ContainerDied","Data":"ad3d7800b15aeb482e1f77cd764aeed417388112e418a2145627373ca625b9c3"} Dec 10 23:19:27 crc kubenswrapper[4791]: I1210 23:19:27.689002 4791 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-rccn4" event={"ID":"05961153-0a72-4c04-bea9-f622c3b54460","Type":"ContainerStarted","Data":"cb8927435bb2dcfc29b2dba2d7aad6a7c27c0dbae9577df995407e6df54520b7"} Dec 10 23:19:27 crc kubenswrapper[4791]: I1210 23:19:27.692210 4791 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-qz8bj" event={"ID":"ee228122-ccd4-4d15-9ba8-a8b75ad66fa3","Type":"ContainerStarted","Data":"c7ca4939ee4c83b45d5fc2612f48a07321af74b9b5d450337fbe5b6645255644"} Dec 10 23:19:27 crc kubenswrapper[4791]: I1210 23:19:27.735556 4791 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/community-operators-qz8bj" podStartSLOduration=2.23353717 podStartE2EDuration="4.735535492s" podCreationTimestamp="2025-12-10 23:19:23 +0000 UTC" firstStartedPulling="2025-12-10 23:19:24.660710942 +0000 UTC m=+1799.090328555" lastFinishedPulling="2025-12-10 23:19:27.162709264 +0000 UTC m=+1801.592326877" observedRunningTime="2025-12-10 23:19:27.729308362 +0000 UTC m=+1802.158925985" watchObservedRunningTime="2025-12-10 23:19:27.735535492 +0000 UTC m=+1802.165153105" Dec 10 23:19:29 crc kubenswrapper[4791]: I1210 23:19:29.717913 4791 generic.go:334] "Generic (PLEG): container finished" podID="05961153-0a72-4c04-bea9-f622c3b54460" containerID="ab42257b441b8fbb32360ade40dae6a726fce47274f676ca283fce653172adde" exitCode=0 Dec 10 23:19:29 crc kubenswrapper[4791]: I1210 23:19:29.717969 4791 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-rccn4" event={"ID":"05961153-0a72-4c04-bea9-f622c3b54460","Type":"ContainerDied","Data":"ab42257b441b8fbb32360ade40dae6a726fce47274f676ca283fce653172adde"} Dec 10 23:19:31 crc kubenswrapper[4791]: I1210 23:19:31.741294 4791 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-rccn4" event={"ID":"05961153-0a72-4c04-bea9-f622c3b54460","Type":"ContainerStarted","Data":"d56168a0a555acfb35cd404aa3d4b8e24e977697cd2cfc2f706730d6fea96578"} Dec 10 23:19:32 crc kubenswrapper[4791]: I1210 23:19:32.777894 4791 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-operators-rccn4" podStartSLOduration=3.631091557 podStartE2EDuration="6.777877321s" podCreationTimestamp="2025-12-10 23:19:26 +0000 UTC" firstStartedPulling="2025-12-10 23:19:27.690669429 +0000 UTC m=+1802.120287052" lastFinishedPulling="2025-12-10 23:19:30.837455203 +0000 UTC m=+1805.267072816" observedRunningTime="2025-12-10 23:19:32.776510102 +0000 UTC m=+1807.206127725" watchObservedRunningTime="2025-12-10 23:19:32.777877321 +0000 UTC m=+1807.207494934" Dec 10 23:19:33 crc kubenswrapper[4791]: I1210 23:19:33.475039 4791 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/community-operators-qz8bj" Dec 10 23:19:33 crc kubenswrapper[4791]: I1210 23:19:33.475499 4791 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/community-operators-qz8bj" Dec 10 23:19:33 crc kubenswrapper[4791]: I1210 23:19:33.523722 4791 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/community-operators-qz8bj" Dec 10 23:19:33 crc kubenswrapper[4791]: I1210 23:19:33.819543 4791 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/community-operators-qz8bj" Dec 10 23:19:33 crc kubenswrapper[4791]: I1210 23:19:33.884871 4791 scope.go:117] "RemoveContainer" containerID="f12779a795c9f6f63a93027d53c67c89d0179d87b9df554c25bfc607f06caff1" Dec 10 23:19:33 crc kubenswrapper[4791]: E1210 23:19:33.885127 4791 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-5rb5l_openshift-machine-config-operator(ba35653c-6e06-4cee-a4d6-137764090d18)\"" pod="openshift-machine-config-operator/machine-config-daemon-5rb5l" podUID="ba35653c-6e06-4cee-a4d6-137764090d18" Dec 10 23:19:36 crc kubenswrapper[4791]: I1210 23:19:36.110783 4791 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-qz8bj"] Dec 10 23:19:36 crc kubenswrapper[4791]: I1210 23:19:36.111306 4791 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/community-operators-qz8bj" podUID="ee228122-ccd4-4d15-9ba8-a8b75ad66fa3" containerName="registry-server" containerID="cri-o://c7ca4939ee4c83b45d5fc2612f48a07321af74b9b5d450337fbe5b6645255644" gracePeriod=2 Dec 10 23:19:36 crc kubenswrapper[4791]: E1210 23:19:36.313728 4791 cadvisor_stats_provider.go:516] "Partial failure issuing cadvisor.ContainerInfoV2" err="partial failures: [\"/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podee228122_ccd4_4d15_9ba8_a8b75ad66fa3.slice/crio-conmon-c7ca4939ee4c83b45d5fc2612f48a07321af74b9b5d450337fbe5b6645255644.scope\": RecentStats: unable to find data in memory cache]" Dec 10 23:19:36 crc kubenswrapper[4791]: I1210 23:19:36.467991 4791 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-operators-rccn4" Dec 10 23:19:36 crc kubenswrapper[4791]: I1210 23:19:36.468379 4791 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-operators-rccn4" Dec 10 23:19:36 crc kubenswrapper[4791]: I1210 23:19:36.591116 4791 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-qz8bj" Dec 10 23:19:36 crc kubenswrapper[4791]: I1210 23:19:36.765638 4791 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-6h9vw\" (UniqueName: \"kubernetes.io/projected/ee228122-ccd4-4d15-9ba8-a8b75ad66fa3-kube-api-access-6h9vw\") pod \"ee228122-ccd4-4d15-9ba8-a8b75ad66fa3\" (UID: \"ee228122-ccd4-4d15-9ba8-a8b75ad66fa3\") " Dec 10 23:19:36 crc kubenswrapper[4791]: I1210 23:19:36.765885 4791 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/ee228122-ccd4-4d15-9ba8-a8b75ad66fa3-catalog-content\") pod \"ee228122-ccd4-4d15-9ba8-a8b75ad66fa3\" (UID: \"ee228122-ccd4-4d15-9ba8-a8b75ad66fa3\") " Dec 10 23:19:36 crc kubenswrapper[4791]: I1210 23:19:36.765943 4791 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/ee228122-ccd4-4d15-9ba8-a8b75ad66fa3-utilities\") pod \"ee228122-ccd4-4d15-9ba8-a8b75ad66fa3\" (UID: \"ee228122-ccd4-4d15-9ba8-a8b75ad66fa3\") " Dec 10 23:19:36 crc kubenswrapper[4791]: I1210 23:19:36.767157 4791 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/ee228122-ccd4-4d15-9ba8-a8b75ad66fa3-utilities" (OuterVolumeSpecName: "utilities") pod "ee228122-ccd4-4d15-9ba8-a8b75ad66fa3" (UID: "ee228122-ccd4-4d15-9ba8-a8b75ad66fa3"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 10 23:19:36 crc kubenswrapper[4791]: I1210 23:19:36.774463 4791 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/ee228122-ccd4-4d15-9ba8-a8b75ad66fa3-kube-api-access-6h9vw" (OuterVolumeSpecName: "kube-api-access-6h9vw") pod "ee228122-ccd4-4d15-9ba8-a8b75ad66fa3" (UID: "ee228122-ccd4-4d15-9ba8-a8b75ad66fa3"). InnerVolumeSpecName "kube-api-access-6h9vw". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 10 23:19:36 crc kubenswrapper[4791]: I1210 23:19:36.806516 4791 generic.go:334] "Generic (PLEG): container finished" podID="ee228122-ccd4-4d15-9ba8-a8b75ad66fa3" containerID="c7ca4939ee4c83b45d5fc2612f48a07321af74b9b5d450337fbe5b6645255644" exitCode=0 Dec 10 23:19:36 crc kubenswrapper[4791]: I1210 23:19:36.806569 4791 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-qz8bj" event={"ID":"ee228122-ccd4-4d15-9ba8-a8b75ad66fa3","Type":"ContainerDied","Data":"c7ca4939ee4c83b45d5fc2612f48a07321af74b9b5d450337fbe5b6645255644"} Dec 10 23:19:36 crc kubenswrapper[4791]: I1210 23:19:36.806579 4791 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-qz8bj" Dec 10 23:19:36 crc kubenswrapper[4791]: I1210 23:19:36.806599 4791 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-qz8bj" event={"ID":"ee228122-ccd4-4d15-9ba8-a8b75ad66fa3","Type":"ContainerDied","Data":"63584d75a8fe1b4d2b3c54068243512a37acff57a1317417984c17b75c64bd7d"} Dec 10 23:19:36 crc kubenswrapper[4791]: I1210 23:19:36.806619 4791 scope.go:117] "RemoveContainer" containerID="c7ca4939ee4c83b45d5fc2612f48a07321af74b9b5d450337fbe5b6645255644" Dec 10 23:19:36 crc kubenswrapper[4791]: I1210 23:19:36.815029 4791 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/ee228122-ccd4-4d15-9ba8-a8b75ad66fa3-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "ee228122-ccd4-4d15-9ba8-a8b75ad66fa3" (UID: "ee228122-ccd4-4d15-9ba8-a8b75ad66fa3"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 10 23:19:36 crc kubenswrapper[4791]: I1210 23:19:36.850592 4791 scope.go:117] "RemoveContainer" containerID="1f601b495c523b51fe65ba26775c6406db8f2e7dce6d9a7cb184f03fa76f2d69" Dec 10 23:19:36 crc kubenswrapper[4791]: I1210 23:19:36.867621 4791 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-6h9vw\" (UniqueName: \"kubernetes.io/projected/ee228122-ccd4-4d15-9ba8-a8b75ad66fa3-kube-api-access-6h9vw\") on node \"crc\" DevicePath \"\"" Dec 10 23:19:36 crc kubenswrapper[4791]: I1210 23:19:36.867652 4791 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/ee228122-ccd4-4d15-9ba8-a8b75ad66fa3-catalog-content\") on node \"crc\" DevicePath \"\"" Dec 10 23:19:36 crc kubenswrapper[4791]: I1210 23:19:36.867661 4791 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/ee228122-ccd4-4d15-9ba8-a8b75ad66fa3-utilities\") on node \"crc\" DevicePath \"\"" Dec 10 23:19:36 crc kubenswrapper[4791]: I1210 23:19:36.880475 4791 scope.go:117] "RemoveContainer" containerID="178371a38a9a8ff14eb3a60f7dfd8af3af6a7bb75b3c70b04412521d48453387" Dec 10 23:19:36 crc kubenswrapper[4791]: I1210 23:19:36.911710 4791 scope.go:117] "RemoveContainer" containerID="c7ca4939ee4c83b45d5fc2612f48a07321af74b9b5d450337fbe5b6645255644" Dec 10 23:19:36 crc kubenswrapper[4791]: E1210 23:19:36.912452 4791 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"c7ca4939ee4c83b45d5fc2612f48a07321af74b9b5d450337fbe5b6645255644\": container with ID starting with c7ca4939ee4c83b45d5fc2612f48a07321af74b9b5d450337fbe5b6645255644 not found: ID does not exist" containerID="c7ca4939ee4c83b45d5fc2612f48a07321af74b9b5d450337fbe5b6645255644" Dec 10 23:19:36 crc kubenswrapper[4791]: I1210 23:19:36.912505 4791 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"c7ca4939ee4c83b45d5fc2612f48a07321af74b9b5d450337fbe5b6645255644"} err="failed to get container status \"c7ca4939ee4c83b45d5fc2612f48a07321af74b9b5d450337fbe5b6645255644\": rpc error: code = NotFound desc = could not find container \"c7ca4939ee4c83b45d5fc2612f48a07321af74b9b5d450337fbe5b6645255644\": container with ID starting with c7ca4939ee4c83b45d5fc2612f48a07321af74b9b5d450337fbe5b6645255644 not found: ID does not exist" Dec 10 23:19:36 crc kubenswrapper[4791]: I1210 23:19:36.912534 4791 scope.go:117] "RemoveContainer" containerID="1f601b495c523b51fe65ba26775c6406db8f2e7dce6d9a7cb184f03fa76f2d69" Dec 10 23:19:36 crc kubenswrapper[4791]: E1210 23:19:36.913739 4791 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"1f601b495c523b51fe65ba26775c6406db8f2e7dce6d9a7cb184f03fa76f2d69\": container with ID starting with 1f601b495c523b51fe65ba26775c6406db8f2e7dce6d9a7cb184f03fa76f2d69 not found: ID does not exist" containerID="1f601b495c523b51fe65ba26775c6406db8f2e7dce6d9a7cb184f03fa76f2d69" Dec 10 23:19:36 crc kubenswrapper[4791]: I1210 23:19:36.913789 4791 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"1f601b495c523b51fe65ba26775c6406db8f2e7dce6d9a7cb184f03fa76f2d69"} err="failed to get container status \"1f601b495c523b51fe65ba26775c6406db8f2e7dce6d9a7cb184f03fa76f2d69\": rpc error: code = NotFound desc = could not find container \"1f601b495c523b51fe65ba26775c6406db8f2e7dce6d9a7cb184f03fa76f2d69\": container with ID starting with 1f601b495c523b51fe65ba26775c6406db8f2e7dce6d9a7cb184f03fa76f2d69 not found: ID does not exist" Dec 10 23:19:36 crc kubenswrapper[4791]: I1210 23:19:36.913820 4791 scope.go:117] "RemoveContainer" containerID="178371a38a9a8ff14eb3a60f7dfd8af3af6a7bb75b3c70b04412521d48453387" Dec 10 23:19:36 crc kubenswrapper[4791]: E1210 23:19:36.914117 4791 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"178371a38a9a8ff14eb3a60f7dfd8af3af6a7bb75b3c70b04412521d48453387\": container with ID starting with 178371a38a9a8ff14eb3a60f7dfd8af3af6a7bb75b3c70b04412521d48453387 not found: ID does not exist" containerID="178371a38a9a8ff14eb3a60f7dfd8af3af6a7bb75b3c70b04412521d48453387" Dec 10 23:19:36 crc kubenswrapper[4791]: I1210 23:19:36.914146 4791 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"178371a38a9a8ff14eb3a60f7dfd8af3af6a7bb75b3c70b04412521d48453387"} err="failed to get container status \"178371a38a9a8ff14eb3a60f7dfd8af3af6a7bb75b3c70b04412521d48453387\": rpc error: code = NotFound desc = could not find container \"178371a38a9a8ff14eb3a60f7dfd8af3af6a7bb75b3c70b04412521d48453387\": container with ID starting with 178371a38a9a8ff14eb3a60f7dfd8af3af6a7bb75b3c70b04412521d48453387 not found: ID does not exist" Dec 10 23:19:37 crc kubenswrapper[4791]: I1210 23:19:37.148881 4791 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-qz8bj"] Dec 10 23:19:37 crc kubenswrapper[4791]: I1210 23:19:37.160468 4791 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/community-operators-qz8bj"] Dec 10 23:19:37 crc kubenswrapper[4791]: I1210 23:19:37.531997 4791 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-marketplace/redhat-operators-rccn4" podUID="05961153-0a72-4c04-bea9-f622c3b54460" containerName="registry-server" probeResult="failure" output=< Dec 10 23:19:37 crc kubenswrapper[4791]: timeout: failed to connect service ":50051" within 1s Dec 10 23:19:37 crc kubenswrapper[4791]: > Dec 10 23:19:37 crc kubenswrapper[4791]: I1210 23:19:37.902158 4791 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="ee228122-ccd4-4d15-9ba8-a8b75ad66fa3" path="/var/lib/kubelet/pods/ee228122-ccd4-4d15-9ba8-a8b75ad66fa3/volumes" Dec 10 23:19:46 crc kubenswrapper[4791]: I1210 23:19:46.544945 4791 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-operators-rccn4" Dec 10 23:19:46 crc kubenswrapper[4791]: I1210 23:19:46.633411 4791 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-operators-rccn4" Dec 10 23:19:46 crc kubenswrapper[4791]: I1210 23:19:46.796607 4791 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-rccn4"] Dec 10 23:19:47 crc kubenswrapper[4791]: I1210 23:19:47.912937 4791 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-operators-rccn4" podUID="05961153-0a72-4c04-bea9-f622c3b54460" containerName="registry-server" containerID="cri-o://d56168a0a555acfb35cd404aa3d4b8e24e977697cd2cfc2f706730d6fea96578" gracePeriod=2 Dec 10 23:19:48 crc kubenswrapper[4791]: I1210 23:19:48.355180 4791 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-rccn4" Dec 10 23:19:48 crc kubenswrapper[4791]: I1210 23:19:48.500286 4791 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/05961153-0a72-4c04-bea9-f622c3b54460-catalog-content\") pod \"05961153-0a72-4c04-bea9-f622c3b54460\" (UID: \"05961153-0a72-4c04-bea9-f622c3b54460\") " Dec 10 23:19:48 crc kubenswrapper[4791]: I1210 23:19:48.500507 4791 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-bqkkb\" (UniqueName: \"kubernetes.io/projected/05961153-0a72-4c04-bea9-f622c3b54460-kube-api-access-bqkkb\") pod \"05961153-0a72-4c04-bea9-f622c3b54460\" (UID: \"05961153-0a72-4c04-bea9-f622c3b54460\") " Dec 10 23:19:48 crc kubenswrapper[4791]: I1210 23:19:48.500550 4791 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/05961153-0a72-4c04-bea9-f622c3b54460-utilities\") pod \"05961153-0a72-4c04-bea9-f622c3b54460\" (UID: \"05961153-0a72-4c04-bea9-f622c3b54460\") " Dec 10 23:19:48 crc kubenswrapper[4791]: I1210 23:19:48.501512 4791 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/05961153-0a72-4c04-bea9-f622c3b54460-utilities" (OuterVolumeSpecName: "utilities") pod "05961153-0a72-4c04-bea9-f622c3b54460" (UID: "05961153-0a72-4c04-bea9-f622c3b54460"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 10 23:19:48 crc kubenswrapper[4791]: I1210 23:19:48.508383 4791 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/05961153-0a72-4c04-bea9-f622c3b54460-kube-api-access-bqkkb" (OuterVolumeSpecName: "kube-api-access-bqkkb") pod "05961153-0a72-4c04-bea9-f622c3b54460" (UID: "05961153-0a72-4c04-bea9-f622c3b54460"). InnerVolumeSpecName "kube-api-access-bqkkb". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 10 23:19:48 crc kubenswrapper[4791]: I1210 23:19:48.602848 4791 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-bqkkb\" (UniqueName: \"kubernetes.io/projected/05961153-0a72-4c04-bea9-f622c3b54460-kube-api-access-bqkkb\") on node \"crc\" DevicePath \"\"" Dec 10 23:19:48 crc kubenswrapper[4791]: I1210 23:19:48.602884 4791 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/05961153-0a72-4c04-bea9-f622c3b54460-utilities\") on node \"crc\" DevicePath \"\"" Dec 10 23:19:48 crc kubenswrapper[4791]: I1210 23:19:48.647824 4791 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/05961153-0a72-4c04-bea9-f622c3b54460-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "05961153-0a72-4c04-bea9-f622c3b54460" (UID: "05961153-0a72-4c04-bea9-f622c3b54460"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 10 23:19:48 crc kubenswrapper[4791]: I1210 23:19:48.704079 4791 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/05961153-0a72-4c04-bea9-f622c3b54460-catalog-content\") on node \"crc\" DevicePath \"\"" Dec 10 23:19:48 crc kubenswrapper[4791]: I1210 23:19:48.885735 4791 scope.go:117] "RemoveContainer" containerID="f12779a795c9f6f63a93027d53c67c89d0179d87b9df554c25bfc607f06caff1" Dec 10 23:19:48 crc kubenswrapper[4791]: E1210 23:19:48.885986 4791 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-5rb5l_openshift-machine-config-operator(ba35653c-6e06-4cee-a4d6-137764090d18)\"" pod="openshift-machine-config-operator/machine-config-daemon-5rb5l" podUID="ba35653c-6e06-4cee-a4d6-137764090d18" Dec 10 23:19:48 crc kubenswrapper[4791]: I1210 23:19:48.929557 4791 generic.go:334] "Generic (PLEG): container finished" podID="05961153-0a72-4c04-bea9-f622c3b54460" containerID="d56168a0a555acfb35cd404aa3d4b8e24e977697cd2cfc2f706730d6fea96578" exitCode=0 Dec 10 23:19:48 crc kubenswrapper[4791]: I1210 23:19:48.929599 4791 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-rccn4" event={"ID":"05961153-0a72-4c04-bea9-f622c3b54460","Type":"ContainerDied","Data":"d56168a0a555acfb35cd404aa3d4b8e24e977697cd2cfc2f706730d6fea96578"} Dec 10 23:19:48 crc kubenswrapper[4791]: I1210 23:19:48.929625 4791 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-rccn4" event={"ID":"05961153-0a72-4c04-bea9-f622c3b54460","Type":"ContainerDied","Data":"cb8927435bb2dcfc29b2dba2d7aad6a7c27c0dbae9577df995407e6df54520b7"} Dec 10 23:19:48 crc kubenswrapper[4791]: I1210 23:19:48.929641 4791 scope.go:117] "RemoveContainer" containerID="d56168a0a555acfb35cd404aa3d4b8e24e977697cd2cfc2f706730d6fea96578" Dec 10 23:19:48 crc kubenswrapper[4791]: I1210 23:19:48.929752 4791 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-rccn4" Dec 10 23:19:48 crc kubenswrapper[4791]: I1210 23:19:48.965742 4791 scope.go:117] "RemoveContainer" containerID="ab42257b441b8fbb32360ade40dae6a726fce47274f676ca283fce653172adde" Dec 10 23:19:48 crc kubenswrapper[4791]: I1210 23:19:48.968033 4791 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-rccn4"] Dec 10 23:19:48 crc kubenswrapper[4791]: I1210 23:19:48.977387 4791 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-operators-rccn4"] Dec 10 23:19:49 crc kubenswrapper[4791]: I1210 23:19:49.001412 4791 scope.go:117] "RemoveContainer" containerID="ad3d7800b15aeb482e1f77cd764aeed417388112e418a2145627373ca625b9c3" Dec 10 23:19:49 crc kubenswrapper[4791]: I1210 23:19:49.052652 4791 scope.go:117] "RemoveContainer" containerID="d56168a0a555acfb35cd404aa3d4b8e24e977697cd2cfc2f706730d6fea96578" Dec 10 23:19:49 crc kubenswrapper[4791]: E1210 23:19:49.053314 4791 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"d56168a0a555acfb35cd404aa3d4b8e24e977697cd2cfc2f706730d6fea96578\": container with ID starting with d56168a0a555acfb35cd404aa3d4b8e24e977697cd2cfc2f706730d6fea96578 not found: ID does not exist" containerID="d56168a0a555acfb35cd404aa3d4b8e24e977697cd2cfc2f706730d6fea96578" Dec 10 23:19:49 crc kubenswrapper[4791]: I1210 23:19:49.053395 4791 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"d56168a0a555acfb35cd404aa3d4b8e24e977697cd2cfc2f706730d6fea96578"} err="failed to get container status \"d56168a0a555acfb35cd404aa3d4b8e24e977697cd2cfc2f706730d6fea96578\": rpc error: code = NotFound desc = could not find container \"d56168a0a555acfb35cd404aa3d4b8e24e977697cd2cfc2f706730d6fea96578\": container with ID starting with d56168a0a555acfb35cd404aa3d4b8e24e977697cd2cfc2f706730d6fea96578 not found: ID does not exist" Dec 10 23:19:49 crc kubenswrapper[4791]: I1210 23:19:49.053427 4791 scope.go:117] "RemoveContainer" containerID="ab42257b441b8fbb32360ade40dae6a726fce47274f676ca283fce653172adde" Dec 10 23:19:49 crc kubenswrapper[4791]: E1210 23:19:49.053888 4791 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"ab42257b441b8fbb32360ade40dae6a726fce47274f676ca283fce653172adde\": container with ID starting with ab42257b441b8fbb32360ade40dae6a726fce47274f676ca283fce653172adde not found: ID does not exist" containerID="ab42257b441b8fbb32360ade40dae6a726fce47274f676ca283fce653172adde" Dec 10 23:19:49 crc kubenswrapper[4791]: I1210 23:19:49.053933 4791 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"ab42257b441b8fbb32360ade40dae6a726fce47274f676ca283fce653172adde"} err="failed to get container status \"ab42257b441b8fbb32360ade40dae6a726fce47274f676ca283fce653172adde\": rpc error: code = NotFound desc = could not find container \"ab42257b441b8fbb32360ade40dae6a726fce47274f676ca283fce653172adde\": container with ID starting with ab42257b441b8fbb32360ade40dae6a726fce47274f676ca283fce653172adde not found: ID does not exist" Dec 10 23:19:49 crc kubenswrapper[4791]: I1210 23:19:49.053951 4791 scope.go:117] "RemoveContainer" containerID="ad3d7800b15aeb482e1f77cd764aeed417388112e418a2145627373ca625b9c3" Dec 10 23:19:49 crc kubenswrapper[4791]: E1210 23:19:49.054381 4791 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"ad3d7800b15aeb482e1f77cd764aeed417388112e418a2145627373ca625b9c3\": container with ID starting with ad3d7800b15aeb482e1f77cd764aeed417388112e418a2145627373ca625b9c3 not found: ID does not exist" containerID="ad3d7800b15aeb482e1f77cd764aeed417388112e418a2145627373ca625b9c3" Dec 10 23:19:49 crc kubenswrapper[4791]: I1210 23:19:49.054471 4791 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"ad3d7800b15aeb482e1f77cd764aeed417388112e418a2145627373ca625b9c3"} err="failed to get container status \"ad3d7800b15aeb482e1f77cd764aeed417388112e418a2145627373ca625b9c3\": rpc error: code = NotFound desc = could not find container \"ad3d7800b15aeb482e1f77cd764aeed417388112e418a2145627373ca625b9c3\": container with ID starting with ad3d7800b15aeb482e1f77cd764aeed417388112e418a2145627373ca625b9c3 not found: ID does not exist" Dec 10 23:19:49 crc kubenswrapper[4791]: I1210 23:19:49.908727 4791 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="05961153-0a72-4c04-bea9-f622c3b54460" path="/var/lib/kubelet/pods/05961153-0a72-4c04-bea9-f622c3b54460/volumes" Dec 10 23:19:50 crc kubenswrapper[4791]: I1210 23:19:50.058326 4791 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-api-db-create-cvjnf"] Dec 10 23:19:50 crc kubenswrapper[4791]: I1210 23:19:50.067129 4791 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-cell1-b007-account-create-update-fkwxb"] Dec 10 23:19:50 crc kubenswrapper[4791]: I1210 23:19:50.076565 4791 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-api-db-create-cvjnf"] Dec 10 23:19:50 crc kubenswrapper[4791]: I1210 23:19:50.086172 4791 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-cell1-b007-account-create-update-fkwxb"] Dec 10 23:19:51 crc kubenswrapper[4791]: I1210 23:19:51.047089 4791 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-cell0-db-create-z2v58"] Dec 10 23:19:51 crc kubenswrapper[4791]: I1210 23:19:51.074104 4791 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-api-8832-account-create-update-wp7p5"] Dec 10 23:19:51 crc kubenswrapper[4791]: I1210 23:19:51.097088 4791 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-cell0-e207-account-create-update-hlmck"] Dec 10 23:19:51 crc kubenswrapper[4791]: I1210 23:19:51.106120 4791 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-cell1-db-create-tz8dr"] Dec 10 23:19:51 crc kubenswrapper[4791]: I1210 23:19:51.115429 4791 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-cell0-db-create-z2v58"] Dec 10 23:19:51 crc kubenswrapper[4791]: I1210 23:19:51.123411 4791 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-api-8832-account-create-update-wp7p5"] Dec 10 23:19:51 crc kubenswrapper[4791]: I1210 23:19:51.130814 4791 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-cell1-db-create-tz8dr"] Dec 10 23:19:51 crc kubenswrapper[4791]: I1210 23:19:51.137304 4791 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-cell0-e207-account-create-update-hlmck"] Dec 10 23:19:51 crc kubenswrapper[4791]: I1210 23:19:51.897311 4791 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="0a94efe1-86a8-4768-9199-f47051935bcf" path="/var/lib/kubelet/pods/0a94efe1-86a8-4768-9199-f47051935bcf/volumes" Dec 10 23:19:51 crc kubenswrapper[4791]: I1210 23:19:51.898215 4791 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="100aa755-c120-4c5a-9135-23bb98661f3d" path="/var/lib/kubelet/pods/100aa755-c120-4c5a-9135-23bb98661f3d/volumes" Dec 10 23:19:51 crc kubenswrapper[4791]: I1210 23:19:51.899027 4791 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="18c84e26-d0bb-4cd5-ba6a-49bf21155889" path="/var/lib/kubelet/pods/18c84e26-d0bb-4cd5-ba6a-49bf21155889/volumes" Dec 10 23:19:51 crc kubenswrapper[4791]: I1210 23:19:51.899699 4791 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="83ddd1be-5119-45a6-a080-1b63f0c8f7a1" path="/var/lib/kubelet/pods/83ddd1be-5119-45a6-a080-1b63f0c8f7a1/volumes" Dec 10 23:19:51 crc kubenswrapper[4791]: I1210 23:19:51.900825 4791 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="b8b6cd29-edfe-4177-bf14-14b8580d86c2" path="/var/lib/kubelet/pods/b8b6cd29-edfe-4177-bf14-14b8580d86c2/volumes" Dec 10 23:19:51 crc kubenswrapper[4791]: I1210 23:19:51.901397 4791 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="cf0595b8-0690-448d-9ef7-a64227bad653" path="/var/lib/kubelet/pods/cf0595b8-0690-448d-9ef7-a64227bad653/volumes" Dec 10 23:19:52 crc kubenswrapper[4791]: I1210 23:19:52.944993 4791 scope.go:117] "RemoveContainer" containerID="6b1df2861daf4c05537a80c73e098b117f3695e83c011f0b5ae42d4c9ac0ddc6" Dec 10 23:19:52 crc kubenswrapper[4791]: I1210 23:19:52.987578 4791 scope.go:117] "RemoveContainer" containerID="c18bd4b8e12d3f0f752d9f47afa76b3cc91ab1912ba9618ed62c16cf23683866" Dec 10 23:19:53 crc kubenswrapper[4791]: I1210 23:19:53.039730 4791 scope.go:117] "RemoveContainer" containerID="e63790203ba9fd9cd2790cabecfb42628c365964dca974e26ff03d91ff75d79d" Dec 10 23:19:53 crc kubenswrapper[4791]: I1210 23:19:53.082855 4791 scope.go:117] "RemoveContainer" containerID="f0eb0592a4a0cdde8425fc8222755e9b651cc41049d8057c83c99f2ae97e73a5" Dec 10 23:19:53 crc kubenswrapper[4791]: I1210 23:19:53.162960 4791 scope.go:117] "RemoveContainer" containerID="d9f21c9e92337da1466d284b85b264870b7e25dac547e118e080479b8e75affa" Dec 10 23:19:53 crc kubenswrapper[4791]: I1210 23:19:53.191117 4791 scope.go:117] "RemoveContainer" containerID="473c10a544d131d442414c7e9d016b5e15a352a6c22d2a9114caba72711977f5" Dec 10 23:19:53 crc kubenswrapper[4791]: I1210 23:19:53.233220 4791 scope.go:117] "RemoveContainer" containerID="8fdda0940a6cb89ce941ba64df5df3bda9bee4bd20519bf14e86899f148c962d" Dec 10 23:19:53 crc kubenswrapper[4791]: I1210 23:19:53.277439 4791 scope.go:117] "RemoveContainer" containerID="25ee7ca4ad03fcd0c892fe328f36b3e3b779fd605dd652f8b53cf1cf97a7e30e" Dec 10 23:19:53 crc kubenswrapper[4791]: I1210 23:19:53.297176 4791 scope.go:117] "RemoveContainer" containerID="96fce98c2678503c67e20f58365f73a09244ac81b06aa07e100ddbf7bb1368d9" Dec 10 23:19:53 crc kubenswrapper[4791]: I1210 23:19:53.329142 4791 scope.go:117] "RemoveContainer" containerID="c8c6f523ef5f1d59b753e569f79692fbe8d9bc4939c81ccec02f0720f1fb8447" Dec 10 23:19:57 crc kubenswrapper[4791]: I1210 23:19:57.020738 4791 generic.go:334] "Generic (PLEG): container finished" podID="852ca005-3b2f-407e-ae86-9b1c5a5fe182" containerID="7a1d98890d7381e4ea9261c3cf3af4edcef1cff3f649a44f2ed84a6ec6d2c0c6" exitCode=0 Dec 10 23:19:57 crc kubenswrapper[4791]: I1210 23:19:57.020839 4791 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/configure-network-edpm-deployment-openstack-edpm-ipam-wf5sq" event={"ID":"852ca005-3b2f-407e-ae86-9b1c5a5fe182","Type":"ContainerDied","Data":"7a1d98890d7381e4ea9261c3cf3af4edcef1cff3f649a44f2ed84a6ec6d2c0c6"} Dec 10 23:19:58 crc kubenswrapper[4791]: I1210 23:19:58.452907 4791 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/configure-network-edpm-deployment-openstack-edpm-ipam-wf5sq" Dec 10 23:19:58 crc kubenswrapper[4791]: I1210 23:19:58.509167 4791 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/852ca005-3b2f-407e-ae86-9b1c5a5fe182-inventory\") pod \"852ca005-3b2f-407e-ae86-9b1c5a5fe182\" (UID: \"852ca005-3b2f-407e-ae86-9b1c5a5fe182\") " Dec 10 23:19:58 crc kubenswrapper[4791]: I1210 23:19:58.509290 4791 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-52twz\" (UniqueName: \"kubernetes.io/projected/852ca005-3b2f-407e-ae86-9b1c5a5fe182-kube-api-access-52twz\") pod \"852ca005-3b2f-407e-ae86-9b1c5a5fe182\" (UID: \"852ca005-3b2f-407e-ae86-9b1c5a5fe182\") " Dec 10 23:19:58 crc kubenswrapper[4791]: I1210 23:19:58.509406 4791 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/852ca005-3b2f-407e-ae86-9b1c5a5fe182-ssh-key\") pod \"852ca005-3b2f-407e-ae86-9b1c5a5fe182\" (UID: \"852ca005-3b2f-407e-ae86-9b1c5a5fe182\") " Dec 10 23:19:58 crc kubenswrapper[4791]: I1210 23:19:58.517330 4791 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/852ca005-3b2f-407e-ae86-9b1c5a5fe182-kube-api-access-52twz" (OuterVolumeSpecName: "kube-api-access-52twz") pod "852ca005-3b2f-407e-ae86-9b1c5a5fe182" (UID: "852ca005-3b2f-407e-ae86-9b1c5a5fe182"). InnerVolumeSpecName "kube-api-access-52twz". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 10 23:19:58 crc kubenswrapper[4791]: I1210 23:19:58.544507 4791 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/852ca005-3b2f-407e-ae86-9b1c5a5fe182-inventory" (OuterVolumeSpecName: "inventory") pod "852ca005-3b2f-407e-ae86-9b1c5a5fe182" (UID: "852ca005-3b2f-407e-ae86-9b1c5a5fe182"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 10 23:19:58 crc kubenswrapper[4791]: I1210 23:19:58.544996 4791 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/852ca005-3b2f-407e-ae86-9b1c5a5fe182-ssh-key" (OuterVolumeSpecName: "ssh-key") pod "852ca005-3b2f-407e-ae86-9b1c5a5fe182" (UID: "852ca005-3b2f-407e-ae86-9b1c5a5fe182"). InnerVolumeSpecName "ssh-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 10 23:19:58 crc kubenswrapper[4791]: I1210 23:19:58.610636 4791 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/852ca005-3b2f-407e-ae86-9b1c5a5fe182-inventory\") on node \"crc\" DevicePath \"\"" Dec 10 23:19:58 crc kubenswrapper[4791]: I1210 23:19:58.610664 4791 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-52twz\" (UniqueName: \"kubernetes.io/projected/852ca005-3b2f-407e-ae86-9b1c5a5fe182-kube-api-access-52twz\") on node \"crc\" DevicePath \"\"" Dec 10 23:19:58 crc kubenswrapper[4791]: I1210 23:19:58.610675 4791 reconciler_common.go:293] "Volume detached for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/852ca005-3b2f-407e-ae86-9b1c5a5fe182-ssh-key\") on node \"crc\" DevicePath \"\"" Dec 10 23:19:59 crc kubenswrapper[4791]: I1210 23:19:59.042399 4791 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/configure-network-edpm-deployment-openstack-edpm-ipam-wf5sq" event={"ID":"852ca005-3b2f-407e-ae86-9b1c5a5fe182","Type":"ContainerDied","Data":"6c0802937126fd1839647890c5e56c806b8a8870202d71af6d0c883fd2bec85a"} Dec 10 23:19:59 crc kubenswrapper[4791]: I1210 23:19:59.042441 4791 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="6c0802937126fd1839647890c5e56c806b8a8870202d71af6d0c883fd2bec85a" Dec 10 23:19:59 crc kubenswrapper[4791]: I1210 23:19:59.042504 4791 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/configure-network-edpm-deployment-openstack-edpm-ipam-wf5sq" Dec 10 23:19:59 crc kubenswrapper[4791]: I1210 23:19:59.160354 4791 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/validate-network-edpm-deployment-openstack-edpm-ipam-gtc8x"] Dec 10 23:19:59 crc kubenswrapper[4791]: E1210 23:19:59.164661 4791 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="05961153-0a72-4c04-bea9-f622c3b54460" containerName="registry-server" Dec 10 23:19:59 crc kubenswrapper[4791]: I1210 23:19:59.164688 4791 state_mem.go:107] "Deleted CPUSet assignment" podUID="05961153-0a72-4c04-bea9-f622c3b54460" containerName="registry-server" Dec 10 23:19:59 crc kubenswrapper[4791]: E1210 23:19:59.164712 4791 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="ee228122-ccd4-4d15-9ba8-a8b75ad66fa3" containerName="extract-utilities" Dec 10 23:19:59 crc kubenswrapper[4791]: I1210 23:19:59.164719 4791 state_mem.go:107] "Deleted CPUSet assignment" podUID="ee228122-ccd4-4d15-9ba8-a8b75ad66fa3" containerName="extract-utilities" Dec 10 23:19:59 crc kubenswrapper[4791]: E1210 23:19:59.164755 4791 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="05961153-0a72-4c04-bea9-f622c3b54460" containerName="extract-content" Dec 10 23:19:59 crc kubenswrapper[4791]: I1210 23:19:59.164761 4791 state_mem.go:107] "Deleted CPUSet assignment" podUID="05961153-0a72-4c04-bea9-f622c3b54460" containerName="extract-content" Dec 10 23:19:59 crc kubenswrapper[4791]: E1210 23:19:59.164779 4791 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="ee228122-ccd4-4d15-9ba8-a8b75ad66fa3" containerName="registry-server" Dec 10 23:19:59 crc kubenswrapper[4791]: I1210 23:19:59.164784 4791 state_mem.go:107] "Deleted CPUSet assignment" podUID="ee228122-ccd4-4d15-9ba8-a8b75ad66fa3" containerName="registry-server" Dec 10 23:19:59 crc kubenswrapper[4791]: E1210 23:19:59.164798 4791 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="ee228122-ccd4-4d15-9ba8-a8b75ad66fa3" containerName="extract-content" Dec 10 23:19:59 crc kubenswrapper[4791]: I1210 23:19:59.164805 4791 state_mem.go:107] "Deleted CPUSet assignment" podUID="ee228122-ccd4-4d15-9ba8-a8b75ad66fa3" containerName="extract-content" Dec 10 23:19:59 crc kubenswrapper[4791]: E1210 23:19:59.164830 4791 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="852ca005-3b2f-407e-ae86-9b1c5a5fe182" containerName="configure-network-edpm-deployment-openstack-edpm-ipam" Dec 10 23:19:59 crc kubenswrapper[4791]: I1210 23:19:59.164845 4791 state_mem.go:107] "Deleted CPUSet assignment" podUID="852ca005-3b2f-407e-ae86-9b1c5a5fe182" containerName="configure-network-edpm-deployment-openstack-edpm-ipam" Dec 10 23:19:59 crc kubenswrapper[4791]: E1210 23:19:59.164864 4791 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="05961153-0a72-4c04-bea9-f622c3b54460" containerName="extract-utilities" Dec 10 23:19:59 crc kubenswrapper[4791]: I1210 23:19:59.164872 4791 state_mem.go:107] "Deleted CPUSet assignment" podUID="05961153-0a72-4c04-bea9-f622c3b54460" containerName="extract-utilities" Dec 10 23:19:59 crc kubenswrapper[4791]: I1210 23:19:59.165639 4791 memory_manager.go:354] "RemoveStaleState removing state" podUID="852ca005-3b2f-407e-ae86-9b1c5a5fe182" containerName="configure-network-edpm-deployment-openstack-edpm-ipam" Dec 10 23:19:59 crc kubenswrapper[4791]: I1210 23:19:59.165671 4791 memory_manager.go:354] "RemoveStaleState removing state" podUID="05961153-0a72-4c04-bea9-f622c3b54460" containerName="registry-server" Dec 10 23:19:59 crc kubenswrapper[4791]: I1210 23:19:59.165691 4791 memory_manager.go:354] "RemoveStaleState removing state" podUID="ee228122-ccd4-4d15-9ba8-a8b75ad66fa3" containerName="registry-server" Dec 10 23:19:59 crc kubenswrapper[4791]: I1210 23:19:59.166609 4791 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/validate-network-edpm-deployment-openstack-edpm-ipam-gtc8x" Dec 10 23:19:59 crc kubenswrapper[4791]: I1210 23:19:59.169248 4791 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-dockercfg-hf8pr" Dec 10 23:19:59 crc kubenswrapper[4791]: I1210 23:19:59.169384 4791 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Dec 10 23:19:59 crc kubenswrapper[4791]: I1210 23:19:59.169894 4791 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-edpm-ipam" Dec 10 23:19:59 crc kubenswrapper[4791]: I1210 23:19:59.170496 4791 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-ansible-ssh-private-key-secret" Dec 10 23:19:59 crc kubenswrapper[4791]: I1210 23:19:59.182934 4791 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/validate-network-edpm-deployment-openstack-edpm-ipam-gtc8x"] Dec 10 23:19:59 crc kubenswrapper[4791]: I1210 23:19:59.223326 4791 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-xqtl2\" (UniqueName: \"kubernetes.io/projected/73316991-2ef6-4257-8118-55378571b7d7-kube-api-access-xqtl2\") pod \"validate-network-edpm-deployment-openstack-edpm-ipam-gtc8x\" (UID: \"73316991-2ef6-4257-8118-55378571b7d7\") " pod="openstack/validate-network-edpm-deployment-openstack-edpm-ipam-gtc8x" Dec 10 23:19:59 crc kubenswrapper[4791]: I1210 23:19:59.223437 4791 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/73316991-2ef6-4257-8118-55378571b7d7-ssh-key\") pod \"validate-network-edpm-deployment-openstack-edpm-ipam-gtc8x\" (UID: \"73316991-2ef6-4257-8118-55378571b7d7\") " pod="openstack/validate-network-edpm-deployment-openstack-edpm-ipam-gtc8x" Dec 10 23:19:59 crc kubenswrapper[4791]: I1210 23:19:59.223568 4791 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/73316991-2ef6-4257-8118-55378571b7d7-inventory\") pod \"validate-network-edpm-deployment-openstack-edpm-ipam-gtc8x\" (UID: \"73316991-2ef6-4257-8118-55378571b7d7\") " pod="openstack/validate-network-edpm-deployment-openstack-edpm-ipam-gtc8x" Dec 10 23:19:59 crc kubenswrapper[4791]: I1210 23:19:59.325152 4791 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-xqtl2\" (UniqueName: \"kubernetes.io/projected/73316991-2ef6-4257-8118-55378571b7d7-kube-api-access-xqtl2\") pod \"validate-network-edpm-deployment-openstack-edpm-ipam-gtc8x\" (UID: \"73316991-2ef6-4257-8118-55378571b7d7\") " pod="openstack/validate-network-edpm-deployment-openstack-edpm-ipam-gtc8x" Dec 10 23:19:59 crc kubenswrapper[4791]: I1210 23:19:59.325634 4791 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/73316991-2ef6-4257-8118-55378571b7d7-ssh-key\") pod \"validate-network-edpm-deployment-openstack-edpm-ipam-gtc8x\" (UID: \"73316991-2ef6-4257-8118-55378571b7d7\") " pod="openstack/validate-network-edpm-deployment-openstack-edpm-ipam-gtc8x" Dec 10 23:19:59 crc kubenswrapper[4791]: I1210 23:19:59.326834 4791 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/73316991-2ef6-4257-8118-55378571b7d7-inventory\") pod \"validate-network-edpm-deployment-openstack-edpm-ipam-gtc8x\" (UID: \"73316991-2ef6-4257-8118-55378571b7d7\") " pod="openstack/validate-network-edpm-deployment-openstack-edpm-ipam-gtc8x" Dec 10 23:19:59 crc kubenswrapper[4791]: I1210 23:19:59.330499 4791 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/73316991-2ef6-4257-8118-55378571b7d7-ssh-key\") pod \"validate-network-edpm-deployment-openstack-edpm-ipam-gtc8x\" (UID: \"73316991-2ef6-4257-8118-55378571b7d7\") " pod="openstack/validate-network-edpm-deployment-openstack-edpm-ipam-gtc8x" Dec 10 23:19:59 crc kubenswrapper[4791]: I1210 23:19:59.335529 4791 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/73316991-2ef6-4257-8118-55378571b7d7-inventory\") pod \"validate-network-edpm-deployment-openstack-edpm-ipam-gtc8x\" (UID: \"73316991-2ef6-4257-8118-55378571b7d7\") " pod="openstack/validate-network-edpm-deployment-openstack-edpm-ipam-gtc8x" Dec 10 23:19:59 crc kubenswrapper[4791]: I1210 23:19:59.345496 4791 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-xqtl2\" (UniqueName: \"kubernetes.io/projected/73316991-2ef6-4257-8118-55378571b7d7-kube-api-access-xqtl2\") pod \"validate-network-edpm-deployment-openstack-edpm-ipam-gtc8x\" (UID: \"73316991-2ef6-4257-8118-55378571b7d7\") " pod="openstack/validate-network-edpm-deployment-openstack-edpm-ipam-gtc8x" Dec 10 23:19:59 crc kubenswrapper[4791]: I1210 23:19:59.494407 4791 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/validate-network-edpm-deployment-openstack-edpm-ipam-gtc8x" Dec 10 23:19:59 crc kubenswrapper[4791]: I1210 23:19:59.885245 4791 scope.go:117] "RemoveContainer" containerID="f12779a795c9f6f63a93027d53c67c89d0179d87b9df554c25bfc607f06caff1" Dec 10 23:19:59 crc kubenswrapper[4791]: E1210 23:19:59.885679 4791 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-5rb5l_openshift-machine-config-operator(ba35653c-6e06-4cee-a4d6-137764090d18)\"" pod="openshift-machine-config-operator/machine-config-daemon-5rb5l" podUID="ba35653c-6e06-4cee-a4d6-137764090d18" Dec 10 23:20:00 crc kubenswrapper[4791]: I1210 23:20:00.052841 4791 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/validate-network-edpm-deployment-openstack-edpm-ipam-gtc8x"] Dec 10 23:20:01 crc kubenswrapper[4791]: I1210 23:20:01.062176 4791 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/validate-network-edpm-deployment-openstack-edpm-ipam-gtc8x" event={"ID":"73316991-2ef6-4257-8118-55378571b7d7","Type":"ContainerStarted","Data":"ba4af7aa18f2a8d3a973173bfdfae5bf12714ebce6a0ad5192306dfb1e81180e"} Dec 10 23:20:02 crc kubenswrapper[4791]: I1210 23:20:02.070854 4791 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/validate-network-edpm-deployment-openstack-edpm-ipam-gtc8x" event={"ID":"73316991-2ef6-4257-8118-55378571b7d7","Type":"ContainerStarted","Data":"06c95eca2b98905df18f4275cb40f66381801d0e4f1060895db92b1aae67fc6c"} Dec 10 23:20:02 crc kubenswrapper[4791]: I1210 23:20:02.102726 4791 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/validate-network-edpm-deployment-openstack-edpm-ipam-gtc8x" podStartSLOduration=2.155914519 podStartE2EDuration="3.102699523s" podCreationTimestamp="2025-12-10 23:19:59 +0000 UTC" firstStartedPulling="2025-12-10 23:20:00.059481642 +0000 UTC m=+1834.489099255" lastFinishedPulling="2025-12-10 23:20:01.006266646 +0000 UTC m=+1835.435884259" observedRunningTime="2025-12-10 23:20:02.089644497 +0000 UTC m=+1836.519262130" watchObservedRunningTime="2025-12-10 23:20:02.102699523 +0000 UTC m=+1836.532317136" Dec 10 23:20:06 crc kubenswrapper[4791]: I1210 23:20:06.118611 4791 generic.go:334] "Generic (PLEG): container finished" podID="73316991-2ef6-4257-8118-55378571b7d7" containerID="06c95eca2b98905df18f4275cb40f66381801d0e4f1060895db92b1aae67fc6c" exitCode=0 Dec 10 23:20:06 crc kubenswrapper[4791]: I1210 23:20:06.118675 4791 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/validate-network-edpm-deployment-openstack-edpm-ipam-gtc8x" event={"ID":"73316991-2ef6-4257-8118-55378571b7d7","Type":"ContainerDied","Data":"06c95eca2b98905df18f4275cb40f66381801d0e4f1060895db92b1aae67fc6c"} Dec 10 23:20:07 crc kubenswrapper[4791]: I1210 23:20:07.562040 4791 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/validate-network-edpm-deployment-openstack-edpm-ipam-gtc8x" Dec 10 23:20:07 crc kubenswrapper[4791]: I1210 23:20:07.603318 4791 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/73316991-2ef6-4257-8118-55378571b7d7-ssh-key\") pod \"73316991-2ef6-4257-8118-55378571b7d7\" (UID: \"73316991-2ef6-4257-8118-55378571b7d7\") " Dec 10 23:20:07 crc kubenswrapper[4791]: I1210 23:20:07.603457 4791 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-xqtl2\" (UniqueName: \"kubernetes.io/projected/73316991-2ef6-4257-8118-55378571b7d7-kube-api-access-xqtl2\") pod \"73316991-2ef6-4257-8118-55378571b7d7\" (UID: \"73316991-2ef6-4257-8118-55378571b7d7\") " Dec 10 23:20:07 crc kubenswrapper[4791]: I1210 23:20:07.603511 4791 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/73316991-2ef6-4257-8118-55378571b7d7-inventory\") pod \"73316991-2ef6-4257-8118-55378571b7d7\" (UID: \"73316991-2ef6-4257-8118-55378571b7d7\") " Dec 10 23:20:07 crc kubenswrapper[4791]: I1210 23:20:07.633953 4791 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/73316991-2ef6-4257-8118-55378571b7d7-kube-api-access-xqtl2" (OuterVolumeSpecName: "kube-api-access-xqtl2") pod "73316991-2ef6-4257-8118-55378571b7d7" (UID: "73316991-2ef6-4257-8118-55378571b7d7"). InnerVolumeSpecName "kube-api-access-xqtl2". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 10 23:20:07 crc kubenswrapper[4791]: I1210 23:20:07.671102 4791 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/73316991-2ef6-4257-8118-55378571b7d7-inventory" (OuterVolumeSpecName: "inventory") pod "73316991-2ef6-4257-8118-55378571b7d7" (UID: "73316991-2ef6-4257-8118-55378571b7d7"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 10 23:20:07 crc kubenswrapper[4791]: I1210 23:20:07.672549 4791 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/73316991-2ef6-4257-8118-55378571b7d7-ssh-key" (OuterVolumeSpecName: "ssh-key") pod "73316991-2ef6-4257-8118-55378571b7d7" (UID: "73316991-2ef6-4257-8118-55378571b7d7"). InnerVolumeSpecName "ssh-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 10 23:20:07 crc kubenswrapper[4791]: I1210 23:20:07.704934 4791 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-xqtl2\" (UniqueName: \"kubernetes.io/projected/73316991-2ef6-4257-8118-55378571b7d7-kube-api-access-xqtl2\") on node \"crc\" DevicePath \"\"" Dec 10 23:20:07 crc kubenswrapper[4791]: I1210 23:20:07.704969 4791 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/73316991-2ef6-4257-8118-55378571b7d7-inventory\") on node \"crc\" DevicePath \"\"" Dec 10 23:20:07 crc kubenswrapper[4791]: I1210 23:20:07.704978 4791 reconciler_common.go:293] "Volume detached for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/73316991-2ef6-4257-8118-55378571b7d7-ssh-key\") on node \"crc\" DevicePath \"\"" Dec 10 23:20:08 crc kubenswrapper[4791]: I1210 23:20:08.137235 4791 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/validate-network-edpm-deployment-openstack-edpm-ipam-gtc8x" event={"ID":"73316991-2ef6-4257-8118-55378571b7d7","Type":"ContainerDied","Data":"ba4af7aa18f2a8d3a973173bfdfae5bf12714ebce6a0ad5192306dfb1e81180e"} Dec 10 23:20:08 crc kubenswrapper[4791]: I1210 23:20:08.137279 4791 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="ba4af7aa18f2a8d3a973173bfdfae5bf12714ebce6a0ad5192306dfb1e81180e" Dec 10 23:20:08 crc kubenswrapper[4791]: I1210 23:20:08.137295 4791 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/validate-network-edpm-deployment-openstack-edpm-ipam-gtc8x" Dec 10 23:20:08 crc kubenswrapper[4791]: I1210 23:20:08.208943 4791 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/install-os-edpm-deployment-openstack-edpm-ipam-c6nvg"] Dec 10 23:20:08 crc kubenswrapper[4791]: E1210 23:20:08.209403 4791 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="73316991-2ef6-4257-8118-55378571b7d7" containerName="validate-network-edpm-deployment-openstack-edpm-ipam" Dec 10 23:20:08 crc kubenswrapper[4791]: I1210 23:20:08.209420 4791 state_mem.go:107] "Deleted CPUSet assignment" podUID="73316991-2ef6-4257-8118-55378571b7d7" containerName="validate-network-edpm-deployment-openstack-edpm-ipam" Dec 10 23:20:08 crc kubenswrapper[4791]: I1210 23:20:08.209601 4791 memory_manager.go:354] "RemoveStaleState removing state" podUID="73316991-2ef6-4257-8118-55378571b7d7" containerName="validate-network-edpm-deployment-openstack-edpm-ipam" Dec 10 23:20:08 crc kubenswrapper[4791]: I1210 23:20:08.210236 4791 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/install-os-edpm-deployment-openstack-edpm-ipam-c6nvg" Dec 10 23:20:08 crc kubenswrapper[4791]: I1210 23:20:08.214726 4791 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-ansible-ssh-private-key-secret" Dec 10 23:20:08 crc kubenswrapper[4791]: I1210 23:20:08.214962 4791 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-edpm-ipam" Dec 10 23:20:08 crc kubenswrapper[4791]: I1210 23:20:08.215848 4791 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Dec 10 23:20:08 crc kubenswrapper[4791]: I1210 23:20:08.215987 4791 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-dockercfg-hf8pr" Dec 10 23:20:08 crc kubenswrapper[4791]: I1210 23:20:08.217771 4791 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/19add645-abcb-46e3-a103-e39305c0be1c-inventory\") pod \"install-os-edpm-deployment-openstack-edpm-ipam-c6nvg\" (UID: \"19add645-abcb-46e3-a103-e39305c0be1c\") " pod="openstack/install-os-edpm-deployment-openstack-edpm-ipam-c6nvg" Dec 10 23:20:08 crc kubenswrapper[4791]: I1210 23:20:08.218004 4791 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/19add645-abcb-46e3-a103-e39305c0be1c-ssh-key\") pod \"install-os-edpm-deployment-openstack-edpm-ipam-c6nvg\" (UID: \"19add645-abcb-46e3-a103-e39305c0be1c\") " pod="openstack/install-os-edpm-deployment-openstack-edpm-ipam-c6nvg" Dec 10 23:20:08 crc kubenswrapper[4791]: I1210 23:20:08.218077 4791 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-2pnxl\" (UniqueName: \"kubernetes.io/projected/19add645-abcb-46e3-a103-e39305c0be1c-kube-api-access-2pnxl\") pod \"install-os-edpm-deployment-openstack-edpm-ipam-c6nvg\" (UID: \"19add645-abcb-46e3-a103-e39305c0be1c\") " pod="openstack/install-os-edpm-deployment-openstack-edpm-ipam-c6nvg" Dec 10 23:20:08 crc kubenswrapper[4791]: I1210 23:20:08.234228 4791 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/install-os-edpm-deployment-openstack-edpm-ipam-c6nvg"] Dec 10 23:20:08 crc kubenswrapper[4791]: I1210 23:20:08.320472 4791 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/19add645-abcb-46e3-a103-e39305c0be1c-ssh-key\") pod \"install-os-edpm-deployment-openstack-edpm-ipam-c6nvg\" (UID: \"19add645-abcb-46e3-a103-e39305c0be1c\") " pod="openstack/install-os-edpm-deployment-openstack-edpm-ipam-c6nvg" Dec 10 23:20:08 crc kubenswrapper[4791]: I1210 23:20:08.320536 4791 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-2pnxl\" (UniqueName: \"kubernetes.io/projected/19add645-abcb-46e3-a103-e39305c0be1c-kube-api-access-2pnxl\") pod \"install-os-edpm-deployment-openstack-edpm-ipam-c6nvg\" (UID: \"19add645-abcb-46e3-a103-e39305c0be1c\") " pod="openstack/install-os-edpm-deployment-openstack-edpm-ipam-c6nvg" Dec 10 23:20:08 crc kubenswrapper[4791]: I1210 23:20:08.320641 4791 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/19add645-abcb-46e3-a103-e39305c0be1c-inventory\") pod \"install-os-edpm-deployment-openstack-edpm-ipam-c6nvg\" (UID: \"19add645-abcb-46e3-a103-e39305c0be1c\") " pod="openstack/install-os-edpm-deployment-openstack-edpm-ipam-c6nvg" Dec 10 23:20:08 crc kubenswrapper[4791]: I1210 23:20:08.325433 4791 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/19add645-abcb-46e3-a103-e39305c0be1c-ssh-key\") pod \"install-os-edpm-deployment-openstack-edpm-ipam-c6nvg\" (UID: \"19add645-abcb-46e3-a103-e39305c0be1c\") " pod="openstack/install-os-edpm-deployment-openstack-edpm-ipam-c6nvg" Dec 10 23:20:08 crc kubenswrapper[4791]: I1210 23:20:08.335832 4791 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/19add645-abcb-46e3-a103-e39305c0be1c-inventory\") pod \"install-os-edpm-deployment-openstack-edpm-ipam-c6nvg\" (UID: \"19add645-abcb-46e3-a103-e39305c0be1c\") " pod="openstack/install-os-edpm-deployment-openstack-edpm-ipam-c6nvg" Dec 10 23:20:08 crc kubenswrapper[4791]: I1210 23:20:08.344239 4791 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-2pnxl\" (UniqueName: \"kubernetes.io/projected/19add645-abcb-46e3-a103-e39305c0be1c-kube-api-access-2pnxl\") pod \"install-os-edpm-deployment-openstack-edpm-ipam-c6nvg\" (UID: \"19add645-abcb-46e3-a103-e39305c0be1c\") " pod="openstack/install-os-edpm-deployment-openstack-edpm-ipam-c6nvg" Dec 10 23:20:08 crc kubenswrapper[4791]: I1210 23:20:08.538597 4791 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/install-os-edpm-deployment-openstack-edpm-ipam-c6nvg" Dec 10 23:20:09 crc kubenswrapper[4791]: I1210 23:20:09.107030 4791 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/install-os-edpm-deployment-openstack-edpm-ipam-c6nvg"] Dec 10 23:20:09 crc kubenswrapper[4791]: I1210 23:20:09.145258 4791 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/install-os-edpm-deployment-openstack-edpm-ipam-c6nvg" event={"ID":"19add645-abcb-46e3-a103-e39305c0be1c","Type":"ContainerStarted","Data":"8cc932a4a47e0ce7b6e6252cc22f5aca75c98b6f19bf5c27ee0add3947d97389"} Dec 10 23:20:10 crc kubenswrapper[4791]: I1210 23:20:10.159147 4791 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/install-os-edpm-deployment-openstack-edpm-ipam-c6nvg" event={"ID":"19add645-abcb-46e3-a103-e39305c0be1c","Type":"ContainerStarted","Data":"a634c5c1284a57ec1965d056155daa2ef4c9cb85536659d223ab8f93ba603b67"} Dec 10 23:20:10 crc kubenswrapper[4791]: I1210 23:20:10.185989 4791 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/install-os-edpm-deployment-openstack-edpm-ipam-c6nvg" podStartSLOduration=1.731182051 podStartE2EDuration="2.185960187s" podCreationTimestamp="2025-12-10 23:20:08 +0000 UTC" firstStartedPulling="2025-12-10 23:20:09.118432732 +0000 UTC m=+1843.548050335" lastFinishedPulling="2025-12-10 23:20:09.573210838 +0000 UTC m=+1844.002828471" observedRunningTime="2025-12-10 23:20:10.18191835 +0000 UTC m=+1844.611535973" watchObservedRunningTime="2025-12-10 23:20:10.185960187 +0000 UTC m=+1844.615577840" Dec 10 23:20:12 crc kubenswrapper[4791]: I1210 23:20:12.885137 4791 scope.go:117] "RemoveContainer" containerID="f12779a795c9f6f63a93027d53c67c89d0179d87b9df554c25bfc607f06caff1" Dec 10 23:20:12 crc kubenswrapper[4791]: E1210 23:20:12.885706 4791 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-5rb5l_openshift-machine-config-operator(ba35653c-6e06-4cee-a4d6-137764090d18)\"" pod="openshift-machine-config-operator/machine-config-daemon-5rb5l" podUID="ba35653c-6e06-4cee-a4d6-137764090d18" Dec 10 23:20:23 crc kubenswrapper[4791]: I1210 23:20:23.071749 4791 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-cell0-conductor-db-sync-jff9b"] Dec 10 23:20:23 crc kubenswrapper[4791]: I1210 23:20:23.081096 4791 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-cell0-conductor-db-sync-jff9b"] Dec 10 23:20:23 crc kubenswrapper[4791]: I1210 23:20:23.886472 4791 scope.go:117] "RemoveContainer" containerID="f12779a795c9f6f63a93027d53c67c89d0179d87b9df554c25bfc607f06caff1" Dec 10 23:20:23 crc kubenswrapper[4791]: E1210 23:20:23.886844 4791 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-5rb5l_openshift-machine-config-operator(ba35653c-6e06-4cee-a4d6-137764090d18)\"" pod="openshift-machine-config-operator/machine-config-daemon-5rb5l" podUID="ba35653c-6e06-4cee-a4d6-137764090d18" Dec 10 23:20:23 crc kubenswrapper[4791]: I1210 23:20:23.905298 4791 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="075728ce-048e-452d-adff-0e3d167d6673" path="/var/lib/kubelet/pods/075728ce-048e-452d-adff-0e3d167d6673/volumes" Dec 10 23:20:34 crc kubenswrapper[4791]: I1210 23:20:34.885456 4791 scope.go:117] "RemoveContainer" containerID="f12779a795c9f6f63a93027d53c67c89d0179d87b9df554c25bfc607f06caff1" Dec 10 23:20:35 crc kubenswrapper[4791]: I1210 23:20:35.441250 4791 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-5rb5l" event={"ID":"ba35653c-6e06-4cee-a4d6-137764090d18","Type":"ContainerStarted","Data":"deb0e97b02d82d70f9f1de2c3e0794a3b407a099d09fc6ee2a1d821d42a42e21"} Dec 10 23:20:47 crc kubenswrapper[4791]: I1210 23:20:47.040786 4791 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-cell1-conductor-db-sync-brjck"] Dec 10 23:20:47 crc kubenswrapper[4791]: I1210 23:20:47.049556 4791 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-cell0-cell-mapping-k64vq"] Dec 10 23:20:47 crc kubenswrapper[4791]: I1210 23:20:47.062548 4791 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-cell1-conductor-db-sync-brjck"] Dec 10 23:20:47 crc kubenswrapper[4791]: I1210 23:20:47.071270 4791 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-cell0-cell-mapping-k64vq"] Dec 10 23:20:47 crc kubenswrapper[4791]: I1210 23:20:47.895002 4791 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="660af60d-b638-42d4-86d1-3a7a07b1b936" path="/var/lib/kubelet/pods/660af60d-b638-42d4-86d1-3a7a07b1b936/volumes" Dec 10 23:20:47 crc kubenswrapper[4791]: I1210 23:20:47.895938 4791 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="da9f584f-c289-4d67-ae91-8d6bf7507fb6" path="/var/lib/kubelet/pods/da9f584f-c289-4d67-ae91-8d6bf7507fb6/volumes" Dec 10 23:20:48 crc kubenswrapper[4791]: I1210 23:20:48.575051 4791 generic.go:334] "Generic (PLEG): container finished" podID="19add645-abcb-46e3-a103-e39305c0be1c" containerID="a634c5c1284a57ec1965d056155daa2ef4c9cb85536659d223ab8f93ba603b67" exitCode=0 Dec 10 23:20:48 crc kubenswrapper[4791]: I1210 23:20:48.575098 4791 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/install-os-edpm-deployment-openstack-edpm-ipam-c6nvg" event={"ID":"19add645-abcb-46e3-a103-e39305c0be1c","Type":"ContainerDied","Data":"a634c5c1284a57ec1965d056155daa2ef4c9cb85536659d223ab8f93ba603b67"} Dec 10 23:20:50 crc kubenswrapper[4791]: I1210 23:20:50.094942 4791 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/install-os-edpm-deployment-openstack-edpm-ipam-c6nvg" Dec 10 23:20:50 crc kubenswrapper[4791]: I1210 23:20:50.148935 4791 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/19add645-abcb-46e3-a103-e39305c0be1c-inventory\") pod \"19add645-abcb-46e3-a103-e39305c0be1c\" (UID: \"19add645-abcb-46e3-a103-e39305c0be1c\") " Dec 10 23:20:50 crc kubenswrapper[4791]: I1210 23:20:50.149116 4791 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-2pnxl\" (UniqueName: \"kubernetes.io/projected/19add645-abcb-46e3-a103-e39305c0be1c-kube-api-access-2pnxl\") pod \"19add645-abcb-46e3-a103-e39305c0be1c\" (UID: \"19add645-abcb-46e3-a103-e39305c0be1c\") " Dec 10 23:20:50 crc kubenswrapper[4791]: I1210 23:20:50.149209 4791 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/19add645-abcb-46e3-a103-e39305c0be1c-ssh-key\") pod \"19add645-abcb-46e3-a103-e39305c0be1c\" (UID: \"19add645-abcb-46e3-a103-e39305c0be1c\") " Dec 10 23:20:50 crc kubenswrapper[4791]: I1210 23:20:50.157532 4791 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/19add645-abcb-46e3-a103-e39305c0be1c-kube-api-access-2pnxl" (OuterVolumeSpecName: "kube-api-access-2pnxl") pod "19add645-abcb-46e3-a103-e39305c0be1c" (UID: "19add645-abcb-46e3-a103-e39305c0be1c"). InnerVolumeSpecName "kube-api-access-2pnxl". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 10 23:20:50 crc kubenswrapper[4791]: I1210 23:20:50.183066 4791 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/19add645-abcb-46e3-a103-e39305c0be1c-inventory" (OuterVolumeSpecName: "inventory") pod "19add645-abcb-46e3-a103-e39305c0be1c" (UID: "19add645-abcb-46e3-a103-e39305c0be1c"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 10 23:20:50 crc kubenswrapper[4791]: I1210 23:20:50.197518 4791 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/19add645-abcb-46e3-a103-e39305c0be1c-ssh-key" (OuterVolumeSpecName: "ssh-key") pod "19add645-abcb-46e3-a103-e39305c0be1c" (UID: "19add645-abcb-46e3-a103-e39305c0be1c"). InnerVolumeSpecName "ssh-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 10 23:20:50 crc kubenswrapper[4791]: I1210 23:20:50.252196 4791 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/19add645-abcb-46e3-a103-e39305c0be1c-inventory\") on node \"crc\" DevicePath \"\"" Dec 10 23:20:50 crc kubenswrapper[4791]: I1210 23:20:50.252252 4791 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-2pnxl\" (UniqueName: \"kubernetes.io/projected/19add645-abcb-46e3-a103-e39305c0be1c-kube-api-access-2pnxl\") on node \"crc\" DevicePath \"\"" Dec 10 23:20:50 crc kubenswrapper[4791]: I1210 23:20:50.252266 4791 reconciler_common.go:293] "Volume detached for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/19add645-abcb-46e3-a103-e39305c0be1c-ssh-key\") on node \"crc\" DevicePath \"\"" Dec 10 23:20:50 crc kubenswrapper[4791]: I1210 23:20:50.601405 4791 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/install-os-edpm-deployment-openstack-edpm-ipam-c6nvg" event={"ID":"19add645-abcb-46e3-a103-e39305c0be1c","Type":"ContainerDied","Data":"8cc932a4a47e0ce7b6e6252cc22f5aca75c98b6f19bf5c27ee0add3947d97389"} Dec 10 23:20:50 crc kubenswrapper[4791]: I1210 23:20:50.601469 4791 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="8cc932a4a47e0ce7b6e6252cc22f5aca75c98b6f19bf5c27ee0add3947d97389" Dec 10 23:20:50 crc kubenswrapper[4791]: I1210 23:20:50.601524 4791 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/install-os-edpm-deployment-openstack-edpm-ipam-c6nvg" Dec 10 23:20:50 crc kubenswrapper[4791]: I1210 23:20:50.740936 4791 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/configure-os-edpm-deployment-openstack-edpm-ipam-cb6z5"] Dec 10 23:20:50 crc kubenswrapper[4791]: E1210 23:20:50.742354 4791 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="19add645-abcb-46e3-a103-e39305c0be1c" containerName="install-os-edpm-deployment-openstack-edpm-ipam" Dec 10 23:20:50 crc kubenswrapper[4791]: I1210 23:20:50.742437 4791 state_mem.go:107] "Deleted CPUSet assignment" podUID="19add645-abcb-46e3-a103-e39305c0be1c" containerName="install-os-edpm-deployment-openstack-edpm-ipam" Dec 10 23:20:50 crc kubenswrapper[4791]: I1210 23:20:50.742789 4791 memory_manager.go:354] "RemoveStaleState removing state" podUID="19add645-abcb-46e3-a103-e39305c0be1c" containerName="install-os-edpm-deployment-openstack-edpm-ipam" Dec 10 23:20:50 crc kubenswrapper[4791]: I1210 23:20:50.743709 4791 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-cb6z5" Dec 10 23:20:50 crc kubenswrapper[4791]: I1210 23:20:50.747824 4791 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-ansible-ssh-private-key-secret" Dec 10 23:20:50 crc kubenswrapper[4791]: I1210 23:20:50.748141 4791 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Dec 10 23:20:50 crc kubenswrapper[4791]: I1210 23:20:50.748507 4791 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-edpm-ipam" Dec 10 23:20:50 crc kubenswrapper[4791]: I1210 23:20:50.748694 4791 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-dockercfg-hf8pr" Dec 10 23:20:50 crc kubenswrapper[4791]: I1210 23:20:50.764655 4791 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/configure-os-edpm-deployment-openstack-edpm-ipam-cb6z5"] Dec 10 23:20:50 crc kubenswrapper[4791]: I1210 23:20:50.765163 4791 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/cbb47bf3-4d3a-4e7e-bc54-c8a3a2ec98c6-ssh-key\") pod \"configure-os-edpm-deployment-openstack-edpm-ipam-cb6z5\" (UID: \"cbb47bf3-4d3a-4e7e-bc54-c8a3a2ec98c6\") " pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-cb6z5" Dec 10 23:20:50 crc kubenswrapper[4791]: I1210 23:20:50.765497 4791 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-7xnl8\" (UniqueName: \"kubernetes.io/projected/cbb47bf3-4d3a-4e7e-bc54-c8a3a2ec98c6-kube-api-access-7xnl8\") pod \"configure-os-edpm-deployment-openstack-edpm-ipam-cb6z5\" (UID: \"cbb47bf3-4d3a-4e7e-bc54-c8a3a2ec98c6\") " pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-cb6z5" Dec 10 23:20:50 crc kubenswrapper[4791]: I1210 23:20:50.765568 4791 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/cbb47bf3-4d3a-4e7e-bc54-c8a3a2ec98c6-inventory\") pod \"configure-os-edpm-deployment-openstack-edpm-ipam-cb6z5\" (UID: \"cbb47bf3-4d3a-4e7e-bc54-c8a3a2ec98c6\") " pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-cb6z5" Dec 10 23:20:50 crc kubenswrapper[4791]: I1210 23:20:50.867203 4791 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-7xnl8\" (UniqueName: \"kubernetes.io/projected/cbb47bf3-4d3a-4e7e-bc54-c8a3a2ec98c6-kube-api-access-7xnl8\") pod \"configure-os-edpm-deployment-openstack-edpm-ipam-cb6z5\" (UID: \"cbb47bf3-4d3a-4e7e-bc54-c8a3a2ec98c6\") " pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-cb6z5" Dec 10 23:20:50 crc kubenswrapper[4791]: I1210 23:20:50.867249 4791 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/cbb47bf3-4d3a-4e7e-bc54-c8a3a2ec98c6-inventory\") pod \"configure-os-edpm-deployment-openstack-edpm-ipam-cb6z5\" (UID: \"cbb47bf3-4d3a-4e7e-bc54-c8a3a2ec98c6\") " pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-cb6z5" Dec 10 23:20:50 crc kubenswrapper[4791]: I1210 23:20:50.867285 4791 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/cbb47bf3-4d3a-4e7e-bc54-c8a3a2ec98c6-ssh-key\") pod \"configure-os-edpm-deployment-openstack-edpm-ipam-cb6z5\" (UID: \"cbb47bf3-4d3a-4e7e-bc54-c8a3a2ec98c6\") " pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-cb6z5" Dec 10 23:20:50 crc kubenswrapper[4791]: I1210 23:20:50.874921 4791 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/cbb47bf3-4d3a-4e7e-bc54-c8a3a2ec98c6-ssh-key\") pod \"configure-os-edpm-deployment-openstack-edpm-ipam-cb6z5\" (UID: \"cbb47bf3-4d3a-4e7e-bc54-c8a3a2ec98c6\") " pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-cb6z5" Dec 10 23:20:50 crc kubenswrapper[4791]: I1210 23:20:50.874928 4791 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/cbb47bf3-4d3a-4e7e-bc54-c8a3a2ec98c6-inventory\") pod \"configure-os-edpm-deployment-openstack-edpm-ipam-cb6z5\" (UID: \"cbb47bf3-4d3a-4e7e-bc54-c8a3a2ec98c6\") " pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-cb6z5" Dec 10 23:20:50 crc kubenswrapper[4791]: I1210 23:20:50.883300 4791 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-7xnl8\" (UniqueName: \"kubernetes.io/projected/cbb47bf3-4d3a-4e7e-bc54-c8a3a2ec98c6-kube-api-access-7xnl8\") pod \"configure-os-edpm-deployment-openstack-edpm-ipam-cb6z5\" (UID: \"cbb47bf3-4d3a-4e7e-bc54-c8a3a2ec98c6\") " pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-cb6z5" Dec 10 23:20:51 crc kubenswrapper[4791]: I1210 23:20:51.072691 4791 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-cb6z5" Dec 10 23:20:51 crc kubenswrapper[4791]: I1210 23:20:51.398750 4791 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/configure-os-edpm-deployment-openstack-edpm-ipam-cb6z5"] Dec 10 23:20:51 crc kubenswrapper[4791]: I1210 23:20:51.616431 4791 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-cb6z5" event={"ID":"cbb47bf3-4d3a-4e7e-bc54-c8a3a2ec98c6","Type":"ContainerStarted","Data":"1738cbe30d55ec88eb25914b1c16b2a95cc89097414102e72208f9cda9793a2e"} Dec 10 23:20:52 crc kubenswrapper[4791]: I1210 23:20:52.625459 4791 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-cb6z5" event={"ID":"cbb47bf3-4d3a-4e7e-bc54-c8a3a2ec98c6","Type":"ContainerStarted","Data":"276faabe829972f30256afb2f4d609f11dd8276cf4d8b3719ee79a01c4448576"} Dec 10 23:20:52 crc kubenswrapper[4791]: I1210 23:20:52.643689 4791 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-cb6z5" podStartSLOduration=2.020790019 podStartE2EDuration="2.643672609s" podCreationTimestamp="2025-12-10 23:20:50 +0000 UTC" firstStartedPulling="2025-12-10 23:20:51.405116196 +0000 UTC m=+1885.834733809" lastFinishedPulling="2025-12-10 23:20:52.027998736 +0000 UTC m=+1886.457616399" observedRunningTime="2025-12-10 23:20:52.642198167 +0000 UTC m=+1887.071815780" watchObservedRunningTime="2025-12-10 23:20:52.643672609 +0000 UTC m=+1887.073290222" Dec 10 23:20:53 crc kubenswrapper[4791]: I1210 23:20:53.561049 4791 scope.go:117] "RemoveContainer" containerID="51879f033bb3a1c8fd84acf13c94ff25334fae72bdf0c18287e1e1c3aeda56a7" Dec 10 23:20:53 crc kubenswrapper[4791]: I1210 23:20:53.622392 4791 scope.go:117] "RemoveContainer" containerID="ccf60cdfa54857941fadb4022be27fb785d133c0ce8f11676f66323e635c7b71" Dec 10 23:20:53 crc kubenswrapper[4791]: I1210 23:20:53.670050 4791 scope.go:117] "RemoveContainer" containerID="b94a8cbc00ff61fc1de29573492cf4b269b745351fa58d04b9a77143fae5b87d" Dec 10 23:21:32 crc kubenswrapper[4791]: I1210 23:21:32.059824 4791 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-cell1-cell-mapping-zcfpp"] Dec 10 23:21:32 crc kubenswrapper[4791]: I1210 23:21:32.072243 4791 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-cell1-cell-mapping-zcfpp"] Dec 10 23:21:33 crc kubenswrapper[4791]: I1210 23:21:33.906010 4791 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="430975e4-f09a-4c3f-9c1d-0cab25b587b1" path="/var/lib/kubelet/pods/430975e4-f09a-4c3f-9c1d-0cab25b587b1/volumes" Dec 10 23:21:40 crc kubenswrapper[4791]: I1210 23:21:40.131965 4791 generic.go:334] "Generic (PLEG): container finished" podID="cbb47bf3-4d3a-4e7e-bc54-c8a3a2ec98c6" containerID="276faabe829972f30256afb2f4d609f11dd8276cf4d8b3719ee79a01c4448576" exitCode=0 Dec 10 23:21:40 crc kubenswrapper[4791]: I1210 23:21:40.132058 4791 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-cb6z5" event={"ID":"cbb47bf3-4d3a-4e7e-bc54-c8a3a2ec98c6","Type":"ContainerDied","Data":"276faabe829972f30256afb2f4d609f11dd8276cf4d8b3719ee79a01c4448576"} Dec 10 23:21:41 crc kubenswrapper[4791]: I1210 23:21:41.608163 4791 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-cb6z5" Dec 10 23:21:41 crc kubenswrapper[4791]: I1210 23:21:41.645731 4791 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-7xnl8\" (UniqueName: \"kubernetes.io/projected/cbb47bf3-4d3a-4e7e-bc54-c8a3a2ec98c6-kube-api-access-7xnl8\") pod \"cbb47bf3-4d3a-4e7e-bc54-c8a3a2ec98c6\" (UID: \"cbb47bf3-4d3a-4e7e-bc54-c8a3a2ec98c6\") " Dec 10 23:21:41 crc kubenswrapper[4791]: I1210 23:21:41.646264 4791 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/cbb47bf3-4d3a-4e7e-bc54-c8a3a2ec98c6-ssh-key\") pod \"cbb47bf3-4d3a-4e7e-bc54-c8a3a2ec98c6\" (UID: \"cbb47bf3-4d3a-4e7e-bc54-c8a3a2ec98c6\") " Dec 10 23:21:41 crc kubenswrapper[4791]: I1210 23:21:41.646441 4791 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/cbb47bf3-4d3a-4e7e-bc54-c8a3a2ec98c6-inventory\") pod \"cbb47bf3-4d3a-4e7e-bc54-c8a3a2ec98c6\" (UID: \"cbb47bf3-4d3a-4e7e-bc54-c8a3a2ec98c6\") " Dec 10 23:21:41 crc kubenswrapper[4791]: I1210 23:21:41.652780 4791 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/cbb47bf3-4d3a-4e7e-bc54-c8a3a2ec98c6-kube-api-access-7xnl8" (OuterVolumeSpecName: "kube-api-access-7xnl8") pod "cbb47bf3-4d3a-4e7e-bc54-c8a3a2ec98c6" (UID: "cbb47bf3-4d3a-4e7e-bc54-c8a3a2ec98c6"). InnerVolumeSpecName "kube-api-access-7xnl8". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 10 23:21:41 crc kubenswrapper[4791]: I1210 23:21:41.684208 4791 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/cbb47bf3-4d3a-4e7e-bc54-c8a3a2ec98c6-ssh-key" (OuterVolumeSpecName: "ssh-key") pod "cbb47bf3-4d3a-4e7e-bc54-c8a3a2ec98c6" (UID: "cbb47bf3-4d3a-4e7e-bc54-c8a3a2ec98c6"). InnerVolumeSpecName "ssh-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 10 23:21:41 crc kubenswrapper[4791]: I1210 23:21:41.686502 4791 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/cbb47bf3-4d3a-4e7e-bc54-c8a3a2ec98c6-inventory" (OuterVolumeSpecName: "inventory") pod "cbb47bf3-4d3a-4e7e-bc54-c8a3a2ec98c6" (UID: "cbb47bf3-4d3a-4e7e-bc54-c8a3a2ec98c6"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 10 23:21:41 crc kubenswrapper[4791]: I1210 23:21:41.748588 4791 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-7xnl8\" (UniqueName: \"kubernetes.io/projected/cbb47bf3-4d3a-4e7e-bc54-c8a3a2ec98c6-kube-api-access-7xnl8\") on node \"crc\" DevicePath \"\"" Dec 10 23:21:41 crc kubenswrapper[4791]: I1210 23:21:41.748646 4791 reconciler_common.go:293] "Volume detached for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/cbb47bf3-4d3a-4e7e-bc54-c8a3a2ec98c6-ssh-key\") on node \"crc\" DevicePath \"\"" Dec 10 23:21:41 crc kubenswrapper[4791]: I1210 23:21:41.748669 4791 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/cbb47bf3-4d3a-4e7e-bc54-c8a3a2ec98c6-inventory\") on node \"crc\" DevicePath \"\"" Dec 10 23:21:42 crc kubenswrapper[4791]: I1210 23:21:42.155812 4791 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-cb6z5" event={"ID":"cbb47bf3-4d3a-4e7e-bc54-c8a3a2ec98c6","Type":"ContainerDied","Data":"1738cbe30d55ec88eb25914b1c16b2a95cc89097414102e72208f9cda9793a2e"} Dec 10 23:21:42 crc kubenswrapper[4791]: I1210 23:21:42.156119 4791 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="1738cbe30d55ec88eb25914b1c16b2a95cc89097414102e72208f9cda9793a2e" Dec 10 23:21:42 crc kubenswrapper[4791]: I1210 23:21:42.155871 4791 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-cb6z5" Dec 10 23:21:42 crc kubenswrapper[4791]: I1210 23:21:42.280707 4791 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ssh-known-hosts-edpm-deployment-8zwc6"] Dec 10 23:21:42 crc kubenswrapper[4791]: E1210 23:21:42.281244 4791 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="cbb47bf3-4d3a-4e7e-bc54-c8a3a2ec98c6" containerName="configure-os-edpm-deployment-openstack-edpm-ipam" Dec 10 23:21:42 crc kubenswrapper[4791]: I1210 23:21:42.281271 4791 state_mem.go:107] "Deleted CPUSet assignment" podUID="cbb47bf3-4d3a-4e7e-bc54-c8a3a2ec98c6" containerName="configure-os-edpm-deployment-openstack-edpm-ipam" Dec 10 23:21:42 crc kubenswrapper[4791]: I1210 23:21:42.281593 4791 memory_manager.go:354] "RemoveStaleState removing state" podUID="cbb47bf3-4d3a-4e7e-bc54-c8a3a2ec98c6" containerName="configure-os-edpm-deployment-openstack-edpm-ipam" Dec 10 23:21:42 crc kubenswrapper[4791]: I1210 23:21:42.282393 4791 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ssh-known-hosts-edpm-deployment-8zwc6" Dec 10 23:21:42 crc kubenswrapper[4791]: I1210 23:21:42.284144 4791 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-edpm-ipam" Dec 10 23:21:42 crc kubenswrapper[4791]: I1210 23:21:42.285373 4791 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-dockercfg-hf8pr" Dec 10 23:21:42 crc kubenswrapper[4791]: I1210 23:21:42.285746 4791 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-ansible-ssh-private-key-secret" Dec 10 23:21:42 crc kubenswrapper[4791]: I1210 23:21:42.285789 4791 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Dec 10 23:21:42 crc kubenswrapper[4791]: I1210 23:21:42.308391 4791 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ssh-known-hosts-edpm-deployment-8zwc6"] Dec 10 23:21:42 crc kubenswrapper[4791]: I1210 23:21:42.359721 4791 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory-0\" (UniqueName: \"kubernetes.io/secret/12dfad91-cb70-4237-be65-427b3e919a05-inventory-0\") pod \"ssh-known-hosts-edpm-deployment-8zwc6\" (UID: \"12dfad91-cb70-4237-be65-427b3e919a05\") " pod="openstack/ssh-known-hosts-edpm-deployment-8zwc6" Dec 10 23:21:42 crc kubenswrapper[4791]: I1210 23:21:42.359883 4791 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-5s8wt\" (UniqueName: \"kubernetes.io/projected/12dfad91-cb70-4237-be65-427b3e919a05-kube-api-access-5s8wt\") pod \"ssh-known-hosts-edpm-deployment-8zwc6\" (UID: \"12dfad91-cb70-4237-be65-427b3e919a05\") " pod="openstack/ssh-known-hosts-edpm-deployment-8zwc6" Dec 10 23:21:42 crc kubenswrapper[4791]: I1210 23:21:42.360038 4791 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/12dfad91-cb70-4237-be65-427b3e919a05-ssh-key-openstack-edpm-ipam\") pod \"ssh-known-hosts-edpm-deployment-8zwc6\" (UID: \"12dfad91-cb70-4237-be65-427b3e919a05\") " pod="openstack/ssh-known-hosts-edpm-deployment-8zwc6" Dec 10 23:21:42 crc kubenswrapper[4791]: I1210 23:21:42.461291 4791 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-5s8wt\" (UniqueName: \"kubernetes.io/projected/12dfad91-cb70-4237-be65-427b3e919a05-kube-api-access-5s8wt\") pod \"ssh-known-hosts-edpm-deployment-8zwc6\" (UID: \"12dfad91-cb70-4237-be65-427b3e919a05\") " pod="openstack/ssh-known-hosts-edpm-deployment-8zwc6" Dec 10 23:21:42 crc kubenswrapper[4791]: I1210 23:21:42.461588 4791 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/12dfad91-cb70-4237-be65-427b3e919a05-ssh-key-openstack-edpm-ipam\") pod \"ssh-known-hosts-edpm-deployment-8zwc6\" (UID: \"12dfad91-cb70-4237-be65-427b3e919a05\") " pod="openstack/ssh-known-hosts-edpm-deployment-8zwc6" Dec 10 23:21:42 crc kubenswrapper[4791]: I1210 23:21:42.461714 4791 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory-0\" (UniqueName: \"kubernetes.io/secret/12dfad91-cb70-4237-be65-427b3e919a05-inventory-0\") pod \"ssh-known-hosts-edpm-deployment-8zwc6\" (UID: \"12dfad91-cb70-4237-be65-427b3e919a05\") " pod="openstack/ssh-known-hosts-edpm-deployment-8zwc6" Dec 10 23:21:42 crc kubenswrapper[4791]: I1210 23:21:42.466255 4791 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory-0\" (UniqueName: \"kubernetes.io/secret/12dfad91-cb70-4237-be65-427b3e919a05-inventory-0\") pod \"ssh-known-hosts-edpm-deployment-8zwc6\" (UID: \"12dfad91-cb70-4237-be65-427b3e919a05\") " pod="openstack/ssh-known-hosts-edpm-deployment-8zwc6" Dec 10 23:21:42 crc kubenswrapper[4791]: I1210 23:21:42.467749 4791 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/12dfad91-cb70-4237-be65-427b3e919a05-ssh-key-openstack-edpm-ipam\") pod \"ssh-known-hosts-edpm-deployment-8zwc6\" (UID: \"12dfad91-cb70-4237-be65-427b3e919a05\") " pod="openstack/ssh-known-hosts-edpm-deployment-8zwc6" Dec 10 23:21:42 crc kubenswrapper[4791]: I1210 23:21:42.482167 4791 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-5s8wt\" (UniqueName: \"kubernetes.io/projected/12dfad91-cb70-4237-be65-427b3e919a05-kube-api-access-5s8wt\") pod \"ssh-known-hosts-edpm-deployment-8zwc6\" (UID: \"12dfad91-cb70-4237-be65-427b3e919a05\") " pod="openstack/ssh-known-hosts-edpm-deployment-8zwc6" Dec 10 23:21:42 crc kubenswrapper[4791]: I1210 23:21:42.605070 4791 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ssh-known-hosts-edpm-deployment-8zwc6" Dec 10 23:21:43 crc kubenswrapper[4791]: I1210 23:21:43.226619 4791 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ssh-known-hosts-edpm-deployment-8zwc6"] Dec 10 23:21:44 crc kubenswrapper[4791]: I1210 23:21:44.190447 4791 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ssh-known-hosts-edpm-deployment-8zwc6" event={"ID":"12dfad91-cb70-4237-be65-427b3e919a05","Type":"ContainerStarted","Data":"ce8129fa086da0b69394ea283996ae21bda82c0f04f86c033504a98af2e82d2a"} Dec 10 23:21:46 crc kubenswrapper[4791]: I1210 23:21:46.220299 4791 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ssh-known-hosts-edpm-deployment-8zwc6" event={"ID":"12dfad91-cb70-4237-be65-427b3e919a05","Type":"ContainerStarted","Data":"c3f95263458970b8c0473a5b31776fe3f3c3682ccbec2f373d65fa5c5bd842dd"} Dec 10 23:21:46 crc kubenswrapper[4791]: I1210 23:21:46.241290 4791 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ssh-known-hosts-edpm-deployment-8zwc6" podStartSLOduration=2.000927963 podStartE2EDuration="4.241271247s" podCreationTimestamp="2025-12-10 23:21:42 +0000 UTC" firstStartedPulling="2025-12-10 23:21:43.231043759 +0000 UTC m=+1937.660661372" lastFinishedPulling="2025-12-10 23:21:45.471387033 +0000 UTC m=+1939.901004656" observedRunningTime="2025-12-10 23:21:46.235522161 +0000 UTC m=+1940.665139824" watchObservedRunningTime="2025-12-10 23:21:46.241271247 +0000 UTC m=+1940.670888860" Dec 10 23:21:53 crc kubenswrapper[4791]: I1210 23:21:53.357065 4791 generic.go:334] "Generic (PLEG): container finished" podID="12dfad91-cb70-4237-be65-427b3e919a05" containerID="c3f95263458970b8c0473a5b31776fe3f3c3682ccbec2f373d65fa5c5bd842dd" exitCode=0 Dec 10 23:21:53 crc kubenswrapper[4791]: I1210 23:21:53.357176 4791 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ssh-known-hosts-edpm-deployment-8zwc6" event={"ID":"12dfad91-cb70-4237-be65-427b3e919a05","Type":"ContainerDied","Data":"c3f95263458970b8c0473a5b31776fe3f3c3682ccbec2f373d65fa5c5bd842dd"} Dec 10 23:21:53 crc kubenswrapper[4791]: I1210 23:21:53.801150 4791 scope.go:117] "RemoveContainer" containerID="8ad39df5144e867e9c45cb797ca31cbba3a462a23ea8808478880b36524e0c12" Dec 10 23:21:54 crc kubenswrapper[4791]: I1210 23:21:54.885010 4791 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ssh-known-hosts-edpm-deployment-8zwc6" Dec 10 23:21:55 crc kubenswrapper[4791]: I1210 23:21:55.066728 4791 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-5s8wt\" (UniqueName: \"kubernetes.io/projected/12dfad91-cb70-4237-be65-427b3e919a05-kube-api-access-5s8wt\") pod \"12dfad91-cb70-4237-be65-427b3e919a05\" (UID: \"12dfad91-cb70-4237-be65-427b3e919a05\") " Dec 10 23:21:55 crc kubenswrapper[4791]: I1210 23:21:55.067174 4791 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/12dfad91-cb70-4237-be65-427b3e919a05-ssh-key-openstack-edpm-ipam\") pod \"12dfad91-cb70-4237-be65-427b3e919a05\" (UID: \"12dfad91-cb70-4237-be65-427b3e919a05\") " Dec 10 23:21:55 crc kubenswrapper[4791]: I1210 23:21:55.067265 4791 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory-0\" (UniqueName: \"kubernetes.io/secret/12dfad91-cb70-4237-be65-427b3e919a05-inventory-0\") pod \"12dfad91-cb70-4237-be65-427b3e919a05\" (UID: \"12dfad91-cb70-4237-be65-427b3e919a05\") " Dec 10 23:21:55 crc kubenswrapper[4791]: I1210 23:21:55.073313 4791 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/12dfad91-cb70-4237-be65-427b3e919a05-kube-api-access-5s8wt" (OuterVolumeSpecName: "kube-api-access-5s8wt") pod "12dfad91-cb70-4237-be65-427b3e919a05" (UID: "12dfad91-cb70-4237-be65-427b3e919a05"). InnerVolumeSpecName "kube-api-access-5s8wt". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 10 23:21:55 crc kubenswrapper[4791]: I1210 23:21:55.095557 4791 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/12dfad91-cb70-4237-be65-427b3e919a05-ssh-key-openstack-edpm-ipam" (OuterVolumeSpecName: "ssh-key-openstack-edpm-ipam") pod "12dfad91-cb70-4237-be65-427b3e919a05" (UID: "12dfad91-cb70-4237-be65-427b3e919a05"). InnerVolumeSpecName "ssh-key-openstack-edpm-ipam". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 10 23:21:55 crc kubenswrapper[4791]: I1210 23:21:55.118086 4791 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/12dfad91-cb70-4237-be65-427b3e919a05-inventory-0" (OuterVolumeSpecName: "inventory-0") pod "12dfad91-cb70-4237-be65-427b3e919a05" (UID: "12dfad91-cb70-4237-be65-427b3e919a05"). InnerVolumeSpecName "inventory-0". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 10 23:21:55 crc kubenswrapper[4791]: I1210 23:21:55.169891 4791 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-5s8wt\" (UniqueName: \"kubernetes.io/projected/12dfad91-cb70-4237-be65-427b3e919a05-kube-api-access-5s8wt\") on node \"crc\" DevicePath \"\"" Dec 10 23:21:55 crc kubenswrapper[4791]: I1210 23:21:55.169930 4791 reconciler_common.go:293] "Volume detached for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/12dfad91-cb70-4237-be65-427b3e919a05-ssh-key-openstack-edpm-ipam\") on node \"crc\" DevicePath \"\"" Dec 10 23:21:55 crc kubenswrapper[4791]: I1210 23:21:55.169948 4791 reconciler_common.go:293] "Volume detached for volume \"inventory-0\" (UniqueName: \"kubernetes.io/secret/12dfad91-cb70-4237-be65-427b3e919a05-inventory-0\") on node \"crc\" DevicePath \"\"" Dec 10 23:21:55 crc kubenswrapper[4791]: I1210 23:21:55.384946 4791 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ssh-known-hosts-edpm-deployment-8zwc6" event={"ID":"12dfad91-cb70-4237-be65-427b3e919a05","Type":"ContainerDied","Data":"ce8129fa086da0b69394ea283996ae21bda82c0f04f86c033504a98af2e82d2a"} Dec 10 23:21:55 crc kubenswrapper[4791]: I1210 23:21:55.385007 4791 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="ce8129fa086da0b69394ea283996ae21bda82c0f04f86c033504a98af2e82d2a" Dec 10 23:21:55 crc kubenswrapper[4791]: I1210 23:21:55.385481 4791 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ssh-known-hosts-edpm-deployment-8zwc6" Dec 10 23:21:55 crc kubenswrapper[4791]: I1210 23:21:55.490421 4791 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/run-os-edpm-deployment-openstack-edpm-ipam-5w7qc"] Dec 10 23:21:55 crc kubenswrapper[4791]: E1210 23:21:55.490950 4791 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="12dfad91-cb70-4237-be65-427b3e919a05" containerName="ssh-known-hosts-edpm-deployment" Dec 10 23:21:55 crc kubenswrapper[4791]: I1210 23:21:55.490969 4791 state_mem.go:107] "Deleted CPUSet assignment" podUID="12dfad91-cb70-4237-be65-427b3e919a05" containerName="ssh-known-hosts-edpm-deployment" Dec 10 23:21:55 crc kubenswrapper[4791]: I1210 23:21:55.491186 4791 memory_manager.go:354] "RemoveStaleState removing state" podUID="12dfad91-cb70-4237-be65-427b3e919a05" containerName="ssh-known-hosts-edpm-deployment" Dec 10 23:21:55 crc kubenswrapper[4791]: I1210 23:21:55.492027 4791 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/run-os-edpm-deployment-openstack-edpm-ipam-5w7qc" Dec 10 23:21:55 crc kubenswrapper[4791]: I1210 23:21:55.497095 4791 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-dockercfg-hf8pr" Dec 10 23:21:55 crc kubenswrapper[4791]: I1210 23:21:55.497158 4791 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-edpm-ipam" Dec 10 23:21:55 crc kubenswrapper[4791]: I1210 23:21:55.497116 4791 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-ansible-ssh-private-key-secret" Dec 10 23:21:55 crc kubenswrapper[4791]: I1210 23:21:55.497211 4791 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Dec 10 23:21:55 crc kubenswrapper[4791]: I1210 23:21:55.503354 4791 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/run-os-edpm-deployment-openstack-edpm-ipam-5w7qc"] Dec 10 23:21:55 crc kubenswrapper[4791]: I1210 23:21:55.581609 4791 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/62988878-9637-47ad-9877-fd81f32199d2-inventory\") pod \"run-os-edpm-deployment-openstack-edpm-ipam-5w7qc\" (UID: \"62988878-9637-47ad-9877-fd81f32199d2\") " pod="openstack/run-os-edpm-deployment-openstack-edpm-ipam-5w7qc" Dec 10 23:21:55 crc kubenswrapper[4791]: I1210 23:21:55.581747 4791 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/62988878-9637-47ad-9877-fd81f32199d2-ssh-key\") pod \"run-os-edpm-deployment-openstack-edpm-ipam-5w7qc\" (UID: \"62988878-9637-47ad-9877-fd81f32199d2\") " pod="openstack/run-os-edpm-deployment-openstack-edpm-ipam-5w7qc" Dec 10 23:21:55 crc kubenswrapper[4791]: I1210 23:21:55.581785 4791 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-tnhhz\" (UniqueName: \"kubernetes.io/projected/62988878-9637-47ad-9877-fd81f32199d2-kube-api-access-tnhhz\") pod \"run-os-edpm-deployment-openstack-edpm-ipam-5w7qc\" (UID: \"62988878-9637-47ad-9877-fd81f32199d2\") " pod="openstack/run-os-edpm-deployment-openstack-edpm-ipam-5w7qc" Dec 10 23:21:55 crc kubenswrapper[4791]: I1210 23:21:55.712882 4791 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/62988878-9637-47ad-9877-fd81f32199d2-ssh-key\") pod \"run-os-edpm-deployment-openstack-edpm-ipam-5w7qc\" (UID: \"62988878-9637-47ad-9877-fd81f32199d2\") " pod="openstack/run-os-edpm-deployment-openstack-edpm-ipam-5w7qc" Dec 10 23:21:55 crc kubenswrapper[4791]: I1210 23:21:55.712954 4791 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-tnhhz\" (UniqueName: \"kubernetes.io/projected/62988878-9637-47ad-9877-fd81f32199d2-kube-api-access-tnhhz\") pod \"run-os-edpm-deployment-openstack-edpm-ipam-5w7qc\" (UID: \"62988878-9637-47ad-9877-fd81f32199d2\") " pod="openstack/run-os-edpm-deployment-openstack-edpm-ipam-5w7qc" Dec 10 23:21:55 crc kubenswrapper[4791]: I1210 23:21:55.713020 4791 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/62988878-9637-47ad-9877-fd81f32199d2-inventory\") pod \"run-os-edpm-deployment-openstack-edpm-ipam-5w7qc\" (UID: \"62988878-9637-47ad-9877-fd81f32199d2\") " pod="openstack/run-os-edpm-deployment-openstack-edpm-ipam-5w7qc" Dec 10 23:21:55 crc kubenswrapper[4791]: I1210 23:21:55.718898 4791 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/62988878-9637-47ad-9877-fd81f32199d2-inventory\") pod \"run-os-edpm-deployment-openstack-edpm-ipam-5w7qc\" (UID: \"62988878-9637-47ad-9877-fd81f32199d2\") " pod="openstack/run-os-edpm-deployment-openstack-edpm-ipam-5w7qc" Dec 10 23:21:55 crc kubenswrapper[4791]: I1210 23:21:55.730953 4791 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/62988878-9637-47ad-9877-fd81f32199d2-ssh-key\") pod \"run-os-edpm-deployment-openstack-edpm-ipam-5w7qc\" (UID: \"62988878-9637-47ad-9877-fd81f32199d2\") " pod="openstack/run-os-edpm-deployment-openstack-edpm-ipam-5w7qc" Dec 10 23:21:55 crc kubenswrapper[4791]: I1210 23:21:55.735185 4791 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-tnhhz\" (UniqueName: \"kubernetes.io/projected/62988878-9637-47ad-9877-fd81f32199d2-kube-api-access-tnhhz\") pod \"run-os-edpm-deployment-openstack-edpm-ipam-5w7qc\" (UID: \"62988878-9637-47ad-9877-fd81f32199d2\") " pod="openstack/run-os-edpm-deployment-openstack-edpm-ipam-5w7qc" Dec 10 23:21:55 crc kubenswrapper[4791]: I1210 23:21:55.832319 4791 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/run-os-edpm-deployment-openstack-edpm-ipam-5w7qc" Dec 10 23:21:56 crc kubenswrapper[4791]: I1210 23:21:56.393209 4791 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/run-os-edpm-deployment-openstack-edpm-ipam-5w7qc"] Dec 10 23:21:57 crc kubenswrapper[4791]: I1210 23:21:57.409676 4791 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/run-os-edpm-deployment-openstack-edpm-ipam-5w7qc" event={"ID":"62988878-9637-47ad-9877-fd81f32199d2","Type":"ContainerStarted","Data":"9d70c3e2a9ab803f11542bc3b33eb2af421d23b6f2b3689037efeac60f0b28a6"} Dec 10 23:21:57 crc kubenswrapper[4791]: I1210 23:21:57.410090 4791 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/run-os-edpm-deployment-openstack-edpm-ipam-5w7qc" event={"ID":"62988878-9637-47ad-9877-fd81f32199d2","Type":"ContainerStarted","Data":"3a730d4eeef6ececca1dd817b7d0b1493953a24a06249ddd56fb040bd429d50b"} Dec 10 23:21:57 crc kubenswrapper[4791]: I1210 23:21:57.445889 4791 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/run-os-edpm-deployment-openstack-edpm-ipam-5w7qc" podStartSLOduration=1.954179763 podStartE2EDuration="2.445861357s" podCreationTimestamp="2025-12-10 23:21:55 +0000 UTC" firstStartedPulling="2025-12-10 23:21:56.397599026 +0000 UTC m=+1950.827216649" lastFinishedPulling="2025-12-10 23:21:56.88928059 +0000 UTC m=+1951.318898243" observedRunningTime="2025-12-10 23:21:57.431547754 +0000 UTC m=+1951.861165427" watchObservedRunningTime="2025-12-10 23:21:57.445861357 +0000 UTC m=+1951.875479010" Dec 10 23:22:05 crc kubenswrapper[4791]: I1210 23:22:05.578429 4791 generic.go:334] "Generic (PLEG): container finished" podID="62988878-9637-47ad-9877-fd81f32199d2" containerID="9d70c3e2a9ab803f11542bc3b33eb2af421d23b6f2b3689037efeac60f0b28a6" exitCode=0 Dec 10 23:22:05 crc kubenswrapper[4791]: I1210 23:22:05.578498 4791 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/run-os-edpm-deployment-openstack-edpm-ipam-5w7qc" event={"ID":"62988878-9637-47ad-9877-fd81f32199d2","Type":"ContainerDied","Data":"9d70c3e2a9ab803f11542bc3b33eb2af421d23b6f2b3689037efeac60f0b28a6"} Dec 10 23:22:07 crc kubenswrapper[4791]: I1210 23:22:07.077209 4791 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/run-os-edpm-deployment-openstack-edpm-ipam-5w7qc" Dec 10 23:22:07 crc kubenswrapper[4791]: I1210 23:22:07.200864 4791 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/62988878-9637-47ad-9877-fd81f32199d2-inventory\") pod \"62988878-9637-47ad-9877-fd81f32199d2\" (UID: \"62988878-9637-47ad-9877-fd81f32199d2\") " Dec 10 23:22:07 crc kubenswrapper[4791]: I1210 23:22:07.200995 4791 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-tnhhz\" (UniqueName: \"kubernetes.io/projected/62988878-9637-47ad-9877-fd81f32199d2-kube-api-access-tnhhz\") pod \"62988878-9637-47ad-9877-fd81f32199d2\" (UID: \"62988878-9637-47ad-9877-fd81f32199d2\") " Dec 10 23:22:07 crc kubenswrapper[4791]: I1210 23:22:07.201082 4791 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/62988878-9637-47ad-9877-fd81f32199d2-ssh-key\") pod \"62988878-9637-47ad-9877-fd81f32199d2\" (UID: \"62988878-9637-47ad-9877-fd81f32199d2\") " Dec 10 23:22:07 crc kubenswrapper[4791]: I1210 23:22:07.207576 4791 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/62988878-9637-47ad-9877-fd81f32199d2-kube-api-access-tnhhz" (OuterVolumeSpecName: "kube-api-access-tnhhz") pod "62988878-9637-47ad-9877-fd81f32199d2" (UID: "62988878-9637-47ad-9877-fd81f32199d2"). InnerVolumeSpecName "kube-api-access-tnhhz". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 10 23:22:07 crc kubenswrapper[4791]: I1210 23:22:07.250228 4791 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/62988878-9637-47ad-9877-fd81f32199d2-ssh-key" (OuterVolumeSpecName: "ssh-key") pod "62988878-9637-47ad-9877-fd81f32199d2" (UID: "62988878-9637-47ad-9877-fd81f32199d2"). InnerVolumeSpecName "ssh-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 10 23:22:07 crc kubenswrapper[4791]: I1210 23:22:07.255719 4791 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/62988878-9637-47ad-9877-fd81f32199d2-inventory" (OuterVolumeSpecName: "inventory") pod "62988878-9637-47ad-9877-fd81f32199d2" (UID: "62988878-9637-47ad-9877-fd81f32199d2"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 10 23:22:07 crc kubenswrapper[4791]: I1210 23:22:07.304510 4791 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/62988878-9637-47ad-9877-fd81f32199d2-inventory\") on node \"crc\" DevicePath \"\"" Dec 10 23:22:07 crc kubenswrapper[4791]: I1210 23:22:07.304559 4791 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-tnhhz\" (UniqueName: \"kubernetes.io/projected/62988878-9637-47ad-9877-fd81f32199d2-kube-api-access-tnhhz\") on node \"crc\" DevicePath \"\"" Dec 10 23:22:07 crc kubenswrapper[4791]: I1210 23:22:07.304581 4791 reconciler_common.go:293] "Volume detached for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/62988878-9637-47ad-9877-fd81f32199d2-ssh-key\") on node \"crc\" DevicePath \"\"" Dec 10 23:22:07 crc kubenswrapper[4791]: I1210 23:22:07.713898 4791 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/run-os-edpm-deployment-openstack-edpm-ipam-5w7qc" event={"ID":"62988878-9637-47ad-9877-fd81f32199d2","Type":"ContainerDied","Data":"3a730d4eeef6ececca1dd817b7d0b1493953a24a06249ddd56fb040bd429d50b"} Dec 10 23:22:07 crc kubenswrapper[4791]: I1210 23:22:07.714572 4791 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="3a730d4eeef6ececca1dd817b7d0b1493953a24a06249ddd56fb040bd429d50b" Dec 10 23:22:07 crc kubenswrapper[4791]: I1210 23:22:07.713974 4791 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/run-os-edpm-deployment-openstack-edpm-ipam-5w7qc" Dec 10 23:22:07 crc kubenswrapper[4791]: I1210 23:22:07.778444 4791 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-fbsgr"] Dec 10 23:22:07 crc kubenswrapper[4791]: E1210 23:22:07.778889 4791 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="62988878-9637-47ad-9877-fd81f32199d2" containerName="run-os-edpm-deployment-openstack-edpm-ipam" Dec 10 23:22:07 crc kubenswrapper[4791]: I1210 23:22:07.778907 4791 state_mem.go:107] "Deleted CPUSet assignment" podUID="62988878-9637-47ad-9877-fd81f32199d2" containerName="run-os-edpm-deployment-openstack-edpm-ipam" Dec 10 23:22:07 crc kubenswrapper[4791]: I1210 23:22:07.779139 4791 memory_manager.go:354] "RemoveStaleState removing state" podUID="62988878-9637-47ad-9877-fd81f32199d2" containerName="run-os-edpm-deployment-openstack-edpm-ipam" Dec 10 23:22:07 crc kubenswrapper[4791]: I1210 23:22:07.779943 4791 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-fbsgr" Dec 10 23:22:07 crc kubenswrapper[4791]: I1210 23:22:07.783301 4791 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-ansible-ssh-private-key-secret" Dec 10 23:22:07 crc kubenswrapper[4791]: I1210 23:22:07.783683 4791 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Dec 10 23:22:07 crc kubenswrapper[4791]: I1210 23:22:07.783751 4791 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-edpm-ipam" Dec 10 23:22:07 crc kubenswrapper[4791]: I1210 23:22:07.783882 4791 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-dockercfg-hf8pr" Dec 10 23:22:07 crc kubenswrapper[4791]: I1210 23:22:07.787805 4791 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-fbsgr"] Dec 10 23:22:07 crc kubenswrapper[4791]: I1210 23:22:07.815741 4791 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/72f75339-ac62-4e10-a6a2-634ac54461cd-inventory\") pod \"reboot-os-edpm-deployment-openstack-edpm-ipam-fbsgr\" (UID: \"72f75339-ac62-4e10-a6a2-634ac54461cd\") " pod="openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-fbsgr" Dec 10 23:22:07 crc kubenswrapper[4791]: I1210 23:22:07.815900 4791 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-hfdwt\" (UniqueName: \"kubernetes.io/projected/72f75339-ac62-4e10-a6a2-634ac54461cd-kube-api-access-hfdwt\") pod \"reboot-os-edpm-deployment-openstack-edpm-ipam-fbsgr\" (UID: \"72f75339-ac62-4e10-a6a2-634ac54461cd\") " pod="openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-fbsgr" Dec 10 23:22:07 crc kubenswrapper[4791]: I1210 23:22:07.815922 4791 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/72f75339-ac62-4e10-a6a2-634ac54461cd-ssh-key\") pod \"reboot-os-edpm-deployment-openstack-edpm-ipam-fbsgr\" (UID: \"72f75339-ac62-4e10-a6a2-634ac54461cd\") " pod="openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-fbsgr" Dec 10 23:22:07 crc kubenswrapper[4791]: I1210 23:22:07.918578 4791 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-hfdwt\" (UniqueName: \"kubernetes.io/projected/72f75339-ac62-4e10-a6a2-634ac54461cd-kube-api-access-hfdwt\") pod \"reboot-os-edpm-deployment-openstack-edpm-ipam-fbsgr\" (UID: \"72f75339-ac62-4e10-a6a2-634ac54461cd\") " pod="openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-fbsgr" Dec 10 23:22:07 crc kubenswrapper[4791]: I1210 23:22:07.918703 4791 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/72f75339-ac62-4e10-a6a2-634ac54461cd-ssh-key\") pod \"reboot-os-edpm-deployment-openstack-edpm-ipam-fbsgr\" (UID: \"72f75339-ac62-4e10-a6a2-634ac54461cd\") " pod="openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-fbsgr" Dec 10 23:22:07 crc kubenswrapper[4791]: I1210 23:22:07.918898 4791 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/72f75339-ac62-4e10-a6a2-634ac54461cd-inventory\") pod \"reboot-os-edpm-deployment-openstack-edpm-ipam-fbsgr\" (UID: \"72f75339-ac62-4e10-a6a2-634ac54461cd\") " pod="openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-fbsgr" Dec 10 23:22:07 crc kubenswrapper[4791]: I1210 23:22:07.922877 4791 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/72f75339-ac62-4e10-a6a2-634ac54461cd-inventory\") pod \"reboot-os-edpm-deployment-openstack-edpm-ipam-fbsgr\" (UID: \"72f75339-ac62-4e10-a6a2-634ac54461cd\") " pod="openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-fbsgr" Dec 10 23:22:07 crc kubenswrapper[4791]: I1210 23:22:07.925934 4791 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/72f75339-ac62-4e10-a6a2-634ac54461cd-ssh-key\") pod \"reboot-os-edpm-deployment-openstack-edpm-ipam-fbsgr\" (UID: \"72f75339-ac62-4e10-a6a2-634ac54461cd\") " pod="openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-fbsgr" Dec 10 23:22:07 crc kubenswrapper[4791]: I1210 23:22:07.935742 4791 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-hfdwt\" (UniqueName: \"kubernetes.io/projected/72f75339-ac62-4e10-a6a2-634ac54461cd-kube-api-access-hfdwt\") pod \"reboot-os-edpm-deployment-openstack-edpm-ipam-fbsgr\" (UID: \"72f75339-ac62-4e10-a6a2-634ac54461cd\") " pod="openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-fbsgr" Dec 10 23:22:08 crc kubenswrapper[4791]: I1210 23:22:08.101365 4791 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-fbsgr" Dec 10 23:22:08 crc kubenswrapper[4791]: I1210 23:22:08.631128 4791 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-fbsgr"] Dec 10 23:22:08 crc kubenswrapper[4791]: W1210 23:22:08.671247 4791 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod72f75339_ac62_4e10_a6a2_634ac54461cd.slice/crio-23736085b5ef9774c314a77a4c4a1529c58671df51120b4552e6415dc071957a WatchSource:0}: Error finding container 23736085b5ef9774c314a77a4c4a1529c58671df51120b4552e6415dc071957a: Status 404 returned error can't find the container with id 23736085b5ef9774c314a77a4c4a1529c58671df51120b4552e6415dc071957a Dec 10 23:22:08 crc kubenswrapper[4791]: I1210 23:22:08.675517 4791 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Dec 10 23:22:08 crc kubenswrapper[4791]: I1210 23:22:08.733909 4791 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-fbsgr" event={"ID":"72f75339-ac62-4e10-a6a2-634ac54461cd","Type":"ContainerStarted","Data":"23736085b5ef9774c314a77a4c4a1529c58671df51120b4552e6415dc071957a"} Dec 10 23:22:10 crc kubenswrapper[4791]: I1210 23:22:10.760359 4791 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-fbsgr" event={"ID":"72f75339-ac62-4e10-a6a2-634ac54461cd","Type":"ContainerStarted","Data":"4809d2d6989d6aa058afb74515b5c277f038ba612c2f2112f74fb64348fb8f15"} Dec 10 23:22:10 crc kubenswrapper[4791]: I1210 23:22:10.789624 4791 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-fbsgr" podStartSLOduration=2.9813729970000002 podStartE2EDuration="3.789577597s" podCreationTimestamp="2025-12-10 23:22:07 +0000 UTC" firstStartedPulling="2025-12-10 23:22:08.675103048 +0000 UTC m=+1963.104720671" lastFinishedPulling="2025-12-10 23:22:09.483307618 +0000 UTC m=+1963.912925271" observedRunningTime="2025-12-10 23:22:10.780566437 +0000 UTC m=+1965.210184060" watchObservedRunningTime="2025-12-10 23:22:10.789577597 +0000 UTC m=+1965.219195250" Dec 10 23:22:18 crc kubenswrapper[4791]: I1210 23:22:18.849446 4791 generic.go:334] "Generic (PLEG): container finished" podID="72f75339-ac62-4e10-a6a2-634ac54461cd" containerID="4809d2d6989d6aa058afb74515b5c277f038ba612c2f2112f74fb64348fb8f15" exitCode=0 Dec 10 23:22:18 crc kubenswrapper[4791]: I1210 23:22:18.849721 4791 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-fbsgr" event={"ID":"72f75339-ac62-4e10-a6a2-634ac54461cd","Type":"ContainerDied","Data":"4809d2d6989d6aa058afb74515b5c277f038ba612c2f2112f74fb64348fb8f15"} Dec 10 23:22:20 crc kubenswrapper[4791]: I1210 23:22:20.263231 4791 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-fbsgr" Dec 10 23:22:20 crc kubenswrapper[4791]: I1210 23:22:20.402225 4791 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/72f75339-ac62-4e10-a6a2-634ac54461cd-ssh-key\") pod \"72f75339-ac62-4e10-a6a2-634ac54461cd\" (UID: \"72f75339-ac62-4e10-a6a2-634ac54461cd\") " Dec 10 23:22:20 crc kubenswrapper[4791]: I1210 23:22:20.402607 4791 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/72f75339-ac62-4e10-a6a2-634ac54461cd-inventory\") pod \"72f75339-ac62-4e10-a6a2-634ac54461cd\" (UID: \"72f75339-ac62-4e10-a6a2-634ac54461cd\") " Dec 10 23:22:20 crc kubenswrapper[4791]: I1210 23:22:20.402745 4791 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-hfdwt\" (UniqueName: \"kubernetes.io/projected/72f75339-ac62-4e10-a6a2-634ac54461cd-kube-api-access-hfdwt\") pod \"72f75339-ac62-4e10-a6a2-634ac54461cd\" (UID: \"72f75339-ac62-4e10-a6a2-634ac54461cd\") " Dec 10 23:22:20 crc kubenswrapper[4791]: I1210 23:22:20.409896 4791 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/72f75339-ac62-4e10-a6a2-634ac54461cd-kube-api-access-hfdwt" (OuterVolumeSpecName: "kube-api-access-hfdwt") pod "72f75339-ac62-4e10-a6a2-634ac54461cd" (UID: "72f75339-ac62-4e10-a6a2-634ac54461cd"). InnerVolumeSpecName "kube-api-access-hfdwt". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 10 23:22:20 crc kubenswrapper[4791]: I1210 23:22:20.448423 4791 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/72f75339-ac62-4e10-a6a2-634ac54461cd-ssh-key" (OuterVolumeSpecName: "ssh-key") pod "72f75339-ac62-4e10-a6a2-634ac54461cd" (UID: "72f75339-ac62-4e10-a6a2-634ac54461cd"). InnerVolumeSpecName "ssh-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 10 23:22:20 crc kubenswrapper[4791]: I1210 23:22:20.456576 4791 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/72f75339-ac62-4e10-a6a2-634ac54461cd-inventory" (OuterVolumeSpecName: "inventory") pod "72f75339-ac62-4e10-a6a2-634ac54461cd" (UID: "72f75339-ac62-4e10-a6a2-634ac54461cd"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 10 23:22:20 crc kubenswrapper[4791]: I1210 23:22:20.505926 4791 reconciler_common.go:293] "Volume detached for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/72f75339-ac62-4e10-a6a2-634ac54461cd-ssh-key\") on node \"crc\" DevicePath \"\"" Dec 10 23:22:20 crc kubenswrapper[4791]: I1210 23:22:20.505955 4791 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/72f75339-ac62-4e10-a6a2-634ac54461cd-inventory\") on node \"crc\" DevicePath \"\"" Dec 10 23:22:20 crc kubenswrapper[4791]: I1210 23:22:20.505966 4791 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-hfdwt\" (UniqueName: \"kubernetes.io/projected/72f75339-ac62-4e10-a6a2-634ac54461cd-kube-api-access-hfdwt\") on node \"crc\" DevicePath \"\"" Dec 10 23:22:20 crc kubenswrapper[4791]: I1210 23:22:20.870791 4791 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-fbsgr" event={"ID":"72f75339-ac62-4e10-a6a2-634ac54461cd","Type":"ContainerDied","Data":"23736085b5ef9774c314a77a4c4a1529c58671df51120b4552e6415dc071957a"} Dec 10 23:22:20 crc kubenswrapper[4791]: I1210 23:22:20.870826 4791 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="23736085b5ef9774c314a77a4c4a1529c58671df51120b4552e6415dc071957a" Dec 10 23:22:20 crc kubenswrapper[4791]: I1210 23:22:20.870909 4791 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-fbsgr" Dec 10 23:22:20 crc kubenswrapper[4791]: I1210 23:22:20.966763 4791 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/install-certs-edpm-deployment-openstack-edpm-ipam-n9w4z"] Dec 10 23:22:20 crc kubenswrapper[4791]: E1210 23:22:20.967244 4791 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="72f75339-ac62-4e10-a6a2-634ac54461cd" containerName="reboot-os-edpm-deployment-openstack-edpm-ipam" Dec 10 23:22:20 crc kubenswrapper[4791]: I1210 23:22:20.967273 4791 state_mem.go:107] "Deleted CPUSet assignment" podUID="72f75339-ac62-4e10-a6a2-634ac54461cd" containerName="reboot-os-edpm-deployment-openstack-edpm-ipam" Dec 10 23:22:20 crc kubenswrapper[4791]: I1210 23:22:20.967509 4791 memory_manager.go:354] "RemoveStaleState removing state" podUID="72f75339-ac62-4e10-a6a2-634ac54461cd" containerName="reboot-os-edpm-deployment-openstack-edpm-ipam" Dec 10 23:22:20 crc kubenswrapper[4791]: I1210 23:22:20.968203 4791 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-n9w4z" Dec 10 23:22:20 crc kubenswrapper[4791]: I1210 23:22:20.970432 4791 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-dockercfg-hf8pr" Dec 10 23:22:20 crc kubenswrapper[4791]: I1210 23:22:20.970683 4791 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-libvirt-default-certs-0" Dec 10 23:22:20 crc kubenswrapper[4791]: I1210 23:22:20.970754 4791 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-edpm-ipam" Dec 10 23:22:20 crc kubenswrapper[4791]: I1210 23:22:20.970865 4791 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-telemetry-default-certs-0" Dec 10 23:22:20 crc kubenswrapper[4791]: I1210 23:22:20.971219 4791 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-neutron-metadata-default-certs-0" Dec 10 23:22:20 crc kubenswrapper[4791]: I1210 23:22:20.971402 4791 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-ansible-ssh-private-key-secret" Dec 10 23:22:20 crc kubenswrapper[4791]: I1210 23:22:20.972688 4791 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Dec 10 23:22:20 crc kubenswrapper[4791]: I1210 23:22:20.972764 4791 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-ovn-default-certs-0" Dec 10 23:22:20 crc kubenswrapper[4791]: I1210 23:22:20.992447 4791 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/install-certs-edpm-deployment-openstack-edpm-ipam-n9w4z"] Dec 10 23:22:21 crc kubenswrapper[4791]: I1210 23:22:21.016542 4791 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"repo-setup-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/7eb11eed-a935-4ce6-b348-2c7c8f6ef16a-repo-setup-combined-ca-bundle\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-n9w4z\" (UID: \"7eb11eed-a935-4ce6-b348-2c7c8f6ef16a\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-n9w4z" Dec 10 23:22:21 crc kubenswrapper[4791]: I1210 23:22:21.016620 4791 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"libvirt-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/7eb11eed-a935-4ce6-b348-2c7c8f6ef16a-libvirt-combined-ca-bundle\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-n9w4z\" (UID: \"7eb11eed-a935-4ce6-b348-2c7c8f6ef16a\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-n9w4z" Dec 10 23:22:21 crc kubenswrapper[4791]: I1210 23:22:21.016699 4791 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"neutron-metadata-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/7eb11eed-a935-4ce6-b348-2c7c8f6ef16a-neutron-metadata-combined-ca-bundle\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-n9w4z\" (UID: \"7eb11eed-a935-4ce6-b348-2c7c8f6ef16a\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-n9w4z" Dec 10 23:22:21 crc kubenswrapper[4791]: I1210 23:22:21.016911 4791 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"openstack-edpm-ipam-telemetry-default-certs-0\" (UniqueName: \"kubernetes.io/projected/7eb11eed-a935-4ce6-b348-2c7c8f6ef16a-openstack-edpm-ipam-telemetry-default-certs-0\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-n9w4z\" (UID: \"7eb11eed-a935-4ce6-b348-2c7c8f6ef16a\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-n9w4z" Dec 10 23:22:21 crc kubenswrapper[4791]: I1210 23:22:21.016968 4791 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovn-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/7eb11eed-a935-4ce6-b348-2c7c8f6ef16a-ovn-combined-ca-bundle\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-n9w4z\" (UID: \"7eb11eed-a935-4ce6-b348-2c7c8f6ef16a\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-n9w4z" Dec 10 23:22:21 crc kubenswrapper[4791]: I1210 23:22:21.017009 4791 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"openstack-edpm-ipam-ovn-default-certs-0\" (UniqueName: \"kubernetes.io/projected/7eb11eed-a935-4ce6-b348-2c7c8f6ef16a-openstack-edpm-ipam-ovn-default-certs-0\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-n9w4z\" (UID: \"7eb11eed-a935-4ce6-b348-2c7c8f6ef16a\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-n9w4z" Dec 10 23:22:21 crc kubenswrapper[4791]: I1210 23:22:21.017114 4791 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"nova-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/7eb11eed-a935-4ce6-b348-2c7c8f6ef16a-nova-combined-ca-bundle\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-n9w4z\" (UID: \"7eb11eed-a935-4ce6-b348-2c7c8f6ef16a\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-n9w4z" Dec 10 23:22:21 crc kubenswrapper[4791]: I1210 23:22:21.017170 4791 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"openstack-edpm-ipam-neutron-metadata-default-certs-0\" (UniqueName: \"kubernetes.io/projected/7eb11eed-a935-4ce6-b348-2c7c8f6ef16a-openstack-edpm-ipam-neutron-metadata-default-certs-0\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-n9w4z\" (UID: \"7eb11eed-a935-4ce6-b348-2c7c8f6ef16a\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-n9w4z" Dec 10 23:22:21 crc kubenswrapper[4791]: I1210 23:22:21.017369 4791 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/7eb11eed-a935-4ce6-b348-2c7c8f6ef16a-inventory\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-n9w4z\" (UID: \"7eb11eed-a935-4ce6-b348-2c7c8f6ef16a\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-n9w4z" Dec 10 23:22:21 crc kubenswrapper[4791]: I1210 23:22:21.017428 4791 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-pjn9h\" (UniqueName: \"kubernetes.io/projected/7eb11eed-a935-4ce6-b348-2c7c8f6ef16a-kube-api-access-pjn9h\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-n9w4z\" (UID: \"7eb11eed-a935-4ce6-b348-2c7c8f6ef16a\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-n9w4z" Dec 10 23:22:21 crc kubenswrapper[4791]: I1210 23:22:21.017549 4791 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"openstack-edpm-ipam-libvirt-default-certs-0\" (UniqueName: \"kubernetes.io/projected/7eb11eed-a935-4ce6-b348-2c7c8f6ef16a-openstack-edpm-ipam-libvirt-default-certs-0\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-n9w4z\" (UID: \"7eb11eed-a935-4ce6-b348-2c7c8f6ef16a\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-n9w4z" Dec 10 23:22:21 crc kubenswrapper[4791]: I1210 23:22:21.017633 4791 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/7eb11eed-a935-4ce6-b348-2c7c8f6ef16a-ssh-key\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-n9w4z\" (UID: \"7eb11eed-a935-4ce6-b348-2c7c8f6ef16a\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-n9w4z" Dec 10 23:22:21 crc kubenswrapper[4791]: I1210 23:22:21.017698 4791 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"bootstrap-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/7eb11eed-a935-4ce6-b348-2c7c8f6ef16a-bootstrap-combined-ca-bundle\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-n9w4z\" (UID: \"7eb11eed-a935-4ce6-b348-2c7c8f6ef16a\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-n9w4z" Dec 10 23:22:21 crc kubenswrapper[4791]: I1210 23:22:21.017797 4791 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"telemetry-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/7eb11eed-a935-4ce6-b348-2c7c8f6ef16a-telemetry-combined-ca-bundle\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-n9w4z\" (UID: \"7eb11eed-a935-4ce6-b348-2c7c8f6ef16a\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-n9w4z" Dec 10 23:22:21 crc kubenswrapper[4791]: I1210 23:22:21.119487 4791 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"neutron-metadata-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/7eb11eed-a935-4ce6-b348-2c7c8f6ef16a-neutron-metadata-combined-ca-bundle\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-n9w4z\" (UID: \"7eb11eed-a935-4ce6-b348-2c7c8f6ef16a\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-n9w4z" Dec 10 23:22:21 crc kubenswrapper[4791]: I1210 23:22:21.119562 4791 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"openstack-edpm-ipam-telemetry-default-certs-0\" (UniqueName: \"kubernetes.io/projected/7eb11eed-a935-4ce6-b348-2c7c8f6ef16a-openstack-edpm-ipam-telemetry-default-certs-0\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-n9w4z\" (UID: \"7eb11eed-a935-4ce6-b348-2c7c8f6ef16a\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-n9w4z" Dec 10 23:22:21 crc kubenswrapper[4791]: I1210 23:22:21.119601 4791 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovn-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/7eb11eed-a935-4ce6-b348-2c7c8f6ef16a-ovn-combined-ca-bundle\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-n9w4z\" (UID: \"7eb11eed-a935-4ce6-b348-2c7c8f6ef16a\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-n9w4z" Dec 10 23:22:21 crc kubenswrapper[4791]: I1210 23:22:21.119629 4791 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"openstack-edpm-ipam-ovn-default-certs-0\" (UniqueName: \"kubernetes.io/projected/7eb11eed-a935-4ce6-b348-2c7c8f6ef16a-openstack-edpm-ipam-ovn-default-certs-0\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-n9w4z\" (UID: \"7eb11eed-a935-4ce6-b348-2c7c8f6ef16a\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-n9w4z" Dec 10 23:22:21 crc kubenswrapper[4791]: I1210 23:22:21.119696 4791 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nova-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/7eb11eed-a935-4ce6-b348-2c7c8f6ef16a-nova-combined-ca-bundle\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-n9w4z\" (UID: \"7eb11eed-a935-4ce6-b348-2c7c8f6ef16a\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-n9w4z" Dec 10 23:22:21 crc kubenswrapper[4791]: I1210 23:22:21.119721 4791 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"openstack-edpm-ipam-neutron-metadata-default-certs-0\" (UniqueName: \"kubernetes.io/projected/7eb11eed-a935-4ce6-b348-2c7c8f6ef16a-openstack-edpm-ipam-neutron-metadata-default-certs-0\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-n9w4z\" (UID: \"7eb11eed-a935-4ce6-b348-2c7c8f6ef16a\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-n9w4z" Dec 10 23:22:21 crc kubenswrapper[4791]: I1210 23:22:21.119813 4791 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/7eb11eed-a935-4ce6-b348-2c7c8f6ef16a-inventory\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-n9w4z\" (UID: \"7eb11eed-a935-4ce6-b348-2c7c8f6ef16a\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-n9w4z" Dec 10 23:22:21 crc kubenswrapper[4791]: I1210 23:22:21.119844 4791 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-pjn9h\" (UniqueName: \"kubernetes.io/projected/7eb11eed-a935-4ce6-b348-2c7c8f6ef16a-kube-api-access-pjn9h\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-n9w4z\" (UID: \"7eb11eed-a935-4ce6-b348-2c7c8f6ef16a\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-n9w4z" Dec 10 23:22:21 crc kubenswrapper[4791]: I1210 23:22:21.119896 4791 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"openstack-edpm-ipam-libvirt-default-certs-0\" (UniqueName: \"kubernetes.io/projected/7eb11eed-a935-4ce6-b348-2c7c8f6ef16a-openstack-edpm-ipam-libvirt-default-certs-0\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-n9w4z\" (UID: \"7eb11eed-a935-4ce6-b348-2c7c8f6ef16a\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-n9w4z" Dec 10 23:22:21 crc kubenswrapper[4791]: I1210 23:22:21.119931 4791 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/7eb11eed-a935-4ce6-b348-2c7c8f6ef16a-ssh-key\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-n9w4z\" (UID: \"7eb11eed-a935-4ce6-b348-2c7c8f6ef16a\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-n9w4z" Dec 10 23:22:21 crc kubenswrapper[4791]: I1210 23:22:21.119962 4791 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"bootstrap-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/7eb11eed-a935-4ce6-b348-2c7c8f6ef16a-bootstrap-combined-ca-bundle\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-n9w4z\" (UID: \"7eb11eed-a935-4ce6-b348-2c7c8f6ef16a\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-n9w4z" Dec 10 23:22:21 crc kubenswrapper[4791]: I1210 23:22:21.120004 4791 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"telemetry-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/7eb11eed-a935-4ce6-b348-2c7c8f6ef16a-telemetry-combined-ca-bundle\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-n9w4z\" (UID: \"7eb11eed-a935-4ce6-b348-2c7c8f6ef16a\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-n9w4z" Dec 10 23:22:21 crc kubenswrapper[4791]: I1210 23:22:21.120040 4791 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"repo-setup-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/7eb11eed-a935-4ce6-b348-2c7c8f6ef16a-repo-setup-combined-ca-bundle\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-n9w4z\" (UID: \"7eb11eed-a935-4ce6-b348-2c7c8f6ef16a\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-n9w4z" Dec 10 23:22:21 crc kubenswrapper[4791]: I1210 23:22:21.120070 4791 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"libvirt-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/7eb11eed-a935-4ce6-b348-2c7c8f6ef16a-libvirt-combined-ca-bundle\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-n9w4z\" (UID: \"7eb11eed-a935-4ce6-b348-2c7c8f6ef16a\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-n9w4z" Dec 10 23:22:21 crc kubenswrapper[4791]: I1210 23:22:21.125615 4791 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"telemetry-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/7eb11eed-a935-4ce6-b348-2c7c8f6ef16a-telemetry-combined-ca-bundle\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-n9w4z\" (UID: \"7eb11eed-a935-4ce6-b348-2c7c8f6ef16a\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-n9w4z" Dec 10 23:22:21 crc kubenswrapper[4791]: I1210 23:22:21.125959 4791 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovn-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/7eb11eed-a935-4ce6-b348-2c7c8f6ef16a-ovn-combined-ca-bundle\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-n9w4z\" (UID: \"7eb11eed-a935-4ce6-b348-2c7c8f6ef16a\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-n9w4z" Dec 10 23:22:21 crc kubenswrapper[4791]: I1210 23:22:21.126241 4791 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"openstack-edpm-ipam-libvirt-default-certs-0\" (UniqueName: \"kubernetes.io/projected/7eb11eed-a935-4ce6-b348-2c7c8f6ef16a-openstack-edpm-ipam-libvirt-default-certs-0\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-n9w4z\" (UID: \"7eb11eed-a935-4ce6-b348-2c7c8f6ef16a\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-n9w4z" Dec 10 23:22:21 crc kubenswrapper[4791]: I1210 23:22:21.126246 4791 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"libvirt-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/7eb11eed-a935-4ce6-b348-2c7c8f6ef16a-libvirt-combined-ca-bundle\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-n9w4z\" (UID: \"7eb11eed-a935-4ce6-b348-2c7c8f6ef16a\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-n9w4z" Dec 10 23:22:21 crc kubenswrapper[4791]: I1210 23:22:21.126398 4791 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"openstack-edpm-ipam-neutron-metadata-default-certs-0\" (UniqueName: \"kubernetes.io/projected/7eb11eed-a935-4ce6-b348-2c7c8f6ef16a-openstack-edpm-ipam-neutron-metadata-default-certs-0\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-n9w4z\" (UID: \"7eb11eed-a935-4ce6-b348-2c7c8f6ef16a\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-n9w4z" Dec 10 23:22:21 crc kubenswrapper[4791]: I1210 23:22:21.126819 4791 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/7eb11eed-a935-4ce6-b348-2c7c8f6ef16a-ssh-key\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-n9w4z\" (UID: \"7eb11eed-a935-4ce6-b348-2c7c8f6ef16a\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-n9w4z" Dec 10 23:22:21 crc kubenswrapper[4791]: I1210 23:22:21.127922 4791 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"bootstrap-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/7eb11eed-a935-4ce6-b348-2c7c8f6ef16a-bootstrap-combined-ca-bundle\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-n9w4z\" (UID: \"7eb11eed-a935-4ce6-b348-2c7c8f6ef16a\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-n9w4z" Dec 10 23:22:21 crc kubenswrapper[4791]: I1210 23:22:21.128190 4791 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"nova-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/7eb11eed-a935-4ce6-b348-2c7c8f6ef16a-nova-combined-ca-bundle\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-n9w4z\" (UID: \"7eb11eed-a935-4ce6-b348-2c7c8f6ef16a\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-n9w4z" Dec 10 23:22:21 crc kubenswrapper[4791]: I1210 23:22:21.128908 4791 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"openstack-edpm-ipam-telemetry-default-certs-0\" (UniqueName: \"kubernetes.io/projected/7eb11eed-a935-4ce6-b348-2c7c8f6ef16a-openstack-edpm-ipam-telemetry-default-certs-0\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-n9w4z\" (UID: \"7eb11eed-a935-4ce6-b348-2c7c8f6ef16a\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-n9w4z" Dec 10 23:22:21 crc kubenswrapper[4791]: I1210 23:22:21.129199 4791 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"openstack-edpm-ipam-ovn-default-certs-0\" (UniqueName: \"kubernetes.io/projected/7eb11eed-a935-4ce6-b348-2c7c8f6ef16a-openstack-edpm-ipam-ovn-default-certs-0\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-n9w4z\" (UID: \"7eb11eed-a935-4ce6-b348-2c7c8f6ef16a\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-n9w4z" Dec 10 23:22:21 crc kubenswrapper[4791]: I1210 23:22:21.129667 4791 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/7eb11eed-a935-4ce6-b348-2c7c8f6ef16a-inventory\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-n9w4z\" (UID: \"7eb11eed-a935-4ce6-b348-2c7c8f6ef16a\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-n9w4z" Dec 10 23:22:21 crc kubenswrapper[4791]: I1210 23:22:21.130589 4791 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"neutron-metadata-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/7eb11eed-a935-4ce6-b348-2c7c8f6ef16a-neutron-metadata-combined-ca-bundle\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-n9w4z\" (UID: \"7eb11eed-a935-4ce6-b348-2c7c8f6ef16a\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-n9w4z" Dec 10 23:22:21 crc kubenswrapper[4791]: I1210 23:22:21.130735 4791 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"repo-setup-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/7eb11eed-a935-4ce6-b348-2c7c8f6ef16a-repo-setup-combined-ca-bundle\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-n9w4z\" (UID: \"7eb11eed-a935-4ce6-b348-2c7c8f6ef16a\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-n9w4z" Dec 10 23:22:21 crc kubenswrapper[4791]: I1210 23:22:21.141401 4791 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-pjn9h\" (UniqueName: \"kubernetes.io/projected/7eb11eed-a935-4ce6-b348-2c7c8f6ef16a-kube-api-access-pjn9h\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-n9w4z\" (UID: \"7eb11eed-a935-4ce6-b348-2c7c8f6ef16a\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-n9w4z" Dec 10 23:22:21 crc kubenswrapper[4791]: I1210 23:22:21.298786 4791 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-n9w4z" Dec 10 23:22:21 crc kubenswrapper[4791]: I1210 23:22:21.640513 4791 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/install-certs-edpm-deployment-openstack-edpm-ipam-n9w4z"] Dec 10 23:22:21 crc kubenswrapper[4791]: I1210 23:22:21.896979 4791 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-n9w4z" event={"ID":"7eb11eed-a935-4ce6-b348-2c7c8f6ef16a","Type":"ContainerStarted","Data":"098aa9721c3acc56edf71bede6f7aeb3062578b367a6f6a6f03c56ea31425ff0"} Dec 10 23:22:22 crc kubenswrapper[4791]: I1210 23:22:22.899494 4791 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-n9w4z" event={"ID":"7eb11eed-a935-4ce6-b348-2c7c8f6ef16a","Type":"ContainerStarted","Data":"fa03e3370a2bb78cadeae1c03012a618bdf24914f3535eec31929bd1c5353a6e"} Dec 10 23:22:22 crc kubenswrapper[4791]: I1210 23:22:22.940715 4791 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-n9w4z" podStartSLOduration=2.2402738429999998 podStartE2EDuration="2.940683752s" podCreationTimestamp="2025-12-10 23:22:20 +0000 UTC" firstStartedPulling="2025-12-10 23:22:21.650528468 +0000 UTC m=+1976.080146081" lastFinishedPulling="2025-12-10 23:22:22.350938337 +0000 UTC m=+1976.780555990" observedRunningTime="2025-12-10 23:22:22.932579958 +0000 UTC m=+1977.362197631" watchObservedRunningTime="2025-12-10 23:22:22.940683752 +0000 UTC m=+1977.370301395" Dec 10 23:22:55 crc kubenswrapper[4791]: I1210 23:22:55.040990 4791 patch_prober.go:28] interesting pod/machine-config-daemon-5rb5l container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 10 23:22:55 crc kubenswrapper[4791]: I1210 23:22:55.041864 4791 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-5rb5l" podUID="ba35653c-6e06-4cee-a4d6-137764090d18" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 10 23:23:00 crc kubenswrapper[4791]: I1210 23:23:00.399408 4791 generic.go:334] "Generic (PLEG): container finished" podID="7eb11eed-a935-4ce6-b348-2c7c8f6ef16a" containerID="fa03e3370a2bb78cadeae1c03012a618bdf24914f3535eec31929bd1c5353a6e" exitCode=0 Dec 10 23:23:00 crc kubenswrapper[4791]: I1210 23:23:00.399776 4791 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-n9w4z" event={"ID":"7eb11eed-a935-4ce6-b348-2c7c8f6ef16a","Type":"ContainerDied","Data":"fa03e3370a2bb78cadeae1c03012a618bdf24914f3535eec31929bd1c5353a6e"} Dec 10 23:23:01 crc kubenswrapper[4791]: I1210 23:23:01.896624 4791 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-n9w4z" Dec 10 23:23:02 crc kubenswrapper[4791]: I1210 23:23:02.064486 4791 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"libvirt-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/7eb11eed-a935-4ce6-b348-2c7c8f6ef16a-libvirt-combined-ca-bundle\") pod \"7eb11eed-a935-4ce6-b348-2c7c8f6ef16a\" (UID: \"7eb11eed-a935-4ce6-b348-2c7c8f6ef16a\") " Dec 10 23:23:02 crc kubenswrapper[4791]: I1210 23:23:02.065195 4791 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"nova-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/7eb11eed-a935-4ce6-b348-2c7c8f6ef16a-nova-combined-ca-bundle\") pod \"7eb11eed-a935-4ce6-b348-2c7c8f6ef16a\" (UID: \"7eb11eed-a935-4ce6-b348-2c7c8f6ef16a\") " Dec 10 23:23:02 crc kubenswrapper[4791]: I1210 23:23:02.065239 4791 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/7eb11eed-a935-4ce6-b348-2c7c8f6ef16a-inventory\") pod \"7eb11eed-a935-4ce6-b348-2c7c8f6ef16a\" (UID: \"7eb11eed-a935-4ce6-b348-2c7c8f6ef16a\") " Dec 10 23:23:02 crc kubenswrapper[4791]: I1210 23:23:02.065312 4791 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"openstack-edpm-ipam-neutron-metadata-default-certs-0\" (UniqueName: \"kubernetes.io/projected/7eb11eed-a935-4ce6-b348-2c7c8f6ef16a-openstack-edpm-ipam-neutron-metadata-default-certs-0\") pod \"7eb11eed-a935-4ce6-b348-2c7c8f6ef16a\" (UID: \"7eb11eed-a935-4ce6-b348-2c7c8f6ef16a\") " Dec 10 23:23:02 crc kubenswrapper[4791]: I1210 23:23:02.065392 4791 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"openstack-edpm-ipam-libvirt-default-certs-0\" (UniqueName: \"kubernetes.io/projected/7eb11eed-a935-4ce6-b348-2c7c8f6ef16a-openstack-edpm-ipam-libvirt-default-certs-0\") pod \"7eb11eed-a935-4ce6-b348-2c7c8f6ef16a\" (UID: \"7eb11eed-a935-4ce6-b348-2c7c8f6ef16a\") " Dec 10 23:23:02 crc kubenswrapper[4791]: I1210 23:23:02.065458 4791 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"neutron-metadata-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/7eb11eed-a935-4ce6-b348-2c7c8f6ef16a-neutron-metadata-combined-ca-bundle\") pod \"7eb11eed-a935-4ce6-b348-2c7c8f6ef16a\" (UID: \"7eb11eed-a935-4ce6-b348-2c7c8f6ef16a\") " Dec 10 23:23:02 crc kubenswrapper[4791]: I1210 23:23:02.065482 4791 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovn-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/7eb11eed-a935-4ce6-b348-2c7c8f6ef16a-ovn-combined-ca-bundle\") pod \"7eb11eed-a935-4ce6-b348-2c7c8f6ef16a\" (UID: \"7eb11eed-a935-4ce6-b348-2c7c8f6ef16a\") " Dec 10 23:23:02 crc kubenswrapper[4791]: I1210 23:23:02.065524 4791 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"bootstrap-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/7eb11eed-a935-4ce6-b348-2c7c8f6ef16a-bootstrap-combined-ca-bundle\") pod \"7eb11eed-a935-4ce6-b348-2c7c8f6ef16a\" (UID: \"7eb11eed-a935-4ce6-b348-2c7c8f6ef16a\") " Dec 10 23:23:02 crc kubenswrapper[4791]: I1210 23:23:02.065578 4791 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"telemetry-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/7eb11eed-a935-4ce6-b348-2c7c8f6ef16a-telemetry-combined-ca-bundle\") pod \"7eb11eed-a935-4ce6-b348-2c7c8f6ef16a\" (UID: \"7eb11eed-a935-4ce6-b348-2c7c8f6ef16a\") " Dec 10 23:23:02 crc kubenswrapper[4791]: I1210 23:23:02.065626 4791 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"openstack-edpm-ipam-telemetry-default-certs-0\" (UniqueName: \"kubernetes.io/projected/7eb11eed-a935-4ce6-b348-2c7c8f6ef16a-openstack-edpm-ipam-telemetry-default-certs-0\") pod \"7eb11eed-a935-4ce6-b348-2c7c8f6ef16a\" (UID: \"7eb11eed-a935-4ce6-b348-2c7c8f6ef16a\") " Dec 10 23:23:02 crc kubenswrapper[4791]: I1210 23:23:02.065667 4791 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/7eb11eed-a935-4ce6-b348-2c7c8f6ef16a-ssh-key\") pod \"7eb11eed-a935-4ce6-b348-2c7c8f6ef16a\" (UID: \"7eb11eed-a935-4ce6-b348-2c7c8f6ef16a\") " Dec 10 23:23:02 crc kubenswrapper[4791]: I1210 23:23:02.065710 4791 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"openstack-edpm-ipam-ovn-default-certs-0\" (UniqueName: \"kubernetes.io/projected/7eb11eed-a935-4ce6-b348-2c7c8f6ef16a-openstack-edpm-ipam-ovn-default-certs-0\") pod \"7eb11eed-a935-4ce6-b348-2c7c8f6ef16a\" (UID: \"7eb11eed-a935-4ce6-b348-2c7c8f6ef16a\") " Dec 10 23:23:02 crc kubenswrapper[4791]: I1210 23:23:02.065744 4791 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-pjn9h\" (UniqueName: \"kubernetes.io/projected/7eb11eed-a935-4ce6-b348-2c7c8f6ef16a-kube-api-access-pjn9h\") pod \"7eb11eed-a935-4ce6-b348-2c7c8f6ef16a\" (UID: \"7eb11eed-a935-4ce6-b348-2c7c8f6ef16a\") " Dec 10 23:23:02 crc kubenswrapper[4791]: I1210 23:23:02.065768 4791 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"repo-setup-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/7eb11eed-a935-4ce6-b348-2c7c8f6ef16a-repo-setup-combined-ca-bundle\") pod \"7eb11eed-a935-4ce6-b348-2c7c8f6ef16a\" (UID: \"7eb11eed-a935-4ce6-b348-2c7c8f6ef16a\") " Dec 10 23:23:02 crc kubenswrapper[4791]: I1210 23:23:02.071959 4791 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/7eb11eed-a935-4ce6-b348-2c7c8f6ef16a-neutron-metadata-combined-ca-bundle" (OuterVolumeSpecName: "neutron-metadata-combined-ca-bundle") pod "7eb11eed-a935-4ce6-b348-2c7c8f6ef16a" (UID: "7eb11eed-a935-4ce6-b348-2c7c8f6ef16a"). InnerVolumeSpecName "neutron-metadata-combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 10 23:23:02 crc kubenswrapper[4791]: I1210 23:23:02.071985 4791 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/7eb11eed-a935-4ce6-b348-2c7c8f6ef16a-openstack-edpm-ipam-ovn-default-certs-0" (OuterVolumeSpecName: "openstack-edpm-ipam-ovn-default-certs-0") pod "7eb11eed-a935-4ce6-b348-2c7c8f6ef16a" (UID: "7eb11eed-a935-4ce6-b348-2c7c8f6ef16a"). InnerVolumeSpecName "openstack-edpm-ipam-ovn-default-certs-0". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 10 23:23:02 crc kubenswrapper[4791]: I1210 23:23:02.072273 4791 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/7eb11eed-a935-4ce6-b348-2c7c8f6ef16a-openstack-edpm-ipam-neutron-metadata-default-certs-0" (OuterVolumeSpecName: "openstack-edpm-ipam-neutron-metadata-default-certs-0") pod "7eb11eed-a935-4ce6-b348-2c7c8f6ef16a" (UID: "7eb11eed-a935-4ce6-b348-2c7c8f6ef16a"). InnerVolumeSpecName "openstack-edpm-ipam-neutron-metadata-default-certs-0". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 10 23:23:02 crc kubenswrapper[4791]: I1210 23:23:02.072915 4791 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/7eb11eed-a935-4ce6-b348-2c7c8f6ef16a-openstack-edpm-ipam-libvirt-default-certs-0" (OuterVolumeSpecName: "openstack-edpm-ipam-libvirt-default-certs-0") pod "7eb11eed-a935-4ce6-b348-2c7c8f6ef16a" (UID: "7eb11eed-a935-4ce6-b348-2c7c8f6ef16a"). InnerVolumeSpecName "openstack-edpm-ipam-libvirt-default-certs-0". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 10 23:23:02 crc kubenswrapper[4791]: I1210 23:23:02.073848 4791 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/7eb11eed-a935-4ce6-b348-2c7c8f6ef16a-nova-combined-ca-bundle" (OuterVolumeSpecName: "nova-combined-ca-bundle") pod "7eb11eed-a935-4ce6-b348-2c7c8f6ef16a" (UID: "7eb11eed-a935-4ce6-b348-2c7c8f6ef16a"). InnerVolumeSpecName "nova-combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 10 23:23:02 crc kubenswrapper[4791]: I1210 23:23:02.074176 4791 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/7eb11eed-a935-4ce6-b348-2c7c8f6ef16a-libvirt-combined-ca-bundle" (OuterVolumeSpecName: "libvirt-combined-ca-bundle") pod "7eb11eed-a935-4ce6-b348-2c7c8f6ef16a" (UID: "7eb11eed-a935-4ce6-b348-2c7c8f6ef16a"). InnerVolumeSpecName "libvirt-combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 10 23:23:02 crc kubenswrapper[4791]: I1210 23:23:02.075292 4791 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/7eb11eed-a935-4ce6-b348-2c7c8f6ef16a-ovn-combined-ca-bundle" (OuterVolumeSpecName: "ovn-combined-ca-bundle") pod "7eb11eed-a935-4ce6-b348-2c7c8f6ef16a" (UID: "7eb11eed-a935-4ce6-b348-2c7c8f6ef16a"). InnerVolumeSpecName "ovn-combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 10 23:23:02 crc kubenswrapper[4791]: I1210 23:23:02.075570 4791 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/7eb11eed-a935-4ce6-b348-2c7c8f6ef16a-bootstrap-combined-ca-bundle" (OuterVolumeSpecName: "bootstrap-combined-ca-bundle") pod "7eb11eed-a935-4ce6-b348-2c7c8f6ef16a" (UID: "7eb11eed-a935-4ce6-b348-2c7c8f6ef16a"). InnerVolumeSpecName "bootstrap-combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 10 23:23:02 crc kubenswrapper[4791]: I1210 23:23:02.075729 4791 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/7eb11eed-a935-4ce6-b348-2c7c8f6ef16a-telemetry-combined-ca-bundle" (OuterVolumeSpecName: "telemetry-combined-ca-bundle") pod "7eb11eed-a935-4ce6-b348-2c7c8f6ef16a" (UID: "7eb11eed-a935-4ce6-b348-2c7c8f6ef16a"). InnerVolumeSpecName "telemetry-combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 10 23:23:02 crc kubenswrapper[4791]: I1210 23:23:02.075767 4791 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/7eb11eed-a935-4ce6-b348-2c7c8f6ef16a-kube-api-access-pjn9h" (OuterVolumeSpecName: "kube-api-access-pjn9h") pod "7eb11eed-a935-4ce6-b348-2c7c8f6ef16a" (UID: "7eb11eed-a935-4ce6-b348-2c7c8f6ef16a"). InnerVolumeSpecName "kube-api-access-pjn9h". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 10 23:23:02 crc kubenswrapper[4791]: I1210 23:23:02.076985 4791 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/7eb11eed-a935-4ce6-b348-2c7c8f6ef16a-repo-setup-combined-ca-bundle" (OuterVolumeSpecName: "repo-setup-combined-ca-bundle") pod "7eb11eed-a935-4ce6-b348-2c7c8f6ef16a" (UID: "7eb11eed-a935-4ce6-b348-2c7c8f6ef16a"). InnerVolumeSpecName "repo-setup-combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 10 23:23:02 crc kubenswrapper[4791]: I1210 23:23:02.079551 4791 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/7eb11eed-a935-4ce6-b348-2c7c8f6ef16a-openstack-edpm-ipam-telemetry-default-certs-0" (OuterVolumeSpecName: "openstack-edpm-ipam-telemetry-default-certs-0") pod "7eb11eed-a935-4ce6-b348-2c7c8f6ef16a" (UID: "7eb11eed-a935-4ce6-b348-2c7c8f6ef16a"). InnerVolumeSpecName "openstack-edpm-ipam-telemetry-default-certs-0". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 10 23:23:02 crc kubenswrapper[4791]: I1210 23:23:02.100036 4791 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/7eb11eed-a935-4ce6-b348-2c7c8f6ef16a-inventory" (OuterVolumeSpecName: "inventory") pod "7eb11eed-a935-4ce6-b348-2c7c8f6ef16a" (UID: "7eb11eed-a935-4ce6-b348-2c7c8f6ef16a"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 10 23:23:02 crc kubenswrapper[4791]: I1210 23:23:02.102056 4791 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/7eb11eed-a935-4ce6-b348-2c7c8f6ef16a-ssh-key" (OuterVolumeSpecName: "ssh-key") pod "7eb11eed-a935-4ce6-b348-2c7c8f6ef16a" (UID: "7eb11eed-a935-4ce6-b348-2c7c8f6ef16a"). InnerVolumeSpecName "ssh-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 10 23:23:02 crc kubenswrapper[4791]: I1210 23:23:02.168818 4791 reconciler_common.go:293] "Volume detached for volume \"telemetry-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/7eb11eed-a935-4ce6-b348-2c7c8f6ef16a-telemetry-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 10 23:23:02 crc kubenswrapper[4791]: I1210 23:23:02.169123 4791 reconciler_common.go:293] "Volume detached for volume \"openstack-edpm-ipam-telemetry-default-certs-0\" (UniqueName: \"kubernetes.io/projected/7eb11eed-a935-4ce6-b348-2c7c8f6ef16a-openstack-edpm-ipam-telemetry-default-certs-0\") on node \"crc\" DevicePath \"\"" Dec 10 23:23:02 crc kubenswrapper[4791]: I1210 23:23:02.169145 4791 reconciler_common.go:293] "Volume detached for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/7eb11eed-a935-4ce6-b348-2c7c8f6ef16a-ssh-key\") on node \"crc\" DevicePath \"\"" Dec 10 23:23:02 crc kubenswrapper[4791]: I1210 23:23:02.169166 4791 reconciler_common.go:293] "Volume detached for volume \"openstack-edpm-ipam-ovn-default-certs-0\" (UniqueName: \"kubernetes.io/projected/7eb11eed-a935-4ce6-b348-2c7c8f6ef16a-openstack-edpm-ipam-ovn-default-certs-0\") on node \"crc\" DevicePath \"\"" Dec 10 23:23:02 crc kubenswrapper[4791]: I1210 23:23:02.169187 4791 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-pjn9h\" (UniqueName: \"kubernetes.io/projected/7eb11eed-a935-4ce6-b348-2c7c8f6ef16a-kube-api-access-pjn9h\") on node \"crc\" DevicePath \"\"" Dec 10 23:23:02 crc kubenswrapper[4791]: I1210 23:23:02.169205 4791 reconciler_common.go:293] "Volume detached for volume \"repo-setup-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/7eb11eed-a935-4ce6-b348-2c7c8f6ef16a-repo-setup-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 10 23:23:02 crc kubenswrapper[4791]: I1210 23:23:02.169223 4791 reconciler_common.go:293] "Volume detached for volume \"libvirt-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/7eb11eed-a935-4ce6-b348-2c7c8f6ef16a-libvirt-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 10 23:23:02 crc kubenswrapper[4791]: I1210 23:23:02.169240 4791 reconciler_common.go:293] "Volume detached for volume \"nova-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/7eb11eed-a935-4ce6-b348-2c7c8f6ef16a-nova-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 10 23:23:02 crc kubenswrapper[4791]: I1210 23:23:02.169257 4791 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/7eb11eed-a935-4ce6-b348-2c7c8f6ef16a-inventory\") on node \"crc\" DevicePath \"\"" Dec 10 23:23:02 crc kubenswrapper[4791]: I1210 23:23:02.169275 4791 reconciler_common.go:293] "Volume detached for volume \"openstack-edpm-ipam-neutron-metadata-default-certs-0\" (UniqueName: \"kubernetes.io/projected/7eb11eed-a935-4ce6-b348-2c7c8f6ef16a-openstack-edpm-ipam-neutron-metadata-default-certs-0\") on node \"crc\" DevicePath \"\"" Dec 10 23:23:02 crc kubenswrapper[4791]: I1210 23:23:02.169295 4791 reconciler_common.go:293] "Volume detached for volume \"openstack-edpm-ipam-libvirt-default-certs-0\" (UniqueName: \"kubernetes.io/projected/7eb11eed-a935-4ce6-b348-2c7c8f6ef16a-openstack-edpm-ipam-libvirt-default-certs-0\") on node \"crc\" DevicePath \"\"" Dec 10 23:23:02 crc kubenswrapper[4791]: I1210 23:23:02.169316 4791 reconciler_common.go:293] "Volume detached for volume \"ovn-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/7eb11eed-a935-4ce6-b348-2c7c8f6ef16a-ovn-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 10 23:23:02 crc kubenswrapper[4791]: I1210 23:23:02.169377 4791 reconciler_common.go:293] "Volume detached for volume \"neutron-metadata-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/7eb11eed-a935-4ce6-b348-2c7c8f6ef16a-neutron-metadata-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 10 23:23:02 crc kubenswrapper[4791]: I1210 23:23:02.169395 4791 reconciler_common.go:293] "Volume detached for volume \"bootstrap-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/7eb11eed-a935-4ce6-b348-2c7c8f6ef16a-bootstrap-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 10 23:23:02 crc kubenswrapper[4791]: I1210 23:23:02.422994 4791 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-n9w4z" event={"ID":"7eb11eed-a935-4ce6-b348-2c7c8f6ef16a","Type":"ContainerDied","Data":"098aa9721c3acc56edf71bede6f7aeb3062578b367a6f6a6f03c56ea31425ff0"} Dec 10 23:23:02 crc kubenswrapper[4791]: I1210 23:23:02.423050 4791 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="098aa9721c3acc56edf71bede6f7aeb3062578b367a6f6a6f03c56ea31425ff0" Dec 10 23:23:02 crc kubenswrapper[4791]: I1210 23:23:02.423137 4791 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-n9w4z" Dec 10 23:23:02 crc kubenswrapper[4791]: I1210 23:23:02.643158 4791 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ovn-edpm-deployment-openstack-edpm-ipam-ql7sj"] Dec 10 23:23:02 crc kubenswrapper[4791]: E1210 23:23:02.643649 4791 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="7eb11eed-a935-4ce6-b348-2c7c8f6ef16a" containerName="install-certs-edpm-deployment-openstack-edpm-ipam" Dec 10 23:23:02 crc kubenswrapper[4791]: I1210 23:23:02.643674 4791 state_mem.go:107] "Deleted CPUSet assignment" podUID="7eb11eed-a935-4ce6-b348-2c7c8f6ef16a" containerName="install-certs-edpm-deployment-openstack-edpm-ipam" Dec 10 23:23:02 crc kubenswrapper[4791]: I1210 23:23:02.643935 4791 memory_manager.go:354] "RemoveStaleState removing state" podUID="7eb11eed-a935-4ce6-b348-2c7c8f6ef16a" containerName="install-certs-edpm-deployment-openstack-edpm-ipam" Dec 10 23:23:02 crc kubenswrapper[4791]: I1210 23:23:02.644894 4791 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-ql7sj" Dec 10 23:23:02 crc kubenswrapper[4791]: I1210 23:23:02.653302 4791 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-ansible-ssh-private-key-secret" Dec 10 23:23:02 crc kubenswrapper[4791]: I1210 23:23:02.653535 4791 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Dec 10 23:23:02 crc kubenswrapper[4791]: I1210 23:23:02.653548 4791 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"ovncontroller-config" Dec 10 23:23:02 crc kubenswrapper[4791]: I1210 23:23:02.655028 4791 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-edpm-ipam" Dec 10 23:23:02 crc kubenswrapper[4791]: I1210 23:23:02.655080 4791 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-dockercfg-hf8pr" Dec 10 23:23:02 crc kubenswrapper[4791]: I1210 23:23:02.661707 4791 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovn-edpm-deployment-openstack-edpm-ipam-ql7sj"] Dec 10 23:23:02 crc kubenswrapper[4791]: I1210 23:23:02.781363 4791 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovn-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/9b2b641e-b123-4328-9151-f4c95e1405f2-ovn-combined-ca-bundle\") pod \"ovn-edpm-deployment-openstack-edpm-ipam-ql7sj\" (UID: \"9b2b641e-b123-4328-9151-f4c95e1405f2\") " pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-ql7sj" Dec 10 23:23:02 crc kubenswrapper[4791]: I1210 23:23:02.781441 4791 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovncontroller-config-0\" (UniqueName: \"kubernetes.io/configmap/9b2b641e-b123-4328-9151-f4c95e1405f2-ovncontroller-config-0\") pod \"ovn-edpm-deployment-openstack-edpm-ipam-ql7sj\" (UID: \"9b2b641e-b123-4328-9151-f4c95e1405f2\") " pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-ql7sj" Dec 10 23:23:02 crc kubenswrapper[4791]: I1210 23:23:02.781491 4791 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/9b2b641e-b123-4328-9151-f4c95e1405f2-inventory\") pod \"ovn-edpm-deployment-openstack-edpm-ipam-ql7sj\" (UID: \"9b2b641e-b123-4328-9151-f4c95e1405f2\") " pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-ql7sj" Dec 10 23:23:02 crc kubenswrapper[4791]: I1210 23:23:02.781518 4791 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/9b2b641e-b123-4328-9151-f4c95e1405f2-ssh-key\") pod \"ovn-edpm-deployment-openstack-edpm-ipam-ql7sj\" (UID: \"9b2b641e-b123-4328-9151-f4c95e1405f2\") " pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-ql7sj" Dec 10 23:23:02 crc kubenswrapper[4791]: I1210 23:23:02.781749 4791 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-pdpcv\" (UniqueName: \"kubernetes.io/projected/9b2b641e-b123-4328-9151-f4c95e1405f2-kube-api-access-pdpcv\") pod \"ovn-edpm-deployment-openstack-edpm-ipam-ql7sj\" (UID: \"9b2b641e-b123-4328-9151-f4c95e1405f2\") " pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-ql7sj" Dec 10 23:23:02 crc kubenswrapper[4791]: I1210 23:23:02.885048 4791 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovn-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/9b2b641e-b123-4328-9151-f4c95e1405f2-ovn-combined-ca-bundle\") pod \"ovn-edpm-deployment-openstack-edpm-ipam-ql7sj\" (UID: \"9b2b641e-b123-4328-9151-f4c95e1405f2\") " pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-ql7sj" Dec 10 23:23:02 crc kubenswrapper[4791]: I1210 23:23:02.885141 4791 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovncontroller-config-0\" (UniqueName: \"kubernetes.io/configmap/9b2b641e-b123-4328-9151-f4c95e1405f2-ovncontroller-config-0\") pod \"ovn-edpm-deployment-openstack-edpm-ipam-ql7sj\" (UID: \"9b2b641e-b123-4328-9151-f4c95e1405f2\") " pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-ql7sj" Dec 10 23:23:02 crc kubenswrapper[4791]: I1210 23:23:02.885183 4791 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/9b2b641e-b123-4328-9151-f4c95e1405f2-inventory\") pod \"ovn-edpm-deployment-openstack-edpm-ipam-ql7sj\" (UID: \"9b2b641e-b123-4328-9151-f4c95e1405f2\") " pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-ql7sj" Dec 10 23:23:02 crc kubenswrapper[4791]: I1210 23:23:02.885216 4791 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/9b2b641e-b123-4328-9151-f4c95e1405f2-ssh-key\") pod \"ovn-edpm-deployment-openstack-edpm-ipam-ql7sj\" (UID: \"9b2b641e-b123-4328-9151-f4c95e1405f2\") " pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-ql7sj" Dec 10 23:23:02 crc kubenswrapper[4791]: I1210 23:23:02.885265 4791 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-pdpcv\" (UniqueName: \"kubernetes.io/projected/9b2b641e-b123-4328-9151-f4c95e1405f2-kube-api-access-pdpcv\") pod \"ovn-edpm-deployment-openstack-edpm-ipam-ql7sj\" (UID: \"9b2b641e-b123-4328-9151-f4c95e1405f2\") " pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-ql7sj" Dec 10 23:23:02 crc kubenswrapper[4791]: I1210 23:23:02.886861 4791 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovncontroller-config-0\" (UniqueName: \"kubernetes.io/configmap/9b2b641e-b123-4328-9151-f4c95e1405f2-ovncontroller-config-0\") pod \"ovn-edpm-deployment-openstack-edpm-ipam-ql7sj\" (UID: \"9b2b641e-b123-4328-9151-f4c95e1405f2\") " pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-ql7sj" Dec 10 23:23:02 crc kubenswrapper[4791]: I1210 23:23:02.891853 4791 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/9b2b641e-b123-4328-9151-f4c95e1405f2-inventory\") pod \"ovn-edpm-deployment-openstack-edpm-ipam-ql7sj\" (UID: \"9b2b641e-b123-4328-9151-f4c95e1405f2\") " pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-ql7sj" Dec 10 23:23:02 crc kubenswrapper[4791]: I1210 23:23:02.891931 4791 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovn-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/9b2b641e-b123-4328-9151-f4c95e1405f2-ovn-combined-ca-bundle\") pod \"ovn-edpm-deployment-openstack-edpm-ipam-ql7sj\" (UID: \"9b2b641e-b123-4328-9151-f4c95e1405f2\") " pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-ql7sj" Dec 10 23:23:02 crc kubenswrapper[4791]: I1210 23:23:02.894059 4791 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/9b2b641e-b123-4328-9151-f4c95e1405f2-ssh-key\") pod \"ovn-edpm-deployment-openstack-edpm-ipam-ql7sj\" (UID: \"9b2b641e-b123-4328-9151-f4c95e1405f2\") " pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-ql7sj" Dec 10 23:23:02 crc kubenswrapper[4791]: I1210 23:23:02.914800 4791 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-pdpcv\" (UniqueName: \"kubernetes.io/projected/9b2b641e-b123-4328-9151-f4c95e1405f2-kube-api-access-pdpcv\") pod \"ovn-edpm-deployment-openstack-edpm-ipam-ql7sj\" (UID: \"9b2b641e-b123-4328-9151-f4c95e1405f2\") " pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-ql7sj" Dec 10 23:23:02 crc kubenswrapper[4791]: I1210 23:23:02.967235 4791 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-ql7sj" Dec 10 23:23:03 crc kubenswrapper[4791]: I1210 23:23:03.588288 4791 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovn-edpm-deployment-openstack-edpm-ipam-ql7sj"] Dec 10 23:23:04 crc kubenswrapper[4791]: I1210 23:23:04.453412 4791 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-ql7sj" event={"ID":"9b2b641e-b123-4328-9151-f4c95e1405f2","Type":"ContainerStarted","Data":"6051a42190b5d9d20259d1faa467e7f3bf33a036ed772457ddede4392cec087c"} Dec 10 23:23:05 crc kubenswrapper[4791]: I1210 23:23:05.468559 4791 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-ql7sj" event={"ID":"9b2b641e-b123-4328-9151-f4c95e1405f2","Type":"ContainerStarted","Data":"3d942f55b5707ccaa90f1c98a9171f76b7a1310b5104c5990208622a534de978"} Dec 10 23:23:05 crc kubenswrapper[4791]: I1210 23:23:05.495885 4791 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-ql7sj" podStartSLOduration=2.60493997 podStartE2EDuration="3.495852639s" podCreationTimestamp="2025-12-10 23:23:02 +0000 UTC" firstStartedPulling="2025-12-10 23:23:03.58617761 +0000 UTC m=+2018.015795263" lastFinishedPulling="2025-12-10 23:23:04.477090279 +0000 UTC m=+2018.906707932" observedRunningTime="2025-12-10 23:23:05.490216566 +0000 UTC m=+2019.919834239" watchObservedRunningTime="2025-12-10 23:23:05.495852639 +0000 UTC m=+2019.925470312" Dec 10 23:23:25 crc kubenswrapper[4791]: I1210 23:23:25.041013 4791 patch_prober.go:28] interesting pod/machine-config-daemon-5rb5l container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 10 23:23:25 crc kubenswrapper[4791]: I1210 23:23:25.041552 4791 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-5rb5l" podUID="ba35653c-6e06-4cee-a4d6-137764090d18" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 10 23:23:55 crc kubenswrapper[4791]: I1210 23:23:55.038911 4791 patch_prober.go:28] interesting pod/machine-config-daemon-5rb5l container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 10 23:23:55 crc kubenswrapper[4791]: I1210 23:23:55.039842 4791 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-5rb5l" podUID="ba35653c-6e06-4cee-a4d6-137764090d18" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 10 23:23:55 crc kubenswrapper[4791]: I1210 23:23:55.039930 4791 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-5rb5l" Dec 10 23:23:55 crc kubenswrapper[4791]: I1210 23:23:55.041132 4791 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"deb0e97b02d82d70f9f1de2c3e0794a3b407a099d09fc6ee2a1d821d42a42e21"} pod="openshift-machine-config-operator/machine-config-daemon-5rb5l" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Dec 10 23:23:55 crc kubenswrapper[4791]: I1210 23:23:55.041256 4791 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-5rb5l" podUID="ba35653c-6e06-4cee-a4d6-137764090d18" containerName="machine-config-daemon" containerID="cri-o://deb0e97b02d82d70f9f1de2c3e0794a3b407a099d09fc6ee2a1d821d42a42e21" gracePeriod=600 Dec 10 23:23:56 crc kubenswrapper[4791]: I1210 23:23:56.143614 4791 generic.go:334] "Generic (PLEG): container finished" podID="ba35653c-6e06-4cee-a4d6-137764090d18" containerID="deb0e97b02d82d70f9f1de2c3e0794a3b407a099d09fc6ee2a1d821d42a42e21" exitCode=0 Dec 10 23:23:56 crc kubenswrapper[4791]: I1210 23:23:56.143719 4791 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-5rb5l" event={"ID":"ba35653c-6e06-4cee-a4d6-137764090d18","Type":"ContainerDied","Data":"deb0e97b02d82d70f9f1de2c3e0794a3b407a099d09fc6ee2a1d821d42a42e21"} Dec 10 23:23:56 crc kubenswrapper[4791]: I1210 23:23:56.144418 4791 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-5rb5l" event={"ID":"ba35653c-6e06-4cee-a4d6-137764090d18","Type":"ContainerStarted","Data":"4229e7e5318c8150a4b76b463eed8dfb67fd267a558642295fc4d4f697411536"} Dec 10 23:23:56 crc kubenswrapper[4791]: I1210 23:23:56.144482 4791 scope.go:117] "RemoveContainer" containerID="f12779a795c9f6f63a93027d53c67c89d0179d87b9df554c25bfc607f06caff1" Dec 10 23:24:08 crc kubenswrapper[4791]: I1210 23:24:08.277963 4791 generic.go:334] "Generic (PLEG): container finished" podID="9b2b641e-b123-4328-9151-f4c95e1405f2" containerID="3d942f55b5707ccaa90f1c98a9171f76b7a1310b5104c5990208622a534de978" exitCode=0 Dec 10 23:24:08 crc kubenswrapper[4791]: I1210 23:24:08.278098 4791 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-ql7sj" event={"ID":"9b2b641e-b123-4328-9151-f4c95e1405f2","Type":"ContainerDied","Data":"3d942f55b5707ccaa90f1c98a9171f76b7a1310b5104c5990208622a534de978"} Dec 10 23:24:09 crc kubenswrapper[4791]: I1210 23:24:09.780303 4791 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-ql7sj" Dec 10 23:24:09 crc kubenswrapper[4791]: I1210 23:24:09.870519 4791 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovncontroller-config-0\" (UniqueName: \"kubernetes.io/configmap/9b2b641e-b123-4328-9151-f4c95e1405f2-ovncontroller-config-0\") pod \"9b2b641e-b123-4328-9151-f4c95e1405f2\" (UID: \"9b2b641e-b123-4328-9151-f4c95e1405f2\") " Dec 10 23:24:09 crc kubenswrapper[4791]: I1210 23:24:09.870594 4791 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/9b2b641e-b123-4328-9151-f4c95e1405f2-inventory\") pod \"9b2b641e-b123-4328-9151-f4c95e1405f2\" (UID: \"9b2b641e-b123-4328-9151-f4c95e1405f2\") " Dec 10 23:24:09 crc kubenswrapper[4791]: I1210 23:24:09.870712 4791 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovn-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/9b2b641e-b123-4328-9151-f4c95e1405f2-ovn-combined-ca-bundle\") pod \"9b2b641e-b123-4328-9151-f4c95e1405f2\" (UID: \"9b2b641e-b123-4328-9151-f4c95e1405f2\") " Dec 10 23:24:09 crc kubenswrapper[4791]: I1210 23:24:09.870791 4791 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-pdpcv\" (UniqueName: \"kubernetes.io/projected/9b2b641e-b123-4328-9151-f4c95e1405f2-kube-api-access-pdpcv\") pod \"9b2b641e-b123-4328-9151-f4c95e1405f2\" (UID: \"9b2b641e-b123-4328-9151-f4c95e1405f2\") " Dec 10 23:24:09 crc kubenswrapper[4791]: I1210 23:24:09.870879 4791 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/9b2b641e-b123-4328-9151-f4c95e1405f2-ssh-key\") pod \"9b2b641e-b123-4328-9151-f4c95e1405f2\" (UID: \"9b2b641e-b123-4328-9151-f4c95e1405f2\") " Dec 10 23:24:09 crc kubenswrapper[4791]: I1210 23:24:09.886609 4791 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/9b2b641e-b123-4328-9151-f4c95e1405f2-kube-api-access-pdpcv" (OuterVolumeSpecName: "kube-api-access-pdpcv") pod "9b2b641e-b123-4328-9151-f4c95e1405f2" (UID: "9b2b641e-b123-4328-9151-f4c95e1405f2"). InnerVolumeSpecName "kube-api-access-pdpcv". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 10 23:24:09 crc kubenswrapper[4791]: I1210 23:24:09.900828 4791 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/9b2b641e-b123-4328-9151-f4c95e1405f2-ovn-combined-ca-bundle" (OuterVolumeSpecName: "ovn-combined-ca-bundle") pod "9b2b641e-b123-4328-9151-f4c95e1405f2" (UID: "9b2b641e-b123-4328-9151-f4c95e1405f2"). InnerVolumeSpecName "ovn-combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 10 23:24:09 crc kubenswrapper[4791]: I1210 23:24:09.934534 4791 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/9b2b641e-b123-4328-9151-f4c95e1405f2-ssh-key" (OuterVolumeSpecName: "ssh-key") pod "9b2b641e-b123-4328-9151-f4c95e1405f2" (UID: "9b2b641e-b123-4328-9151-f4c95e1405f2"). InnerVolumeSpecName "ssh-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 10 23:24:09 crc kubenswrapper[4791]: I1210 23:24:09.943390 4791 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/9b2b641e-b123-4328-9151-f4c95e1405f2-inventory" (OuterVolumeSpecName: "inventory") pod "9b2b641e-b123-4328-9151-f4c95e1405f2" (UID: "9b2b641e-b123-4328-9151-f4c95e1405f2"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 10 23:24:09 crc kubenswrapper[4791]: I1210 23:24:09.962076 4791 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/9b2b641e-b123-4328-9151-f4c95e1405f2-ovncontroller-config-0" (OuterVolumeSpecName: "ovncontroller-config-0") pod "9b2b641e-b123-4328-9151-f4c95e1405f2" (UID: "9b2b641e-b123-4328-9151-f4c95e1405f2"). InnerVolumeSpecName "ovncontroller-config-0". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 10 23:24:09 crc kubenswrapper[4791]: I1210 23:24:09.974581 4791 reconciler_common.go:293] "Volume detached for volume \"ovn-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/9b2b641e-b123-4328-9151-f4c95e1405f2-ovn-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 10 23:24:09 crc kubenswrapper[4791]: I1210 23:24:09.974623 4791 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-pdpcv\" (UniqueName: \"kubernetes.io/projected/9b2b641e-b123-4328-9151-f4c95e1405f2-kube-api-access-pdpcv\") on node \"crc\" DevicePath \"\"" Dec 10 23:24:09 crc kubenswrapper[4791]: I1210 23:24:09.974634 4791 reconciler_common.go:293] "Volume detached for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/9b2b641e-b123-4328-9151-f4c95e1405f2-ssh-key\") on node \"crc\" DevicePath \"\"" Dec 10 23:24:09 crc kubenswrapper[4791]: I1210 23:24:09.974645 4791 reconciler_common.go:293] "Volume detached for volume \"ovncontroller-config-0\" (UniqueName: \"kubernetes.io/configmap/9b2b641e-b123-4328-9151-f4c95e1405f2-ovncontroller-config-0\") on node \"crc\" DevicePath \"\"" Dec 10 23:24:09 crc kubenswrapper[4791]: I1210 23:24:09.974657 4791 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/9b2b641e-b123-4328-9151-f4c95e1405f2-inventory\") on node \"crc\" DevicePath \"\"" Dec 10 23:24:10 crc kubenswrapper[4791]: I1210 23:24:10.342015 4791 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-ql7sj" event={"ID":"9b2b641e-b123-4328-9151-f4c95e1405f2","Type":"ContainerDied","Data":"6051a42190b5d9d20259d1faa467e7f3bf33a036ed772457ddede4392cec087c"} Dec 10 23:24:10 crc kubenswrapper[4791]: I1210 23:24:10.342106 4791 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="6051a42190b5d9d20259d1faa467e7f3bf33a036ed772457ddede4392cec087c" Dec 10 23:24:10 crc kubenswrapper[4791]: I1210 23:24:10.342116 4791 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-ql7sj" Dec 10 23:24:10 crc kubenswrapper[4791]: I1210 23:24:10.442053 4791 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-nnggl"] Dec 10 23:24:10 crc kubenswrapper[4791]: E1210 23:24:10.442880 4791 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="9b2b641e-b123-4328-9151-f4c95e1405f2" containerName="ovn-edpm-deployment-openstack-edpm-ipam" Dec 10 23:24:10 crc kubenswrapper[4791]: I1210 23:24:10.443439 4791 state_mem.go:107] "Deleted CPUSet assignment" podUID="9b2b641e-b123-4328-9151-f4c95e1405f2" containerName="ovn-edpm-deployment-openstack-edpm-ipam" Dec 10 23:24:10 crc kubenswrapper[4791]: I1210 23:24:10.443731 4791 memory_manager.go:354] "RemoveStaleState removing state" podUID="9b2b641e-b123-4328-9151-f4c95e1405f2" containerName="ovn-edpm-deployment-openstack-edpm-ipam" Dec 10 23:24:10 crc kubenswrapper[4791]: I1210 23:24:10.445146 4791 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-nnggl" Dec 10 23:24:10 crc kubenswrapper[4791]: I1210 23:24:10.449218 4791 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-dockercfg-hf8pr" Dec 10 23:24:10 crc kubenswrapper[4791]: I1210 23:24:10.449479 4791 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Dec 10 23:24:10 crc kubenswrapper[4791]: I1210 23:24:10.449718 4791 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-ansible-ssh-private-key-secret" Dec 10 23:24:10 crc kubenswrapper[4791]: I1210 23:24:10.449881 4791 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-edpm-ipam" Dec 10 23:24:10 crc kubenswrapper[4791]: I1210 23:24:10.450026 4791 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-metadata-neutron-config" Dec 10 23:24:10 crc kubenswrapper[4791]: I1210 23:24:10.450185 4791 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"neutron-ovn-metadata-agent-neutron-config" Dec 10 23:24:10 crc kubenswrapper[4791]: I1210 23:24:10.469022 4791 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-nnggl"] Dec 10 23:24:10 crc kubenswrapper[4791]: I1210 23:24:10.589401 4791 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"neutron-metadata-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/9b204a02-650a-4f54-b257-5791c8399bf6-neutron-metadata-combined-ca-bundle\") pod \"neutron-metadata-edpm-deployment-openstack-edpm-ipam-nnggl\" (UID: \"9b204a02-650a-4f54-b257-5791c8399bf6\") " pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-nnggl" Dec 10 23:24:10 crc kubenswrapper[4791]: I1210 23:24:10.589758 4791 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"nova-metadata-neutron-config-0\" (UniqueName: \"kubernetes.io/secret/9b204a02-650a-4f54-b257-5791c8399bf6-nova-metadata-neutron-config-0\") pod \"neutron-metadata-edpm-deployment-openstack-edpm-ipam-nnggl\" (UID: \"9b204a02-650a-4f54-b257-5791c8399bf6\") " pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-nnggl" Dec 10 23:24:10 crc kubenswrapper[4791]: I1210 23:24:10.589913 4791 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/9b204a02-650a-4f54-b257-5791c8399bf6-ssh-key\") pod \"neutron-metadata-edpm-deployment-openstack-edpm-ipam-nnggl\" (UID: \"9b204a02-650a-4f54-b257-5791c8399bf6\") " pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-nnggl" Dec 10 23:24:10 crc kubenswrapper[4791]: I1210 23:24:10.590025 4791 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/9b204a02-650a-4f54-b257-5791c8399bf6-inventory\") pod \"neutron-metadata-edpm-deployment-openstack-edpm-ipam-nnggl\" (UID: \"9b204a02-650a-4f54-b257-5791c8399bf6\") " pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-nnggl" Dec 10 23:24:10 crc kubenswrapper[4791]: I1210 23:24:10.590184 4791 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-q5bc6\" (UniqueName: \"kubernetes.io/projected/9b204a02-650a-4f54-b257-5791c8399bf6-kube-api-access-q5bc6\") pod \"neutron-metadata-edpm-deployment-openstack-edpm-ipam-nnggl\" (UID: \"9b204a02-650a-4f54-b257-5791c8399bf6\") " pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-nnggl" Dec 10 23:24:10 crc kubenswrapper[4791]: I1210 23:24:10.590327 4791 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"neutron-ovn-metadata-agent-neutron-config-0\" (UniqueName: \"kubernetes.io/secret/9b204a02-650a-4f54-b257-5791c8399bf6-neutron-ovn-metadata-agent-neutron-config-0\") pod \"neutron-metadata-edpm-deployment-openstack-edpm-ipam-nnggl\" (UID: \"9b204a02-650a-4f54-b257-5791c8399bf6\") " pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-nnggl" Dec 10 23:24:10 crc kubenswrapper[4791]: I1210 23:24:10.692655 4791 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"neutron-metadata-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/9b204a02-650a-4f54-b257-5791c8399bf6-neutron-metadata-combined-ca-bundle\") pod \"neutron-metadata-edpm-deployment-openstack-edpm-ipam-nnggl\" (UID: \"9b204a02-650a-4f54-b257-5791c8399bf6\") " pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-nnggl" Dec 10 23:24:10 crc kubenswrapper[4791]: I1210 23:24:10.693143 4791 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nova-metadata-neutron-config-0\" (UniqueName: \"kubernetes.io/secret/9b204a02-650a-4f54-b257-5791c8399bf6-nova-metadata-neutron-config-0\") pod \"neutron-metadata-edpm-deployment-openstack-edpm-ipam-nnggl\" (UID: \"9b204a02-650a-4f54-b257-5791c8399bf6\") " pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-nnggl" Dec 10 23:24:10 crc kubenswrapper[4791]: I1210 23:24:10.693359 4791 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/9b204a02-650a-4f54-b257-5791c8399bf6-ssh-key\") pod \"neutron-metadata-edpm-deployment-openstack-edpm-ipam-nnggl\" (UID: \"9b204a02-650a-4f54-b257-5791c8399bf6\") " pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-nnggl" Dec 10 23:24:10 crc kubenswrapper[4791]: I1210 23:24:10.693523 4791 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/9b204a02-650a-4f54-b257-5791c8399bf6-inventory\") pod \"neutron-metadata-edpm-deployment-openstack-edpm-ipam-nnggl\" (UID: \"9b204a02-650a-4f54-b257-5791c8399bf6\") " pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-nnggl" Dec 10 23:24:10 crc kubenswrapper[4791]: I1210 23:24:10.693747 4791 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-q5bc6\" (UniqueName: \"kubernetes.io/projected/9b204a02-650a-4f54-b257-5791c8399bf6-kube-api-access-q5bc6\") pod \"neutron-metadata-edpm-deployment-openstack-edpm-ipam-nnggl\" (UID: \"9b204a02-650a-4f54-b257-5791c8399bf6\") " pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-nnggl" Dec 10 23:24:10 crc kubenswrapper[4791]: I1210 23:24:10.694285 4791 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"neutron-ovn-metadata-agent-neutron-config-0\" (UniqueName: \"kubernetes.io/secret/9b204a02-650a-4f54-b257-5791c8399bf6-neutron-ovn-metadata-agent-neutron-config-0\") pod \"neutron-metadata-edpm-deployment-openstack-edpm-ipam-nnggl\" (UID: \"9b204a02-650a-4f54-b257-5791c8399bf6\") " pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-nnggl" Dec 10 23:24:10 crc kubenswrapper[4791]: I1210 23:24:10.698858 4791 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"neutron-metadata-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/9b204a02-650a-4f54-b257-5791c8399bf6-neutron-metadata-combined-ca-bundle\") pod \"neutron-metadata-edpm-deployment-openstack-edpm-ipam-nnggl\" (UID: \"9b204a02-650a-4f54-b257-5791c8399bf6\") " pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-nnggl" Dec 10 23:24:10 crc kubenswrapper[4791]: I1210 23:24:10.699731 4791 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"nova-metadata-neutron-config-0\" (UniqueName: \"kubernetes.io/secret/9b204a02-650a-4f54-b257-5791c8399bf6-nova-metadata-neutron-config-0\") pod \"neutron-metadata-edpm-deployment-openstack-edpm-ipam-nnggl\" (UID: \"9b204a02-650a-4f54-b257-5791c8399bf6\") " pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-nnggl" Dec 10 23:24:10 crc kubenswrapper[4791]: I1210 23:24:10.699745 4791 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/9b204a02-650a-4f54-b257-5791c8399bf6-inventory\") pod \"neutron-metadata-edpm-deployment-openstack-edpm-ipam-nnggl\" (UID: \"9b204a02-650a-4f54-b257-5791c8399bf6\") " pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-nnggl" Dec 10 23:24:10 crc kubenswrapper[4791]: I1210 23:24:10.699951 4791 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/9b204a02-650a-4f54-b257-5791c8399bf6-ssh-key\") pod \"neutron-metadata-edpm-deployment-openstack-edpm-ipam-nnggl\" (UID: \"9b204a02-650a-4f54-b257-5791c8399bf6\") " pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-nnggl" Dec 10 23:24:10 crc kubenswrapper[4791]: I1210 23:24:10.700925 4791 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"neutron-ovn-metadata-agent-neutron-config-0\" (UniqueName: \"kubernetes.io/secret/9b204a02-650a-4f54-b257-5791c8399bf6-neutron-ovn-metadata-agent-neutron-config-0\") pod \"neutron-metadata-edpm-deployment-openstack-edpm-ipam-nnggl\" (UID: \"9b204a02-650a-4f54-b257-5791c8399bf6\") " pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-nnggl" Dec 10 23:24:10 crc kubenswrapper[4791]: I1210 23:24:10.719876 4791 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-q5bc6\" (UniqueName: \"kubernetes.io/projected/9b204a02-650a-4f54-b257-5791c8399bf6-kube-api-access-q5bc6\") pod \"neutron-metadata-edpm-deployment-openstack-edpm-ipam-nnggl\" (UID: \"9b204a02-650a-4f54-b257-5791c8399bf6\") " pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-nnggl" Dec 10 23:24:10 crc kubenswrapper[4791]: I1210 23:24:10.771194 4791 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-nnggl" Dec 10 23:24:11 crc kubenswrapper[4791]: I1210 23:24:11.340248 4791 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-nnggl"] Dec 10 23:24:12 crc kubenswrapper[4791]: I1210 23:24:12.361216 4791 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-nnggl" event={"ID":"9b204a02-650a-4f54-b257-5791c8399bf6","Type":"ContainerStarted","Data":"f84b68df9e737fcd3fd830632d44d22ac22783390613983e284430da58146819"} Dec 10 23:24:12 crc kubenswrapper[4791]: I1210 23:24:12.361619 4791 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-nnggl" event={"ID":"9b204a02-650a-4f54-b257-5791c8399bf6","Type":"ContainerStarted","Data":"0c686bf5830cc4f52673759b5738f8499c48bbed1a9d1b3e84f9abc070775a56"} Dec 10 23:24:12 crc kubenswrapper[4791]: I1210 23:24:12.392884 4791 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-nnggl" podStartSLOduration=1.715159999 podStartE2EDuration="2.392839012s" podCreationTimestamp="2025-12-10 23:24:10 +0000 UTC" firstStartedPulling="2025-12-10 23:24:11.346225509 +0000 UTC m=+2085.775843142" lastFinishedPulling="2025-12-10 23:24:12.023904542 +0000 UTC m=+2086.453522155" observedRunningTime="2025-12-10 23:24:12.381061762 +0000 UTC m=+2086.810679405" watchObservedRunningTime="2025-12-10 23:24:12.392839012 +0000 UTC m=+2086.822456635" Dec 10 23:24:41 crc kubenswrapper[4791]: I1210 23:24:41.540898 4791 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-marketplace-x5zr4"] Dec 10 23:24:41 crc kubenswrapper[4791]: I1210 23:24:41.544755 4791 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-x5zr4" Dec 10 23:24:41 crc kubenswrapper[4791]: I1210 23:24:41.567077 4791 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-x5zr4"] Dec 10 23:24:41 crc kubenswrapper[4791]: I1210 23:24:41.641824 4791 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/25e796fc-d6b1-48e3-af3b-e79924d347ba-utilities\") pod \"redhat-marketplace-x5zr4\" (UID: \"25e796fc-d6b1-48e3-af3b-e79924d347ba\") " pod="openshift-marketplace/redhat-marketplace-x5zr4" Dec 10 23:24:41 crc kubenswrapper[4791]: I1210 23:24:41.641906 4791 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/25e796fc-d6b1-48e3-af3b-e79924d347ba-catalog-content\") pod \"redhat-marketplace-x5zr4\" (UID: \"25e796fc-d6b1-48e3-af3b-e79924d347ba\") " pod="openshift-marketplace/redhat-marketplace-x5zr4" Dec 10 23:24:41 crc kubenswrapper[4791]: I1210 23:24:41.641994 4791 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-nzrsz\" (UniqueName: \"kubernetes.io/projected/25e796fc-d6b1-48e3-af3b-e79924d347ba-kube-api-access-nzrsz\") pod \"redhat-marketplace-x5zr4\" (UID: \"25e796fc-d6b1-48e3-af3b-e79924d347ba\") " pod="openshift-marketplace/redhat-marketplace-x5zr4" Dec 10 23:24:41 crc kubenswrapper[4791]: I1210 23:24:41.743658 4791 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/25e796fc-d6b1-48e3-af3b-e79924d347ba-catalog-content\") pod \"redhat-marketplace-x5zr4\" (UID: \"25e796fc-d6b1-48e3-af3b-e79924d347ba\") " pod="openshift-marketplace/redhat-marketplace-x5zr4" Dec 10 23:24:41 crc kubenswrapper[4791]: I1210 23:24:41.743712 4791 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-nzrsz\" (UniqueName: \"kubernetes.io/projected/25e796fc-d6b1-48e3-af3b-e79924d347ba-kube-api-access-nzrsz\") pod \"redhat-marketplace-x5zr4\" (UID: \"25e796fc-d6b1-48e3-af3b-e79924d347ba\") " pod="openshift-marketplace/redhat-marketplace-x5zr4" Dec 10 23:24:41 crc kubenswrapper[4791]: I1210 23:24:41.744574 4791 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/25e796fc-d6b1-48e3-af3b-e79924d347ba-catalog-content\") pod \"redhat-marketplace-x5zr4\" (UID: \"25e796fc-d6b1-48e3-af3b-e79924d347ba\") " pod="openshift-marketplace/redhat-marketplace-x5zr4" Dec 10 23:24:41 crc kubenswrapper[4791]: I1210 23:24:41.745131 4791 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/25e796fc-d6b1-48e3-af3b-e79924d347ba-utilities\") pod \"redhat-marketplace-x5zr4\" (UID: \"25e796fc-d6b1-48e3-af3b-e79924d347ba\") " pod="openshift-marketplace/redhat-marketplace-x5zr4" Dec 10 23:24:41 crc kubenswrapper[4791]: I1210 23:24:41.798474 4791 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/25e796fc-d6b1-48e3-af3b-e79924d347ba-utilities\") pod \"redhat-marketplace-x5zr4\" (UID: \"25e796fc-d6b1-48e3-af3b-e79924d347ba\") " pod="openshift-marketplace/redhat-marketplace-x5zr4" Dec 10 23:24:41 crc kubenswrapper[4791]: I1210 23:24:41.830672 4791 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-nzrsz\" (UniqueName: \"kubernetes.io/projected/25e796fc-d6b1-48e3-af3b-e79924d347ba-kube-api-access-nzrsz\") pod \"redhat-marketplace-x5zr4\" (UID: \"25e796fc-d6b1-48e3-af3b-e79924d347ba\") " pod="openshift-marketplace/redhat-marketplace-x5zr4" Dec 10 23:24:41 crc kubenswrapper[4791]: I1210 23:24:41.868215 4791 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-x5zr4" Dec 10 23:24:42 crc kubenswrapper[4791]: I1210 23:24:42.329890 4791 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-x5zr4"] Dec 10 23:24:42 crc kubenswrapper[4791]: I1210 23:24:42.736035 4791 generic.go:334] "Generic (PLEG): container finished" podID="25e796fc-d6b1-48e3-af3b-e79924d347ba" containerID="2cf77d8de69b59ff8329e95022f7b4af1e6f2be58acbd006b36abe435f225a48" exitCode=0 Dec 10 23:24:42 crc kubenswrapper[4791]: I1210 23:24:42.736119 4791 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-x5zr4" event={"ID":"25e796fc-d6b1-48e3-af3b-e79924d347ba","Type":"ContainerDied","Data":"2cf77d8de69b59ff8329e95022f7b4af1e6f2be58acbd006b36abe435f225a48"} Dec 10 23:24:42 crc kubenswrapper[4791]: I1210 23:24:42.736642 4791 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-x5zr4" event={"ID":"25e796fc-d6b1-48e3-af3b-e79924d347ba","Type":"ContainerStarted","Data":"b62e9e3851d593c84327379bc2c00a909e8a5bea7c24079c9c1aa40211ba3107"} Dec 10 23:24:43 crc kubenswrapper[4791]: I1210 23:24:43.751219 4791 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-x5zr4" event={"ID":"25e796fc-d6b1-48e3-af3b-e79924d347ba","Type":"ContainerStarted","Data":"f98df9a7aecebf7ce7ebdfddc719bad5de4696329078fb826f508f37a2f813ae"} Dec 10 23:24:44 crc kubenswrapper[4791]: I1210 23:24:44.766780 4791 generic.go:334] "Generic (PLEG): container finished" podID="25e796fc-d6b1-48e3-af3b-e79924d347ba" containerID="f98df9a7aecebf7ce7ebdfddc719bad5de4696329078fb826f508f37a2f813ae" exitCode=0 Dec 10 23:24:44 crc kubenswrapper[4791]: I1210 23:24:44.767001 4791 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-x5zr4" event={"ID":"25e796fc-d6b1-48e3-af3b-e79924d347ba","Type":"ContainerDied","Data":"f98df9a7aecebf7ce7ebdfddc719bad5de4696329078fb826f508f37a2f813ae"} Dec 10 23:24:47 crc kubenswrapper[4791]: I1210 23:24:47.401508 4791 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-x5zr4" event={"ID":"25e796fc-d6b1-48e3-af3b-e79924d347ba","Type":"ContainerStarted","Data":"92f00c531fc8fe0200f0b49e2e6c3b3982350310b9222e8a60490e6f144ee4f9"} Dec 10 23:24:47 crc kubenswrapper[4791]: I1210 23:24:47.428963 4791 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-marketplace-x5zr4" podStartSLOduration=2.838284167 podStartE2EDuration="6.42891946s" podCreationTimestamp="2025-12-10 23:24:41 +0000 UTC" firstStartedPulling="2025-12-10 23:24:42.73920238 +0000 UTC m=+2117.168820023" lastFinishedPulling="2025-12-10 23:24:46.329837693 +0000 UTC m=+2120.759455316" observedRunningTime="2025-12-10 23:24:47.425480881 +0000 UTC m=+2121.855098494" watchObservedRunningTime="2025-12-10 23:24:47.42891946 +0000 UTC m=+2121.858537083" Dec 10 23:24:51 crc kubenswrapper[4791]: I1210 23:24:51.868617 4791 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-marketplace-x5zr4" Dec 10 23:24:51 crc kubenswrapper[4791]: I1210 23:24:51.869238 4791 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-marketplace-x5zr4" Dec 10 23:24:51 crc kubenswrapper[4791]: I1210 23:24:51.921096 4791 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-marketplace-x5zr4" Dec 10 23:24:52 crc kubenswrapper[4791]: I1210 23:24:52.499484 4791 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-marketplace-x5zr4" Dec 10 23:24:52 crc kubenswrapper[4791]: I1210 23:24:52.548755 4791 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-x5zr4"] Dec 10 23:24:54 crc kubenswrapper[4791]: I1210 23:24:54.474763 4791 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-marketplace-x5zr4" podUID="25e796fc-d6b1-48e3-af3b-e79924d347ba" containerName="registry-server" containerID="cri-o://92f00c531fc8fe0200f0b49e2e6c3b3982350310b9222e8a60490e6f144ee4f9" gracePeriod=2 Dec 10 23:24:54 crc kubenswrapper[4791]: I1210 23:24:54.949465 4791 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-x5zr4" Dec 10 23:24:55 crc kubenswrapper[4791]: I1210 23:24:55.063411 4791 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/25e796fc-d6b1-48e3-af3b-e79924d347ba-catalog-content\") pod \"25e796fc-d6b1-48e3-af3b-e79924d347ba\" (UID: \"25e796fc-d6b1-48e3-af3b-e79924d347ba\") " Dec 10 23:24:55 crc kubenswrapper[4791]: I1210 23:24:55.063965 4791 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-nzrsz\" (UniqueName: \"kubernetes.io/projected/25e796fc-d6b1-48e3-af3b-e79924d347ba-kube-api-access-nzrsz\") pod \"25e796fc-d6b1-48e3-af3b-e79924d347ba\" (UID: \"25e796fc-d6b1-48e3-af3b-e79924d347ba\") " Dec 10 23:24:55 crc kubenswrapper[4791]: I1210 23:24:55.064122 4791 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/25e796fc-d6b1-48e3-af3b-e79924d347ba-utilities\") pod \"25e796fc-d6b1-48e3-af3b-e79924d347ba\" (UID: \"25e796fc-d6b1-48e3-af3b-e79924d347ba\") " Dec 10 23:24:55 crc kubenswrapper[4791]: I1210 23:24:55.065285 4791 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/25e796fc-d6b1-48e3-af3b-e79924d347ba-utilities" (OuterVolumeSpecName: "utilities") pod "25e796fc-d6b1-48e3-af3b-e79924d347ba" (UID: "25e796fc-d6b1-48e3-af3b-e79924d347ba"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 10 23:24:55 crc kubenswrapper[4791]: I1210 23:24:55.075410 4791 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/25e796fc-d6b1-48e3-af3b-e79924d347ba-kube-api-access-nzrsz" (OuterVolumeSpecName: "kube-api-access-nzrsz") pod "25e796fc-d6b1-48e3-af3b-e79924d347ba" (UID: "25e796fc-d6b1-48e3-af3b-e79924d347ba"). InnerVolumeSpecName "kube-api-access-nzrsz". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 10 23:24:55 crc kubenswrapper[4791]: I1210 23:24:55.084134 4791 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/25e796fc-d6b1-48e3-af3b-e79924d347ba-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "25e796fc-d6b1-48e3-af3b-e79924d347ba" (UID: "25e796fc-d6b1-48e3-af3b-e79924d347ba"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 10 23:24:55 crc kubenswrapper[4791]: I1210 23:24:55.167497 4791 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-nzrsz\" (UniqueName: \"kubernetes.io/projected/25e796fc-d6b1-48e3-af3b-e79924d347ba-kube-api-access-nzrsz\") on node \"crc\" DevicePath \"\"" Dec 10 23:24:55 crc kubenswrapper[4791]: I1210 23:24:55.167573 4791 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/25e796fc-d6b1-48e3-af3b-e79924d347ba-utilities\") on node \"crc\" DevicePath \"\"" Dec 10 23:24:55 crc kubenswrapper[4791]: I1210 23:24:55.167599 4791 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/25e796fc-d6b1-48e3-af3b-e79924d347ba-catalog-content\") on node \"crc\" DevicePath \"\"" Dec 10 23:24:55 crc kubenswrapper[4791]: I1210 23:24:55.486887 4791 generic.go:334] "Generic (PLEG): container finished" podID="25e796fc-d6b1-48e3-af3b-e79924d347ba" containerID="92f00c531fc8fe0200f0b49e2e6c3b3982350310b9222e8a60490e6f144ee4f9" exitCode=0 Dec 10 23:24:55 crc kubenswrapper[4791]: I1210 23:24:55.486937 4791 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-x5zr4" event={"ID":"25e796fc-d6b1-48e3-af3b-e79924d347ba","Type":"ContainerDied","Data":"92f00c531fc8fe0200f0b49e2e6c3b3982350310b9222e8a60490e6f144ee4f9"} Dec 10 23:24:55 crc kubenswrapper[4791]: I1210 23:24:55.486963 4791 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-x5zr4" event={"ID":"25e796fc-d6b1-48e3-af3b-e79924d347ba","Type":"ContainerDied","Data":"b62e9e3851d593c84327379bc2c00a909e8a5bea7c24079c9c1aa40211ba3107"} Dec 10 23:24:55 crc kubenswrapper[4791]: I1210 23:24:55.486965 4791 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-x5zr4" Dec 10 23:24:55 crc kubenswrapper[4791]: I1210 23:24:55.486980 4791 scope.go:117] "RemoveContainer" containerID="92f00c531fc8fe0200f0b49e2e6c3b3982350310b9222e8a60490e6f144ee4f9" Dec 10 23:24:55 crc kubenswrapper[4791]: I1210 23:24:55.538046 4791 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-x5zr4"] Dec 10 23:24:55 crc kubenswrapper[4791]: I1210 23:24:55.549188 4791 scope.go:117] "RemoveContainer" containerID="f98df9a7aecebf7ce7ebdfddc719bad5de4696329078fb826f508f37a2f813ae" Dec 10 23:24:55 crc kubenswrapper[4791]: I1210 23:24:55.553335 4791 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-marketplace-x5zr4"] Dec 10 23:24:55 crc kubenswrapper[4791]: I1210 23:24:55.573222 4791 scope.go:117] "RemoveContainer" containerID="2cf77d8de69b59ff8329e95022f7b4af1e6f2be58acbd006b36abe435f225a48" Dec 10 23:24:55 crc kubenswrapper[4791]: I1210 23:24:55.619920 4791 scope.go:117] "RemoveContainer" containerID="92f00c531fc8fe0200f0b49e2e6c3b3982350310b9222e8a60490e6f144ee4f9" Dec 10 23:24:55 crc kubenswrapper[4791]: E1210 23:24:55.620460 4791 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"92f00c531fc8fe0200f0b49e2e6c3b3982350310b9222e8a60490e6f144ee4f9\": container with ID starting with 92f00c531fc8fe0200f0b49e2e6c3b3982350310b9222e8a60490e6f144ee4f9 not found: ID does not exist" containerID="92f00c531fc8fe0200f0b49e2e6c3b3982350310b9222e8a60490e6f144ee4f9" Dec 10 23:24:55 crc kubenswrapper[4791]: I1210 23:24:55.620494 4791 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"92f00c531fc8fe0200f0b49e2e6c3b3982350310b9222e8a60490e6f144ee4f9"} err="failed to get container status \"92f00c531fc8fe0200f0b49e2e6c3b3982350310b9222e8a60490e6f144ee4f9\": rpc error: code = NotFound desc = could not find container \"92f00c531fc8fe0200f0b49e2e6c3b3982350310b9222e8a60490e6f144ee4f9\": container with ID starting with 92f00c531fc8fe0200f0b49e2e6c3b3982350310b9222e8a60490e6f144ee4f9 not found: ID does not exist" Dec 10 23:24:55 crc kubenswrapper[4791]: I1210 23:24:55.620514 4791 scope.go:117] "RemoveContainer" containerID="f98df9a7aecebf7ce7ebdfddc719bad5de4696329078fb826f508f37a2f813ae" Dec 10 23:24:55 crc kubenswrapper[4791]: E1210 23:24:55.620914 4791 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"f98df9a7aecebf7ce7ebdfddc719bad5de4696329078fb826f508f37a2f813ae\": container with ID starting with f98df9a7aecebf7ce7ebdfddc719bad5de4696329078fb826f508f37a2f813ae not found: ID does not exist" containerID="f98df9a7aecebf7ce7ebdfddc719bad5de4696329078fb826f508f37a2f813ae" Dec 10 23:24:55 crc kubenswrapper[4791]: I1210 23:24:55.621005 4791 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"f98df9a7aecebf7ce7ebdfddc719bad5de4696329078fb826f508f37a2f813ae"} err="failed to get container status \"f98df9a7aecebf7ce7ebdfddc719bad5de4696329078fb826f508f37a2f813ae\": rpc error: code = NotFound desc = could not find container \"f98df9a7aecebf7ce7ebdfddc719bad5de4696329078fb826f508f37a2f813ae\": container with ID starting with f98df9a7aecebf7ce7ebdfddc719bad5de4696329078fb826f508f37a2f813ae not found: ID does not exist" Dec 10 23:24:55 crc kubenswrapper[4791]: I1210 23:24:55.621049 4791 scope.go:117] "RemoveContainer" containerID="2cf77d8de69b59ff8329e95022f7b4af1e6f2be58acbd006b36abe435f225a48" Dec 10 23:24:55 crc kubenswrapper[4791]: E1210 23:24:55.621615 4791 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"2cf77d8de69b59ff8329e95022f7b4af1e6f2be58acbd006b36abe435f225a48\": container with ID starting with 2cf77d8de69b59ff8329e95022f7b4af1e6f2be58acbd006b36abe435f225a48 not found: ID does not exist" containerID="2cf77d8de69b59ff8329e95022f7b4af1e6f2be58acbd006b36abe435f225a48" Dec 10 23:24:55 crc kubenswrapper[4791]: I1210 23:24:55.621655 4791 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"2cf77d8de69b59ff8329e95022f7b4af1e6f2be58acbd006b36abe435f225a48"} err="failed to get container status \"2cf77d8de69b59ff8329e95022f7b4af1e6f2be58acbd006b36abe435f225a48\": rpc error: code = NotFound desc = could not find container \"2cf77d8de69b59ff8329e95022f7b4af1e6f2be58acbd006b36abe435f225a48\": container with ID starting with 2cf77d8de69b59ff8329e95022f7b4af1e6f2be58acbd006b36abe435f225a48 not found: ID does not exist" Dec 10 23:24:55 crc kubenswrapper[4791]: I1210 23:24:55.897395 4791 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="25e796fc-d6b1-48e3-af3b-e79924d347ba" path="/var/lib/kubelet/pods/25e796fc-d6b1-48e3-af3b-e79924d347ba/volumes" Dec 10 23:24:59 crc kubenswrapper[4791]: I1210 23:24:59.556844 4791 generic.go:334] "Generic (PLEG): container finished" podID="9b204a02-650a-4f54-b257-5791c8399bf6" containerID="f84b68df9e737fcd3fd830632d44d22ac22783390613983e284430da58146819" exitCode=0 Dec 10 23:24:59 crc kubenswrapper[4791]: I1210 23:24:59.556922 4791 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-nnggl" event={"ID":"9b204a02-650a-4f54-b257-5791c8399bf6","Type":"ContainerDied","Data":"f84b68df9e737fcd3fd830632d44d22ac22783390613983e284430da58146819"} Dec 10 23:25:01 crc kubenswrapper[4791]: I1210 23:25:01.080806 4791 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-nnggl" Dec 10 23:25:01 crc kubenswrapper[4791]: I1210 23:25:01.103730 4791 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/9b204a02-650a-4f54-b257-5791c8399bf6-ssh-key\") pod \"9b204a02-650a-4f54-b257-5791c8399bf6\" (UID: \"9b204a02-650a-4f54-b257-5791c8399bf6\") " Dec 10 23:25:01 crc kubenswrapper[4791]: I1210 23:25:01.103778 4791 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"neutron-ovn-metadata-agent-neutron-config-0\" (UniqueName: \"kubernetes.io/secret/9b204a02-650a-4f54-b257-5791c8399bf6-neutron-ovn-metadata-agent-neutron-config-0\") pod \"9b204a02-650a-4f54-b257-5791c8399bf6\" (UID: \"9b204a02-650a-4f54-b257-5791c8399bf6\") " Dec 10 23:25:01 crc kubenswrapper[4791]: I1210 23:25:01.103823 4791 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/9b204a02-650a-4f54-b257-5791c8399bf6-inventory\") pod \"9b204a02-650a-4f54-b257-5791c8399bf6\" (UID: \"9b204a02-650a-4f54-b257-5791c8399bf6\") " Dec 10 23:25:01 crc kubenswrapper[4791]: I1210 23:25:01.103993 4791 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"nova-metadata-neutron-config-0\" (UniqueName: \"kubernetes.io/secret/9b204a02-650a-4f54-b257-5791c8399bf6-nova-metadata-neutron-config-0\") pod \"9b204a02-650a-4f54-b257-5791c8399bf6\" (UID: \"9b204a02-650a-4f54-b257-5791c8399bf6\") " Dec 10 23:25:01 crc kubenswrapper[4791]: I1210 23:25:01.104024 4791 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-q5bc6\" (UniqueName: \"kubernetes.io/projected/9b204a02-650a-4f54-b257-5791c8399bf6-kube-api-access-q5bc6\") pod \"9b204a02-650a-4f54-b257-5791c8399bf6\" (UID: \"9b204a02-650a-4f54-b257-5791c8399bf6\") " Dec 10 23:25:01 crc kubenswrapper[4791]: I1210 23:25:01.104092 4791 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"neutron-metadata-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/9b204a02-650a-4f54-b257-5791c8399bf6-neutron-metadata-combined-ca-bundle\") pod \"9b204a02-650a-4f54-b257-5791c8399bf6\" (UID: \"9b204a02-650a-4f54-b257-5791c8399bf6\") " Dec 10 23:25:01 crc kubenswrapper[4791]: I1210 23:25:01.110622 4791 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/9b204a02-650a-4f54-b257-5791c8399bf6-kube-api-access-q5bc6" (OuterVolumeSpecName: "kube-api-access-q5bc6") pod "9b204a02-650a-4f54-b257-5791c8399bf6" (UID: "9b204a02-650a-4f54-b257-5791c8399bf6"). InnerVolumeSpecName "kube-api-access-q5bc6". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 10 23:25:01 crc kubenswrapper[4791]: I1210 23:25:01.116573 4791 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/9b204a02-650a-4f54-b257-5791c8399bf6-neutron-metadata-combined-ca-bundle" (OuterVolumeSpecName: "neutron-metadata-combined-ca-bundle") pod "9b204a02-650a-4f54-b257-5791c8399bf6" (UID: "9b204a02-650a-4f54-b257-5791c8399bf6"). InnerVolumeSpecName "neutron-metadata-combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 10 23:25:01 crc kubenswrapper[4791]: I1210 23:25:01.131975 4791 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/9b204a02-650a-4f54-b257-5791c8399bf6-inventory" (OuterVolumeSpecName: "inventory") pod "9b204a02-650a-4f54-b257-5791c8399bf6" (UID: "9b204a02-650a-4f54-b257-5791c8399bf6"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 10 23:25:01 crc kubenswrapper[4791]: I1210 23:25:01.140005 4791 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/9b204a02-650a-4f54-b257-5791c8399bf6-nova-metadata-neutron-config-0" (OuterVolumeSpecName: "nova-metadata-neutron-config-0") pod "9b204a02-650a-4f54-b257-5791c8399bf6" (UID: "9b204a02-650a-4f54-b257-5791c8399bf6"). InnerVolumeSpecName "nova-metadata-neutron-config-0". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 10 23:25:01 crc kubenswrapper[4791]: I1210 23:25:01.157318 4791 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/9b204a02-650a-4f54-b257-5791c8399bf6-ssh-key" (OuterVolumeSpecName: "ssh-key") pod "9b204a02-650a-4f54-b257-5791c8399bf6" (UID: "9b204a02-650a-4f54-b257-5791c8399bf6"). InnerVolumeSpecName "ssh-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 10 23:25:01 crc kubenswrapper[4791]: I1210 23:25:01.171858 4791 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/9b204a02-650a-4f54-b257-5791c8399bf6-neutron-ovn-metadata-agent-neutron-config-0" (OuterVolumeSpecName: "neutron-ovn-metadata-agent-neutron-config-0") pod "9b204a02-650a-4f54-b257-5791c8399bf6" (UID: "9b204a02-650a-4f54-b257-5791c8399bf6"). InnerVolumeSpecName "neutron-ovn-metadata-agent-neutron-config-0". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 10 23:25:01 crc kubenswrapper[4791]: I1210 23:25:01.206561 4791 reconciler_common.go:293] "Volume detached for volume \"nova-metadata-neutron-config-0\" (UniqueName: \"kubernetes.io/secret/9b204a02-650a-4f54-b257-5791c8399bf6-nova-metadata-neutron-config-0\") on node \"crc\" DevicePath \"\"" Dec 10 23:25:01 crc kubenswrapper[4791]: I1210 23:25:01.206620 4791 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-q5bc6\" (UniqueName: \"kubernetes.io/projected/9b204a02-650a-4f54-b257-5791c8399bf6-kube-api-access-q5bc6\") on node \"crc\" DevicePath \"\"" Dec 10 23:25:01 crc kubenswrapper[4791]: I1210 23:25:01.206634 4791 reconciler_common.go:293] "Volume detached for volume \"neutron-metadata-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/9b204a02-650a-4f54-b257-5791c8399bf6-neutron-metadata-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 10 23:25:01 crc kubenswrapper[4791]: I1210 23:25:01.206653 4791 reconciler_common.go:293] "Volume detached for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/9b204a02-650a-4f54-b257-5791c8399bf6-ssh-key\") on node \"crc\" DevicePath \"\"" Dec 10 23:25:01 crc kubenswrapper[4791]: I1210 23:25:01.206669 4791 reconciler_common.go:293] "Volume detached for volume \"neutron-ovn-metadata-agent-neutron-config-0\" (UniqueName: \"kubernetes.io/secret/9b204a02-650a-4f54-b257-5791c8399bf6-neutron-ovn-metadata-agent-neutron-config-0\") on node \"crc\" DevicePath \"\"" Dec 10 23:25:01 crc kubenswrapper[4791]: I1210 23:25:01.206683 4791 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/9b204a02-650a-4f54-b257-5791c8399bf6-inventory\") on node \"crc\" DevicePath \"\"" Dec 10 23:25:01 crc kubenswrapper[4791]: I1210 23:25:01.578221 4791 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-nnggl" event={"ID":"9b204a02-650a-4f54-b257-5791c8399bf6","Type":"ContainerDied","Data":"0c686bf5830cc4f52673759b5738f8499c48bbed1a9d1b3e84f9abc070775a56"} Dec 10 23:25:01 crc kubenswrapper[4791]: I1210 23:25:01.578607 4791 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="0c686bf5830cc4f52673759b5738f8499c48bbed1a9d1b3e84f9abc070775a56" Dec 10 23:25:01 crc kubenswrapper[4791]: I1210 23:25:01.578322 4791 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-nnggl" Dec 10 23:25:01 crc kubenswrapper[4791]: I1210 23:25:01.706566 4791 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/libvirt-edpm-deployment-openstack-edpm-ipam-9czng"] Dec 10 23:25:01 crc kubenswrapper[4791]: E1210 23:25:01.707056 4791 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="9b204a02-650a-4f54-b257-5791c8399bf6" containerName="neutron-metadata-edpm-deployment-openstack-edpm-ipam" Dec 10 23:25:01 crc kubenswrapper[4791]: I1210 23:25:01.707090 4791 state_mem.go:107] "Deleted CPUSet assignment" podUID="9b204a02-650a-4f54-b257-5791c8399bf6" containerName="neutron-metadata-edpm-deployment-openstack-edpm-ipam" Dec 10 23:25:01 crc kubenswrapper[4791]: E1210 23:25:01.707120 4791 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="25e796fc-d6b1-48e3-af3b-e79924d347ba" containerName="extract-content" Dec 10 23:25:01 crc kubenswrapper[4791]: I1210 23:25:01.707139 4791 state_mem.go:107] "Deleted CPUSet assignment" podUID="25e796fc-d6b1-48e3-af3b-e79924d347ba" containerName="extract-content" Dec 10 23:25:01 crc kubenswrapper[4791]: E1210 23:25:01.707162 4791 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="25e796fc-d6b1-48e3-af3b-e79924d347ba" containerName="extract-utilities" Dec 10 23:25:01 crc kubenswrapper[4791]: I1210 23:25:01.707170 4791 state_mem.go:107] "Deleted CPUSet assignment" podUID="25e796fc-d6b1-48e3-af3b-e79924d347ba" containerName="extract-utilities" Dec 10 23:25:01 crc kubenswrapper[4791]: E1210 23:25:01.707206 4791 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="25e796fc-d6b1-48e3-af3b-e79924d347ba" containerName="registry-server" Dec 10 23:25:01 crc kubenswrapper[4791]: I1210 23:25:01.707214 4791 state_mem.go:107] "Deleted CPUSet assignment" podUID="25e796fc-d6b1-48e3-af3b-e79924d347ba" containerName="registry-server" Dec 10 23:25:01 crc kubenswrapper[4791]: I1210 23:25:01.707447 4791 memory_manager.go:354] "RemoveStaleState removing state" podUID="9b204a02-650a-4f54-b257-5791c8399bf6" containerName="neutron-metadata-edpm-deployment-openstack-edpm-ipam" Dec 10 23:25:01 crc kubenswrapper[4791]: I1210 23:25:01.707484 4791 memory_manager.go:354] "RemoveStaleState removing state" podUID="25e796fc-d6b1-48e3-af3b-e79924d347ba" containerName="registry-server" Dec 10 23:25:01 crc kubenswrapper[4791]: I1210 23:25:01.708298 4791 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-9czng" Dec 10 23:25:01 crc kubenswrapper[4791]: I1210 23:25:01.710707 4791 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"libvirt-secret" Dec 10 23:25:01 crc kubenswrapper[4791]: I1210 23:25:01.710960 4791 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-edpm-ipam" Dec 10 23:25:01 crc kubenswrapper[4791]: I1210 23:25:01.711632 4791 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-ansible-ssh-private-key-secret" Dec 10 23:25:01 crc kubenswrapper[4791]: I1210 23:25:01.711899 4791 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-dockercfg-hf8pr" Dec 10 23:25:01 crc kubenswrapper[4791]: I1210 23:25:01.713271 4791 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Dec 10 23:25:01 crc kubenswrapper[4791]: I1210 23:25:01.714725 4791 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"libvirt-secret-0\" (UniqueName: \"kubernetes.io/secret/b180a2cd-d91a-4203-b0f8-7de5c1be8226-libvirt-secret-0\") pod \"libvirt-edpm-deployment-openstack-edpm-ipam-9czng\" (UID: \"b180a2cd-d91a-4203-b0f8-7de5c1be8226\") " pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-9czng" Dec 10 23:25:01 crc kubenswrapper[4791]: I1210 23:25:01.714791 4791 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-7s6z9\" (UniqueName: \"kubernetes.io/projected/b180a2cd-d91a-4203-b0f8-7de5c1be8226-kube-api-access-7s6z9\") pod \"libvirt-edpm-deployment-openstack-edpm-ipam-9czng\" (UID: \"b180a2cd-d91a-4203-b0f8-7de5c1be8226\") " pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-9czng" Dec 10 23:25:01 crc kubenswrapper[4791]: I1210 23:25:01.714867 4791 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/b180a2cd-d91a-4203-b0f8-7de5c1be8226-inventory\") pod \"libvirt-edpm-deployment-openstack-edpm-ipam-9czng\" (UID: \"b180a2cd-d91a-4203-b0f8-7de5c1be8226\") " pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-9czng" Dec 10 23:25:01 crc kubenswrapper[4791]: I1210 23:25:01.715037 4791 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/b180a2cd-d91a-4203-b0f8-7de5c1be8226-ssh-key\") pod \"libvirt-edpm-deployment-openstack-edpm-ipam-9czng\" (UID: \"b180a2cd-d91a-4203-b0f8-7de5c1be8226\") " pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-9czng" Dec 10 23:25:01 crc kubenswrapper[4791]: I1210 23:25:01.715272 4791 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"libvirt-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/b180a2cd-d91a-4203-b0f8-7de5c1be8226-libvirt-combined-ca-bundle\") pod \"libvirt-edpm-deployment-openstack-edpm-ipam-9czng\" (UID: \"b180a2cd-d91a-4203-b0f8-7de5c1be8226\") " pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-9czng" Dec 10 23:25:01 crc kubenswrapper[4791]: I1210 23:25:01.740160 4791 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/libvirt-edpm-deployment-openstack-edpm-ipam-9czng"] Dec 10 23:25:01 crc kubenswrapper[4791]: I1210 23:25:01.817174 4791 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"libvirt-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/b180a2cd-d91a-4203-b0f8-7de5c1be8226-libvirt-combined-ca-bundle\") pod \"libvirt-edpm-deployment-openstack-edpm-ipam-9czng\" (UID: \"b180a2cd-d91a-4203-b0f8-7de5c1be8226\") " pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-9czng" Dec 10 23:25:01 crc kubenswrapper[4791]: I1210 23:25:01.817325 4791 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"libvirt-secret-0\" (UniqueName: \"kubernetes.io/secret/b180a2cd-d91a-4203-b0f8-7de5c1be8226-libvirt-secret-0\") pod \"libvirt-edpm-deployment-openstack-edpm-ipam-9czng\" (UID: \"b180a2cd-d91a-4203-b0f8-7de5c1be8226\") " pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-9czng" Dec 10 23:25:01 crc kubenswrapper[4791]: I1210 23:25:01.817380 4791 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-7s6z9\" (UniqueName: \"kubernetes.io/projected/b180a2cd-d91a-4203-b0f8-7de5c1be8226-kube-api-access-7s6z9\") pod \"libvirt-edpm-deployment-openstack-edpm-ipam-9czng\" (UID: \"b180a2cd-d91a-4203-b0f8-7de5c1be8226\") " pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-9czng" Dec 10 23:25:01 crc kubenswrapper[4791]: I1210 23:25:01.817411 4791 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/b180a2cd-d91a-4203-b0f8-7de5c1be8226-inventory\") pod \"libvirt-edpm-deployment-openstack-edpm-ipam-9czng\" (UID: \"b180a2cd-d91a-4203-b0f8-7de5c1be8226\") " pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-9czng" Dec 10 23:25:01 crc kubenswrapper[4791]: I1210 23:25:01.817460 4791 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/b180a2cd-d91a-4203-b0f8-7de5c1be8226-ssh-key\") pod \"libvirt-edpm-deployment-openstack-edpm-ipam-9czng\" (UID: \"b180a2cd-d91a-4203-b0f8-7de5c1be8226\") " pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-9czng" Dec 10 23:25:01 crc kubenswrapper[4791]: I1210 23:25:01.822312 4791 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/b180a2cd-d91a-4203-b0f8-7de5c1be8226-inventory\") pod \"libvirt-edpm-deployment-openstack-edpm-ipam-9czng\" (UID: \"b180a2cd-d91a-4203-b0f8-7de5c1be8226\") " pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-9czng" Dec 10 23:25:01 crc kubenswrapper[4791]: I1210 23:25:01.822753 4791 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"libvirt-secret-0\" (UniqueName: \"kubernetes.io/secret/b180a2cd-d91a-4203-b0f8-7de5c1be8226-libvirt-secret-0\") pod \"libvirt-edpm-deployment-openstack-edpm-ipam-9czng\" (UID: \"b180a2cd-d91a-4203-b0f8-7de5c1be8226\") " pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-9czng" Dec 10 23:25:01 crc kubenswrapper[4791]: I1210 23:25:01.823599 4791 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/b180a2cd-d91a-4203-b0f8-7de5c1be8226-ssh-key\") pod \"libvirt-edpm-deployment-openstack-edpm-ipam-9czng\" (UID: \"b180a2cd-d91a-4203-b0f8-7de5c1be8226\") " pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-9czng" Dec 10 23:25:01 crc kubenswrapper[4791]: I1210 23:25:01.824913 4791 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"libvirt-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/b180a2cd-d91a-4203-b0f8-7de5c1be8226-libvirt-combined-ca-bundle\") pod \"libvirt-edpm-deployment-openstack-edpm-ipam-9czng\" (UID: \"b180a2cd-d91a-4203-b0f8-7de5c1be8226\") " pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-9czng" Dec 10 23:25:01 crc kubenswrapper[4791]: I1210 23:25:01.838794 4791 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-7s6z9\" (UniqueName: \"kubernetes.io/projected/b180a2cd-d91a-4203-b0f8-7de5c1be8226-kube-api-access-7s6z9\") pod \"libvirt-edpm-deployment-openstack-edpm-ipam-9czng\" (UID: \"b180a2cd-d91a-4203-b0f8-7de5c1be8226\") " pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-9czng" Dec 10 23:25:02 crc kubenswrapper[4791]: I1210 23:25:02.037216 4791 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-9czng" Dec 10 23:25:02 crc kubenswrapper[4791]: I1210 23:25:02.613298 4791 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/libvirt-edpm-deployment-openstack-edpm-ipam-9czng"] Dec 10 23:25:03 crc kubenswrapper[4791]: I1210 23:25:03.611420 4791 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-9czng" event={"ID":"b180a2cd-d91a-4203-b0f8-7de5c1be8226","Type":"ContainerStarted","Data":"ebcb2c70e3397a8dc344f89ac539376137604b6194e602f99b4c58365f6f9a82"} Dec 10 23:25:03 crc kubenswrapper[4791]: I1210 23:25:03.611723 4791 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-9czng" event={"ID":"b180a2cd-d91a-4203-b0f8-7de5c1be8226","Type":"ContainerStarted","Data":"e8d74c85bbb3d3753dd9d4e130cdbfb9cc99ac769edf9f393c59ac4704be8e23"} Dec 10 23:25:03 crc kubenswrapper[4791]: I1210 23:25:03.637954 4791 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-9czng" podStartSLOduration=1.9405800050000002 podStartE2EDuration="2.637918425s" podCreationTimestamp="2025-12-10 23:25:01 +0000 UTC" firstStartedPulling="2025-12-10 23:25:02.60129094 +0000 UTC m=+2137.030908573" lastFinishedPulling="2025-12-10 23:25:03.29862938 +0000 UTC m=+2137.728246993" observedRunningTime="2025-12-10 23:25:03.630216073 +0000 UTC m=+2138.059833726" watchObservedRunningTime="2025-12-10 23:25:03.637918425 +0000 UTC m=+2138.067536078" Dec 10 23:25:55 crc kubenswrapper[4791]: I1210 23:25:55.039031 4791 patch_prober.go:28] interesting pod/machine-config-daemon-5rb5l container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 10 23:25:55 crc kubenswrapper[4791]: I1210 23:25:55.039921 4791 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-5rb5l" podUID="ba35653c-6e06-4cee-a4d6-137764090d18" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 10 23:26:25 crc kubenswrapper[4791]: I1210 23:26:25.038525 4791 patch_prober.go:28] interesting pod/machine-config-daemon-5rb5l container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 10 23:26:25 crc kubenswrapper[4791]: I1210 23:26:25.039073 4791 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-5rb5l" podUID="ba35653c-6e06-4cee-a4d6-137764090d18" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 10 23:26:55 crc kubenswrapper[4791]: I1210 23:26:55.037658 4791 patch_prober.go:28] interesting pod/machine-config-daemon-5rb5l container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 10 23:26:55 crc kubenswrapper[4791]: I1210 23:26:55.038203 4791 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-5rb5l" podUID="ba35653c-6e06-4cee-a4d6-137764090d18" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 10 23:26:55 crc kubenswrapper[4791]: I1210 23:26:55.038256 4791 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-5rb5l" Dec 10 23:26:55 crc kubenswrapper[4791]: I1210 23:26:55.039064 4791 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"4229e7e5318c8150a4b76b463eed8dfb67fd267a558642295fc4d4f697411536"} pod="openshift-machine-config-operator/machine-config-daemon-5rb5l" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Dec 10 23:26:55 crc kubenswrapper[4791]: I1210 23:26:55.039119 4791 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-5rb5l" podUID="ba35653c-6e06-4cee-a4d6-137764090d18" containerName="machine-config-daemon" containerID="cri-o://4229e7e5318c8150a4b76b463eed8dfb67fd267a558642295fc4d4f697411536" gracePeriod=600 Dec 10 23:26:55 crc kubenswrapper[4791]: E1210 23:26:55.168187 4791 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-5rb5l_openshift-machine-config-operator(ba35653c-6e06-4cee-a4d6-137764090d18)\"" pod="openshift-machine-config-operator/machine-config-daemon-5rb5l" podUID="ba35653c-6e06-4cee-a4d6-137764090d18" Dec 10 23:26:55 crc kubenswrapper[4791]: I1210 23:26:55.946443 4791 generic.go:334] "Generic (PLEG): container finished" podID="ba35653c-6e06-4cee-a4d6-137764090d18" containerID="4229e7e5318c8150a4b76b463eed8dfb67fd267a558642295fc4d4f697411536" exitCode=0 Dec 10 23:26:55 crc kubenswrapper[4791]: I1210 23:26:55.946539 4791 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-5rb5l" event={"ID":"ba35653c-6e06-4cee-a4d6-137764090d18","Type":"ContainerDied","Data":"4229e7e5318c8150a4b76b463eed8dfb67fd267a558642295fc4d4f697411536"} Dec 10 23:26:55 crc kubenswrapper[4791]: I1210 23:26:55.946688 4791 scope.go:117] "RemoveContainer" containerID="deb0e97b02d82d70f9f1de2c3e0794a3b407a099d09fc6ee2a1d821d42a42e21" Dec 10 23:26:55 crc kubenswrapper[4791]: I1210 23:26:55.955358 4791 scope.go:117] "RemoveContainer" containerID="4229e7e5318c8150a4b76b463eed8dfb67fd267a558642295fc4d4f697411536" Dec 10 23:26:55 crc kubenswrapper[4791]: E1210 23:26:55.956008 4791 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-5rb5l_openshift-machine-config-operator(ba35653c-6e06-4cee-a4d6-137764090d18)\"" pod="openshift-machine-config-operator/machine-config-daemon-5rb5l" podUID="ba35653c-6e06-4cee-a4d6-137764090d18" Dec 10 23:27:08 crc kubenswrapper[4791]: I1210 23:27:08.885714 4791 scope.go:117] "RemoveContainer" containerID="4229e7e5318c8150a4b76b463eed8dfb67fd267a558642295fc4d4f697411536" Dec 10 23:27:08 crc kubenswrapper[4791]: E1210 23:27:08.886647 4791 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-5rb5l_openshift-machine-config-operator(ba35653c-6e06-4cee-a4d6-137764090d18)\"" pod="openshift-machine-config-operator/machine-config-daemon-5rb5l" podUID="ba35653c-6e06-4cee-a4d6-137764090d18" Dec 10 23:27:23 crc kubenswrapper[4791]: I1210 23:27:23.885406 4791 scope.go:117] "RemoveContainer" containerID="4229e7e5318c8150a4b76b463eed8dfb67fd267a558642295fc4d4f697411536" Dec 10 23:27:23 crc kubenswrapper[4791]: E1210 23:27:23.888079 4791 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-5rb5l_openshift-machine-config-operator(ba35653c-6e06-4cee-a4d6-137764090d18)\"" pod="openshift-machine-config-operator/machine-config-daemon-5rb5l" podUID="ba35653c-6e06-4cee-a4d6-137764090d18" Dec 10 23:27:37 crc kubenswrapper[4791]: I1210 23:27:37.885871 4791 scope.go:117] "RemoveContainer" containerID="4229e7e5318c8150a4b76b463eed8dfb67fd267a558642295fc4d4f697411536" Dec 10 23:27:37 crc kubenswrapper[4791]: E1210 23:27:37.887249 4791 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-5rb5l_openshift-machine-config-operator(ba35653c-6e06-4cee-a4d6-137764090d18)\"" pod="openshift-machine-config-operator/machine-config-daemon-5rb5l" podUID="ba35653c-6e06-4cee-a4d6-137764090d18" Dec 10 23:27:48 crc kubenswrapper[4791]: I1210 23:27:48.885678 4791 scope.go:117] "RemoveContainer" containerID="4229e7e5318c8150a4b76b463eed8dfb67fd267a558642295fc4d4f697411536" Dec 10 23:27:48 crc kubenswrapper[4791]: E1210 23:27:48.886755 4791 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-5rb5l_openshift-machine-config-operator(ba35653c-6e06-4cee-a4d6-137764090d18)\"" pod="openshift-machine-config-operator/machine-config-daemon-5rb5l" podUID="ba35653c-6e06-4cee-a4d6-137764090d18" Dec 10 23:28:00 crc kubenswrapper[4791]: I1210 23:28:00.885493 4791 scope.go:117] "RemoveContainer" containerID="4229e7e5318c8150a4b76b463eed8dfb67fd267a558642295fc4d4f697411536" Dec 10 23:28:00 crc kubenswrapper[4791]: E1210 23:28:00.886356 4791 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-5rb5l_openshift-machine-config-operator(ba35653c-6e06-4cee-a4d6-137764090d18)\"" pod="openshift-machine-config-operator/machine-config-daemon-5rb5l" podUID="ba35653c-6e06-4cee-a4d6-137764090d18" Dec 10 23:28:11 crc kubenswrapper[4791]: I1210 23:28:11.885590 4791 scope.go:117] "RemoveContainer" containerID="4229e7e5318c8150a4b76b463eed8dfb67fd267a558642295fc4d4f697411536" Dec 10 23:28:11 crc kubenswrapper[4791]: E1210 23:28:11.886379 4791 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-5rb5l_openshift-machine-config-operator(ba35653c-6e06-4cee-a4d6-137764090d18)\"" pod="openshift-machine-config-operator/machine-config-daemon-5rb5l" podUID="ba35653c-6e06-4cee-a4d6-137764090d18" Dec 10 23:28:25 crc kubenswrapper[4791]: I1210 23:28:25.897791 4791 scope.go:117] "RemoveContainer" containerID="4229e7e5318c8150a4b76b463eed8dfb67fd267a558642295fc4d4f697411536" Dec 10 23:28:25 crc kubenswrapper[4791]: E1210 23:28:25.898732 4791 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-5rb5l_openshift-machine-config-operator(ba35653c-6e06-4cee-a4d6-137764090d18)\"" pod="openshift-machine-config-operator/machine-config-daemon-5rb5l" podUID="ba35653c-6e06-4cee-a4d6-137764090d18" Dec 10 23:28:40 crc kubenswrapper[4791]: I1210 23:28:40.885578 4791 scope.go:117] "RemoveContainer" containerID="4229e7e5318c8150a4b76b463eed8dfb67fd267a558642295fc4d4f697411536" Dec 10 23:28:40 crc kubenswrapper[4791]: E1210 23:28:40.887882 4791 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-5rb5l_openshift-machine-config-operator(ba35653c-6e06-4cee-a4d6-137764090d18)\"" pod="openshift-machine-config-operator/machine-config-daemon-5rb5l" podUID="ba35653c-6e06-4cee-a4d6-137764090d18" Dec 10 23:28:51 crc kubenswrapper[4791]: I1210 23:28:51.884764 4791 scope.go:117] "RemoveContainer" containerID="4229e7e5318c8150a4b76b463eed8dfb67fd267a558642295fc4d4f697411536" Dec 10 23:28:51 crc kubenswrapper[4791]: E1210 23:28:51.886968 4791 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-5rb5l_openshift-machine-config-operator(ba35653c-6e06-4cee-a4d6-137764090d18)\"" pod="openshift-machine-config-operator/machine-config-daemon-5rb5l" podUID="ba35653c-6e06-4cee-a4d6-137764090d18" Dec 10 23:29:02 crc kubenswrapper[4791]: I1210 23:29:02.885531 4791 scope.go:117] "RemoveContainer" containerID="4229e7e5318c8150a4b76b463eed8dfb67fd267a558642295fc4d4f697411536" Dec 10 23:29:02 crc kubenswrapper[4791]: E1210 23:29:02.886645 4791 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-5rb5l_openshift-machine-config-operator(ba35653c-6e06-4cee-a4d6-137764090d18)\"" pod="openshift-machine-config-operator/machine-config-daemon-5rb5l" podUID="ba35653c-6e06-4cee-a4d6-137764090d18" Dec 10 23:29:04 crc kubenswrapper[4791]: I1210 23:29:04.529753 4791 generic.go:334] "Generic (PLEG): container finished" podID="b180a2cd-d91a-4203-b0f8-7de5c1be8226" containerID="ebcb2c70e3397a8dc344f89ac539376137604b6194e602f99b4c58365f6f9a82" exitCode=0 Dec 10 23:29:04 crc kubenswrapper[4791]: I1210 23:29:04.529839 4791 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-9czng" event={"ID":"b180a2cd-d91a-4203-b0f8-7de5c1be8226","Type":"ContainerDied","Data":"ebcb2c70e3397a8dc344f89ac539376137604b6194e602f99b4c58365f6f9a82"} Dec 10 23:29:06 crc kubenswrapper[4791]: I1210 23:29:06.048052 4791 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-9czng" Dec 10 23:29:06 crc kubenswrapper[4791]: I1210 23:29:06.129249 4791 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"libvirt-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/b180a2cd-d91a-4203-b0f8-7de5c1be8226-libvirt-combined-ca-bundle\") pod \"b180a2cd-d91a-4203-b0f8-7de5c1be8226\" (UID: \"b180a2cd-d91a-4203-b0f8-7de5c1be8226\") " Dec 10 23:29:06 crc kubenswrapper[4791]: I1210 23:29:06.129502 4791 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"libvirt-secret-0\" (UniqueName: \"kubernetes.io/secret/b180a2cd-d91a-4203-b0f8-7de5c1be8226-libvirt-secret-0\") pod \"b180a2cd-d91a-4203-b0f8-7de5c1be8226\" (UID: \"b180a2cd-d91a-4203-b0f8-7de5c1be8226\") " Dec 10 23:29:06 crc kubenswrapper[4791]: I1210 23:29:06.129541 4791 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/b180a2cd-d91a-4203-b0f8-7de5c1be8226-inventory\") pod \"b180a2cd-d91a-4203-b0f8-7de5c1be8226\" (UID: \"b180a2cd-d91a-4203-b0f8-7de5c1be8226\") " Dec 10 23:29:06 crc kubenswrapper[4791]: I1210 23:29:06.129596 4791 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/b180a2cd-d91a-4203-b0f8-7de5c1be8226-ssh-key\") pod \"b180a2cd-d91a-4203-b0f8-7de5c1be8226\" (UID: \"b180a2cd-d91a-4203-b0f8-7de5c1be8226\") " Dec 10 23:29:06 crc kubenswrapper[4791]: I1210 23:29:06.129639 4791 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-7s6z9\" (UniqueName: \"kubernetes.io/projected/b180a2cd-d91a-4203-b0f8-7de5c1be8226-kube-api-access-7s6z9\") pod \"b180a2cd-d91a-4203-b0f8-7de5c1be8226\" (UID: \"b180a2cd-d91a-4203-b0f8-7de5c1be8226\") " Dec 10 23:29:06 crc kubenswrapper[4791]: I1210 23:29:06.138557 4791 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/b180a2cd-d91a-4203-b0f8-7de5c1be8226-libvirt-combined-ca-bundle" (OuterVolumeSpecName: "libvirt-combined-ca-bundle") pod "b180a2cd-d91a-4203-b0f8-7de5c1be8226" (UID: "b180a2cd-d91a-4203-b0f8-7de5c1be8226"). InnerVolumeSpecName "libvirt-combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 10 23:29:06 crc kubenswrapper[4791]: I1210 23:29:06.138749 4791 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/b180a2cd-d91a-4203-b0f8-7de5c1be8226-kube-api-access-7s6z9" (OuterVolumeSpecName: "kube-api-access-7s6z9") pod "b180a2cd-d91a-4203-b0f8-7de5c1be8226" (UID: "b180a2cd-d91a-4203-b0f8-7de5c1be8226"). InnerVolumeSpecName "kube-api-access-7s6z9". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 10 23:29:06 crc kubenswrapper[4791]: I1210 23:29:06.158787 4791 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/b180a2cd-d91a-4203-b0f8-7de5c1be8226-libvirt-secret-0" (OuterVolumeSpecName: "libvirt-secret-0") pod "b180a2cd-d91a-4203-b0f8-7de5c1be8226" (UID: "b180a2cd-d91a-4203-b0f8-7de5c1be8226"). InnerVolumeSpecName "libvirt-secret-0". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 10 23:29:06 crc kubenswrapper[4791]: I1210 23:29:06.159616 4791 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/b180a2cd-d91a-4203-b0f8-7de5c1be8226-ssh-key" (OuterVolumeSpecName: "ssh-key") pod "b180a2cd-d91a-4203-b0f8-7de5c1be8226" (UID: "b180a2cd-d91a-4203-b0f8-7de5c1be8226"). InnerVolumeSpecName "ssh-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 10 23:29:06 crc kubenswrapper[4791]: I1210 23:29:06.169554 4791 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/b180a2cd-d91a-4203-b0f8-7de5c1be8226-inventory" (OuterVolumeSpecName: "inventory") pod "b180a2cd-d91a-4203-b0f8-7de5c1be8226" (UID: "b180a2cd-d91a-4203-b0f8-7de5c1be8226"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 10 23:29:06 crc kubenswrapper[4791]: I1210 23:29:06.231658 4791 reconciler_common.go:293] "Volume detached for volume \"libvirt-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/b180a2cd-d91a-4203-b0f8-7de5c1be8226-libvirt-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 10 23:29:06 crc kubenswrapper[4791]: I1210 23:29:06.231693 4791 reconciler_common.go:293] "Volume detached for volume \"libvirt-secret-0\" (UniqueName: \"kubernetes.io/secret/b180a2cd-d91a-4203-b0f8-7de5c1be8226-libvirt-secret-0\") on node \"crc\" DevicePath \"\"" Dec 10 23:29:06 crc kubenswrapper[4791]: I1210 23:29:06.231707 4791 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/b180a2cd-d91a-4203-b0f8-7de5c1be8226-inventory\") on node \"crc\" DevicePath \"\"" Dec 10 23:29:06 crc kubenswrapper[4791]: I1210 23:29:06.231719 4791 reconciler_common.go:293] "Volume detached for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/b180a2cd-d91a-4203-b0f8-7de5c1be8226-ssh-key\") on node \"crc\" DevicePath \"\"" Dec 10 23:29:06 crc kubenswrapper[4791]: I1210 23:29:06.231733 4791 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-7s6z9\" (UniqueName: \"kubernetes.io/projected/b180a2cd-d91a-4203-b0f8-7de5c1be8226-kube-api-access-7s6z9\") on node \"crc\" DevicePath \"\"" Dec 10 23:29:06 crc kubenswrapper[4791]: I1210 23:29:06.556614 4791 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-9czng" event={"ID":"b180a2cd-d91a-4203-b0f8-7de5c1be8226","Type":"ContainerDied","Data":"e8d74c85bbb3d3753dd9d4e130cdbfb9cc99ac769edf9f393c59ac4704be8e23"} Dec 10 23:29:06 crc kubenswrapper[4791]: I1210 23:29:06.556661 4791 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="e8d74c85bbb3d3753dd9d4e130cdbfb9cc99ac769edf9f393c59ac4704be8e23" Dec 10 23:29:06 crc kubenswrapper[4791]: I1210 23:29:06.556725 4791 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-9czng" Dec 10 23:29:06 crc kubenswrapper[4791]: I1210 23:29:06.707368 4791 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-edpm-deployment-openstack-edpm-ipam-5krlr"] Dec 10 23:29:06 crc kubenswrapper[4791]: E1210 23:29:06.708206 4791 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="b180a2cd-d91a-4203-b0f8-7de5c1be8226" containerName="libvirt-edpm-deployment-openstack-edpm-ipam" Dec 10 23:29:06 crc kubenswrapper[4791]: I1210 23:29:06.708253 4791 state_mem.go:107] "Deleted CPUSet assignment" podUID="b180a2cd-d91a-4203-b0f8-7de5c1be8226" containerName="libvirt-edpm-deployment-openstack-edpm-ipam" Dec 10 23:29:06 crc kubenswrapper[4791]: I1210 23:29:06.708788 4791 memory_manager.go:354] "RemoveStaleState removing state" podUID="b180a2cd-d91a-4203-b0f8-7de5c1be8226" containerName="libvirt-edpm-deployment-openstack-edpm-ipam" Dec 10 23:29:06 crc kubenswrapper[4791]: I1210 23:29:06.710242 4791 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-5krlr" Dec 10 23:29:06 crc kubenswrapper[4791]: I1210 23:29:06.719962 4791 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Dec 10 23:29:06 crc kubenswrapper[4791]: I1210 23:29:06.720161 4791 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"nova-extra-config" Dec 10 23:29:06 crc kubenswrapper[4791]: I1210 23:29:06.720395 4791 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-dockercfg-hf8pr" Dec 10 23:29:06 crc kubenswrapper[4791]: I1210 23:29:06.720567 4791 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell1-compute-config" Dec 10 23:29:06 crc kubenswrapper[4791]: I1210 23:29:06.720691 4791 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-migration-ssh-key" Dec 10 23:29:06 crc kubenswrapper[4791]: I1210 23:29:06.720899 4791 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-ansible-ssh-private-key-secret" Dec 10 23:29:06 crc kubenswrapper[4791]: I1210 23:29:06.721243 4791 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-edpm-ipam" Dec 10 23:29:06 crc kubenswrapper[4791]: I1210 23:29:06.735540 4791 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-edpm-deployment-openstack-edpm-ipam-5krlr"] Dec 10 23:29:06 crc kubenswrapper[4791]: I1210 23:29:06.844116 4791 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"nova-migration-ssh-key-0\" (UniqueName: \"kubernetes.io/secret/f9fc3f8e-9635-48a4-8ec3-20034a1bc7ad-nova-migration-ssh-key-0\") pod \"nova-edpm-deployment-openstack-edpm-ipam-5krlr\" (UID: \"f9fc3f8e-9635-48a4-8ec3-20034a1bc7ad\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-5krlr" Dec 10 23:29:06 crc kubenswrapper[4791]: I1210 23:29:06.844174 4791 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"nova-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f9fc3f8e-9635-48a4-8ec3-20034a1bc7ad-nova-combined-ca-bundle\") pod \"nova-edpm-deployment-openstack-edpm-ipam-5krlr\" (UID: \"f9fc3f8e-9635-48a4-8ec3-20034a1bc7ad\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-5krlr" Dec 10 23:29:06 crc kubenswrapper[4791]: I1210 23:29:06.844256 4791 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/f9fc3f8e-9635-48a4-8ec3-20034a1bc7ad-inventory\") pod \"nova-edpm-deployment-openstack-edpm-ipam-5krlr\" (UID: \"f9fc3f8e-9635-48a4-8ec3-20034a1bc7ad\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-5krlr" Dec 10 23:29:06 crc kubenswrapper[4791]: I1210 23:29:06.844295 4791 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"nova-extra-config-0\" (UniqueName: \"kubernetes.io/configmap/f9fc3f8e-9635-48a4-8ec3-20034a1bc7ad-nova-extra-config-0\") pod \"nova-edpm-deployment-openstack-edpm-ipam-5krlr\" (UID: \"f9fc3f8e-9635-48a4-8ec3-20034a1bc7ad\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-5krlr" Dec 10 23:29:06 crc kubenswrapper[4791]: I1210 23:29:06.844311 4791 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/f9fc3f8e-9635-48a4-8ec3-20034a1bc7ad-ssh-key\") pod \"nova-edpm-deployment-openstack-edpm-ipam-5krlr\" (UID: \"f9fc3f8e-9635-48a4-8ec3-20034a1bc7ad\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-5krlr" Dec 10 23:29:06 crc kubenswrapper[4791]: I1210 23:29:06.844335 4791 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-475wv\" (UniqueName: \"kubernetes.io/projected/f9fc3f8e-9635-48a4-8ec3-20034a1bc7ad-kube-api-access-475wv\") pod \"nova-edpm-deployment-openstack-edpm-ipam-5krlr\" (UID: \"f9fc3f8e-9635-48a4-8ec3-20034a1bc7ad\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-5krlr" Dec 10 23:29:06 crc kubenswrapper[4791]: I1210 23:29:06.844374 4791 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"nova-cell1-compute-config-0\" (UniqueName: \"kubernetes.io/secret/f9fc3f8e-9635-48a4-8ec3-20034a1bc7ad-nova-cell1-compute-config-0\") pod \"nova-edpm-deployment-openstack-edpm-ipam-5krlr\" (UID: \"f9fc3f8e-9635-48a4-8ec3-20034a1bc7ad\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-5krlr" Dec 10 23:29:06 crc kubenswrapper[4791]: I1210 23:29:06.844402 4791 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"nova-migration-ssh-key-1\" (UniqueName: \"kubernetes.io/secret/f9fc3f8e-9635-48a4-8ec3-20034a1bc7ad-nova-migration-ssh-key-1\") pod \"nova-edpm-deployment-openstack-edpm-ipam-5krlr\" (UID: \"f9fc3f8e-9635-48a4-8ec3-20034a1bc7ad\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-5krlr" Dec 10 23:29:06 crc kubenswrapper[4791]: I1210 23:29:06.844479 4791 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"nova-cell1-compute-config-1\" (UniqueName: \"kubernetes.io/secret/f9fc3f8e-9635-48a4-8ec3-20034a1bc7ad-nova-cell1-compute-config-1\") pod \"nova-edpm-deployment-openstack-edpm-ipam-5krlr\" (UID: \"f9fc3f8e-9635-48a4-8ec3-20034a1bc7ad\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-5krlr" Dec 10 23:29:06 crc kubenswrapper[4791]: I1210 23:29:06.946034 4791 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nova-cell1-compute-config-0\" (UniqueName: \"kubernetes.io/secret/f9fc3f8e-9635-48a4-8ec3-20034a1bc7ad-nova-cell1-compute-config-0\") pod \"nova-edpm-deployment-openstack-edpm-ipam-5krlr\" (UID: \"f9fc3f8e-9635-48a4-8ec3-20034a1bc7ad\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-5krlr" Dec 10 23:29:06 crc kubenswrapper[4791]: I1210 23:29:06.946204 4791 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nova-migration-ssh-key-1\" (UniqueName: \"kubernetes.io/secret/f9fc3f8e-9635-48a4-8ec3-20034a1bc7ad-nova-migration-ssh-key-1\") pod \"nova-edpm-deployment-openstack-edpm-ipam-5krlr\" (UID: \"f9fc3f8e-9635-48a4-8ec3-20034a1bc7ad\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-5krlr" Dec 10 23:29:06 crc kubenswrapper[4791]: I1210 23:29:06.947067 4791 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nova-cell1-compute-config-1\" (UniqueName: \"kubernetes.io/secret/f9fc3f8e-9635-48a4-8ec3-20034a1bc7ad-nova-cell1-compute-config-1\") pod \"nova-edpm-deployment-openstack-edpm-ipam-5krlr\" (UID: \"f9fc3f8e-9635-48a4-8ec3-20034a1bc7ad\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-5krlr" Dec 10 23:29:06 crc kubenswrapper[4791]: I1210 23:29:06.947138 4791 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nova-migration-ssh-key-0\" (UniqueName: \"kubernetes.io/secret/f9fc3f8e-9635-48a4-8ec3-20034a1bc7ad-nova-migration-ssh-key-0\") pod \"nova-edpm-deployment-openstack-edpm-ipam-5krlr\" (UID: \"f9fc3f8e-9635-48a4-8ec3-20034a1bc7ad\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-5krlr" Dec 10 23:29:06 crc kubenswrapper[4791]: I1210 23:29:06.947196 4791 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nova-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f9fc3f8e-9635-48a4-8ec3-20034a1bc7ad-nova-combined-ca-bundle\") pod \"nova-edpm-deployment-openstack-edpm-ipam-5krlr\" (UID: \"f9fc3f8e-9635-48a4-8ec3-20034a1bc7ad\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-5krlr" Dec 10 23:29:06 crc kubenswrapper[4791]: I1210 23:29:06.948012 4791 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/f9fc3f8e-9635-48a4-8ec3-20034a1bc7ad-inventory\") pod \"nova-edpm-deployment-openstack-edpm-ipam-5krlr\" (UID: \"f9fc3f8e-9635-48a4-8ec3-20034a1bc7ad\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-5krlr" Dec 10 23:29:06 crc kubenswrapper[4791]: I1210 23:29:06.948186 4791 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nova-extra-config-0\" (UniqueName: \"kubernetes.io/configmap/f9fc3f8e-9635-48a4-8ec3-20034a1bc7ad-nova-extra-config-0\") pod \"nova-edpm-deployment-openstack-edpm-ipam-5krlr\" (UID: \"f9fc3f8e-9635-48a4-8ec3-20034a1bc7ad\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-5krlr" Dec 10 23:29:06 crc kubenswrapper[4791]: I1210 23:29:06.948242 4791 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/f9fc3f8e-9635-48a4-8ec3-20034a1bc7ad-ssh-key\") pod \"nova-edpm-deployment-openstack-edpm-ipam-5krlr\" (UID: \"f9fc3f8e-9635-48a4-8ec3-20034a1bc7ad\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-5krlr" Dec 10 23:29:06 crc kubenswrapper[4791]: I1210 23:29:06.948307 4791 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-475wv\" (UniqueName: \"kubernetes.io/projected/f9fc3f8e-9635-48a4-8ec3-20034a1bc7ad-kube-api-access-475wv\") pod \"nova-edpm-deployment-openstack-edpm-ipam-5krlr\" (UID: \"f9fc3f8e-9635-48a4-8ec3-20034a1bc7ad\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-5krlr" Dec 10 23:29:06 crc kubenswrapper[4791]: I1210 23:29:06.948974 4791 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"nova-extra-config-0\" (UniqueName: \"kubernetes.io/configmap/f9fc3f8e-9635-48a4-8ec3-20034a1bc7ad-nova-extra-config-0\") pod \"nova-edpm-deployment-openstack-edpm-ipam-5krlr\" (UID: \"f9fc3f8e-9635-48a4-8ec3-20034a1bc7ad\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-5krlr" Dec 10 23:29:06 crc kubenswrapper[4791]: I1210 23:29:06.950685 4791 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"nova-migration-ssh-key-0\" (UniqueName: \"kubernetes.io/secret/f9fc3f8e-9635-48a4-8ec3-20034a1bc7ad-nova-migration-ssh-key-0\") pod \"nova-edpm-deployment-openstack-edpm-ipam-5krlr\" (UID: \"f9fc3f8e-9635-48a4-8ec3-20034a1bc7ad\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-5krlr" Dec 10 23:29:06 crc kubenswrapper[4791]: I1210 23:29:06.952172 4791 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/f9fc3f8e-9635-48a4-8ec3-20034a1bc7ad-ssh-key\") pod \"nova-edpm-deployment-openstack-edpm-ipam-5krlr\" (UID: \"f9fc3f8e-9635-48a4-8ec3-20034a1bc7ad\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-5krlr" Dec 10 23:29:06 crc kubenswrapper[4791]: I1210 23:29:06.952936 4791 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"nova-cell1-compute-config-0\" (UniqueName: \"kubernetes.io/secret/f9fc3f8e-9635-48a4-8ec3-20034a1bc7ad-nova-cell1-compute-config-0\") pod \"nova-edpm-deployment-openstack-edpm-ipam-5krlr\" (UID: \"f9fc3f8e-9635-48a4-8ec3-20034a1bc7ad\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-5krlr" Dec 10 23:29:06 crc kubenswrapper[4791]: I1210 23:29:06.954016 4791 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"nova-cell1-compute-config-1\" (UniqueName: \"kubernetes.io/secret/f9fc3f8e-9635-48a4-8ec3-20034a1bc7ad-nova-cell1-compute-config-1\") pod \"nova-edpm-deployment-openstack-edpm-ipam-5krlr\" (UID: \"f9fc3f8e-9635-48a4-8ec3-20034a1bc7ad\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-5krlr" Dec 10 23:29:06 crc kubenswrapper[4791]: I1210 23:29:06.954826 4791 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"nova-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f9fc3f8e-9635-48a4-8ec3-20034a1bc7ad-nova-combined-ca-bundle\") pod \"nova-edpm-deployment-openstack-edpm-ipam-5krlr\" (UID: \"f9fc3f8e-9635-48a4-8ec3-20034a1bc7ad\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-5krlr" Dec 10 23:29:06 crc kubenswrapper[4791]: I1210 23:29:06.955921 4791 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"nova-migration-ssh-key-1\" (UniqueName: \"kubernetes.io/secret/f9fc3f8e-9635-48a4-8ec3-20034a1bc7ad-nova-migration-ssh-key-1\") pod \"nova-edpm-deployment-openstack-edpm-ipam-5krlr\" (UID: \"f9fc3f8e-9635-48a4-8ec3-20034a1bc7ad\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-5krlr" Dec 10 23:29:06 crc kubenswrapper[4791]: I1210 23:29:06.957537 4791 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/f9fc3f8e-9635-48a4-8ec3-20034a1bc7ad-inventory\") pod \"nova-edpm-deployment-openstack-edpm-ipam-5krlr\" (UID: \"f9fc3f8e-9635-48a4-8ec3-20034a1bc7ad\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-5krlr" Dec 10 23:29:06 crc kubenswrapper[4791]: I1210 23:29:06.978131 4791 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-475wv\" (UniqueName: \"kubernetes.io/projected/f9fc3f8e-9635-48a4-8ec3-20034a1bc7ad-kube-api-access-475wv\") pod \"nova-edpm-deployment-openstack-edpm-ipam-5krlr\" (UID: \"f9fc3f8e-9635-48a4-8ec3-20034a1bc7ad\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-5krlr" Dec 10 23:29:07 crc kubenswrapper[4791]: I1210 23:29:07.037633 4791 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-5krlr" Dec 10 23:29:07 crc kubenswrapper[4791]: I1210 23:29:07.668885 4791 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-edpm-deployment-openstack-edpm-ipam-5krlr"] Dec 10 23:29:07 crc kubenswrapper[4791]: I1210 23:29:07.677221 4791 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Dec 10 23:29:08 crc kubenswrapper[4791]: I1210 23:29:08.585627 4791 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-5krlr" event={"ID":"f9fc3f8e-9635-48a4-8ec3-20034a1bc7ad","Type":"ContainerStarted","Data":"794f938154b42ad1c47f3b63b7e71723ed55a0d93b0d76114b4958d534874626"} Dec 10 23:29:08 crc kubenswrapper[4791]: I1210 23:29:08.585943 4791 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-5krlr" event={"ID":"f9fc3f8e-9635-48a4-8ec3-20034a1bc7ad","Type":"ContainerStarted","Data":"9202e0c33e0c04504153b8fb76d4d02d9260df574897a27ad6f0bc8f15639b64"} Dec 10 23:29:08 crc kubenswrapper[4791]: I1210 23:29:08.612920 4791 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-5krlr" podStartSLOduration=2.028799688 podStartE2EDuration="2.61288212s" podCreationTimestamp="2025-12-10 23:29:06 +0000 UTC" firstStartedPulling="2025-12-10 23:29:07.676970242 +0000 UTC m=+2382.106587855" lastFinishedPulling="2025-12-10 23:29:08.261052674 +0000 UTC m=+2382.690670287" observedRunningTime="2025-12-10 23:29:08.605428736 +0000 UTC m=+2383.035046349" watchObservedRunningTime="2025-12-10 23:29:08.61288212 +0000 UTC m=+2383.042499753" Dec 10 23:29:16 crc kubenswrapper[4791]: I1210 23:29:16.884679 4791 scope.go:117] "RemoveContainer" containerID="4229e7e5318c8150a4b76b463eed8dfb67fd267a558642295fc4d4f697411536" Dec 10 23:29:16 crc kubenswrapper[4791]: E1210 23:29:16.885482 4791 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-5rb5l_openshift-machine-config-operator(ba35653c-6e06-4cee-a4d6-137764090d18)\"" pod="openshift-machine-config-operator/machine-config-daemon-5rb5l" podUID="ba35653c-6e06-4cee-a4d6-137764090d18" Dec 10 23:29:30 crc kubenswrapper[4791]: I1210 23:29:30.885226 4791 scope.go:117] "RemoveContainer" containerID="4229e7e5318c8150a4b76b463eed8dfb67fd267a558642295fc4d4f697411536" Dec 10 23:29:30 crc kubenswrapper[4791]: E1210 23:29:30.886280 4791 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-5rb5l_openshift-machine-config-operator(ba35653c-6e06-4cee-a4d6-137764090d18)\"" pod="openshift-machine-config-operator/machine-config-daemon-5rb5l" podUID="ba35653c-6e06-4cee-a4d6-137764090d18" Dec 10 23:29:41 crc kubenswrapper[4791]: I1210 23:29:41.884676 4791 scope.go:117] "RemoveContainer" containerID="4229e7e5318c8150a4b76b463eed8dfb67fd267a558642295fc4d4f697411536" Dec 10 23:29:41 crc kubenswrapper[4791]: E1210 23:29:41.885364 4791 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-5rb5l_openshift-machine-config-operator(ba35653c-6e06-4cee-a4d6-137764090d18)\"" pod="openshift-machine-config-operator/machine-config-daemon-5rb5l" podUID="ba35653c-6e06-4cee-a4d6-137764090d18" Dec 10 23:29:55 crc kubenswrapper[4791]: I1210 23:29:55.893180 4791 scope.go:117] "RemoveContainer" containerID="4229e7e5318c8150a4b76b463eed8dfb67fd267a558642295fc4d4f697411536" Dec 10 23:29:55 crc kubenswrapper[4791]: E1210 23:29:55.894115 4791 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-5rb5l_openshift-machine-config-operator(ba35653c-6e06-4cee-a4d6-137764090d18)\"" pod="openshift-machine-config-operator/machine-config-daemon-5rb5l" podUID="ba35653c-6e06-4cee-a4d6-137764090d18" Dec 10 23:30:00 crc kubenswrapper[4791]: I1210 23:30:00.180516 4791 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29423490-wplkh"] Dec 10 23:30:00 crc kubenswrapper[4791]: I1210 23:30:00.195330 4791 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29423490-wplkh" Dec 10 23:30:00 crc kubenswrapper[4791]: I1210 23:30:00.199148 4791 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29423490-wplkh"] Dec 10 23:30:00 crc kubenswrapper[4791]: I1210 23:30:00.199270 4791 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"collect-profiles-dockercfg-kzf4t" Dec 10 23:30:00 crc kubenswrapper[4791]: I1210 23:30:00.202086 4791 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-operator-lifecycle-manager"/"collect-profiles-config" Dec 10 23:30:00 crc kubenswrapper[4791]: I1210 23:30:00.328622 4791 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/978a38ea-5cd4-4e0f-b2db-950f8644cb19-config-volume\") pod \"collect-profiles-29423490-wplkh\" (UID: \"978a38ea-5cd4-4e0f-b2db-950f8644cb19\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29423490-wplkh" Dec 10 23:30:00 crc kubenswrapper[4791]: I1210 23:30:00.328761 4791 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-qwz47\" (UniqueName: \"kubernetes.io/projected/978a38ea-5cd4-4e0f-b2db-950f8644cb19-kube-api-access-qwz47\") pod \"collect-profiles-29423490-wplkh\" (UID: \"978a38ea-5cd4-4e0f-b2db-950f8644cb19\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29423490-wplkh" Dec 10 23:30:00 crc kubenswrapper[4791]: I1210 23:30:00.328803 4791 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/978a38ea-5cd4-4e0f-b2db-950f8644cb19-secret-volume\") pod \"collect-profiles-29423490-wplkh\" (UID: \"978a38ea-5cd4-4e0f-b2db-950f8644cb19\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29423490-wplkh" Dec 10 23:30:00 crc kubenswrapper[4791]: I1210 23:30:00.430393 4791 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-qwz47\" (UniqueName: \"kubernetes.io/projected/978a38ea-5cd4-4e0f-b2db-950f8644cb19-kube-api-access-qwz47\") pod \"collect-profiles-29423490-wplkh\" (UID: \"978a38ea-5cd4-4e0f-b2db-950f8644cb19\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29423490-wplkh" Dec 10 23:30:00 crc kubenswrapper[4791]: I1210 23:30:00.430461 4791 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/978a38ea-5cd4-4e0f-b2db-950f8644cb19-secret-volume\") pod \"collect-profiles-29423490-wplkh\" (UID: \"978a38ea-5cd4-4e0f-b2db-950f8644cb19\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29423490-wplkh" Dec 10 23:30:00 crc kubenswrapper[4791]: I1210 23:30:00.430515 4791 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/978a38ea-5cd4-4e0f-b2db-950f8644cb19-config-volume\") pod \"collect-profiles-29423490-wplkh\" (UID: \"978a38ea-5cd4-4e0f-b2db-950f8644cb19\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29423490-wplkh" Dec 10 23:30:00 crc kubenswrapper[4791]: I1210 23:30:00.431435 4791 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/978a38ea-5cd4-4e0f-b2db-950f8644cb19-config-volume\") pod \"collect-profiles-29423490-wplkh\" (UID: \"978a38ea-5cd4-4e0f-b2db-950f8644cb19\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29423490-wplkh" Dec 10 23:30:00 crc kubenswrapper[4791]: I1210 23:30:00.440160 4791 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/978a38ea-5cd4-4e0f-b2db-950f8644cb19-secret-volume\") pod \"collect-profiles-29423490-wplkh\" (UID: \"978a38ea-5cd4-4e0f-b2db-950f8644cb19\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29423490-wplkh" Dec 10 23:30:00 crc kubenswrapper[4791]: I1210 23:30:00.454245 4791 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-qwz47\" (UniqueName: \"kubernetes.io/projected/978a38ea-5cd4-4e0f-b2db-950f8644cb19-kube-api-access-qwz47\") pod \"collect-profiles-29423490-wplkh\" (UID: \"978a38ea-5cd4-4e0f-b2db-950f8644cb19\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29423490-wplkh" Dec 10 23:30:00 crc kubenswrapper[4791]: I1210 23:30:00.545184 4791 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29423490-wplkh" Dec 10 23:30:01 crc kubenswrapper[4791]: I1210 23:30:01.037325 4791 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29423490-wplkh"] Dec 10 23:30:01 crc kubenswrapper[4791]: I1210 23:30:01.603747 4791 generic.go:334] "Generic (PLEG): container finished" podID="978a38ea-5cd4-4e0f-b2db-950f8644cb19" containerID="6d06e53670014e6efc1d63d85db9ba851e457b6432c2ea71ab7877a2c5b99b8c" exitCode=0 Dec 10 23:30:01 crc kubenswrapper[4791]: I1210 23:30:01.603872 4791 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29423490-wplkh" event={"ID":"978a38ea-5cd4-4e0f-b2db-950f8644cb19","Type":"ContainerDied","Data":"6d06e53670014e6efc1d63d85db9ba851e457b6432c2ea71ab7877a2c5b99b8c"} Dec 10 23:30:01 crc kubenswrapper[4791]: I1210 23:30:01.604299 4791 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29423490-wplkh" event={"ID":"978a38ea-5cd4-4e0f-b2db-950f8644cb19","Type":"ContainerStarted","Data":"59fc4ab2e3269678e1e7d497e259f616c076835711928f20e26dfad0d1eb3da4"} Dec 10 23:30:03 crc kubenswrapper[4791]: I1210 23:30:03.013740 4791 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29423490-wplkh" Dec 10 23:30:03 crc kubenswrapper[4791]: I1210 23:30:03.085983 4791 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/978a38ea-5cd4-4e0f-b2db-950f8644cb19-config-volume\") pod \"978a38ea-5cd4-4e0f-b2db-950f8644cb19\" (UID: \"978a38ea-5cd4-4e0f-b2db-950f8644cb19\") " Dec 10 23:30:03 crc kubenswrapper[4791]: I1210 23:30:03.086047 4791 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/978a38ea-5cd4-4e0f-b2db-950f8644cb19-secret-volume\") pod \"978a38ea-5cd4-4e0f-b2db-950f8644cb19\" (UID: \"978a38ea-5cd4-4e0f-b2db-950f8644cb19\") " Dec 10 23:30:03 crc kubenswrapper[4791]: I1210 23:30:03.086245 4791 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-qwz47\" (UniqueName: \"kubernetes.io/projected/978a38ea-5cd4-4e0f-b2db-950f8644cb19-kube-api-access-qwz47\") pod \"978a38ea-5cd4-4e0f-b2db-950f8644cb19\" (UID: \"978a38ea-5cd4-4e0f-b2db-950f8644cb19\") " Dec 10 23:30:03 crc kubenswrapper[4791]: I1210 23:30:03.086961 4791 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/978a38ea-5cd4-4e0f-b2db-950f8644cb19-config-volume" (OuterVolumeSpecName: "config-volume") pod "978a38ea-5cd4-4e0f-b2db-950f8644cb19" (UID: "978a38ea-5cd4-4e0f-b2db-950f8644cb19"). InnerVolumeSpecName "config-volume". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 10 23:30:03 crc kubenswrapper[4791]: I1210 23:30:03.092368 4791 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/978a38ea-5cd4-4e0f-b2db-950f8644cb19-secret-volume" (OuterVolumeSpecName: "secret-volume") pod "978a38ea-5cd4-4e0f-b2db-950f8644cb19" (UID: "978a38ea-5cd4-4e0f-b2db-950f8644cb19"). InnerVolumeSpecName "secret-volume". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 10 23:30:03 crc kubenswrapper[4791]: I1210 23:30:03.092790 4791 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/978a38ea-5cd4-4e0f-b2db-950f8644cb19-kube-api-access-qwz47" (OuterVolumeSpecName: "kube-api-access-qwz47") pod "978a38ea-5cd4-4e0f-b2db-950f8644cb19" (UID: "978a38ea-5cd4-4e0f-b2db-950f8644cb19"). InnerVolumeSpecName "kube-api-access-qwz47". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 10 23:30:03 crc kubenswrapper[4791]: I1210 23:30:03.188501 4791 reconciler_common.go:293] "Volume detached for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/978a38ea-5cd4-4e0f-b2db-950f8644cb19-config-volume\") on node \"crc\" DevicePath \"\"" Dec 10 23:30:03 crc kubenswrapper[4791]: I1210 23:30:03.188836 4791 reconciler_common.go:293] "Volume detached for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/978a38ea-5cd4-4e0f-b2db-950f8644cb19-secret-volume\") on node \"crc\" DevicePath \"\"" Dec 10 23:30:03 crc kubenswrapper[4791]: I1210 23:30:03.188939 4791 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-qwz47\" (UniqueName: \"kubernetes.io/projected/978a38ea-5cd4-4e0f-b2db-950f8644cb19-kube-api-access-qwz47\") on node \"crc\" DevicePath \"\"" Dec 10 23:30:03 crc kubenswrapper[4791]: I1210 23:30:03.624994 4791 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29423490-wplkh" event={"ID":"978a38ea-5cd4-4e0f-b2db-950f8644cb19","Type":"ContainerDied","Data":"59fc4ab2e3269678e1e7d497e259f616c076835711928f20e26dfad0d1eb3da4"} Dec 10 23:30:03 crc kubenswrapper[4791]: I1210 23:30:03.625514 4791 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="59fc4ab2e3269678e1e7d497e259f616c076835711928f20e26dfad0d1eb3da4" Dec 10 23:30:03 crc kubenswrapper[4791]: I1210 23:30:03.625083 4791 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29423490-wplkh" Dec 10 23:30:04 crc kubenswrapper[4791]: I1210 23:30:04.100536 4791 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29423445-b9cxb"] Dec 10 23:30:04 crc kubenswrapper[4791]: I1210 23:30:04.108920 4791 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29423445-b9cxb"] Dec 10 23:30:05 crc kubenswrapper[4791]: I1210 23:30:05.903407 4791 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="90640468-53d7-42bc-a835-f7acf48a486e" path="/var/lib/kubelet/pods/90640468-53d7-42bc-a835-f7acf48a486e/volumes" Dec 10 23:30:09 crc kubenswrapper[4791]: I1210 23:30:09.885423 4791 scope.go:117] "RemoveContainer" containerID="4229e7e5318c8150a4b76b463eed8dfb67fd267a558642295fc4d4f697411536" Dec 10 23:30:09 crc kubenswrapper[4791]: E1210 23:30:09.886259 4791 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-5rb5l_openshift-machine-config-operator(ba35653c-6e06-4cee-a4d6-137764090d18)\"" pod="openshift-machine-config-operator/machine-config-daemon-5rb5l" podUID="ba35653c-6e06-4cee-a4d6-137764090d18" Dec 10 23:30:20 crc kubenswrapper[4791]: I1210 23:30:20.884988 4791 scope.go:117] "RemoveContainer" containerID="4229e7e5318c8150a4b76b463eed8dfb67fd267a558642295fc4d4f697411536" Dec 10 23:30:20 crc kubenswrapper[4791]: E1210 23:30:20.885787 4791 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-5rb5l_openshift-machine-config-operator(ba35653c-6e06-4cee-a4d6-137764090d18)\"" pod="openshift-machine-config-operator/machine-config-daemon-5rb5l" podUID="ba35653c-6e06-4cee-a4d6-137764090d18" Dec 10 23:30:32 crc kubenswrapper[4791]: I1210 23:30:32.048573 4791 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-operators-p5qd2"] Dec 10 23:30:32 crc kubenswrapper[4791]: E1210 23:30:32.050287 4791 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="978a38ea-5cd4-4e0f-b2db-950f8644cb19" containerName="collect-profiles" Dec 10 23:30:32 crc kubenswrapper[4791]: I1210 23:30:32.050312 4791 state_mem.go:107] "Deleted CPUSet assignment" podUID="978a38ea-5cd4-4e0f-b2db-950f8644cb19" containerName="collect-profiles" Dec 10 23:30:32 crc kubenswrapper[4791]: I1210 23:30:32.065578 4791 memory_manager.go:354] "RemoveStaleState removing state" podUID="978a38ea-5cd4-4e0f-b2db-950f8644cb19" containerName="collect-profiles" Dec 10 23:30:32 crc kubenswrapper[4791]: I1210 23:30:32.068087 4791 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-p5qd2"] Dec 10 23:30:32 crc kubenswrapper[4791]: I1210 23:30:32.068251 4791 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-p5qd2" Dec 10 23:30:32 crc kubenswrapper[4791]: I1210 23:30:32.167377 4791 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/e028c5e3-c73a-4ae7-bd3a-89e861fc7b22-utilities\") pod \"redhat-operators-p5qd2\" (UID: \"e028c5e3-c73a-4ae7-bd3a-89e861fc7b22\") " pod="openshift-marketplace/redhat-operators-p5qd2" Dec 10 23:30:32 crc kubenswrapper[4791]: I1210 23:30:32.167456 4791 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-m8dgk\" (UniqueName: \"kubernetes.io/projected/e028c5e3-c73a-4ae7-bd3a-89e861fc7b22-kube-api-access-m8dgk\") pod \"redhat-operators-p5qd2\" (UID: \"e028c5e3-c73a-4ae7-bd3a-89e861fc7b22\") " pod="openshift-marketplace/redhat-operators-p5qd2" Dec 10 23:30:32 crc kubenswrapper[4791]: I1210 23:30:32.167802 4791 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/e028c5e3-c73a-4ae7-bd3a-89e861fc7b22-catalog-content\") pod \"redhat-operators-p5qd2\" (UID: \"e028c5e3-c73a-4ae7-bd3a-89e861fc7b22\") " pod="openshift-marketplace/redhat-operators-p5qd2" Dec 10 23:30:32 crc kubenswrapper[4791]: I1210 23:30:32.269660 4791 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/e028c5e3-c73a-4ae7-bd3a-89e861fc7b22-utilities\") pod \"redhat-operators-p5qd2\" (UID: \"e028c5e3-c73a-4ae7-bd3a-89e861fc7b22\") " pod="openshift-marketplace/redhat-operators-p5qd2" Dec 10 23:30:32 crc kubenswrapper[4791]: I1210 23:30:32.269721 4791 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-m8dgk\" (UniqueName: \"kubernetes.io/projected/e028c5e3-c73a-4ae7-bd3a-89e861fc7b22-kube-api-access-m8dgk\") pod \"redhat-operators-p5qd2\" (UID: \"e028c5e3-c73a-4ae7-bd3a-89e861fc7b22\") " pod="openshift-marketplace/redhat-operators-p5qd2" Dec 10 23:30:32 crc kubenswrapper[4791]: I1210 23:30:32.269794 4791 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/e028c5e3-c73a-4ae7-bd3a-89e861fc7b22-catalog-content\") pod \"redhat-operators-p5qd2\" (UID: \"e028c5e3-c73a-4ae7-bd3a-89e861fc7b22\") " pod="openshift-marketplace/redhat-operators-p5qd2" Dec 10 23:30:32 crc kubenswrapper[4791]: I1210 23:30:32.270539 4791 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/e028c5e3-c73a-4ae7-bd3a-89e861fc7b22-catalog-content\") pod \"redhat-operators-p5qd2\" (UID: \"e028c5e3-c73a-4ae7-bd3a-89e861fc7b22\") " pod="openshift-marketplace/redhat-operators-p5qd2" Dec 10 23:30:32 crc kubenswrapper[4791]: I1210 23:30:32.270634 4791 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/e028c5e3-c73a-4ae7-bd3a-89e861fc7b22-utilities\") pod \"redhat-operators-p5qd2\" (UID: \"e028c5e3-c73a-4ae7-bd3a-89e861fc7b22\") " pod="openshift-marketplace/redhat-operators-p5qd2" Dec 10 23:30:32 crc kubenswrapper[4791]: I1210 23:30:32.295854 4791 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-m8dgk\" (UniqueName: \"kubernetes.io/projected/e028c5e3-c73a-4ae7-bd3a-89e861fc7b22-kube-api-access-m8dgk\") pod \"redhat-operators-p5qd2\" (UID: \"e028c5e3-c73a-4ae7-bd3a-89e861fc7b22\") " pod="openshift-marketplace/redhat-operators-p5qd2" Dec 10 23:30:32 crc kubenswrapper[4791]: I1210 23:30:32.393780 4791 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-p5qd2" Dec 10 23:30:32 crc kubenswrapper[4791]: I1210 23:30:32.837647 4791 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-p5qd2"] Dec 10 23:30:32 crc kubenswrapper[4791]: I1210 23:30:32.950716 4791 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-p5qd2" event={"ID":"e028c5e3-c73a-4ae7-bd3a-89e861fc7b22","Type":"ContainerStarted","Data":"3396b86d4c9019ae3fbc6bf0ebdb39831ca234c86d6345ee8ef2498e5554a724"} Dec 10 23:30:33 crc kubenswrapper[4791]: I1210 23:30:33.964154 4791 generic.go:334] "Generic (PLEG): container finished" podID="e028c5e3-c73a-4ae7-bd3a-89e861fc7b22" containerID="2f2b1519177f3476c905337934e172dd87b54e82a2f8bbadd5ba1915de0ab1b1" exitCode=0 Dec 10 23:30:33 crc kubenswrapper[4791]: I1210 23:30:33.964421 4791 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-p5qd2" event={"ID":"e028c5e3-c73a-4ae7-bd3a-89e861fc7b22","Type":"ContainerDied","Data":"2f2b1519177f3476c905337934e172dd87b54e82a2f8bbadd5ba1915de0ab1b1"} Dec 10 23:30:35 crc kubenswrapper[4791]: I1210 23:30:35.893679 4791 scope.go:117] "RemoveContainer" containerID="4229e7e5318c8150a4b76b463eed8dfb67fd267a558642295fc4d4f697411536" Dec 10 23:30:35 crc kubenswrapper[4791]: E1210 23:30:35.894257 4791 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-5rb5l_openshift-machine-config-operator(ba35653c-6e06-4cee-a4d6-137764090d18)\"" pod="openshift-machine-config-operator/machine-config-daemon-5rb5l" podUID="ba35653c-6e06-4cee-a4d6-137764090d18" Dec 10 23:30:37 crc kubenswrapper[4791]: I1210 23:30:37.007295 4791 generic.go:334] "Generic (PLEG): container finished" podID="e028c5e3-c73a-4ae7-bd3a-89e861fc7b22" containerID="4a5d0a6ea3bef6116395dec420a1b5bf94858e99d17da457f613a015351046a3" exitCode=0 Dec 10 23:30:37 crc kubenswrapper[4791]: I1210 23:30:37.007408 4791 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-p5qd2" event={"ID":"e028c5e3-c73a-4ae7-bd3a-89e861fc7b22","Type":"ContainerDied","Data":"4a5d0a6ea3bef6116395dec420a1b5bf94858e99d17da457f613a015351046a3"} Dec 10 23:30:38 crc kubenswrapper[4791]: I1210 23:30:38.016564 4791 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-p5qd2" event={"ID":"e028c5e3-c73a-4ae7-bd3a-89e861fc7b22","Type":"ContainerStarted","Data":"b2f74b03c1bc697c2a8cd7279f64792a877a6ebc3aea70576247245c61bfe9e5"} Dec 10 23:30:38 crc kubenswrapper[4791]: I1210 23:30:38.030744 4791 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-operators-p5qd2" podStartSLOduration=2.518479939 podStartE2EDuration="6.030709341s" podCreationTimestamp="2025-12-10 23:30:32 +0000 UTC" firstStartedPulling="2025-12-10 23:30:33.967530989 +0000 UTC m=+2468.397148642" lastFinishedPulling="2025-12-10 23:30:37.479760431 +0000 UTC m=+2471.909378044" observedRunningTime="2025-12-10 23:30:38.030055432 +0000 UTC m=+2472.459673055" watchObservedRunningTime="2025-12-10 23:30:38.030709341 +0000 UTC m=+2472.460326954" Dec 10 23:30:42 crc kubenswrapper[4791]: I1210 23:30:42.395062 4791 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-operators-p5qd2" Dec 10 23:30:42 crc kubenswrapper[4791]: I1210 23:30:42.395590 4791 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-operators-p5qd2" Dec 10 23:30:43 crc kubenswrapper[4791]: I1210 23:30:43.460184 4791 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-marketplace/redhat-operators-p5qd2" podUID="e028c5e3-c73a-4ae7-bd3a-89e861fc7b22" containerName="registry-server" probeResult="failure" output=< Dec 10 23:30:43 crc kubenswrapper[4791]: timeout: failed to connect service ":50051" within 1s Dec 10 23:30:43 crc kubenswrapper[4791]: > Dec 10 23:30:45 crc kubenswrapper[4791]: I1210 23:30:45.938672 4791 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/community-operators-hkrmt"] Dec 10 23:30:45 crc kubenswrapper[4791]: I1210 23:30:45.943037 4791 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-hkrmt" Dec 10 23:30:45 crc kubenswrapper[4791]: I1210 23:30:45.957902 4791 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-hkrmt"] Dec 10 23:30:46 crc kubenswrapper[4791]: I1210 23:30:46.043242 4791 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/2de93151-69a4-4456-9719-16133f9c65a8-catalog-content\") pod \"community-operators-hkrmt\" (UID: \"2de93151-69a4-4456-9719-16133f9c65a8\") " pod="openshift-marketplace/community-operators-hkrmt" Dec 10 23:30:46 crc kubenswrapper[4791]: I1210 23:30:46.043866 4791 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/2de93151-69a4-4456-9719-16133f9c65a8-utilities\") pod \"community-operators-hkrmt\" (UID: \"2de93151-69a4-4456-9719-16133f9c65a8\") " pod="openshift-marketplace/community-operators-hkrmt" Dec 10 23:30:46 crc kubenswrapper[4791]: I1210 23:30:46.044154 4791 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-s4flf\" (UniqueName: \"kubernetes.io/projected/2de93151-69a4-4456-9719-16133f9c65a8-kube-api-access-s4flf\") pod \"community-operators-hkrmt\" (UID: \"2de93151-69a4-4456-9719-16133f9c65a8\") " pod="openshift-marketplace/community-operators-hkrmt" Dec 10 23:30:46 crc kubenswrapper[4791]: I1210 23:30:46.146260 4791 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/2de93151-69a4-4456-9719-16133f9c65a8-utilities\") pod \"community-operators-hkrmt\" (UID: \"2de93151-69a4-4456-9719-16133f9c65a8\") " pod="openshift-marketplace/community-operators-hkrmt" Dec 10 23:30:46 crc kubenswrapper[4791]: I1210 23:30:46.146414 4791 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-s4flf\" (UniqueName: \"kubernetes.io/projected/2de93151-69a4-4456-9719-16133f9c65a8-kube-api-access-s4flf\") pod \"community-operators-hkrmt\" (UID: \"2de93151-69a4-4456-9719-16133f9c65a8\") " pod="openshift-marketplace/community-operators-hkrmt" Dec 10 23:30:46 crc kubenswrapper[4791]: I1210 23:30:46.146465 4791 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/2de93151-69a4-4456-9719-16133f9c65a8-catalog-content\") pod \"community-operators-hkrmt\" (UID: \"2de93151-69a4-4456-9719-16133f9c65a8\") " pod="openshift-marketplace/community-operators-hkrmt" Dec 10 23:30:46 crc kubenswrapper[4791]: I1210 23:30:46.146914 4791 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/2de93151-69a4-4456-9719-16133f9c65a8-utilities\") pod \"community-operators-hkrmt\" (UID: \"2de93151-69a4-4456-9719-16133f9c65a8\") " pod="openshift-marketplace/community-operators-hkrmt" Dec 10 23:30:46 crc kubenswrapper[4791]: I1210 23:30:46.147090 4791 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/2de93151-69a4-4456-9719-16133f9c65a8-catalog-content\") pod \"community-operators-hkrmt\" (UID: \"2de93151-69a4-4456-9719-16133f9c65a8\") " pod="openshift-marketplace/community-operators-hkrmt" Dec 10 23:30:46 crc kubenswrapper[4791]: I1210 23:30:46.170593 4791 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-s4flf\" (UniqueName: \"kubernetes.io/projected/2de93151-69a4-4456-9719-16133f9c65a8-kube-api-access-s4flf\") pod \"community-operators-hkrmt\" (UID: \"2de93151-69a4-4456-9719-16133f9c65a8\") " pod="openshift-marketplace/community-operators-hkrmt" Dec 10 23:30:46 crc kubenswrapper[4791]: I1210 23:30:46.267549 4791 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-hkrmt" Dec 10 23:30:46 crc kubenswrapper[4791]: I1210 23:30:46.825817 4791 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-hkrmt"] Dec 10 23:30:46 crc kubenswrapper[4791]: I1210 23:30:46.885249 4791 scope.go:117] "RemoveContainer" containerID="4229e7e5318c8150a4b76b463eed8dfb67fd267a558642295fc4d4f697411536" Dec 10 23:30:46 crc kubenswrapper[4791]: E1210 23:30:46.885548 4791 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-5rb5l_openshift-machine-config-operator(ba35653c-6e06-4cee-a4d6-137764090d18)\"" pod="openshift-machine-config-operator/machine-config-daemon-5rb5l" podUID="ba35653c-6e06-4cee-a4d6-137764090d18" Dec 10 23:30:47 crc kubenswrapper[4791]: I1210 23:30:47.107081 4791 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-hkrmt" event={"ID":"2de93151-69a4-4456-9719-16133f9c65a8","Type":"ContainerStarted","Data":"cd70397aa736320d0f2ab8b028554ab125e3eb56bdc8d661a1f62d0a5c87c931"} Dec 10 23:30:51 crc kubenswrapper[4791]: I1210 23:30:51.149144 4791 generic.go:334] "Generic (PLEG): container finished" podID="2de93151-69a4-4456-9719-16133f9c65a8" containerID="bb17daf77ca348fff3ef2f906335d665d4dd2a6acab38e977df59982b8681278" exitCode=0 Dec 10 23:30:51 crc kubenswrapper[4791]: I1210 23:30:51.149251 4791 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-hkrmt" event={"ID":"2de93151-69a4-4456-9719-16133f9c65a8","Type":"ContainerDied","Data":"bb17daf77ca348fff3ef2f906335d665d4dd2a6acab38e977df59982b8681278"} Dec 10 23:30:52 crc kubenswrapper[4791]: I1210 23:30:52.450625 4791 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-operators-p5qd2" Dec 10 23:30:52 crc kubenswrapper[4791]: I1210 23:30:52.520171 4791 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-operators-p5qd2" Dec 10 23:30:52 crc kubenswrapper[4791]: I1210 23:30:52.698065 4791 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-p5qd2"] Dec 10 23:30:53 crc kubenswrapper[4791]: I1210 23:30:53.174081 4791 generic.go:334] "Generic (PLEG): container finished" podID="2de93151-69a4-4456-9719-16133f9c65a8" containerID="69494d4b229e99e3f12f305f51c6a9ec5663ded7c7666d937a4cb9874f0d09b1" exitCode=0 Dec 10 23:30:53 crc kubenswrapper[4791]: I1210 23:30:53.174136 4791 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-hkrmt" event={"ID":"2de93151-69a4-4456-9719-16133f9c65a8","Type":"ContainerDied","Data":"69494d4b229e99e3f12f305f51c6a9ec5663ded7c7666d937a4cb9874f0d09b1"} Dec 10 23:30:54 crc kubenswrapper[4791]: I1210 23:30:54.121830 4791 scope.go:117] "RemoveContainer" containerID="e2c74df205f03bdc1e7206068bf864feaf6f07d7ae52dd750d7b9bb336e3c712" Dec 10 23:30:54 crc kubenswrapper[4791]: I1210 23:30:54.185408 4791 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-operators-p5qd2" podUID="e028c5e3-c73a-4ae7-bd3a-89e861fc7b22" containerName="registry-server" containerID="cri-o://b2f74b03c1bc697c2a8cd7279f64792a877a6ebc3aea70576247245c61bfe9e5" gracePeriod=2 Dec 10 23:30:54 crc kubenswrapper[4791]: I1210 23:30:54.672161 4791 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-p5qd2" Dec 10 23:30:54 crc kubenswrapper[4791]: I1210 23:30:54.821229 4791 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/e028c5e3-c73a-4ae7-bd3a-89e861fc7b22-catalog-content\") pod \"e028c5e3-c73a-4ae7-bd3a-89e861fc7b22\" (UID: \"e028c5e3-c73a-4ae7-bd3a-89e861fc7b22\") " Dec 10 23:30:54 crc kubenswrapper[4791]: I1210 23:30:54.821634 4791 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-m8dgk\" (UniqueName: \"kubernetes.io/projected/e028c5e3-c73a-4ae7-bd3a-89e861fc7b22-kube-api-access-m8dgk\") pod \"e028c5e3-c73a-4ae7-bd3a-89e861fc7b22\" (UID: \"e028c5e3-c73a-4ae7-bd3a-89e861fc7b22\") " Dec 10 23:30:54 crc kubenswrapper[4791]: I1210 23:30:54.821832 4791 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/e028c5e3-c73a-4ae7-bd3a-89e861fc7b22-utilities\") pod \"e028c5e3-c73a-4ae7-bd3a-89e861fc7b22\" (UID: \"e028c5e3-c73a-4ae7-bd3a-89e861fc7b22\") " Dec 10 23:30:54 crc kubenswrapper[4791]: I1210 23:30:54.822835 4791 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/e028c5e3-c73a-4ae7-bd3a-89e861fc7b22-utilities" (OuterVolumeSpecName: "utilities") pod "e028c5e3-c73a-4ae7-bd3a-89e861fc7b22" (UID: "e028c5e3-c73a-4ae7-bd3a-89e861fc7b22"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 10 23:30:54 crc kubenswrapper[4791]: I1210 23:30:54.827853 4791 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/e028c5e3-c73a-4ae7-bd3a-89e861fc7b22-kube-api-access-m8dgk" (OuterVolumeSpecName: "kube-api-access-m8dgk") pod "e028c5e3-c73a-4ae7-bd3a-89e861fc7b22" (UID: "e028c5e3-c73a-4ae7-bd3a-89e861fc7b22"). InnerVolumeSpecName "kube-api-access-m8dgk". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 10 23:30:54 crc kubenswrapper[4791]: I1210 23:30:54.925282 4791 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/e028c5e3-c73a-4ae7-bd3a-89e861fc7b22-utilities\") on node \"crc\" DevicePath \"\"" Dec 10 23:30:54 crc kubenswrapper[4791]: I1210 23:30:54.925381 4791 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-m8dgk\" (UniqueName: \"kubernetes.io/projected/e028c5e3-c73a-4ae7-bd3a-89e861fc7b22-kube-api-access-m8dgk\") on node \"crc\" DevicePath \"\"" Dec 10 23:30:54 crc kubenswrapper[4791]: I1210 23:30:54.991824 4791 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/e028c5e3-c73a-4ae7-bd3a-89e861fc7b22-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "e028c5e3-c73a-4ae7-bd3a-89e861fc7b22" (UID: "e028c5e3-c73a-4ae7-bd3a-89e861fc7b22"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 10 23:30:55 crc kubenswrapper[4791]: I1210 23:30:55.027373 4791 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/e028c5e3-c73a-4ae7-bd3a-89e861fc7b22-catalog-content\") on node \"crc\" DevicePath \"\"" Dec 10 23:30:55 crc kubenswrapper[4791]: I1210 23:30:55.195471 4791 generic.go:334] "Generic (PLEG): container finished" podID="e028c5e3-c73a-4ae7-bd3a-89e861fc7b22" containerID="b2f74b03c1bc697c2a8cd7279f64792a877a6ebc3aea70576247245c61bfe9e5" exitCode=0 Dec 10 23:30:55 crc kubenswrapper[4791]: I1210 23:30:55.195513 4791 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-p5qd2" event={"ID":"e028c5e3-c73a-4ae7-bd3a-89e861fc7b22","Type":"ContainerDied","Data":"b2f74b03c1bc697c2a8cd7279f64792a877a6ebc3aea70576247245c61bfe9e5"} Dec 10 23:30:55 crc kubenswrapper[4791]: I1210 23:30:55.195849 4791 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-p5qd2" event={"ID":"e028c5e3-c73a-4ae7-bd3a-89e861fc7b22","Type":"ContainerDied","Data":"3396b86d4c9019ae3fbc6bf0ebdb39831ca234c86d6345ee8ef2498e5554a724"} Dec 10 23:30:55 crc kubenswrapper[4791]: I1210 23:30:55.195562 4791 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-p5qd2" Dec 10 23:30:55 crc kubenswrapper[4791]: I1210 23:30:55.195876 4791 scope.go:117] "RemoveContainer" containerID="b2f74b03c1bc697c2a8cd7279f64792a877a6ebc3aea70576247245c61bfe9e5" Dec 10 23:30:55 crc kubenswrapper[4791]: I1210 23:30:55.200269 4791 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-hkrmt" event={"ID":"2de93151-69a4-4456-9719-16133f9c65a8","Type":"ContainerStarted","Data":"ea52acdc6a3493644638252e18c82a8caab6437b7edeaaa93d63d81602bb70ff"} Dec 10 23:30:55 crc kubenswrapper[4791]: I1210 23:30:55.226270 4791 scope.go:117] "RemoveContainer" containerID="4a5d0a6ea3bef6116395dec420a1b5bf94858e99d17da457f613a015351046a3" Dec 10 23:30:55 crc kubenswrapper[4791]: I1210 23:30:55.228556 4791 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/community-operators-hkrmt" podStartSLOduration=7.067953543 podStartE2EDuration="10.228539315s" podCreationTimestamp="2025-12-10 23:30:45 +0000 UTC" firstStartedPulling="2025-12-10 23:30:51.151270787 +0000 UTC m=+2485.580888410" lastFinishedPulling="2025-12-10 23:30:54.311856539 +0000 UTC m=+2488.741474182" observedRunningTime="2025-12-10 23:30:55.217155578 +0000 UTC m=+2489.646773201" watchObservedRunningTime="2025-12-10 23:30:55.228539315 +0000 UTC m=+2489.658156928" Dec 10 23:30:55 crc kubenswrapper[4791]: I1210 23:30:55.250677 4791 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-p5qd2"] Dec 10 23:30:55 crc kubenswrapper[4791]: I1210 23:30:55.261286 4791 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-operators-p5qd2"] Dec 10 23:30:55 crc kubenswrapper[4791]: I1210 23:30:55.264732 4791 scope.go:117] "RemoveContainer" containerID="2f2b1519177f3476c905337934e172dd87b54e82a2f8bbadd5ba1915de0ab1b1" Dec 10 23:30:55 crc kubenswrapper[4791]: I1210 23:30:55.298925 4791 scope.go:117] "RemoveContainer" containerID="b2f74b03c1bc697c2a8cd7279f64792a877a6ebc3aea70576247245c61bfe9e5" Dec 10 23:30:55 crc kubenswrapper[4791]: E1210 23:30:55.299415 4791 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"b2f74b03c1bc697c2a8cd7279f64792a877a6ebc3aea70576247245c61bfe9e5\": container with ID starting with b2f74b03c1bc697c2a8cd7279f64792a877a6ebc3aea70576247245c61bfe9e5 not found: ID does not exist" containerID="b2f74b03c1bc697c2a8cd7279f64792a877a6ebc3aea70576247245c61bfe9e5" Dec 10 23:30:55 crc kubenswrapper[4791]: I1210 23:30:55.299496 4791 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"b2f74b03c1bc697c2a8cd7279f64792a877a6ebc3aea70576247245c61bfe9e5"} err="failed to get container status \"b2f74b03c1bc697c2a8cd7279f64792a877a6ebc3aea70576247245c61bfe9e5\": rpc error: code = NotFound desc = could not find container \"b2f74b03c1bc697c2a8cd7279f64792a877a6ebc3aea70576247245c61bfe9e5\": container with ID starting with b2f74b03c1bc697c2a8cd7279f64792a877a6ebc3aea70576247245c61bfe9e5 not found: ID does not exist" Dec 10 23:30:55 crc kubenswrapper[4791]: I1210 23:30:55.299537 4791 scope.go:117] "RemoveContainer" containerID="4a5d0a6ea3bef6116395dec420a1b5bf94858e99d17da457f613a015351046a3" Dec 10 23:30:55 crc kubenswrapper[4791]: E1210 23:30:55.299850 4791 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"4a5d0a6ea3bef6116395dec420a1b5bf94858e99d17da457f613a015351046a3\": container with ID starting with 4a5d0a6ea3bef6116395dec420a1b5bf94858e99d17da457f613a015351046a3 not found: ID does not exist" containerID="4a5d0a6ea3bef6116395dec420a1b5bf94858e99d17da457f613a015351046a3" Dec 10 23:30:55 crc kubenswrapper[4791]: I1210 23:30:55.299872 4791 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"4a5d0a6ea3bef6116395dec420a1b5bf94858e99d17da457f613a015351046a3"} err="failed to get container status \"4a5d0a6ea3bef6116395dec420a1b5bf94858e99d17da457f613a015351046a3\": rpc error: code = NotFound desc = could not find container \"4a5d0a6ea3bef6116395dec420a1b5bf94858e99d17da457f613a015351046a3\": container with ID starting with 4a5d0a6ea3bef6116395dec420a1b5bf94858e99d17da457f613a015351046a3 not found: ID does not exist" Dec 10 23:30:55 crc kubenswrapper[4791]: I1210 23:30:55.299891 4791 scope.go:117] "RemoveContainer" containerID="2f2b1519177f3476c905337934e172dd87b54e82a2f8bbadd5ba1915de0ab1b1" Dec 10 23:30:55 crc kubenswrapper[4791]: E1210 23:30:55.300308 4791 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"2f2b1519177f3476c905337934e172dd87b54e82a2f8bbadd5ba1915de0ab1b1\": container with ID starting with 2f2b1519177f3476c905337934e172dd87b54e82a2f8bbadd5ba1915de0ab1b1 not found: ID does not exist" containerID="2f2b1519177f3476c905337934e172dd87b54e82a2f8bbadd5ba1915de0ab1b1" Dec 10 23:30:55 crc kubenswrapper[4791]: I1210 23:30:55.300353 4791 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"2f2b1519177f3476c905337934e172dd87b54e82a2f8bbadd5ba1915de0ab1b1"} err="failed to get container status \"2f2b1519177f3476c905337934e172dd87b54e82a2f8bbadd5ba1915de0ab1b1\": rpc error: code = NotFound desc = could not find container \"2f2b1519177f3476c905337934e172dd87b54e82a2f8bbadd5ba1915de0ab1b1\": container with ID starting with 2f2b1519177f3476c905337934e172dd87b54e82a2f8bbadd5ba1915de0ab1b1 not found: ID does not exist" Dec 10 23:30:55 crc kubenswrapper[4791]: I1210 23:30:55.901762 4791 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="e028c5e3-c73a-4ae7-bd3a-89e861fc7b22" path="/var/lib/kubelet/pods/e028c5e3-c73a-4ae7-bd3a-89e861fc7b22/volumes" Dec 10 23:30:56 crc kubenswrapper[4791]: I1210 23:30:56.268016 4791 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/community-operators-hkrmt" Dec 10 23:30:56 crc kubenswrapper[4791]: I1210 23:30:56.268400 4791 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/community-operators-hkrmt" Dec 10 23:30:57 crc kubenswrapper[4791]: I1210 23:30:57.325926 4791 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-marketplace/community-operators-hkrmt" podUID="2de93151-69a4-4456-9719-16133f9c65a8" containerName="registry-server" probeResult="failure" output=< Dec 10 23:30:57 crc kubenswrapper[4791]: timeout: failed to connect service ":50051" within 1s Dec 10 23:30:57 crc kubenswrapper[4791]: > Dec 10 23:30:59 crc kubenswrapper[4791]: I1210 23:30:59.885011 4791 scope.go:117] "RemoveContainer" containerID="4229e7e5318c8150a4b76b463eed8dfb67fd267a558642295fc4d4f697411536" Dec 10 23:30:59 crc kubenswrapper[4791]: E1210 23:30:59.885658 4791 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-5rb5l_openshift-machine-config-operator(ba35653c-6e06-4cee-a4d6-137764090d18)\"" pod="openshift-machine-config-operator/machine-config-daemon-5rb5l" podUID="ba35653c-6e06-4cee-a4d6-137764090d18" Dec 10 23:31:06 crc kubenswrapper[4791]: I1210 23:31:06.338014 4791 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/community-operators-hkrmt" Dec 10 23:31:06 crc kubenswrapper[4791]: I1210 23:31:06.414471 4791 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/community-operators-hkrmt" Dec 10 23:31:06 crc kubenswrapper[4791]: I1210 23:31:06.589849 4791 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-hkrmt"] Dec 10 23:31:08 crc kubenswrapper[4791]: I1210 23:31:08.328271 4791 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/community-operators-hkrmt" podUID="2de93151-69a4-4456-9719-16133f9c65a8" containerName="registry-server" containerID="cri-o://ea52acdc6a3493644638252e18c82a8caab6437b7edeaaa93d63d81602bb70ff" gracePeriod=2 Dec 10 23:31:08 crc kubenswrapper[4791]: I1210 23:31:08.883519 4791 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-hkrmt" Dec 10 23:31:09 crc kubenswrapper[4791]: I1210 23:31:09.000943 4791 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/2de93151-69a4-4456-9719-16133f9c65a8-utilities\") pod \"2de93151-69a4-4456-9719-16133f9c65a8\" (UID: \"2de93151-69a4-4456-9719-16133f9c65a8\") " Dec 10 23:31:09 crc kubenswrapper[4791]: I1210 23:31:09.001233 4791 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/2de93151-69a4-4456-9719-16133f9c65a8-catalog-content\") pod \"2de93151-69a4-4456-9719-16133f9c65a8\" (UID: \"2de93151-69a4-4456-9719-16133f9c65a8\") " Dec 10 23:31:09 crc kubenswrapper[4791]: I1210 23:31:09.001270 4791 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-s4flf\" (UniqueName: \"kubernetes.io/projected/2de93151-69a4-4456-9719-16133f9c65a8-kube-api-access-s4flf\") pod \"2de93151-69a4-4456-9719-16133f9c65a8\" (UID: \"2de93151-69a4-4456-9719-16133f9c65a8\") " Dec 10 23:31:09 crc kubenswrapper[4791]: I1210 23:31:09.003635 4791 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/2de93151-69a4-4456-9719-16133f9c65a8-utilities" (OuterVolumeSpecName: "utilities") pod "2de93151-69a4-4456-9719-16133f9c65a8" (UID: "2de93151-69a4-4456-9719-16133f9c65a8"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 10 23:31:09 crc kubenswrapper[4791]: I1210 23:31:09.009192 4791 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/2de93151-69a4-4456-9719-16133f9c65a8-kube-api-access-s4flf" (OuterVolumeSpecName: "kube-api-access-s4flf") pod "2de93151-69a4-4456-9719-16133f9c65a8" (UID: "2de93151-69a4-4456-9719-16133f9c65a8"). InnerVolumeSpecName "kube-api-access-s4flf". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 10 23:31:09 crc kubenswrapper[4791]: I1210 23:31:09.062656 4791 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/2de93151-69a4-4456-9719-16133f9c65a8-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "2de93151-69a4-4456-9719-16133f9c65a8" (UID: "2de93151-69a4-4456-9719-16133f9c65a8"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 10 23:31:09 crc kubenswrapper[4791]: I1210 23:31:09.103464 4791 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/2de93151-69a4-4456-9719-16133f9c65a8-catalog-content\") on node \"crc\" DevicePath \"\"" Dec 10 23:31:09 crc kubenswrapper[4791]: I1210 23:31:09.103501 4791 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-s4flf\" (UniqueName: \"kubernetes.io/projected/2de93151-69a4-4456-9719-16133f9c65a8-kube-api-access-s4flf\") on node \"crc\" DevicePath \"\"" Dec 10 23:31:09 crc kubenswrapper[4791]: I1210 23:31:09.103513 4791 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/2de93151-69a4-4456-9719-16133f9c65a8-utilities\") on node \"crc\" DevicePath \"\"" Dec 10 23:31:09 crc kubenswrapper[4791]: I1210 23:31:09.337486 4791 generic.go:334] "Generic (PLEG): container finished" podID="2de93151-69a4-4456-9719-16133f9c65a8" containerID="ea52acdc6a3493644638252e18c82a8caab6437b7edeaaa93d63d81602bb70ff" exitCode=0 Dec 10 23:31:09 crc kubenswrapper[4791]: I1210 23:31:09.337523 4791 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-hkrmt" event={"ID":"2de93151-69a4-4456-9719-16133f9c65a8","Type":"ContainerDied","Data":"ea52acdc6a3493644638252e18c82a8caab6437b7edeaaa93d63d81602bb70ff"} Dec 10 23:31:09 crc kubenswrapper[4791]: I1210 23:31:09.337547 4791 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-hkrmt" event={"ID":"2de93151-69a4-4456-9719-16133f9c65a8","Type":"ContainerDied","Data":"cd70397aa736320d0f2ab8b028554ab125e3eb56bdc8d661a1f62d0a5c87c931"} Dec 10 23:31:09 crc kubenswrapper[4791]: I1210 23:31:09.337566 4791 scope.go:117] "RemoveContainer" containerID="ea52acdc6a3493644638252e18c82a8caab6437b7edeaaa93d63d81602bb70ff" Dec 10 23:31:09 crc kubenswrapper[4791]: I1210 23:31:09.337587 4791 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-hkrmt" Dec 10 23:31:09 crc kubenswrapper[4791]: I1210 23:31:09.358828 4791 scope.go:117] "RemoveContainer" containerID="69494d4b229e99e3f12f305f51c6a9ec5663ded7c7666d937a4cb9874f0d09b1" Dec 10 23:31:09 crc kubenswrapper[4791]: I1210 23:31:09.373702 4791 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-hkrmt"] Dec 10 23:31:09 crc kubenswrapper[4791]: I1210 23:31:09.381275 4791 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/community-operators-hkrmt"] Dec 10 23:31:09 crc kubenswrapper[4791]: I1210 23:31:09.405588 4791 scope.go:117] "RemoveContainer" containerID="bb17daf77ca348fff3ef2f906335d665d4dd2a6acab38e977df59982b8681278" Dec 10 23:31:09 crc kubenswrapper[4791]: I1210 23:31:09.445307 4791 scope.go:117] "RemoveContainer" containerID="ea52acdc6a3493644638252e18c82a8caab6437b7edeaaa93d63d81602bb70ff" Dec 10 23:31:09 crc kubenswrapper[4791]: E1210 23:31:09.446079 4791 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"ea52acdc6a3493644638252e18c82a8caab6437b7edeaaa93d63d81602bb70ff\": container with ID starting with ea52acdc6a3493644638252e18c82a8caab6437b7edeaaa93d63d81602bb70ff not found: ID does not exist" containerID="ea52acdc6a3493644638252e18c82a8caab6437b7edeaaa93d63d81602bb70ff" Dec 10 23:31:09 crc kubenswrapper[4791]: I1210 23:31:09.446121 4791 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"ea52acdc6a3493644638252e18c82a8caab6437b7edeaaa93d63d81602bb70ff"} err="failed to get container status \"ea52acdc6a3493644638252e18c82a8caab6437b7edeaaa93d63d81602bb70ff\": rpc error: code = NotFound desc = could not find container \"ea52acdc6a3493644638252e18c82a8caab6437b7edeaaa93d63d81602bb70ff\": container with ID starting with ea52acdc6a3493644638252e18c82a8caab6437b7edeaaa93d63d81602bb70ff not found: ID does not exist" Dec 10 23:31:09 crc kubenswrapper[4791]: I1210 23:31:09.446144 4791 scope.go:117] "RemoveContainer" containerID="69494d4b229e99e3f12f305f51c6a9ec5663ded7c7666d937a4cb9874f0d09b1" Dec 10 23:31:09 crc kubenswrapper[4791]: E1210 23:31:09.446444 4791 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"69494d4b229e99e3f12f305f51c6a9ec5663ded7c7666d937a4cb9874f0d09b1\": container with ID starting with 69494d4b229e99e3f12f305f51c6a9ec5663ded7c7666d937a4cb9874f0d09b1 not found: ID does not exist" containerID="69494d4b229e99e3f12f305f51c6a9ec5663ded7c7666d937a4cb9874f0d09b1" Dec 10 23:31:09 crc kubenswrapper[4791]: I1210 23:31:09.446482 4791 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"69494d4b229e99e3f12f305f51c6a9ec5663ded7c7666d937a4cb9874f0d09b1"} err="failed to get container status \"69494d4b229e99e3f12f305f51c6a9ec5663ded7c7666d937a4cb9874f0d09b1\": rpc error: code = NotFound desc = could not find container \"69494d4b229e99e3f12f305f51c6a9ec5663ded7c7666d937a4cb9874f0d09b1\": container with ID starting with 69494d4b229e99e3f12f305f51c6a9ec5663ded7c7666d937a4cb9874f0d09b1 not found: ID does not exist" Dec 10 23:31:09 crc kubenswrapper[4791]: I1210 23:31:09.446503 4791 scope.go:117] "RemoveContainer" containerID="bb17daf77ca348fff3ef2f906335d665d4dd2a6acab38e977df59982b8681278" Dec 10 23:31:09 crc kubenswrapper[4791]: E1210 23:31:09.446797 4791 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"bb17daf77ca348fff3ef2f906335d665d4dd2a6acab38e977df59982b8681278\": container with ID starting with bb17daf77ca348fff3ef2f906335d665d4dd2a6acab38e977df59982b8681278 not found: ID does not exist" containerID="bb17daf77ca348fff3ef2f906335d665d4dd2a6acab38e977df59982b8681278" Dec 10 23:31:09 crc kubenswrapper[4791]: I1210 23:31:09.446826 4791 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"bb17daf77ca348fff3ef2f906335d665d4dd2a6acab38e977df59982b8681278"} err="failed to get container status \"bb17daf77ca348fff3ef2f906335d665d4dd2a6acab38e977df59982b8681278\": rpc error: code = NotFound desc = could not find container \"bb17daf77ca348fff3ef2f906335d665d4dd2a6acab38e977df59982b8681278\": container with ID starting with bb17daf77ca348fff3ef2f906335d665d4dd2a6acab38e977df59982b8681278 not found: ID does not exist" Dec 10 23:31:09 crc kubenswrapper[4791]: I1210 23:31:09.901760 4791 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="2de93151-69a4-4456-9719-16133f9c65a8" path="/var/lib/kubelet/pods/2de93151-69a4-4456-9719-16133f9c65a8/volumes" Dec 10 23:31:13 crc kubenswrapper[4791]: I1210 23:31:13.884968 4791 scope.go:117] "RemoveContainer" containerID="4229e7e5318c8150a4b76b463eed8dfb67fd267a558642295fc4d4f697411536" Dec 10 23:31:13 crc kubenswrapper[4791]: E1210 23:31:13.885653 4791 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-5rb5l_openshift-machine-config-operator(ba35653c-6e06-4cee-a4d6-137764090d18)\"" pod="openshift-machine-config-operator/machine-config-daemon-5rb5l" podUID="ba35653c-6e06-4cee-a4d6-137764090d18" Dec 10 23:31:24 crc kubenswrapper[4791]: I1210 23:31:24.884751 4791 scope.go:117] "RemoveContainer" containerID="4229e7e5318c8150a4b76b463eed8dfb67fd267a558642295fc4d4f697411536" Dec 10 23:31:24 crc kubenswrapper[4791]: E1210 23:31:24.885436 4791 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-5rb5l_openshift-machine-config-operator(ba35653c-6e06-4cee-a4d6-137764090d18)\"" pod="openshift-machine-config-operator/machine-config-daemon-5rb5l" podUID="ba35653c-6e06-4cee-a4d6-137764090d18" Dec 10 23:31:37 crc kubenswrapper[4791]: I1210 23:31:37.885959 4791 scope.go:117] "RemoveContainer" containerID="4229e7e5318c8150a4b76b463eed8dfb67fd267a558642295fc4d4f697411536" Dec 10 23:31:37 crc kubenswrapper[4791]: E1210 23:31:37.886969 4791 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-5rb5l_openshift-machine-config-operator(ba35653c-6e06-4cee-a4d6-137764090d18)\"" pod="openshift-machine-config-operator/machine-config-daemon-5rb5l" podUID="ba35653c-6e06-4cee-a4d6-137764090d18" Dec 10 23:31:44 crc kubenswrapper[4791]: I1210 23:31:44.731634 4791 generic.go:334] "Generic (PLEG): container finished" podID="f9fc3f8e-9635-48a4-8ec3-20034a1bc7ad" containerID="794f938154b42ad1c47f3b63b7e71723ed55a0d93b0d76114b4958d534874626" exitCode=0 Dec 10 23:31:44 crc kubenswrapper[4791]: I1210 23:31:44.732143 4791 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-5krlr" event={"ID":"f9fc3f8e-9635-48a4-8ec3-20034a1bc7ad","Type":"ContainerDied","Data":"794f938154b42ad1c47f3b63b7e71723ed55a0d93b0d76114b4958d534874626"} Dec 10 23:31:45 crc kubenswrapper[4791]: I1210 23:31:45.359549 4791 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/certified-operators-qwr9t"] Dec 10 23:31:45 crc kubenswrapper[4791]: E1210 23:31:45.360265 4791 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="e028c5e3-c73a-4ae7-bd3a-89e861fc7b22" containerName="extract-content" Dec 10 23:31:45 crc kubenswrapper[4791]: I1210 23:31:45.360485 4791 state_mem.go:107] "Deleted CPUSet assignment" podUID="e028c5e3-c73a-4ae7-bd3a-89e861fc7b22" containerName="extract-content" Dec 10 23:31:45 crc kubenswrapper[4791]: E1210 23:31:45.360501 4791 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="2de93151-69a4-4456-9719-16133f9c65a8" containerName="registry-server" Dec 10 23:31:45 crc kubenswrapper[4791]: I1210 23:31:45.360507 4791 state_mem.go:107] "Deleted CPUSet assignment" podUID="2de93151-69a4-4456-9719-16133f9c65a8" containerName="registry-server" Dec 10 23:31:45 crc kubenswrapper[4791]: E1210 23:31:45.360521 4791 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="e028c5e3-c73a-4ae7-bd3a-89e861fc7b22" containerName="registry-server" Dec 10 23:31:45 crc kubenswrapper[4791]: I1210 23:31:45.360527 4791 state_mem.go:107] "Deleted CPUSet assignment" podUID="e028c5e3-c73a-4ae7-bd3a-89e861fc7b22" containerName="registry-server" Dec 10 23:31:45 crc kubenswrapper[4791]: E1210 23:31:45.360541 4791 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="2de93151-69a4-4456-9719-16133f9c65a8" containerName="extract-content" Dec 10 23:31:45 crc kubenswrapper[4791]: I1210 23:31:45.360549 4791 state_mem.go:107] "Deleted CPUSet assignment" podUID="2de93151-69a4-4456-9719-16133f9c65a8" containerName="extract-content" Dec 10 23:31:45 crc kubenswrapper[4791]: E1210 23:31:45.360573 4791 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="e028c5e3-c73a-4ae7-bd3a-89e861fc7b22" containerName="extract-utilities" Dec 10 23:31:45 crc kubenswrapper[4791]: I1210 23:31:45.360579 4791 state_mem.go:107] "Deleted CPUSet assignment" podUID="e028c5e3-c73a-4ae7-bd3a-89e861fc7b22" containerName="extract-utilities" Dec 10 23:31:45 crc kubenswrapper[4791]: E1210 23:31:45.360593 4791 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="2de93151-69a4-4456-9719-16133f9c65a8" containerName="extract-utilities" Dec 10 23:31:45 crc kubenswrapper[4791]: I1210 23:31:45.360600 4791 state_mem.go:107] "Deleted CPUSet assignment" podUID="2de93151-69a4-4456-9719-16133f9c65a8" containerName="extract-utilities" Dec 10 23:31:45 crc kubenswrapper[4791]: I1210 23:31:45.360791 4791 memory_manager.go:354] "RemoveStaleState removing state" podUID="2de93151-69a4-4456-9719-16133f9c65a8" containerName="registry-server" Dec 10 23:31:45 crc kubenswrapper[4791]: I1210 23:31:45.360815 4791 memory_manager.go:354] "RemoveStaleState removing state" podUID="e028c5e3-c73a-4ae7-bd3a-89e861fc7b22" containerName="registry-server" Dec 10 23:31:45 crc kubenswrapper[4791]: I1210 23:31:45.362628 4791 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-qwr9t" Dec 10 23:31:45 crc kubenswrapper[4791]: I1210 23:31:45.389427 4791 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-qwr9t"] Dec 10 23:31:45 crc kubenswrapper[4791]: I1210 23:31:45.445776 4791 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/a7300d58-0959-4835-b910-5dd33161afb8-catalog-content\") pod \"certified-operators-qwr9t\" (UID: \"a7300d58-0959-4835-b910-5dd33161afb8\") " pod="openshift-marketplace/certified-operators-qwr9t" Dec 10 23:31:45 crc kubenswrapper[4791]: I1210 23:31:45.446013 4791 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/a7300d58-0959-4835-b910-5dd33161afb8-utilities\") pod \"certified-operators-qwr9t\" (UID: \"a7300d58-0959-4835-b910-5dd33161afb8\") " pod="openshift-marketplace/certified-operators-qwr9t" Dec 10 23:31:45 crc kubenswrapper[4791]: I1210 23:31:45.446180 4791 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-vlwfs\" (UniqueName: \"kubernetes.io/projected/a7300d58-0959-4835-b910-5dd33161afb8-kube-api-access-vlwfs\") pod \"certified-operators-qwr9t\" (UID: \"a7300d58-0959-4835-b910-5dd33161afb8\") " pod="openshift-marketplace/certified-operators-qwr9t" Dec 10 23:31:45 crc kubenswrapper[4791]: I1210 23:31:45.547960 4791 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/a7300d58-0959-4835-b910-5dd33161afb8-catalog-content\") pod \"certified-operators-qwr9t\" (UID: \"a7300d58-0959-4835-b910-5dd33161afb8\") " pod="openshift-marketplace/certified-operators-qwr9t" Dec 10 23:31:45 crc kubenswrapper[4791]: I1210 23:31:45.548464 4791 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/a7300d58-0959-4835-b910-5dd33161afb8-catalog-content\") pod \"certified-operators-qwr9t\" (UID: \"a7300d58-0959-4835-b910-5dd33161afb8\") " pod="openshift-marketplace/certified-operators-qwr9t" Dec 10 23:31:45 crc kubenswrapper[4791]: I1210 23:31:45.548846 4791 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/a7300d58-0959-4835-b910-5dd33161afb8-utilities\") pod \"certified-operators-qwr9t\" (UID: \"a7300d58-0959-4835-b910-5dd33161afb8\") " pod="openshift-marketplace/certified-operators-qwr9t" Dec 10 23:31:45 crc kubenswrapper[4791]: I1210 23:31:45.548879 4791 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/a7300d58-0959-4835-b910-5dd33161afb8-utilities\") pod \"certified-operators-qwr9t\" (UID: \"a7300d58-0959-4835-b910-5dd33161afb8\") " pod="openshift-marketplace/certified-operators-qwr9t" Dec 10 23:31:45 crc kubenswrapper[4791]: I1210 23:31:45.548963 4791 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-vlwfs\" (UniqueName: \"kubernetes.io/projected/a7300d58-0959-4835-b910-5dd33161afb8-kube-api-access-vlwfs\") pod \"certified-operators-qwr9t\" (UID: \"a7300d58-0959-4835-b910-5dd33161afb8\") " pod="openshift-marketplace/certified-operators-qwr9t" Dec 10 23:31:45 crc kubenswrapper[4791]: I1210 23:31:45.569673 4791 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-vlwfs\" (UniqueName: \"kubernetes.io/projected/a7300d58-0959-4835-b910-5dd33161afb8-kube-api-access-vlwfs\") pod \"certified-operators-qwr9t\" (UID: \"a7300d58-0959-4835-b910-5dd33161afb8\") " pod="openshift-marketplace/certified-operators-qwr9t" Dec 10 23:31:45 crc kubenswrapper[4791]: I1210 23:31:45.680587 4791 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-qwr9t" Dec 10 23:31:46 crc kubenswrapper[4791]: I1210 23:31:46.173649 4791 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-qwr9t"] Dec 10 23:31:46 crc kubenswrapper[4791]: I1210 23:31:46.284725 4791 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-5krlr" Dec 10 23:31:46 crc kubenswrapper[4791]: I1210 23:31:46.364333 4791 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"nova-extra-config-0\" (UniqueName: \"kubernetes.io/configmap/f9fc3f8e-9635-48a4-8ec3-20034a1bc7ad-nova-extra-config-0\") pod \"f9fc3f8e-9635-48a4-8ec3-20034a1bc7ad\" (UID: \"f9fc3f8e-9635-48a4-8ec3-20034a1bc7ad\") " Dec 10 23:31:46 crc kubenswrapper[4791]: I1210 23:31:46.364432 4791 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"nova-cell1-compute-config-1\" (UniqueName: \"kubernetes.io/secret/f9fc3f8e-9635-48a4-8ec3-20034a1bc7ad-nova-cell1-compute-config-1\") pod \"f9fc3f8e-9635-48a4-8ec3-20034a1bc7ad\" (UID: \"f9fc3f8e-9635-48a4-8ec3-20034a1bc7ad\") " Dec 10 23:31:46 crc kubenswrapper[4791]: I1210 23:31:46.364509 4791 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/f9fc3f8e-9635-48a4-8ec3-20034a1bc7ad-ssh-key\") pod \"f9fc3f8e-9635-48a4-8ec3-20034a1bc7ad\" (UID: \"f9fc3f8e-9635-48a4-8ec3-20034a1bc7ad\") " Dec 10 23:31:46 crc kubenswrapper[4791]: I1210 23:31:46.364560 4791 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-475wv\" (UniqueName: \"kubernetes.io/projected/f9fc3f8e-9635-48a4-8ec3-20034a1bc7ad-kube-api-access-475wv\") pod \"f9fc3f8e-9635-48a4-8ec3-20034a1bc7ad\" (UID: \"f9fc3f8e-9635-48a4-8ec3-20034a1bc7ad\") " Dec 10 23:31:46 crc kubenswrapper[4791]: I1210 23:31:46.364619 4791 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"nova-migration-ssh-key-0\" (UniqueName: \"kubernetes.io/secret/f9fc3f8e-9635-48a4-8ec3-20034a1bc7ad-nova-migration-ssh-key-0\") pod \"f9fc3f8e-9635-48a4-8ec3-20034a1bc7ad\" (UID: \"f9fc3f8e-9635-48a4-8ec3-20034a1bc7ad\") " Dec 10 23:31:46 crc kubenswrapper[4791]: I1210 23:31:46.364655 4791 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"nova-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f9fc3f8e-9635-48a4-8ec3-20034a1bc7ad-nova-combined-ca-bundle\") pod \"f9fc3f8e-9635-48a4-8ec3-20034a1bc7ad\" (UID: \"f9fc3f8e-9635-48a4-8ec3-20034a1bc7ad\") " Dec 10 23:31:46 crc kubenswrapper[4791]: I1210 23:31:46.364678 4791 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"nova-migration-ssh-key-1\" (UniqueName: \"kubernetes.io/secret/f9fc3f8e-9635-48a4-8ec3-20034a1bc7ad-nova-migration-ssh-key-1\") pod \"f9fc3f8e-9635-48a4-8ec3-20034a1bc7ad\" (UID: \"f9fc3f8e-9635-48a4-8ec3-20034a1bc7ad\") " Dec 10 23:31:46 crc kubenswrapper[4791]: I1210 23:31:46.364736 4791 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/f9fc3f8e-9635-48a4-8ec3-20034a1bc7ad-inventory\") pod \"f9fc3f8e-9635-48a4-8ec3-20034a1bc7ad\" (UID: \"f9fc3f8e-9635-48a4-8ec3-20034a1bc7ad\") " Dec 10 23:31:46 crc kubenswrapper[4791]: I1210 23:31:46.364760 4791 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"nova-cell1-compute-config-0\" (UniqueName: \"kubernetes.io/secret/f9fc3f8e-9635-48a4-8ec3-20034a1bc7ad-nova-cell1-compute-config-0\") pod \"f9fc3f8e-9635-48a4-8ec3-20034a1bc7ad\" (UID: \"f9fc3f8e-9635-48a4-8ec3-20034a1bc7ad\") " Dec 10 23:31:46 crc kubenswrapper[4791]: I1210 23:31:46.371816 4791 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/f9fc3f8e-9635-48a4-8ec3-20034a1bc7ad-kube-api-access-475wv" (OuterVolumeSpecName: "kube-api-access-475wv") pod "f9fc3f8e-9635-48a4-8ec3-20034a1bc7ad" (UID: "f9fc3f8e-9635-48a4-8ec3-20034a1bc7ad"). InnerVolumeSpecName "kube-api-access-475wv". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 10 23:31:46 crc kubenswrapper[4791]: I1210 23:31:46.374575 4791 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/f9fc3f8e-9635-48a4-8ec3-20034a1bc7ad-nova-combined-ca-bundle" (OuterVolumeSpecName: "nova-combined-ca-bundle") pod "f9fc3f8e-9635-48a4-8ec3-20034a1bc7ad" (UID: "f9fc3f8e-9635-48a4-8ec3-20034a1bc7ad"). InnerVolumeSpecName "nova-combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 10 23:31:46 crc kubenswrapper[4791]: I1210 23:31:46.396881 4791 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/f9fc3f8e-9635-48a4-8ec3-20034a1bc7ad-nova-cell1-compute-config-0" (OuterVolumeSpecName: "nova-cell1-compute-config-0") pod "f9fc3f8e-9635-48a4-8ec3-20034a1bc7ad" (UID: "f9fc3f8e-9635-48a4-8ec3-20034a1bc7ad"). InnerVolumeSpecName "nova-cell1-compute-config-0". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 10 23:31:46 crc kubenswrapper[4791]: I1210 23:31:46.410861 4791 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/f9fc3f8e-9635-48a4-8ec3-20034a1bc7ad-nova-migration-ssh-key-0" (OuterVolumeSpecName: "nova-migration-ssh-key-0") pod "f9fc3f8e-9635-48a4-8ec3-20034a1bc7ad" (UID: "f9fc3f8e-9635-48a4-8ec3-20034a1bc7ad"). InnerVolumeSpecName "nova-migration-ssh-key-0". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 10 23:31:46 crc kubenswrapper[4791]: I1210 23:31:46.412185 4791 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/f9fc3f8e-9635-48a4-8ec3-20034a1bc7ad-nova-cell1-compute-config-1" (OuterVolumeSpecName: "nova-cell1-compute-config-1") pod "f9fc3f8e-9635-48a4-8ec3-20034a1bc7ad" (UID: "f9fc3f8e-9635-48a4-8ec3-20034a1bc7ad"). InnerVolumeSpecName "nova-cell1-compute-config-1". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 10 23:31:46 crc kubenswrapper[4791]: I1210 23:31:46.417610 4791 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/f9fc3f8e-9635-48a4-8ec3-20034a1bc7ad-ssh-key" (OuterVolumeSpecName: "ssh-key") pod "f9fc3f8e-9635-48a4-8ec3-20034a1bc7ad" (UID: "f9fc3f8e-9635-48a4-8ec3-20034a1bc7ad"). InnerVolumeSpecName "ssh-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 10 23:31:46 crc kubenswrapper[4791]: I1210 23:31:46.422837 4791 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/f9fc3f8e-9635-48a4-8ec3-20034a1bc7ad-nova-extra-config-0" (OuterVolumeSpecName: "nova-extra-config-0") pod "f9fc3f8e-9635-48a4-8ec3-20034a1bc7ad" (UID: "f9fc3f8e-9635-48a4-8ec3-20034a1bc7ad"). InnerVolumeSpecName "nova-extra-config-0". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 10 23:31:46 crc kubenswrapper[4791]: I1210 23:31:46.423181 4791 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/f9fc3f8e-9635-48a4-8ec3-20034a1bc7ad-inventory" (OuterVolumeSpecName: "inventory") pod "f9fc3f8e-9635-48a4-8ec3-20034a1bc7ad" (UID: "f9fc3f8e-9635-48a4-8ec3-20034a1bc7ad"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 10 23:31:46 crc kubenswrapper[4791]: I1210 23:31:46.429189 4791 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/f9fc3f8e-9635-48a4-8ec3-20034a1bc7ad-nova-migration-ssh-key-1" (OuterVolumeSpecName: "nova-migration-ssh-key-1") pod "f9fc3f8e-9635-48a4-8ec3-20034a1bc7ad" (UID: "f9fc3f8e-9635-48a4-8ec3-20034a1bc7ad"). InnerVolumeSpecName "nova-migration-ssh-key-1". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 10 23:31:46 crc kubenswrapper[4791]: I1210 23:31:46.466761 4791 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/f9fc3f8e-9635-48a4-8ec3-20034a1bc7ad-inventory\") on node \"crc\" DevicePath \"\"" Dec 10 23:31:46 crc kubenswrapper[4791]: I1210 23:31:46.466808 4791 reconciler_common.go:293] "Volume detached for volume \"nova-cell1-compute-config-0\" (UniqueName: \"kubernetes.io/secret/f9fc3f8e-9635-48a4-8ec3-20034a1bc7ad-nova-cell1-compute-config-0\") on node \"crc\" DevicePath \"\"" Dec 10 23:31:46 crc kubenswrapper[4791]: I1210 23:31:46.466828 4791 reconciler_common.go:293] "Volume detached for volume \"nova-extra-config-0\" (UniqueName: \"kubernetes.io/configmap/f9fc3f8e-9635-48a4-8ec3-20034a1bc7ad-nova-extra-config-0\") on node \"crc\" DevicePath \"\"" Dec 10 23:31:46 crc kubenswrapper[4791]: I1210 23:31:46.466851 4791 reconciler_common.go:293] "Volume detached for volume \"nova-cell1-compute-config-1\" (UniqueName: \"kubernetes.io/secret/f9fc3f8e-9635-48a4-8ec3-20034a1bc7ad-nova-cell1-compute-config-1\") on node \"crc\" DevicePath \"\"" Dec 10 23:31:46 crc kubenswrapper[4791]: I1210 23:31:46.466861 4791 reconciler_common.go:293] "Volume detached for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/f9fc3f8e-9635-48a4-8ec3-20034a1bc7ad-ssh-key\") on node \"crc\" DevicePath \"\"" Dec 10 23:31:46 crc kubenswrapper[4791]: I1210 23:31:46.466870 4791 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-475wv\" (UniqueName: \"kubernetes.io/projected/f9fc3f8e-9635-48a4-8ec3-20034a1bc7ad-kube-api-access-475wv\") on node \"crc\" DevicePath \"\"" Dec 10 23:31:46 crc kubenswrapper[4791]: I1210 23:31:46.466883 4791 reconciler_common.go:293] "Volume detached for volume \"nova-migration-ssh-key-0\" (UniqueName: \"kubernetes.io/secret/f9fc3f8e-9635-48a4-8ec3-20034a1bc7ad-nova-migration-ssh-key-0\") on node \"crc\" DevicePath \"\"" Dec 10 23:31:46 crc kubenswrapper[4791]: I1210 23:31:46.466892 4791 reconciler_common.go:293] "Volume detached for volume \"nova-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f9fc3f8e-9635-48a4-8ec3-20034a1bc7ad-nova-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 10 23:31:46 crc kubenswrapper[4791]: I1210 23:31:46.466901 4791 reconciler_common.go:293] "Volume detached for volume \"nova-migration-ssh-key-1\" (UniqueName: \"kubernetes.io/secret/f9fc3f8e-9635-48a4-8ec3-20034a1bc7ad-nova-migration-ssh-key-1\") on node \"crc\" DevicePath \"\"" Dec 10 23:31:46 crc kubenswrapper[4791]: I1210 23:31:46.755310 4791 generic.go:334] "Generic (PLEG): container finished" podID="a7300d58-0959-4835-b910-5dd33161afb8" containerID="269088158da961ee18243c60db5e35db13df272826622d5f78d410bbcde32f2b" exitCode=0 Dec 10 23:31:46 crc kubenswrapper[4791]: I1210 23:31:46.755921 4791 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-qwr9t" event={"ID":"a7300d58-0959-4835-b910-5dd33161afb8","Type":"ContainerDied","Data":"269088158da961ee18243c60db5e35db13df272826622d5f78d410bbcde32f2b"} Dec 10 23:31:46 crc kubenswrapper[4791]: I1210 23:31:46.755951 4791 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-qwr9t" event={"ID":"a7300d58-0959-4835-b910-5dd33161afb8","Type":"ContainerStarted","Data":"dc25bdfac1545c6b303417b8d0b9fece7bf02b368e465c029c8268a931334bc7"} Dec 10 23:31:46 crc kubenswrapper[4791]: I1210 23:31:46.758200 4791 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-5krlr" event={"ID":"f9fc3f8e-9635-48a4-8ec3-20034a1bc7ad","Type":"ContainerDied","Data":"9202e0c33e0c04504153b8fb76d4d02d9260df574897a27ad6f0bc8f15639b64"} Dec 10 23:31:46 crc kubenswrapper[4791]: I1210 23:31:46.758231 4791 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="9202e0c33e0c04504153b8fb76d4d02d9260df574897a27ad6f0bc8f15639b64" Dec 10 23:31:46 crc kubenswrapper[4791]: I1210 23:31:46.758385 4791 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-5krlr" Dec 10 23:31:46 crc kubenswrapper[4791]: I1210 23:31:46.873211 4791 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/telemetry-edpm-deployment-openstack-edpm-ipam-kxtwt"] Dec 10 23:31:46 crc kubenswrapper[4791]: E1210 23:31:46.873666 4791 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f9fc3f8e-9635-48a4-8ec3-20034a1bc7ad" containerName="nova-edpm-deployment-openstack-edpm-ipam" Dec 10 23:31:46 crc kubenswrapper[4791]: I1210 23:31:46.873688 4791 state_mem.go:107] "Deleted CPUSet assignment" podUID="f9fc3f8e-9635-48a4-8ec3-20034a1bc7ad" containerName="nova-edpm-deployment-openstack-edpm-ipam" Dec 10 23:31:46 crc kubenswrapper[4791]: I1210 23:31:46.873868 4791 memory_manager.go:354] "RemoveStaleState removing state" podUID="f9fc3f8e-9635-48a4-8ec3-20034a1bc7ad" containerName="nova-edpm-deployment-openstack-edpm-ipam" Dec 10 23:31:46 crc kubenswrapper[4791]: I1210 23:31:46.874639 4791 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-kxtwt" Dec 10 23:31:46 crc kubenswrapper[4791]: I1210 23:31:46.876718 4791 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Dec 10 23:31:46 crc kubenswrapper[4791]: I1210 23:31:46.876768 4791 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-ansible-ssh-private-key-secret" Dec 10 23:31:46 crc kubenswrapper[4791]: I1210 23:31:46.876717 4791 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-compute-config-data" Dec 10 23:31:46 crc kubenswrapper[4791]: I1210 23:31:46.877521 4791 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-edpm-ipam" Dec 10 23:31:46 crc kubenswrapper[4791]: I1210 23:31:46.880261 4791 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-dockercfg-hf8pr" Dec 10 23:31:46 crc kubenswrapper[4791]: I1210 23:31:46.882056 4791 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/telemetry-edpm-deployment-openstack-edpm-ipam-kxtwt"] Dec 10 23:31:46 crc kubenswrapper[4791]: I1210 23:31:46.980060 4791 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ceilometer-compute-config-data-1\" (UniqueName: \"kubernetes.io/secret/282a792a-a4a9-4e6d-b7ae-f5cedbe55787-ceilometer-compute-config-data-1\") pod \"telemetry-edpm-deployment-openstack-edpm-ipam-kxtwt\" (UID: \"282a792a-a4a9-4e6d-b7ae-f5cedbe55787\") " pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-kxtwt" Dec 10 23:31:46 crc kubenswrapper[4791]: I1210 23:31:46.980149 4791 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ceilometer-compute-config-data-0\" (UniqueName: \"kubernetes.io/secret/282a792a-a4a9-4e6d-b7ae-f5cedbe55787-ceilometer-compute-config-data-0\") pod \"telemetry-edpm-deployment-openstack-edpm-ipam-kxtwt\" (UID: \"282a792a-a4a9-4e6d-b7ae-f5cedbe55787\") " pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-kxtwt" Dec 10 23:31:46 crc kubenswrapper[4791]: I1210 23:31:46.980179 4791 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/282a792a-a4a9-4e6d-b7ae-f5cedbe55787-ssh-key\") pod \"telemetry-edpm-deployment-openstack-edpm-ipam-kxtwt\" (UID: \"282a792a-a4a9-4e6d-b7ae-f5cedbe55787\") " pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-kxtwt" Dec 10 23:31:46 crc kubenswrapper[4791]: I1210 23:31:46.980256 4791 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-lwfsw\" (UniqueName: \"kubernetes.io/projected/282a792a-a4a9-4e6d-b7ae-f5cedbe55787-kube-api-access-lwfsw\") pod \"telemetry-edpm-deployment-openstack-edpm-ipam-kxtwt\" (UID: \"282a792a-a4a9-4e6d-b7ae-f5cedbe55787\") " pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-kxtwt" Dec 10 23:31:46 crc kubenswrapper[4791]: I1210 23:31:46.980303 4791 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/282a792a-a4a9-4e6d-b7ae-f5cedbe55787-inventory\") pod \"telemetry-edpm-deployment-openstack-edpm-ipam-kxtwt\" (UID: \"282a792a-a4a9-4e6d-b7ae-f5cedbe55787\") " pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-kxtwt" Dec 10 23:31:46 crc kubenswrapper[4791]: I1210 23:31:46.980348 4791 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ceilometer-compute-config-data-2\" (UniqueName: \"kubernetes.io/secret/282a792a-a4a9-4e6d-b7ae-f5cedbe55787-ceilometer-compute-config-data-2\") pod \"telemetry-edpm-deployment-openstack-edpm-ipam-kxtwt\" (UID: \"282a792a-a4a9-4e6d-b7ae-f5cedbe55787\") " pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-kxtwt" Dec 10 23:31:46 crc kubenswrapper[4791]: I1210 23:31:46.980398 4791 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"telemetry-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/282a792a-a4a9-4e6d-b7ae-f5cedbe55787-telemetry-combined-ca-bundle\") pod \"telemetry-edpm-deployment-openstack-edpm-ipam-kxtwt\" (UID: \"282a792a-a4a9-4e6d-b7ae-f5cedbe55787\") " pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-kxtwt" Dec 10 23:31:47 crc kubenswrapper[4791]: I1210 23:31:47.082380 4791 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/282a792a-a4a9-4e6d-b7ae-f5cedbe55787-inventory\") pod \"telemetry-edpm-deployment-openstack-edpm-ipam-kxtwt\" (UID: \"282a792a-a4a9-4e6d-b7ae-f5cedbe55787\") " pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-kxtwt" Dec 10 23:31:47 crc kubenswrapper[4791]: I1210 23:31:47.082452 4791 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ceilometer-compute-config-data-2\" (UniqueName: \"kubernetes.io/secret/282a792a-a4a9-4e6d-b7ae-f5cedbe55787-ceilometer-compute-config-data-2\") pod \"telemetry-edpm-deployment-openstack-edpm-ipam-kxtwt\" (UID: \"282a792a-a4a9-4e6d-b7ae-f5cedbe55787\") " pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-kxtwt" Dec 10 23:31:47 crc kubenswrapper[4791]: I1210 23:31:47.082493 4791 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"telemetry-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/282a792a-a4a9-4e6d-b7ae-f5cedbe55787-telemetry-combined-ca-bundle\") pod \"telemetry-edpm-deployment-openstack-edpm-ipam-kxtwt\" (UID: \"282a792a-a4a9-4e6d-b7ae-f5cedbe55787\") " pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-kxtwt" Dec 10 23:31:47 crc kubenswrapper[4791]: I1210 23:31:47.082582 4791 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ceilometer-compute-config-data-1\" (UniqueName: \"kubernetes.io/secret/282a792a-a4a9-4e6d-b7ae-f5cedbe55787-ceilometer-compute-config-data-1\") pod \"telemetry-edpm-deployment-openstack-edpm-ipam-kxtwt\" (UID: \"282a792a-a4a9-4e6d-b7ae-f5cedbe55787\") " pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-kxtwt" Dec 10 23:31:47 crc kubenswrapper[4791]: I1210 23:31:47.082624 4791 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ceilometer-compute-config-data-0\" (UniqueName: \"kubernetes.io/secret/282a792a-a4a9-4e6d-b7ae-f5cedbe55787-ceilometer-compute-config-data-0\") pod \"telemetry-edpm-deployment-openstack-edpm-ipam-kxtwt\" (UID: \"282a792a-a4a9-4e6d-b7ae-f5cedbe55787\") " pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-kxtwt" Dec 10 23:31:47 crc kubenswrapper[4791]: I1210 23:31:47.082645 4791 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/282a792a-a4a9-4e6d-b7ae-f5cedbe55787-ssh-key\") pod \"telemetry-edpm-deployment-openstack-edpm-ipam-kxtwt\" (UID: \"282a792a-a4a9-4e6d-b7ae-f5cedbe55787\") " pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-kxtwt" Dec 10 23:31:47 crc kubenswrapper[4791]: I1210 23:31:47.082693 4791 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-lwfsw\" (UniqueName: \"kubernetes.io/projected/282a792a-a4a9-4e6d-b7ae-f5cedbe55787-kube-api-access-lwfsw\") pod \"telemetry-edpm-deployment-openstack-edpm-ipam-kxtwt\" (UID: \"282a792a-a4a9-4e6d-b7ae-f5cedbe55787\") " pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-kxtwt" Dec 10 23:31:47 crc kubenswrapper[4791]: I1210 23:31:47.088264 4791 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/282a792a-a4a9-4e6d-b7ae-f5cedbe55787-ssh-key\") pod \"telemetry-edpm-deployment-openstack-edpm-ipam-kxtwt\" (UID: \"282a792a-a4a9-4e6d-b7ae-f5cedbe55787\") " pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-kxtwt" Dec 10 23:31:47 crc kubenswrapper[4791]: I1210 23:31:47.088424 4791 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"telemetry-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/282a792a-a4a9-4e6d-b7ae-f5cedbe55787-telemetry-combined-ca-bundle\") pod \"telemetry-edpm-deployment-openstack-edpm-ipam-kxtwt\" (UID: \"282a792a-a4a9-4e6d-b7ae-f5cedbe55787\") " pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-kxtwt" Dec 10 23:31:47 crc kubenswrapper[4791]: I1210 23:31:47.089558 4791 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/282a792a-a4a9-4e6d-b7ae-f5cedbe55787-inventory\") pod \"telemetry-edpm-deployment-openstack-edpm-ipam-kxtwt\" (UID: \"282a792a-a4a9-4e6d-b7ae-f5cedbe55787\") " pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-kxtwt" Dec 10 23:31:47 crc kubenswrapper[4791]: I1210 23:31:47.089732 4791 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ceilometer-compute-config-data-2\" (UniqueName: \"kubernetes.io/secret/282a792a-a4a9-4e6d-b7ae-f5cedbe55787-ceilometer-compute-config-data-2\") pod \"telemetry-edpm-deployment-openstack-edpm-ipam-kxtwt\" (UID: \"282a792a-a4a9-4e6d-b7ae-f5cedbe55787\") " pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-kxtwt" Dec 10 23:31:47 crc kubenswrapper[4791]: I1210 23:31:47.089843 4791 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ceilometer-compute-config-data-1\" (UniqueName: \"kubernetes.io/secret/282a792a-a4a9-4e6d-b7ae-f5cedbe55787-ceilometer-compute-config-data-1\") pod \"telemetry-edpm-deployment-openstack-edpm-ipam-kxtwt\" (UID: \"282a792a-a4a9-4e6d-b7ae-f5cedbe55787\") " pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-kxtwt" Dec 10 23:31:47 crc kubenswrapper[4791]: I1210 23:31:47.095564 4791 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ceilometer-compute-config-data-0\" (UniqueName: \"kubernetes.io/secret/282a792a-a4a9-4e6d-b7ae-f5cedbe55787-ceilometer-compute-config-data-0\") pod \"telemetry-edpm-deployment-openstack-edpm-ipam-kxtwt\" (UID: \"282a792a-a4a9-4e6d-b7ae-f5cedbe55787\") " pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-kxtwt" Dec 10 23:31:47 crc kubenswrapper[4791]: I1210 23:31:47.101872 4791 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-lwfsw\" (UniqueName: \"kubernetes.io/projected/282a792a-a4a9-4e6d-b7ae-f5cedbe55787-kube-api-access-lwfsw\") pod \"telemetry-edpm-deployment-openstack-edpm-ipam-kxtwt\" (UID: \"282a792a-a4a9-4e6d-b7ae-f5cedbe55787\") " pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-kxtwt" Dec 10 23:31:47 crc kubenswrapper[4791]: I1210 23:31:47.194751 4791 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-kxtwt" Dec 10 23:31:47 crc kubenswrapper[4791]: I1210 23:31:47.721789 4791 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/telemetry-edpm-deployment-openstack-edpm-ipam-kxtwt"] Dec 10 23:31:47 crc kubenswrapper[4791]: W1210 23:31:47.728611 4791 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod282a792a_a4a9_4e6d_b7ae_f5cedbe55787.slice/crio-0ec16ae3d5909dba618dda651fa7d95522684f31f589e263862bb54b5ce1390d WatchSource:0}: Error finding container 0ec16ae3d5909dba618dda651fa7d95522684f31f589e263862bb54b5ce1390d: Status 404 returned error can't find the container with id 0ec16ae3d5909dba618dda651fa7d95522684f31f589e263862bb54b5ce1390d Dec 10 23:31:47 crc kubenswrapper[4791]: I1210 23:31:47.773039 4791 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-kxtwt" event={"ID":"282a792a-a4a9-4e6d-b7ae-f5cedbe55787","Type":"ContainerStarted","Data":"0ec16ae3d5909dba618dda651fa7d95522684f31f589e263862bb54b5ce1390d"} Dec 10 23:31:48 crc kubenswrapper[4791]: I1210 23:31:48.885685 4791 scope.go:117] "RemoveContainer" containerID="4229e7e5318c8150a4b76b463eed8dfb67fd267a558642295fc4d4f697411536" Dec 10 23:31:48 crc kubenswrapper[4791]: E1210 23:31:48.886145 4791 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-5rb5l_openshift-machine-config-operator(ba35653c-6e06-4cee-a4d6-137764090d18)\"" pod="openshift-machine-config-operator/machine-config-daemon-5rb5l" podUID="ba35653c-6e06-4cee-a4d6-137764090d18" Dec 10 23:31:50 crc kubenswrapper[4791]: I1210 23:31:50.801791 4791 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-qwr9t" event={"ID":"a7300d58-0959-4835-b910-5dd33161afb8","Type":"ContainerStarted","Data":"370eb2048033fc3860c5bf2cb9c9853e166458fb51997617b656bc70da2d9eaf"} Dec 10 23:31:50 crc kubenswrapper[4791]: I1210 23:31:50.805631 4791 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-kxtwt" event={"ID":"282a792a-a4a9-4e6d-b7ae-f5cedbe55787","Type":"ContainerStarted","Data":"5d2a0518f345f6a1f817224c7647f2c250fd76fca7f34a509c4e79b7c3e23443"} Dec 10 23:31:50 crc kubenswrapper[4791]: I1210 23:31:50.844646 4791 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-kxtwt" podStartSLOduration=2.591168813 podStartE2EDuration="4.844627693s" podCreationTimestamp="2025-12-10 23:31:46 +0000 UTC" firstStartedPulling="2025-12-10 23:31:47.740949467 +0000 UTC m=+2542.170567080" lastFinishedPulling="2025-12-10 23:31:49.994408347 +0000 UTC m=+2544.424025960" observedRunningTime="2025-12-10 23:31:50.838750204 +0000 UTC m=+2545.268367817" watchObservedRunningTime="2025-12-10 23:31:50.844627693 +0000 UTC m=+2545.274245306" Dec 10 23:31:51 crc kubenswrapper[4791]: I1210 23:31:51.816046 4791 generic.go:334] "Generic (PLEG): container finished" podID="a7300d58-0959-4835-b910-5dd33161afb8" containerID="370eb2048033fc3860c5bf2cb9c9853e166458fb51997617b656bc70da2d9eaf" exitCode=0 Dec 10 23:31:51 crc kubenswrapper[4791]: I1210 23:31:51.816123 4791 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-qwr9t" event={"ID":"a7300d58-0959-4835-b910-5dd33161afb8","Type":"ContainerDied","Data":"370eb2048033fc3860c5bf2cb9c9853e166458fb51997617b656bc70da2d9eaf"} Dec 10 23:31:53 crc kubenswrapper[4791]: I1210 23:31:53.840992 4791 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-qwr9t" event={"ID":"a7300d58-0959-4835-b910-5dd33161afb8","Type":"ContainerStarted","Data":"6c13adbaa4e05b21f6f5bba90540720ace0393764309e4c332b1eca8e9fbf93a"} Dec 10 23:31:53 crc kubenswrapper[4791]: I1210 23:31:53.877731 4791 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/certified-operators-qwr9t" podStartSLOduration=2.978203518 podStartE2EDuration="8.877699008s" podCreationTimestamp="2025-12-10 23:31:45 +0000 UTC" firstStartedPulling="2025-12-10 23:31:46.766056668 +0000 UTC m=+2541.195674281" lastFinishedPulling="2025-12-10 23:31:52.665552118 +0000 UTC m=+2547.095169771" observedRunningTime="2025-12-10 23:31:53.862335636 +0000 UTC m=+2548.292046782" watchObservedRunningTime="2025-12-10 23:31:53.877699008 +0000 UTC m=+2548.307316661" Dec 10 23:31:55 crc kubenswrapper[4791]: I1210 23:31:55.682434 4791 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/certified-operators-qwr9t" Dec 10 23:31:55 crc kubenswrapper[4791]: I1210 23:31:55.682474 4791 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/certified-operators-qwr9t" Dec 10 23:31:55 crc kubenswrapper[4791]: I1210 23:31:55.732094 4791 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/certified-operators-qwr9t" Dec 10 23:32:01 crc kubenswrapper[4791]: I1210 23:32:01.885858 4791 scope.go:117] "RemoveContainer" containerID="4229e7e5318c8150a4b76b463eed8dfb67fd267a558642295fc4d4f697411536" Dec 10 23:32:02 crc kubenswrapper[4791]: I1210 23:32:02.964754 4791 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-5rb5l" event={"ID":"ba35653c-6e06-4cee-a4d6-137764090d18","Type":"ContainerStarted","Data":"6982a96e9311c148f4ebb66ec887daeec48d7fbda61684c0ff78ee7d6d234688"} Dec 10 23:32:05 crc kubenswrapper[4791]: I1210 23:32:05.746791 4791 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/certified-operators-qwr9t" Dec 10 23:32:05 crc kubenswrapper[4791]: I1210 23:32:05.814331 4791 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-qwr9t"] Dec 10 23:32:05 crc kubenswrapper[4791]: I1210 23:32:05.996440 4791 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/certified-operators-qwr9t" podUID="a7300d58-0959-4835-b910-5dd33161afb8" containerName="registry-server" containerID="cri-o://6c13adbaa4e05b21f6f5bba90540720ace0393764309e4c332b1eca8e9fbf93a" gracePeriod=2 Dec 10 23:32:06 crc kubenswrapper[4791]: I1210 23:32:06.549306 4791 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-qwr9t" Dec 10 23:32:06 crc kubenswrapper[4791]: I1210 23:32:06.721267 4791 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/a7300d58-0959-4835-b910-5dd33161afb8-utilities\") pod \"a7300d58-0959-4835-b910-5dd33161afb8\" (UID: \"a7300d58-0959-4835-b910-5dd33161afb8\") " Dec 10 23:32:06 crc kubenswrapper[4791]: I1210 23:32:06.721402 4791 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-vlwfs\" (UniqueName: \"kubernetes.io/projected/a7300d58-0959-4835-b910-5dd33161afb8-kube-api-access-vlwfs\") pod \"a7300d58-0959-4835-b910-5dd33161afb8\" (UID: \"a7300d58-0959-4835-b910-5dd33161afb8\") " Dec 10 23:32:06 crc kubenswrapper[4791]: I1210 23:32:06.721630 4791 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/a7300d58-0959-4835-b910-5dd33161afb8-catalog-content\") pod \"a7300d58-0959-4835-b910-5dd33161afb8\" (UID: \"a7300d58-0959-4835-b910-5dd33161afb8\") " Dec 10 23:32:06 crc kubenswrapper[4791]: I1210 23:32:06.725401 4791 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/a7300d58-0959-4835-b910-5dd33161afb8-utilities" (OuterVolumeSpecName: "utilities") pod "a7300d58-0959-4835-b910-5dd33161afb8" (UID: "a7300d58-0959-4835-b910-5dd33161afb8"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 10 23:32:06 crc kubenswrapper[4791]: I1210 23:32:06.733510 4791 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/a7300d58-0959-4835-b910-5dd33161afb8-kube-api-access-vlwfs" (OuterVolumeSpecName: "kube-api-access-vlwfs") pod "a7300d58-0959-4835-b910-5dd33161afb8" (UID: "a7300d58-0959-4835-b910-5dd33161afb8"). InnerVolumeSpecName "kube-api-access-vlwfs". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 10 23:32:06 crc kubenswrapper[4791]: I1210 23:32:06.774900 4791 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/a7300d58-0959-4835-b910-5dd33161afb8-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "a7300d58-0959-4835-b910-5dd33161afb8" (UID: "a7300d58-0959-4835-b910-5dd33161afb8"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 10 23:32:06 crc kubenswrapper[4791]: I1210 23:32:06.823996 4791 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/a7300d58-0959-4835-b910-5dd33161afb8-catalog-content\") on node \"crc\" DevicePath \"\"" Dec 10 23:32:06 crc kubenswrapper[4791]: I1210 23:32:06.824031 4791 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/a7300d58-0959-4835-b910-5dd33161afb8-utilities\") on node \"crc\" DevicePath \"\"" Dec 10 23:32:06 crc kubenswrapper[4791]: I1210 23:32:06.824042 4791 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-vlwfs\" (UniqueName: \"kubernetes.io/projected/a7300d58-0959-4835-b910-5dd33161afb8-kube-api-access-vlwfs\") on node \"crc\" DevicePath \"\"" Dec 10 23:32:07 crc kubenswrapper[4791]: I1210 23:32:07.011129 4791 generic.go:334] "Generic (PLEG): container finished" podID="a7300d58-0959-4835-b910-5dd33161afb8" containerID="6c13adbaa4e05b21f6f5bba90540720ace0393764309e4c332b1eca8e9fbf93a" exitCode=0 Dec 10 23:32:07 crc kubenswrapper[4791]: I1210 23:32:07.011180 4791 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-qwr9t" event={"ID":"a7300d58-0959-4835-b910-5dd33161afb8","Type":"ContainerDied","Data":"6c13adbaa4e05b21f6f5bba90540720ace0393764309e4c332b1eca8e9fbf93a"} Dec 10 23:32:07 crc kubenswrapper[4791]: I1210 23:32:07.011235 4791 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-qwr9t" event={"ID":"a7300d58-0959-4835-b910-5dd33161afb8","Type":"ContainerDied","Data":"dc25bdfac1545c6b303417b8d0b9fece7bf02b368e465c029c8268a931334bc7"} Dec 10 23:32:07 crc kubenswrapper[4791]: I1210 23:32:07.011270 4791 scope.go:117] "RemoveContainer" containerID="6c13adbaa4e05b21f6f5bba90540720ace0393764309e4c332b1eca8e9fbf93a" Dec 10 23:32:07 crc kubenswrapper[4791]: I1210 23:32:07.011332 4791 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-qwr9t" Dec 10 23:32:07 crc kubenswrapper[4791]: I1210 23:32:07.054406 4791 scope.go:117] "RemoveContainer" containerID="370eb2048033fc3860c5bf2cb9c9853e166458fb51997617b656bc70da2d9eaf" Dec 10 23:32:07 crc kubenswrapper[4791]: I1210 23:32:07.055089 4791 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-qwr9t"] Dec 10 23:32:07 crc kubenswrapper[4791]: I1210 23:32:07.062970 4791 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/certified-operators-qwr9t"] Dec 10 23:32:07 crc kubenswrapper[4791]: I1210 23:32:07.090947 4791 scope.go:117] "RemoveContainer" containerID="269088158da961ee18243c60db5e35db13df272826622d5f78d410bbcde32f2b" Dec 10 23:32:07 crc kubenswrapper[4791]: I1210 23:32:07.134881 4791 scope.go:117] "RemoveContainer" containerID="6c13adbaa4e05b21f6f5bba90540720ace0393764309e4c332b1eca8e9fbf93a" Dec 10 23:32:07 crc kubenswrapper[4791]: E1210 23:32:07.135389 4791 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"6c13adbaa4e05b21f6f5bba90540720ace0393764309e4c332b1eca8e9fbf93a\": container with ID starting with 6c13adbaa4e05b21f6f5bba90540720ace0393764309e4c332b1eca8e9fbf93a not found: ID does not exist" containerID="6c13adbaa4e05b21f6f5bba90540720ace0393764309e4c332b1eca8e9fbf93a" Dec 10 23:32:07 crc kubenswrapper[4791]: I1210 23:32:07.135425 4791 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"6c13adbaa4e05b21f6f5bba90540720ace0393764309e4c332b1eca8e9fbf93a"} err="failed to get container status \"6c13adbaa4e05b21f6f5bba90540720ace0393764309e4c332b1eca8e9fbf93a\": rpc error: code = NotFound desc = could not find container \"6c13adbaa4e05b21f6f5bba90540720ace0393764309e4c332b1eca8e9fbf93a\": container with ID starting with 6c13adbaa4e05b21f6f5bba90540720ace0393764309e4c332b1eca8e9fbf93a not found: ID does not exist" Dec 10 23:32:07 crc kubenswrapper[4791]: I1210 23:32:07.135445 4791 scope.go:117] "RemoveContainer" containerID="370eb2048033fc3860c5bf2cb9c9853e166458fb51997617b656bc70da2d9eaf" Dec 10 23:32:07 crc kubenswrapper[4791]: E1210 23:32:07.135795 4791 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"370eb2048033fc3860c5bf2cb9c9853e166458fb51997617b656bc70da2d9eaf\": container with ID starting with 370eb2048033fc3860c5bf2cb9c9853e166458fb51997617b656bc70da2d9eaf not found: ID does not exist" containerID="370eb2048033fc3860c5bf2cb9c9853e166458fb51997617b656bc70da2d9eaf" Dec 10 23:32:07 crc kubenswrapper[4791]: I1210 23:32:07.135839 4791 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"370eb2048033fc3860c5bf2cb9c9853e166458fb51997617b656bc70da2d9eaf"} err="failed to get container status \"370eb2048033fc3860c5bf2cb9c9853e166458fb51997617b656bc70da2d9eaf\": rpc error: code = NotFound desc = could not find container \"370eb2048033fc3860c5bf2cb9c9853e166458fb51997617b656bc70da2d9eaf\": container with ID starting with 370eb2048033fc3860c5bf2cb9c9853e166458fb51997617b656bc70da2d9eaf not found: ID does not exist" Dec 10 23:32:07 crc kubenswrapper[4791]: I1210 23:32:07.135873 4791 scope.go:117] "RemoveContainer" containerID="269088158da961ee18243c60db5e35db13df272826622d5f78d410bbcde32f2b" Dec 10 23:32:07 crc kubenswrapper[4791]: E1210 23:32:07.136198 4791 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"269088158da961ee18243c60db5e35db13df272826622d5f78d410bbcde32f2b\": container with ID starting with 269088158da961ee18243c60db5e35db13df272826622d5f78d410bbcde32f2b not found: ID does not exist" containerID="269088158da961ee18243c60db5e35db13df272826622d5f78d410bbcde32f2b" Dec 10 23:32:07 crc kubenswrapper[4791]: I1210 23:32:07.136226 4791 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"269088158da961ee18243c60db5e35db13df272826622d5f78d410bbcde32f2b"} err="failed to get container status \"269088158da961ee18243c60db5e35db13df272826622d5f78d410bbcde32f2b\": rpc error: code = NotFound desc = could not find container \"269088158da961ee18243c60db5e35db13df272826622d5f78d410bbcde32f2b\": container with ID starting with 269088158da961ee18243c60db5e35db13df272826622d5f78d410bbcde32f2b not found: ID does not exist" Dec 10 23:32:07 crc kubenswrapper[4791]: I1210 23:32:07.898777 4791 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="a7300d58-0959-4835-b910-5dd33161afb8" path="/var/lib/kubelet/pods/a7300d58-0959-4835-b910-5dd33161afb8/volumes" Dec 10 23:34:05 crc kubenswrapper[4791]: I1210 23:34:05.289451 4791 generic.go:334] "Generic (PLEG): container finished" podID="282a792a-a4a9-4e6d-b7ae-f5cedbe55787" containerID="5d2a0518f345f6a1f817224c7647f2c250fd76fca7f34a509c4e79b7c3e23443" exitCode=0 Dec 10 23:34:05 crc kubenswrapper[4791]: I1210 23:34:05.289516 4791 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-kxtwt" event={"ID":"282a792a-a4a9-4e6d-b7ae-f5cedbe55787","Type":"ContainerDied","Data":"5d2a0518f345f6a1f817224c7647f2c250fd76fca7f34a509c4e79b7c3e23443"} Dec 10 23:34:06 crc kubenswrapper[4791]: I1210 23:34:06.737551 4791 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-kxtwt" Dec 10 23:34:06 crc kubenswrapper[4791]: I1210 23:34:06.868588 4791 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ceilometer-compute-config-data-2\" (UniqueName: \"kubernetes.io/secret/282a792a-a4a9-4e6d-b7ae-f5cedbe55787-ceilometer-compute-config-data-2\") pod \"282a792a-a4a9-4e6d-b7ae-f5cedbe55787\" (UID: \"282a792a-a4a9-4e6d-b7ae-f5cedbe55787\") " Dec 10 23:34:06 crc kubenswrapper[4791]: I1210 23:34:06.868632 4791 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"telemetry-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/282a792a-a4a9-4e6d-b7ae-f5cedbe55787-telemetry-combined-ca-bundle\") pod \"282a792a-a4a9-4e6d-b7ae-f5cedbe55787\" (UID: \"282a792a-a4a9-4e6d-b7ae-f5cedbe55787\") " Dec 10 23:34:06 crc kubenswrapper[4791]: I1210 23:34:06.868709 4791 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-lwfsw\" (UniqueName: \"kubernetes.io/projected/282a792a-a4a9-4e6d-b7ae-f5cedbe55787-kube-api-access-lwfsw\") pod \"282a792a-a4a9-4e6d-b7ae-f5cedbe55787\" (UID: \"282a792a-a4a9-4e6d-b7ae-f5cedbe55787\") " Dec 10 23:34:06 crc kubenswrapper[4791]: I1210 23:34:06.868776 4791 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ceilometer-compute-config-data-1\" (UniqueName: \"kubernetes.io/secret/282a792a-a4a9-4e6d-b7ae-f5cedbe55787-ceilometer-compute-config-data-1\") pod \"282a792a-a4a9-4e6d-b7ae-f5cedbe55787\" (UID: \"282a792a-a4a9-4e6d-b7ae-f5cedbe55787\") " Dec 10 23:34:06 crc kubenswrapper[4791]: I1210 23:34:06.868824 4791 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ceilometer-compute-config-data-0\" (UniqueName: \"kubernetes.io/secret/282a792a-a4a9-4e6d-b7ae-f5cedbe55787-ceilometer-compute-config-data-0\") pod \"282a792a-a4a9-4e6d-b7ae-f5cedbe55787\" (UID: \"282a792a-a4a9-4e6d-b7ae-f5cedbe55787\") " Dec 10 23:34:06 crc kubenswrapper[4791]: I1210 23:34:06.868851 4791 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/282a792a-a4a9-4e6d-b7ae-f5cedbe55787-ssh-key\") pod \"282a792a-a4a9-4e6d-b7ae-f5cedbe55787\" (UID: \"282a792a-a4a9-4e6d-b7ae-f5cedbe55787\") " Dec 10 23:34:06 crc kubenswrapper[4791]: I1210 23:34:06.868895 4791 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/282a792a-a4a9-4e6d-b7ae-f5cedbe55787-inventory\") pod \"282a792a-a4a9-4e6d-b7ae-f5cedbe55787\" (UID: \"282a792a-a4a9-4e6d-b7ae-f5cedbe55787\") " Dec 10 23:34:06 crc kubenswrapper[4791]: I1210 23:34:06.873921 4791 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/282a792a-a4a9-4e6d-b7ae-f5cedbe55787-telemetry-combined-ca-bundle" (OuterVolumeSpecName: "telemetry-combined-ca-bundle") pod "282a792a-a4a9-4e6d-b7ae-f5cedbe55787" (UID: "282a792a-a4a9-4e6d-b7ae-f5cedbe55787"). InnerVolumeSpecName "telemetry-combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 10 23:34:06 crc kubenswrapper[4791]: I1210 23:34:06.874412 4791 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/282a792a-a4a9-4e6d-b7ae-f5cedbe55787-kube-api-access-lwfsw" (OuterVolumeSpecName: "kube-api-access-lwfsw") pod "282a792a-a4a9-4e6d-b7ae-f5cedbe55787" (UID: "282a792a-a4a9-4e6d-b7ae-f5cedbe55787"). InnerVolumeSpecName "kube-api-access-lwfsw". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 10 23:34:06 crc kubenswrapper[4791]: I1210 23:34:06.897698 4791 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/282a792a-a4a9-4e6d-b7ae-f5cedbe55787-ssh-key" (OuterVolumeSpecName: "ssh-key") pod "282a792a-a4a9-4e6d-b7ae-f5cedbe55787" (UID: "282a792a-a4a9-4e6d-b7ae-f5cedbe55787"). InnerVolumeSpecName "ssh-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 10 23:34:06 crc kubenswrapper[4791]: I1210 23:34:06.898116 4791 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/282a792a-a4a9-4e6d-b7ae-f5cedbe55787-ceilometer-compute-config-data-1" (OuterVolumeSpecName: "ceilometer-compute-config-data-1") pod "282a792a-a4a9-4e6d-b7ae-f5cedbe55787" (UID: "282a792a-a4a9-4e6d-b7ae-f5cedbe55787"). InnerVolumeSpecName "ceilometer-compute-config-data-1". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 10 23:34:06 crc kubenswrapper[4791]: I1210 23:34:06.900281 4791 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/282a792a-a4a9-4e6d-b7ae-f5cedbe55787-ceilometer-compute-config-data-2" (OuterVolumeSpecName: "ceilometer-compute-config-data-2") pod "282a792a-a4a9-4e6d-b7ae-f5cedbe55787" (UID: "282a792a-a4a9-4e6d-b7ae-f5cedbe55787"). InnerVolumeSpecName "ceilometer-compute-config-data-2". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 10 23:34:06 crc kubenswrapper[4791]: I1210 23:34:06.904030 4791 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/282a792a-a4a9-4e6d-b7ae-f5cedbe55787-ceilometer-compute-config-data-0" (OuterVolumeSpecName: "ceilometer-compute-config-data-0") pod "282a792a-a4a9-4e6d-b7ae-f5cedbe55787" (UID: "282a792a-a4a9-4e6d-b7ae-f5cedbe55787"). InnerVolumeSpecName "ceilometer-compute-config-data-0". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 10 23:34:06 crc kubenswrapper[4791]: I1210 23:34:06.904205 4791 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/282a792a-a4a9-4e6d-b7ae-f5cedbe55787-inventory" (OuterVolumeSpecName: "inventory") pod "282a792a-a4a9-4e6d-b7ae-f5cedbe55787" (UID: "282a792a-a4a9-4e6d-b7ae-f5cedbe55787"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 10 23:34:06 crc kubenswrapper[4791]: I1210 23:34:06.974678 4791 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/282a792a-a4a9-4e6d-b7ae-f5cedbe55787-inventory\") on node \"crc\" DevicePath \"\"" Dec 10 23:34:06 crc kubenswrapper[4791]: I1210 23:34:06.975020 4791 reconciler_common.go:293] "Volume detached for volume \"ceilometer-compute-config-data-2\" (UniqueName: \"kubernetes.io/secret/282a792a-a4a9-4e6d-b7ae-f5cedbe55787-ceilometer-compute-config-data-2\") on node \"crc\" DevicePath \"\"" Dec 10 23:34:06 crc kubenswrapper[4791]: I1210 23:34:06.975033 4791 reconciler_common.go:293] "Volume detached for volume \"telemetry-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/282a792a-a4a9-4e6d-b7ae-f5cedbe55787-telemetry-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 10 23:34:06 crc kubenswrapper[4791]: I1210 23:34:06.975044 4791 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-lwfsw\" (UniqueName: \"kubernetes.io/projected/282a792a-a4a9-4e6d-b7ae-f5cedbe55787-kube-api-access-lwfsw\") on node \"crc\" DevicePath \"\"" Dec 10 23:34:06 crc kubenswrapper[4791]: I1210 23:34:06.975053 4791 reconciler_common.go:293] "Volume detached for volume \"ceilometer-compute-config-data-1\" (UniqueName: \"kubernetes.io/secret/282a792a-a4a9-4e6d-b7ae-f5cedbe55787-ceilometer-compute-config-data-1\") on node \"crc\" DevicePath \"\"" Dec 10 23:34:06 crc kubenswrapper[4791]: I1210 23:34:06.975061 4791 reconciler_common.go:293] "Volume detached for volume \"ceilometer-compute-config-data-0\" (UniqueName: \"kubernetes.io/secret/282a792a-a4a9-4e6d-b7ae-f5cedbe55787-ceilometer-compute-config-data-0\") on node \"crc\" DevicePath \"\"" Dec 10 23:34:06 crc kubenswrapper[4791]: I1210 23:34:06.975088 4791 reconciler_common.go:293] "Volume detached for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/282a792a-a4a9-4e6d-b7ae-f5cedbe55787-ssh-key\") on node \"crc\" DevicePath \"\"" Dec 10 23:34:07 crc kubenswrapper[4791]: I1210 23:34:07.312865 4791 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-kxtwt" event={"ID":"282a792a-a4a9-4e6d-b7ae-f5cedbe55787","Type":"ContainerDied","Data":"0ec16ae3d5909dba618dda651fa7d95522684f31f589e263862bb54b5ce1390d"} Dec 10 23:34:07 crc kubenswrapper[4791]: I1210 23:34:07.312907 4791 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="0ec16ae3d5909dba618dda651fa7d95522684f31f589e263862bb54b5ce1390d" Dec 10 23:34:07 crc kubenswrapper[4791]: I1210 23:34:07.312913 4791 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-kxtwt" Dec 10 23:34:25 crc kubenswrapper[4791]: I1210 23:34:25.038202 4791 patch_prober.go:28] interesting pod/machine-config-daemon-5rb5l container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 10 23:34:25 crc kubenswrapper[4791]: I1210 23:34:25.038896 4791 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-5rb5l" podUID="ba35653c-6e06-4cee-a4d6-137764090d18" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 10 23:34:55 crc kubenswrapper[4791]: I1210 23:34:55.038735 4791 patch_prober.go:28] interesting pod/machine-config-daemon-5rb5l container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 10 23:34:55 crc kubenswrapper[4791]: I1210 23:34:55.039435 4791 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-5rb5l" podUID="ba35653c-6e06-4cee-a4d6-137764090d18" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 10 23:35:04 crc kubenswrapper[4791]: I1210 23:35:04.788218 4791 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/tempest-tests-tempest"] Dec 10 23:35:04 crc kubenswrapper[4791]: E1210 23:35:04.789988 4791 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="a7300d58-0959-4835-b910-5dd33161afb8" containerName="registry-server" Dec 10 23:35:04 crc kubenswrapper[4791]: I1210 23:35:04.790071 4791 state_mem.go:107] "Deleted CPUSet assignment" podUID="a7300d58-0959-4835-b910-5dd33161afb8" containerName="registry-server" Dec 10 23:35:04 crc kubenswrapper[4791]: E1210 23:35:04.790163 4791 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="a7300d58-0959-4835-b910-5dd33161afb8" containerName="extract-content" Dec 10 23:35:04 crc kubenswrapper[4791]: I1210 23:35:04.790221 4791 state_mem.go:107] "Deleted CPUSet assignment" podUID="a7300d58-0959-4835-b910-5dd33161afb8" containerName="extract-content" Dec 10 23:35:04 crc kubenswrapper[4791]: E1210 23:35:04.790302 4791 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="282a792a-a4a9-4e6d-b7ae-f5cedbe55787" containerName="telemetry-edpm-deployment-openstack-edpm-ipam" Dec 10 23:35:04 crc kubenswrapper[4791]: I1210 23:35:04.790396 4791 state_mem.go:107] "Deleted CPUSet assignment" podUID="282a792a-a4a9-4e6d-b7ae-f5cedbe55787" containerName="telemetry-edpm-deployment-openstack-edpm-ipam" Dec 10 23:35:04 crc kubenswrapper[4791]: E1210 23:35:04.790463 4791 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="a7300d58-0959-4835-b910-5dd33161afb8" containerName="extract-utilities" Dec 10 23:35:04 crc kubenswrapper[4791]: I1210 23:35:04.790518 4791 state_mem.go:107] "Deleted CPUSet assignment" podUID="a7300d58-0959-4835-b910-5dd33161afb8" containerName="extract-utilities" Dec 10 23:35:04 crc kubenswrapper[4791]: I1210 23:35:04.790738 4791 memory_manager.go:354] "RemoveStaleState removing state" podUID="282a792a-a4a9-4e6d-b7ae-f5cedbe55787" containerName="telemetry-edpm-deployment-openstack-edpm-ipam" Dec 10 23:35:04 crc kubenswrapper[4791]: I1210 23:35:04.790809 4791 memory_manager.go:354] "RemoveStaleState removing state" podUID="a7300d58-0959-4835-b910-5dd33161afb8" containerName="registry-server" Dec 10 23:35:04 crc kubenswrapper[4791]: I1210 23:35:04.791504 4791 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/tempest-tests-tempest" Dec 10 23:35:04 crc kubenswrapper[4791]: I1210 23:35:04.794546 4791 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"tempest-tests-tempest-custom-data-s0" Dec 10 23:35:04 crc kubenswrapper[4791]: I1210 23:35:04.795300 4791 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"test-operator-controller-priv-key" Dec 10 23:35:04 crc kubenswrapper[4791]: I1210 23:35:04.796171 4791 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"default-dockercfg-2z5p4" Dec 10 23:35:04 crc kubenswrapper[4791]: I1210 23:35:04.797615 4791 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"tempest-tests-tempest-env-vars-s0" Dec 10 23:35:04 crc kubenswrapper[4791]: I1210 23:35:04.809243 4791 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/tempest-tests-tempest"] Dec 10 23:35:04 crc kubenswrapper[4791]: I1210 23:35:04.949673 4791 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/cf11ce8f-d433-49ff-9390-064a2f4ec60d-ssh-key\") pod \"tempest-tests-tempest\" (UID: \"cf11ce8f-d433-49ff-9390-064a2f4ec60d\") " pod="openstack/tempest-tests-tempest" Dec 10 23:35:04 crc kubenswrapper[4791]: I1210 23:35:04.949728 4791 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ca-certs\" (UniqueName: \"kubernetes.io/secret/cf11ce8f-d433-49ff-9390-064a2f4ec60d-ca-certs\") pod \"tempest-tests-tempest\" (UID: \"cf11ce8f-d433-49ff-9390-064a2f4ec60d\") " pod="openstack/tempest-tests-tempest" Dec 10 23:35:04 crc kubenswrapper[4791]: I1210 23:35:04.949879 4791 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"test-operator-ephemeral-temporary\" (UniqueName: \"kubernetes.io/empty-dir/cf11ce8f-d433-49ff-9390-064a2f4ec60d-test-operator-ephemeral-temporary\") pod \"tempest-tests-tempest\" (UID: \"cf11ce8f-d433-49ff-9390-064a2f4ec60d\") " pod="openstack/tempest-tests-tempest" Dec 10 23:35:04 crc kubenswrapper[4791]: I1210 23:35:04.949933 4791 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"test-operator-ephemeral-workdir\" (UniqueName: \"kubernetes.io/empty-dir/cf11ce8f-d433-49ff-9390-064a2f4ec60d-test-operator-ephemeral-workdir\") pod \"tempest-tests-tempest\" (UID: \"cf11ce8f-d433-49ff-9390-064a2f4ec60d\") " pod="openstack/tempest-tests-tempest" Dec 10 23:35:04 crc kubenswrapper[4791]: I1210 23:35:04.950058 4791 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage05-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage05-crc\") pod \"tempest-tests-tempest\" (UID: \"cf11ce8f-d433-49ff-9390-064a2f4ec60d\") " pod="openstack/tempest-tests-tempest" Dec 10 23:35:04 crc kubenswrapper[4791]: I1210 23:35:04.950201 4791 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-5q7hd\" (UniqueName: \"kubernetes.io/projected/cf11ce8f-d433-49ff-9390-064a2f4ec60d-kube-api-access-5q7hd\") pod \"tempest-tests-tempest\" (UID: \"cf11ce8f-d433-49ff-9390-064a2f4ec60d\") " pod="openstack/tempest-tests-tempest" Dec 10 23:35:04 crc kubenswrapper[4791]: I1210 23:35:04.950226 4791 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"openstack-config\" (UniqueName: \"kubernetes.io/configmap/cf11ce8f-d433-49ff-9390-064a2f4ec60d-openstack-config\") pod \"tempest-tests-tempest\" (UID: \"cf11ce8f-d433-49ff-9390-064a2f4ec60d\") " pod="openstack/tempest-tests-tempest" Dec 10 23:35:04 crc kubenswrapper[4791]: I1210 23:35:04.950240 4791 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"openstack-config-secret\" (UniqueName: \"kubernetes.io/secret/cf11ce8f-d433-49ff-9390-064a2f4ec60d-openstack-config-secret\") pod \"tempest-tests-tempest\" (UID: \"cf11ce8f-d433-49ff-9390-064a2f4ec60d\") " pod="openstack/tempest-tests-tempest" Dec 10 23:35:04 crc kubenswrapper[4791]: I1210 23:35:04.950256 4791 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/cf11ce8f-d433-49ff-9390-064a2f4ec60d-config-data\") pod \"tempest-tests-tempest\" (UID: \"cf11ce8f-d433-49ff-9390-064a2f4ec60d\") " pod="openstack/tempest-tests-tempest" Dec 10 23:35:05 crc kubenswrapper[4791]: I1210 23:35:05.052165 4791 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"test-operator-ephemeral-workdir\" (UniqueName: \"kubernetes.io/empty-dir/cf11ce8f-d433-49ff-9390-064a2f4ec60d-test-operator-ephemeral-workdir\") pod \"tempest-tests-tempest\" (UID: \"cf11ce8f-d433-49ff-9390-064a2f4ec60d\") " pod="openstack/tempest-tests-tempest" Dec 10 23:35:05 crc kubenswrapper[4791]: I1210 23:35:05.052323 4791 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage05-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage05-crc\") pod \"tempest-tests-tempest\" (UID: \"cf11ce8f-d433-49ff-9390-064a2f4ec60d\") " pod="openstack/tempest-tests-tempest" Dec 10 23:35:05 crc kubenswrapper[4791]: I1210 23:35:05.052508 4791 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-5q7hd\" (UniqueName: \"kubernetes.io/projected/cf11ce8f-d433-49ff-9390-064a2f4ec60d-kube-api-access-5q7hd\") pod \"tempest-tests-tempest\" (UID: \"cf11ce8f-d433-49ff-9390-064a2f4ec60d\") " pod="openstack/tempest-tests-tempest" Dec 10 23:35:05 crc kubenswrapper[4791]: I1210 23:35:05.052547 4791 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/cf11ce8f-d433-49ff-9390-064a2f4ec60d-config-data\") pod \"tempest-tests-tempest\" (UID: \"cf11ce8f-d433-49ff-9390-064a2f4ec60d\") " pod="openstack/tempest-tests-tempest" Dec 10 23:35:05 crc kubenswrapper[4791]: I1210 23:35:05.052580 4791 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"openstack-config\" (UniqueName: \"kubernetes.io/configmap/cf11ce8f-d433-49ff-9390-064a2f4ec60d-openstack-config\") pod \"tempest-tests-tempest\" (UID: \"cf11ce8f-d433-49ff-9390-064a2f4ec60d\") " pod="openstack/tempest-tests-tempest" Dec 10 23:35:05 crc kubenswrapper[4791]: I1210 23:35:05.052607 4791 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"openstack-config-secret\" (UniqueName: \"kubernetes.io/secret/cf11ce8f-d433-49ff-9390-064a2f4ec60d-openstack-config-secret\") pod \"tempest-tests-tempest\" (UID: \"cf11ce8f-d433-49ff-9390-064a2f4ec60d\") " pod="openstack/tempest-tests-tempest" Dec 10 23:35:05 crc kubenswrapper[4791]: I1210 23:35:05.052739 4791 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/cf11ce8f-d433-49ff-9390-064a2f4ec60d-ssh-key\") pod \"tempest-tests-tempest\" (UID: \"cf11ce8f-d433-49ff-9390-064a2f4ec60d\") " pod="openstack/tempest-tests-tempest" Dec 10 23:35:05 crc kubenswrapper[4791]: I1210 23:35:05.052785 4791 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ca-certs\" (UniqueName: \"kubernetes.io/secret/cf11ce8f-d433-49ff-9390-064a2f4ec60d-ca-certs\") pod \"tempest-tests-tempest\" (UID: \"cf11ce8f-d433-49ff-9390-064a2f4ec60d\") " pod="openstack/tempest-tests-tempest" Dec 10 23:35:05 crc kubenswrapper[4791]: I1210 23:35:05.052884 4791 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"test-operator-ephemeral-temporary\" (UniqueName: \"kubernetes.io/empty-dir/cf11ce8f-d433-49ff-9390-064a2f4ec60d-test-operator-ephemeral-temporary\") pod \"tempest-tests-tempest\" (UID: \"cf11ce8f-d433-49ff-9390-064a2f4ec60d\") " pod="openstack/tempest-tests-tempest" Dec 10 23:35:05 crc kubenswrapper[4791]: I1210 23:35:05.053189 4791 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage05-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage05-crc\") pod \"tempest-tests-tempest\" (UID: \"cf11ce8f-d433-49ff-9390-064a2f4ec60d\") device mount path \"/mnt/openstack/pv05\"" pod="openstack/tempest-tests-tempest" Dec 10 23:35:05 crc kubenswrapper[4791]: I1210 23:35:05.052783 4791 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"test-operator-ephemeral-workdir\" (UniqueName: \"kubernetes.io/empty-dir/cf11ce8f-d433-49ff-9390-064a2f4ec60d-test-operator-ephemeral-workdir\") pod \"tempest-tests-tempest\" (UID: \"cf11ce8f-d433-49ff-9390-064a2f4ec60d\") " pod="openstack/tempest-tests-tempest" Dec 10 23:35:05 crc kubenswrapper[4791]: I1210 23:35:05.053597 4791 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"test-operator-ephemeral-temporary\" (UniqueName: \"kubernetes.io/empty-dir/cf11ce8f-d433-49ff-9390-064a2f4ec60d-test-operator-ephemeral-temporary\") pod \"tempest-tests-tempest\" (UID: \"cf11ce8f-d433-49ff-9390-064a2f4ec60d\") " pod="openstack/tempest-tests-tempest" Dec 10 23:35:05 crc kubenswrapper[4791]: I1210 23:35:05.053772 4791 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"openstack-config\" (UniqueName: \"kubernetes.io/configmap/cf11ce8f-d433-49ff-9390-064a2f4ec60d-openstack-config\") pod \"tempest-tests-tempest\" (UID: \"cf11ce8f-d433-49ff-9390-064a2f4ec60d\") " pod="openstack/tempest-tests-tempest" Dec 10 23:35:05 crc kubenswrapper[4791]: I1210 23:35:05.054021 4791 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/cf11ce8f-d433-49ff-9390-064a2f4ec60d-config-data\") pod \"tempest-tests-tempest\" (UID: \"cf11ce8f-d433-49ff-9390-064a2f4ec60d\") " pod="openstack/tempest-tests-tempest" Dec 10 23:35:05 crc kubenswrapper[4791]: I1210 23:35:05.059822 4791 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"openstack-config-secret\" (UniqueName: \"kubernetes.io/secret/cf11ce8f-d433-49ff-9390-064a2f4ec60d-openstack-config-secret\") pod \"tempest-tests-tempest\" (UID: \"cf11ce8f-d433-49ff-9390-064a2f4ec60d\") " pod="openstack/tempest-tests-tempest" Dec 10 23:35:05 crc kubenswrapper[4791]: I1210 23:35:05.060008 4791 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ca-certs\" (UniqueName: \"kubernetes.io/secret/cf11ce8f-d433-49ff-9390-064a2f4ec60d-ca-certs\") pod \"tempest-tests-tempest\" (UID: \"cf11ce8f-d433-49ff-9390-064a2f4ec60d\") " pod="openstack/tempest-tests-tempest" Dec 10 23:35:05 crc kubenswrapper[4791]: I1210 23:35:05.060552 4791 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/cf11ce8f-d433-49ff-9390-064a2f4ec60d-ssh-key\") pod \"tempest-tests-tempest\" (UID: \"cf11ce8f-d433-49ff-9390-064a2f4ec60d\") " pod="openstack/tempest-tests-tempest" Dec 10 23:35:05 crc kubenswrapper[4791]: I1210 23:35:05.070678 4791 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-5q7hd\" (UniqueName: \"kubernetes.io/projected/cf11ce8f-d433-49ff-9390-064a2f4ec60d-kube-api-access-5q7hd\") pod \"tempest-tests-tempest\" (UID: \"cf11ce8f-d433-49ff-9390-064a2f4ec60d\") " pod="openstack/tempest-tests-tempest" Dec 10 23:35:05 crc kubenswrapper[4791]: I1210 23:35:05.087434 4791 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage05-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage05-crc\") pod \"tempest-tests-tempest\" (UID: \"cf11ce8f-d433-49ff-9390-064a2f4ec60d\") " pod="openstack/tempest-tests-tempest" Dec 10 23:35:05 crc kubenswrapper[4791]: I1210 23:35:05.109457 4791 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/tempest-tests-tempest" Dec 10 23:35:05 crc kubenswrapper[4791]: I1210 23:35:05.577372 4791 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/tempest-tests-tempest"] Dec 10 23:35:05 crc kubenswrapper[4791]: I1210 23:35:05.581071 4791 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Dec 10 23:35:05 crc kubenswrapper[4791]: I1210 23:35:05.897874 4791 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/tempest-tests-tempest" event={"ID":"cf11ce8f-d433-49ff-9390-064a2f4ec60d","Type":"ContainerStarted","Data":"9f1caab4a7c10473b84a7da85950f86eeabba357974aa29c56bc77427421353d"} Dec 10 23:35:25 crc kubenswrapper[4791]: I1210 23:35:25.038268 4791 patch_prober.go:28] interesting pod/machine-config-daemon-5rb5l container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 10 23:35:25 crc kubenswrapper[4791]: I1210 23:35:25.038945 4791 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-5rb5l" podUID="ba35653c-6e06-4cee-a4d6-137764090d18" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 10 23:35:25 crc kubenswrapper[4791]: I1210 23:35:25.039002 4791 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-5rb5l" Dec 10 23:35:25 crc kubenswrapper[4791]: I1210 23:35:25.040284 4791 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"6982a96e9311c148f4ebb66ec887daeec48d7fbda61684c0ff78ee7d6d234688"} pod="openshift-machine-config-operator/machine-config-daemon-5rb5l" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Dec 10 23:35:25 crc kubenswrapper[4791]: I1210 23:35:25.042837 4791 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-5rb5l" podUID="ba35653c-6e06-4cee-a4d6-137764090d18" containerName="machine-config-daemon" containerID="cri-o://6982a96e9311c148f4ebb66ec887daeec48d7fbda61684c0ff78ee7d6d234688" gracePeriod=600 Dec 10 23:35:26 crc kubenswrapper[4791]: I1210 23:35:26.084377 4791 generic.go:334] "Generic (PLEG): container finished" podID="ba35653c-6e06-4cee-a4d6-137764090d18" containerID="6982a96e9311c148f4ebb66ec887daeec48d7fbda61684c0ff78ee7d6d234688" exitCode=0 Dec 10 23:35:26 crc kubenswrapper[4791]: I1210 23:35:26.084775 4791 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-5rb5l" event={"ID":"ba35653c-6e06-4cee-a4d6-137764090d18","Type":"ContainerDied","Data":"6982a96e9311c148f4ebb66ec887daeec48d7fbda61684c0ff78ee7d6d234688"} Dec 10 23:35:26 crc kubenswrapper[4791]: I1210 23:35:26.084814 4791 scope.go:117] "RemoveContainer" containerID="4229e7e5318c8150a4b76b463eed8dfb67fd267a558642295fc4d4f697411536" Dec 10 23:35:41 crc kubenswrapper[4791]: E1210 23:35:41.691313 4791 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="quay.io/podified-antelope-centos9/openstack-tempest-all:current-podified" Dec 10 23:35:41 crc kubenswrapper[4791]: E1210 23:35:41.692381 4791 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:tempest-tests-tempest-tests-runner,Image:quay.io/podified-antelope-centos9/openstack-tempest-all:current-podified,Command:[],Args:[],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:config-data,ReadOnly:false,MountPath:/etc/test_operator,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:test-operator-ephemeral-workdir,ReadOnly:false,MountPath:/var/lib/tempest,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:test-operator-ephemeral-temporary,ReadOnly:false,MountPath:/tmp,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:test-operator-logs,ReadOnly:false,MountPath:/var/lib/tempest/external_files,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:openstack-config,ReadOnly:true,MountPath:/etc/openstack/clouds.yaml,SubPath:clouds.yaml,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:openstack-config,ReadOnly:true,MountPath:/var/lib/tempest/.config/openstack/clouds.yaml,SubPath:clouds.yaml,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:openstack-config-secret,ReadOnly:false,MountPath:/etc/openstack/secure.yaml,SubPath:secure.yaml,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:ca-certs,ReadOnly:true,MountPath:/etc/pki/ca-trust/extracted/pem/tls-ca-bundle.pem,SubPath:tls-ca-bundle.pem,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:ssh-key,ReadOnly:false,MountPath:/var/lib/tempest/id_ecdsa,SubPath:ssh_key,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-5q7hd,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*42480,RunAsNonRoot:*false,ReadOnlyRootFilesystem:*false,AllowPrivilegeEscalation:*true,RunAsGroup:*42480,ProcMount:nil,WindowsOptions:nil,SeccompProfile:&SeccompProfile{Type:RuntimeDefault,LocalhostProfile:nil,},AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{EnvFromSource{Prefix:,ConfigMapRef:&ConfigMapEnvSource{LocalObjectReference:LocalObjectReference{Name:tempest-tests-tempest-custom-data-s0,},Optional:nil,},SecretRef:nil,},EnvFromSource{Prefix:,ConfigMapRef:&ConfigMapEnvSource{LocalObjectReference:LocalObjectReference{Name:tempest-tests-tempest-env-vars-s0,},Optional:nil,},SecretRef:nil,},},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod tempest-tests-tempest_openstack(cf11ce8f-d433-49ff-9390-064a2f4ec60d): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError" Dec 10 23:35:41 crc kubenswrapper[4791]: E1210 23:35:41.693643 4791 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"tempest-tests-tempest-tests-runner\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\"" pod="openstack/tempest-tests-tempest" podUID="cf11ce8f-d433-49ff-9390-064a2f4ec60d" Dec 10 23:35:42 crc kubenswrapper[4791]: I1210 23:35:42.244185 4791 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-5rb5l" event={"ID":"ba35653c-6e06-4cee-a4d6-137764090d18","Type":"ContainerStarted","Data":"89d61e80933d85eeb2188e63c1702b39d9223c0393e3c83d39da9537a851be64"} Dec 10 23:35:42 crc kubenswrapper[4791]: E1210 23:35:42.245702 4791 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"tempest-tests-tempest-tests-runner\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/podified-antelope-centos9/openstack-tempest-all:current-podified\\\"\"" pod="openstack/tempest-tests-tempest" podUID="cf11ce8f-d433-49ff-9390-064a2f4ec60d" Dec 10 23:35:56 crc kubenswrapper[4791]: I1210 23:35:56.390327 4791 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"tempest-tests-tempest-env-vars-s0" Dec 10 23:35:58 crc kubenswrapper[4791]: I1210 23:35:58.421823 4791 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/tempest-tests-tempest" event={"ID":"cf11ce8f-d433-49ff-9390-064a2f4ec60d","Type":"ContainerStarted","Data":"f73cf0bd24d3f0b1d51a66e7bef480518a1fbb1721aadb250c779b5578569e42"} Dec 10 23:35:58 crc kubenswrapper[4791]: I1210 23:35:58.445300 4791 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/tempest-tests-tempest" podStartSLOduration=4.638325925 podStartE2EDuration="55.445274003s" podCreationTimestamp="2025-12-10 23:35:03 +0000 UTC" firstStartedPulling="2025-12-10 23:35:05.580615038 +0000 UTC m=+2740.010232691" lastFinishedPulling="2025-12-10 23:35:56.387563146 +0000 UTC m=+2790.817180769" observedRunningTime="2025-12-10 23:35:58.438861479 +0000 UTC m=+2792.868479112" watchObservedRunningTime="2025-12-10 23:35:58.445274003 +0000 UTC m=+2792.874891626" Dec 10 23:37:55 crc kubenswrapper[4791]: I1210 23:37:55.039215 4791 patch_prober.go:28] interesting pod/machine-config-daemon-5rb5l container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 10 23:37:55 crc kubenswrapper[4791]: I1210 23:37:55.039808 4791 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-5rb5l" podUID="ba35653c-6e06-4cee-a4d6-137764090d18" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 10 23:38:16 crc kubenswrapper[4791]: I1210 23:38:16.169319 4791 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-marketplace-zktrp"] Dec 10 23:38:16 crc kubenswrapper[4791]: I1210 23:38:16.172479 4791 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-zktrp" Dec 10 23:38:16 crc kubenswrapper[4791]: I1210 23:38:16.193090 4791 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-zktrp"] Dec 10 23:38:16 crc kubenswrapper[4791]: I1210 23:38:16.296258 4791 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-477lg\" (UniqueName: \"kubernetes.io/projected/2f5f8228-fca5-410c-958b-bca1f292a771-kube-api-access-477lg\") pod \"redhat-marketplace-zktrp\" (UID: \"2f5f8228-fca5-410c-958b-bca1f292a771\") " pod="openshift-marketplace/redhat-marketplace-zktrp" Dec 10 23:38:16 crc kubenswrapper[4791]: I1210 23:38:16.296421 4791 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/2f5f8228-fca5-410c-958b-bca1f292a771-utilities\") pod \"redhat-marketplace-zktrp\" (UID: \"2f5f8228-fca5-410c-958b-bca1f292a771\") " pod="openshift-marketplace/redhat-marketplace-zktrp" Dec 10 23:38:16 crc kubenswrapper[4791]: I1210 23:38:16.296477 4791 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/2f5f8228-fca5-410c-958b-bca1f292a771-catalog-content\") pod \"redhat-marketplace-zktrp\" (UID: \"2f5f8228-fca5-410c-958b-bca1f292a771\") " pod="openshift-marketplace/redhat-marketplace-zktrp" Dec 10 23:38:16 crc kubenswrapper[4791]: I1210 23:38:16.397625 4791 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/2f5f8228-fca5-410c-958b-bca1f292a771-utilities\") pod \"redhat-marketplace-zktrp\" (UID: \"2f5f8228-fca5-410c-958b-bca1f292a771\") " pod="openshift-marketplace/redhat-marketplace-zktrp" Dec 10 23:38:16 crc kubenswrapper[4791]: I1210 23:38:16.397907 4791 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/2f5f8228-fca5-410c-958b-bca1f292a771-catalog-content\") pod \"redhat-marketplace-zktrp\" (UID: \"2f5f8228-fca5-410c-958b-bca1f292a771\") " pod="openshift-marketplace/redhat-marketplace-zktrp" Dec 10 23:38:16 crc kubenswrapper[4791]: I1210 23:38:16.398064 4791 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-477lg\" (UniqueName: \"kubernetes.io/projected/2f5f8228-fca5-410c-958b-bca1f292a771-kube-api-access-477lg\") pod \"redhat-marketplace-zktrp\" (UID: \"2f5f8228-fca5-410c-958b-bca1f292a771\") " pod="openshift-marketplace/redhat-marketplace-zktrp" Dec 10 23:38:16 crc kubenswrapper[4791]: I1210 23:38:16.398233 4791 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/2f5f8228-fca5-410c-958b-bca1f292a771-catalog-content\") pod \"redhat-marketplace-zktrp\" (UID: \"2f5f8228-fca5-410c-958b-bca1f292a771\") " pod="openshift-marketplace/redhat-marketplace-zktrp" Dec 10 23:38:16 crc kubenswrapper[4791]: I1210 23:38:16.398095 4791 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/2f5f8228-fca5-410c-958b-bca1f292a771-utilities\") pod \"redhat-marketplace-zktrp\" (UID: \"2f5f8228-fca5-410c-958b-bca1f292a771\") " pod="openshift-marketplace/redhat-marketplace-zktrp" Dec 10 23:38:16 crc kubenswrapper[4791]: I1210 23:38:16.421384 4791 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-477lg\" (UniqueName: \"kubernetes.io/projected/2f5f8228-fca5-410c-958b-bca1f292a771-kube-api-access-477lg\") pod \"redhat-marketplace-zktrp\" (UID: \"2f5f8228-fca5-410c-958b-bca1f292a771\") " pod="openshift-marketplace/redhat-marketplace-zktrp" Dec 10 23:38:16 crc kubenswrapper[4791]: I1210 23:38:16.497692 4791 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-zktrp" Dec 10 23:38:17 crc kubenswrapper[4791]: I1210 23:38:17.097574 4791 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-zktrp"] Dec 10 23:38:17 crc kubenswrapper[4791]: I1210 23:38:17.916577 4791 generic.go:334] "Generic (PLEG): container finished" podID="2f5f8228-fca5-410c-958b-bca1f292a771" containerID="608cd6cd1e54a0af7118016287bd84eaa97d5dd1c91fa02d9e4084a000394f38" exitCode=0 Dec 10 23:38:17 crc kubenswrapper[4791]: I1210 23:38:17.916852 4791 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-zktrp" event={"ID":"2f5f8228-fca5-410c-958b-bca1f292a771","Type":"ContainerDied","Data":"608cd6cd1e54a0af7118016287bd84eaa97d5dd1c91fa02d9e4084a000394f38"} Dec 10 23:38:17 crc kubenswrapper[4791]: I1210 23:38:17.916877 4791 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-zktrp" event={"ID":"2f5f8228-fca5-410c-958b-bca1f292a771","Type":"ContainerStarted","Data":"77725f46626b382f621c86dd5051fe3bac5800c89ff923de67baf0e5664e1cb2"} Dec 10 23:38:18 crc kubenswrapper[4791]: I1210 23:38:18.926943 4791 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-zktrp" event={"ID":"2f5f8228-fca5-410c-958b-bca1f292a771","Type":"ContainerStarted","Data":"e67c08302d3e1933d81b216f77c2b3259ec2871bd379a9b12c1ec2f32e9a8994"} Dec 10 23:38:19 crc kubenswrapper[4791]: I1210 23:38:19.938491 4791 generic.go:334] "Generic (PLEG): container finished" podID="2f5f8228-fca5-410c-958b-bca1f292a771" containerID="e67c08302d3e1933d81b216f77c2b3259ec2871bd379a9b12c1ec2f32e9a8994" exitCode=0 Dec 10 23:38:19 crc kubenswrapper[4791]: I1210 23:38:19.938541 4791 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-zktrp" event={"ID":"2f5f8228-fca5-410c-958b-bca1f292a771","Type":"ContainerDied","Data":"e67c08302d3e1933d81b216f77c2b3259ec2871bd379a9b12c1ec2f32e9a8994"} Dec 10 23:38:20 crc kubenswrapper[4791]: I1210 23:38:20.948939 4791 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-zktrp" event={"ID":"2f5f8228-fca5-410c-958b-bca1f292a771","Type":"ContainerStarted","Data":"cd76b5e7f1614fe877b16e078e13a33c789c270dcffb6c87ece64ddff1edbd73"} Dec 10 23:38:20 crc kubenswrapper[4791]: I1210 23:38:20.968531 4791 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-marketplace-zktrp" podStartSLOduration=2.49285995 podStartE2EDuration="4.968512709s" podCreationTimestamp="2025-12-10 23:38:16 +0000 UTC" firstStartedPulling="2025-12-10 23:38:17.92385333 +0000 UTC m=+2932.353470943" lastFinishedPulling="2025-12-10 23:38:20.399506089 +0000 UTC m=+2934.829123702" observedRunningTime="2025-12-10 23:38:20.965015978 +0000 UTC m=+2935.394633591" watchObservedRunningTime="2025-12-10 23:38:20.968512709 +0000 UTC m=+2935.398130322" Dec 10 23:38:25 crc kubenswrapper[4791]: I1210 23:38:25.038726 4791 patch_prober.go:28] interesting pod/machine-config-daemon-5rb5l container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 10 23:38:25 crc kubenswrapper[4791]: I1210 23:38:25.038810 4791 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-5rb5l" podUID="ba35653c-6e06-4cee-a4d6-137764090d18" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 10 23:38:26 crc kubenswrapper[4791]: I1210 23:38:26.498760 4791 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-marketplace-zktrp" Dec 10 23:38:26 crc kubenswrapper[4791]: I1210 23:38:26.499227 4791 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-marketplace-zktrp" Dec 10 23:38:26 crc kubenswrapper[4791]: I1210 23:38:26.569851 4791 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-marketplace-zktrp" Dec 10 23:38:27 crc kubenswrapper[4791]: I1210 23:38:27.057468 4791 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-marketplace-zktrp" Dec 10 23:38:27 crc kubenswrapper[4791]: I1210 23:38:27.103492 4791 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-zktrp"] Dec 10 23:38:29 crc kubenswrapper[4791]: I1210 23:38:29.025257 4791 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-marketplace-zktrp" podUID="2f5f8228-fca5-410c-958b-bca1f292a771" containerName="registry-server" containerID="cri-o://cd76b5e7f1614fe877b16e078e13a33c789c270dcffb6c87ece64ddff1edbd73" gracePeriod=2 Dec 10 23:38:29 crc kubenswrapper[4791]: I1210 23:38:29.506430 4791 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-zktrp" Dec 10 23:38:29 crc kubenswrapper[4791]: I1210 23:38:29.613879 4791 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/2f5f8228-fca5-410c-958b-bca1f292a771-catalog-content\") pod \"2f5f8228-fca5-410c-958b-bca1f292a771\" (UID: \"2f5f8228-fca5-410c-958b-bca1f292a771\") " Dec 10 23:38:29 crc kubenswrapper[4791]: I1210 23:38:29.613965 4791 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/2f5f8228-fca5-410c-958b-bca1f292a771-utilities\") pod \"2f5f8228-fca5-410c-958b-bca1f292a771\" (UID: \"2f5f8228-fca5-410c-958b-bca1f292a771\") " Dec 10 23:38:29 crc kubenswrapper[4791]: I1210 23:38:29.614135 4791 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-477lg\" (UniqueName: \"kubernetes.io/projected/2f5f8228-fca5-410c-958b-bca1f292a771-kube-api-access-477lg\") pod \"2f5f8228-fca5-410c-958b-bca1f292a771\" (UID: \"2f5f8228-fca5-410c-958b-bca1f292a771\") " Dec 10 23:38:29 crc kubenswrapper[4791]: I1210 23:38:29.615638 4791 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/2f5f8228-fca5-410c-958b-bca1f292a771-utilities" (OuterVolumeSpecName: "utilities") pod "2f5f8228-fca5-410c-958b-bca1f292a771" (UID: "2f5f8228-fca5-410c-958b-bca1f292a771"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 10 23:38:29 crc kubenswrapper[4791]: I1210 23:38:29.621353 4791 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/2f5f8228-fca5-410c-958b-bca1f292a771-kube-api-access-477lg" (OuterVolumeSpecName: "kube-api-access-477lg") pod "2f5f8228-fca5-410c-958b-bca1f292a771" (UID: "2f5f8228-fca5-410c-958b-bca1f292a771"). InnerVolumeSpecName "kube-api-access-477lg". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 10 23:38:29 crc kubenswrapper[4791]: I1210 23:38:29.658318 4791 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/2f5f8228-fca5-410c-958b-bca1f292a771-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "2f5f8228-fca5-410c-958b-bca1f292a771" (UID: "2f5f8228-fca5-410c-958b-bca1f292a771"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 10 23:38:29 crc kubenswrapper[4791]: I1210 23:38:29.716834 4791 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-477lg\" (UniqueName: \"kubernetes.io/projected/2f5f8228-fca5-410c-958b-bca1f292a771-kube-api-access-477lg\") on node \"crc\" DevicePath \"\"" Dec 10 23:38:29 crc kubenswrapper[4791]: I1210 23:38:29.716872 4791 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/2f5f8228-fca5-410c-958b-bca1f292a771-catalog-content\") on node \"crc\" DevicePath \"\"" Dec 10 23:38:29 crc kubenswrapper[4791]: I1210 23:38:29.716883 4791 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/2f5f8228-fca5-410c-958b-bca1f292a771-utilities\") on node \"crc\" DevicePath \"\"" Dec 10 23:38:30 crc kubenswrapper[4791]: I1210 23:38:30.035887 4791 generic.go:334] "Generic (PLEG): container finished" podID="2f5f8228-fca5-410c-958b-bca1f292a771" containerID="cd76b5e7f1614fe877b16e078e13a33c789c270dcffb6c87ece64ddff1edbd73" exitCode=0 Dec 10 23:38:30 crc kubenswrapper[4791]: I1210 23:38:30.036172 4791 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-zktrp" event={"ID":"2f5f8228-fca5-410c-958b-bca1f292a771","Type":"ContainerDied","Data":"cd76b5e7f1614fe877b16e078e13a33c789c270dcffb6c87ece64ddff1edbd73"} Dec 10 23:38:30 crc kubenswrapper[4791]: I1210 23:38:30.036218 4791 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-zktrp" event={"ID":"2f5f8228-fca5-410c-958b-bca1f292a771","Type":"ContainerDied","Data":"77725f46626b382f621c86dd5051fe3bac5800c89ff923de67baf0e5664e1cb2"} Dec 10 23:38:30 crc kubenswrapper[4791]: I1210 23:38:30.036238 4791 scope.go:117] "RemoveContainer" containerID="cd76b5e7f1614fe877b16e078e13a33c789c270dcffb6c87ece64ddff1edbd73" Dec 10 23:38:30 crc kubenswrapper[4791]: I1210 23:38:30.036895 4791 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-zktrp" Dec 10 23:38:30 crc kubenswrapper[4791]: I1210 23:38:30.070565 4791 scope.go:117] "RemoveContainer" containerID="e67c08302d3e1933d81b216f77c2b3259ec2871bd379a9b12c1ec2f32e9a8994" Dec 10 23:38:30 crc kubenswrapper[4791]: I1210 23:38:30.077774 4791 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-zktrp"] Dec 10 23:38:30 crc kubenswrapper[4791]: I1210 23:38:30.091600 4791 scope.go:117] "RemoveContainer" containerID="608cd6cd1e54a0af7118016287bd84eaa97d5dd1c91fa02d9e4084a000394f38" Dec 10 23:38:30 crc kubenswrapper[4791]: I1210 23:38:30.105077 4791 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-marketplace-zktrp"] Dec 10 23:38:30 crc kubenswrapper[4791]: I1210 23:38:30.132854 4791 scope.go:117] "RemoveContainer" containerID="cd76b5e7f1614fe877b16e078e13a33c789c270dcffb6c87ece64ddff1edbd73" Dec 10 23:38:30 crc kubenswrapper[4791]: E1210 23:38:30.133182 4791 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"cd76b5e7f1614fe877b16e078e13a33c789c270dcffb6c87ece64ddff1edbd73\": container with ID starting with cd76b5e7f1614fe877b16e078e13a33c789c270dcffb6c87ece64ddff1edbd73 not found: ID does not exist" containerID="cd76b5e7f1614fe877b16e078e13a33c789c270dcffb6c87ece64ddff1edbd73" Dec 10 23:38:30 crc kubenswrapper[4791]: I1210 23:38:30.133208 4791 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"cd76b5e7f1614fe877b16e078e13a33c789c270dcffb6c87ece64ddff1edbd73"} err="failed to get container status \"cd76b5e7f1614fe877b16e078e13a33c789c270dcffb6c87ece64ddff1edbd73\": rpc error: code = NotFound desc = could not find container \"cd76b5e7f1614fe877b16e078e13a33c789c270dcffb6c87ece64ddff1edbd73\": container with ID starting with cd76b5e7f1614fe877b16e078e13a33c789c270dcffb6c87ece64ddff1edbd73 not found: ID does not exist" Dec 10 23:38:30 crc kubenswrapper[4791]: I1210 23:38:30.133234 4791 scope.go:117] "RemoveContainer" containerID="e67c08302d3e1933d81b216f77c2b3259ec2871bd379a9b12c1ec2f32e9a8994" Dec 10 23:38:30 crc kubenswrapper[4791]: E1210 23:38:30.133721 4791 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"e67c08302d3e1933d81b216f77c2b3259ec2871bd379a9b12c1ec2f32e9a8994\": container with ID starting with e67c08302d3e1933d81b216f77c2b3259ec2871bd379a9b12c1ec2f32e9a8994 not found: ID does not exist" containerID="e67c08302d3e1933d81b216f77c2b3259ec2871bd379a9b12c1ec2f32e9a8994" Dec 10 23:38:30 crc kubenswrapper[4791]: I1210 23:38:30.133770 4791 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"e67c08302d3e1933d81b216f77c2b3259ec2871bd379a9b12c1ec2f32e9a8994"} err="failed to get container status \"e67c08302d3e1933d81b216f77c2b3259ec2871bd379a9b12c1ec2f32e9a8994\": rpc error: code = NotFound desc = could not find container \"e67c08302d3e1933d81b216f77c2b3259ec2871bd379a9b12c1ec2f32e9a8994\": container with ID starting with e67c08302d3e1933d81b216f77c2b3259ec2871bd379a9b12c1ec2f32e9a8994 not found: ID does not exist" Dec 10 23:38:30 crc kubenswrapper[4791]: I1210 23:38:30.133809 4791 scope.go:117] "RemoveContainer" containerID="608cd6cd1e54a0af7118016287bd84eaa97d5dd1c91fa02d9e4084a000394f38" Dec 10 23:38:30 crc kubenswrapper[4791]: E1210 23:38:30.134181 4791 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"608cd6cd1e54a0af7118016287bd84eaa97d5dd1c91fa02d9e4084a000394f38\": container with ID starting with 608cd6cd1e54a0af7118016287bd84eaa97d5dd1c91fa02d9e4084a000394f38 not found: ID does not exist" containerID="608cd6cd1e54a0af7118016287bd84eaa97d5dd1c91fa02d9e4084a000394f38" Dec 10 23:38:30 crc kubenswrapper[4791]: I1210 23:38:30.134205 4791 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"608cd6cd1e54a0af7118016287bd84eaa97d5dd1c91fa02d9e4084a000394f38"} err="failed to get container status \"608cd6cd1e54a0af7118016287bd84eaa97d5dd1c91fa02d9e4084a000394f38\": rpc error: code = NotFound desc = could not find container \"608cd6cd1e54a0af7118016287bd84eaa97d5dd1c91fa02d9e4084a000394f38\": container with ID starting with 608cd6cd1e54a0af7118016287bd84eaa97d5dd1c91fa02d9e4084a000394f38 not found: ID does not exist" Dec 10 23:38:31 crc kubenswrapper[4791]: I1210 23:38:31.900931 4791 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="2f5f8228-fca5-410c-958b-bca1f292a771" path="/var/lib/kubelet/pods/2f5f8228-fca5-410c-958b-bca1f292a771/volumes" Dec 10 23:38:55 crc kubenswrapper[4791]: I1210 23:38:55.038461 4791 patch_prober.go:28] interesting pod/machine-config-daemon-5rb5l container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 10 23:38:55 crc kubenswrapper[4791]: I1210 23:38:55.039053 4791 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-5rb5l" podUID="ba35653c-6e06-4cee-a4d6-137764090d18" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 10 23:38:55 crc kubenswrapper[4791]: I1210 23:38:55.039102 4791 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-5rb5l" Dec 10 23:38:55 crc kubenswrapper[4791]: I1210 23:38:55.039931 4791 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"89d61e80933d85eeb2188e63c1702b39d9223c0393e3c83d39da9537a851be64"} pod="openshift-machine-config-operator/machine-config-daemon-5rb5l" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Dec 10 23:38:55 crc kubenswrapper[4791]: I1210 23:38:55.040083 4791 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-5rb5l" podUID="ba35653c-6e06-4cee-a4d6-137764090d18" containerName="machine-config-daemon" containerID="cri-o://89d61e80933d85eeb2188e63c1702b39d9223c0393e3c83d39da9537a851be64" gracePeriod=600 Dec 10 23:38:55 crc kubenswrapper[4791]: E1210 23:38:55.176787 4791 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-5rb5l_openshift-machine-config-operator(ba35653c-6e06-4cee-a4d6-137764090d18)\"" pod="openshift-machine-config-operator/machine-config-daemon-5rb5l" podUID="ba35653c-6e06-4cee-a4d6-137764090d18" Dec 10 23:38:55 crc kubenswrapper[4791]: I1210 23:38:55.300315 4791 generic.go:334] "Generic (PLEG): container finished" podID="ba35653c-6e06-4cee-a4d6-137764090d18" containerID="89d61e80933d85eeb2188e63c1702b39d9223c0393e3c83d39da9537a851be64" exitCode=0 Dec 10 23:38:55 crc kubenswrapper[4791]: I1210 23:38:55.300390 4791 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-5rb5l" event={"ID":"ba35653c-6e06-4cee-a4d6-137764090d18","Type":"ContainerDied","Data":"89d61e80933d85eeb2188e63c1702b39d9223c0393e3c83d39da9537a851be64"} Dec 10 23:38:55 crc kubenswrapper[4791]: I1210 23:38:55.300474 4791 scope.go:117] "RemoveContainer" containerID="6982a96e9311c148f4ebb66ec887daeec48d7fbda61684c0ff78ee7d6d234688" Dec 10 23:38:55 crc kubenswrapper[4791]: I1210 23:38:55.303508 4791 scope.go:117] "RemoveContainer" containerID="89d61e80933d85eeb2188e63c1702b39d9223c0393e3c83d39da9537a851be64" Dec 10 23:38:55 crc kubenswrapper[4791]: E1210 23:38:55.304199 4791 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-5rb5l_openshift-machine-config-operator(ba35653c-6e06-4cee-a4d6-137764090d18)\"" pod="openshift-machine-config-operator/machine-config-daemon-5rb5l" podUID="ba35653c-6e06-4cee-a4d6-137764090d18" Dec 10 23:39:06 crc kubenswrapper[4791]: I1210 23:39:06.885674 4791 scope.go:117] "RemoveContainer" containerID="89d61e80933d85eeb2188e63c1702b39d9223c0393e3c83d39da9537a851be64" Dec 10 23:39:06 crc kubenswrapper[4791]: E1210 23:39:06.886416 4791 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-5rb5l_openshift-machine-config-operator(ba35653c-6e06-4cee-a4d6-137764090d18)\"" pod="openshift-machine-config-operator/machine-config-daemon-5rb5l" podUID="ba35653c-6e06-4cee-a4d6-137764090d18" Dec 10 23:39:19 crc kubenswrapper[4791]: I1210 23:39:19.886012 4791 scope.go:117] "RemoveContainer" containerID="89d61e80933d85eeb2188e63c1702b39d9223c0393e3c83d39da9537a851be64" Dec 10 23:39:19 crc kubenswrapper[4791]: E1210 23:39:19.886979 4791 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-5rb5l_openshift-machine-config-operator(ba35653c-6e06-4cee-a4d6-137764090d18)\"" pod="openshift-machine-config-operator/machine-config-daemon-5rb5l" podUID="ba35653c-6e06-4cee-a4d6-137764090d18" Dec 10 23:39:32 crc kubenswrapper[4791]: I1210 23:39:32.885337 4791 scope.go:117] "RemoveContainer" containerID="89d61e80933d85eeb2188e63c1702b39d9223c0393e3c83d39da9537a851be64" Dec 10 23:39:32 crc kubenswrapper[4791]: E1210 23:39:32.886621 4791 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-5rb5l_openshift-machine-config-operator(ba35653c-6e06-4cee-a4d6-137764090d18)\"" pod="openshift-machine-config-operator/machine-config-daemon-5rb5l" podUID="ba35653c-6e06-4cee-a4d6-137764090d18" Dec 10 23:39:44 crc kubenswrapper[4791]: I1210 23:39:44.885108 4791 scope.go:117] "RemoveContainer" containerID="89d61e80933d85eeb2188e63c1702b39d9223c0393e3c83d39da9537a851be64" Dec 10 23:39:44 crc kubenswrapper[4791]: E1210 23:39:44.885796 4791 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-5rb5l_openshift-machine-config-operator(ba35653c-6e06-4cee-a4d6-137764090d18)\"" pod="openshift-machine-config-operator/machine-config-daemon-5rb5l" podUID="ba35653c-6e06-4cee-a4d6-137764090d18" Dec 10 23:39:55 crc kubenswrapper[4791]: I1210 23:39:55.900915 4791 scope.go:117] "RemoveContainer" containerID="89d61e80933d85eeb2188e63c1702b39d9223c0393e3c83d39da9537a851be64" Dec 10 23:39:55 crc kubenswrapper[4791]: E1210 23:39:55.902722 4791 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-5rb5l_openshift-machine-config-operator(ba35653c-6e06-4cee-a4d6-137764090d18)\"" pod="openshift-machine-config-operator/machine-config-daemon-5rb5l" podUID="ba35653c-6e06-4cee-a4d6-137764090d18" Dec 10 23:40:06 crc kubenswrapper[4791]: I1210 23:40:06.886559 4791 scope.go:117] "RemoveContainer" containerID="89d61e80933d85eeb2188e63c1702b39d9223c0393e3c83d39da9537a851be64" Dec 10 23:40:06 crc kubenswrapper[4791]: E1210 23:40:06.887910 4791 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-5rb5l_openshift-machine-config-operator(ba35653c-6e06-4cee-a4d6-137764090d18)\"" pod="openshift-machine-config-operator/machine-config-daemon-5rb5l" podUID="ba35653c-6e06-4cee-a4d6-137764090d18" Dec 10 23:40:20 crc kubenswrapper[4791]: I1210 23:40:20.885257 4791 scope.go:117] "RemoveContainer" containerID="89d61e80933d85eeb2188e63c1702b39d9223c0393e3c83d39da9537a851be64" Dec 10 23:40:20 crc kubenswrapper[4791]: E1210 23:40:20.886123 4791 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-5rb5l_openshift-machine-config-operator(ba35653c-6e06-4cee-a4d6-137764090d18)\"" pod="openshift-machine-config-operator/machine-config-daemon-5rb5l" podUID="ba35653c-6e06-4cee-a4d6-137764090d18" Dec 10 23:40:33 crc kubenswrapper[4791]: I1210 23:40:33.885221 4791 scope.go:117] "RemoveContainer" containerID="89d61e80933d85eeb2188e63c1702b39d9223c0393e3c83d39da9537a851be64" Dec 10 23:40:33 crc kubenswrapper[4791]: E1210 23:40:33.886415 4791 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-5rb5l_openshift-machine-config-operator(ba35653c-6e06-4cee-a4d6-137764090d18)\"" pod="openshift-machine-config-operator/machine-config-daemon-5rb5l" podUID="ba35653c-6e06-4cee-a4d6-137764090d18" Dec 10 23:40:45 crc kubenswrapper[4791]: I1210 23:40:45.894502 4791 scope.go:117] "RemoveContainer" containerID="89d61e80933d85eeb2188e63c1702b39d9223c0393e3c83d39da9537a851be64" Dec 10 23:40:45 crc kubenswrapper[4791]: E1210 23:40:45.895309 4791 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-5rb5l_openshift-machine-config-operator(ba35653c-6e06-4cee-a4d6-137764090d18)\"" pod="openshift-machine-config-operator/machine-config-daemon-5rb5l" podUID="ba35653c-6e06-4cee-a4d6-137764090d18" Dec 10 23:40:56 crc kubenswrapper[4791]: I1210 23:40:56.886143 4791 scope.go:117] "RemoveContainer" containerID="89d61e80933d85eeb2188e63c1702b39d9223c0393e3c83d39da9537a851be64" Dec 10 23:40:56 crc kubenswrapper[4791]: E1210 23:40:56.886982 4791 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-5rb5l_openshift-machine-config-operator(ba35653c-6e06-4cee-a4d6-137764090d18)\"" pod="openshift-machine-config-operator/machine-config-daemon-5rb5l" podUID="ba35653c-6e06-4cee-a4d6-137764090d18" Dec 10 23:41:01 crc kubenswrapper[4791]: I1210 23:41:01.956059 4791 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-operators-jfsc6"] Dec 10 23:41:01 crc kubenswrapper[4791]: E1210 23:41:01.957252 4791 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="2f5f8228-fca5-410c-958b-bca1f292a771" containerName="extract-utilities" Dec 10 23:41:01 crc kubenswrapper[4791]: I1210 23:41:01.957310 4791 state_mem.go:107] "Deleted CPUSet assignment" podUID="2f5f8228-fca5-410c-958b-bca1f292a771" containerName="extract-utilities" Dec 10 23:41:01 crc kubenswrapper[4791]: E1210 23:41:01.957326 4791 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="2f5f8228-fca5-410c-958b-bca1f292a771" containerName="extract-content" Dec 10 23:41:01 crc kubenswrapper[4791]: I1210 23:41:01.957334 4791 state_mem.go:107] "Deleted CPUSet assignment" podUID="2f5f8228-fca5-410c-958b-bca1f292a771" containerName="extract-content" Dec 10 23:41:01 crc kubenswrapper[4791]: E1210 23:41:01.957388 4791 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="2f5f8228-fca5-410c-958b-bca1f292a771" containerName="registry-server" Dec 10 23:41:01 crc kubenswrapper[4791]: I1210 23:41:01.957399 4791 state_mem.go:107] "Deleted CPUSet assignment" podUID="2f5f8228-fca5-410c-958b-bca1f292a771" containerName="registry-server" Dec 10 23:41:01 crc kubenswrapper[4791]: I1210 23:41:01.957647 4791 memory_manager.go:354] "RemoveStaleState removing state" podUID="2f5f8228-fca5-410c-958b-bca1f292a771" containerName="registry-server" Dec 10 23:41:01 crc kubenswrapper[4791]: I1210 23:41:01.959320 4791 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-jfsc6" Dec 10 23:41:01 crc kubenswrapper[4791]: I1210 23:41:01.970784 4791 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-jfsc6"] Dec 10 23:41:02 crc kubenswrapper[4791]: I1210 23:41:02.055424 4791 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-xn9tp\" (UniqueName: \"kubernetes.io/projected/fc193084-a565-419a-a1a4-81bc17e23e56-kube-api-access-xn9tp\") pod \"redhat-operators-jfsc6\" (UID: \"fc193084-a565-419a-a1a4-81bc17e23e56\") " pod="openshift-marketplace/redhat-operators-jfsc6" Dec 10 23:41:02 crc kubenswrapper[4791]: I1210 23:41:02.055486 4791 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/fc193084-a565-419a-a1a4-81bc17e23e56-catalog-content\") pod \"redhat-operators-jfsc6\" (UID: \"fc193084-a565-419a-a1a4-81bc17e23e56\") " pod="openshift-marketplace/redhat-operators-jfsc6" Dec 10 23:41:02 crc kubenswrapper[4791]: I1210 23:41:02.055584 4791 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/fc193084-a565-419a-a1a4-81bc17e23e56-utilities\") pod \"redhat-operators-jfsc6\" (UID: \"fc193084-a565-419a-a1a4-81bc17e23e56\") " pod="openshift-marketplace/redhat-operators-jfsc6" Dec 10 23:41:02 crc kubenswrapper[4791]: I1210 23:41:02.158464 4791 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-xn9tp\" (UniqueName: \"kubernetes.io/projected/fc193084-a565-419a-a1a4-81bc17e23e56-kube-api-access-xn9tp\") pod \"redhat-operators-jfsc6\" (UID: \"fc193084-a565-419a-a1a4-81bc17e23e56\") " pod="openshift-marketplace/redhat-operators-jfsc6" Dec 10 23:41:02 crc kubenswrapper[4791]: I1210 23:41:02.158520 4791 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/fc193084-a565-419a-a1a4-81bc17e23e56-catalog-content\") pod \"redhat-operators-jfsc6\" (UID: \"fc193084-a565-419a-a1a4-81bc17e23e56\") " pod="openshift-marketplace/redhat-operators-jfsc6" Dec 10 23:41:02 crc kubenswrapper[4791]: I1210 23:41:02.158582 4791 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/fc193084-a565-419a-a1a4-81bc17e23e56-utilities\") pod \"redhat-operators-jfsc6\" (UID: \"fc193084-a565-419a-a1a4-81bc17e23e56\") " pod="openshift-marketplace/redhat-operators-jfsc6" Dec 10 23:41:02 crc kubenswrapper[4791]: I1210 23:41:02.159209 4791 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/fc193084-a565-419a-a1a4-81bc17e23e56-utilities\") pod \"redhat-operators-jfsc6\" (UID: \"fc193084-a565-419a-a1a4-81bc17e23e56\") " pod="openshift-marketplace/redhat-operators-jfsc6" Dec 10 23:41:02 crc kubenswrapper[4791]: I1210 23:41:02.159285 4791 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/fc193084-a565-419a-a1a4-81bc17e23e56-catalog-content\") pod \"redhat-operators-jfsc6\" (UID: \"fc193084-a565-419a-a1a4-81bc17e23e56\") " pod="openshift-marketplace/redhat-operators-jfsc6" Dec 10 23:41:02 crc kubenswrapper[4791]: I1210 23:41:02.190511 4791 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-xn9tp\" (UniqueName: \"kubernetes.io/projected/fc193084-a565-419a-a1a4-81bc17e23e56-kube-api-access-xn9tp\") pod \"redhat-operators-jfsc6\" (UID: \"fc193084-a565-419a-a1a4-81bc17e23e56\") " pod="openshift-marketplace/redhat-operators-jfsc6" Dec 10 23:41:02 crc kubenswrapper[4791]: I1210 23:41:02.280002 4791 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-jfsc6" Dec 10 23:41:02 crc kubenswrapper[4791]: I1210 23:41:02.767644 4791 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-jfsc6"] Dec 10 23:41:03 crc kubenswrapper[4791]: I1210 23:41:03.687233 4791 generic.go:334] "Generic (PLEG): container finished" podID="fc193084-a565-419a-a1a4-81bc17e23e56" containerID="4e460fded1d3c9998007abb7998b5dfeda0669b4b6574cf05861539cce4c6797" exitCode=0 Dec 10 23:41:03 crc kubenswrapper[4791]: I1210 23:41:03.687296 4791 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-jfsc6" event={"ID":"fc193084-a565-419a-a1a4-81bc17e23e56","Type":"ContainerDied","Data":"4e460fded1d3c9998007abb7998b5dfeda0669b4b6574cf05861539cce4c6797"} Dec 10 23:41:03 crc kubenswrapper[4791]: I1210 23:41:03.688635 4791 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-jfsc6" event={"ID":"fc193084-a565-419a-a1a4-81bc17e23e56","Type":"ContainerStarted","Data":"6ca7e06725cf59ba9741d718c12457a5e1071fb53065b85fece43a74470c3672"} Dec 10 23:41:03 crc kubenswrapper[4791]: I1210 23:41:03.690060 4791 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Dec 10 23:41:04 crc kubenswrapper[4791]: I1210 23:41:04.704118 4791 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-jfsc6" event={"ID":"fc193084-a565-419a-a1a4-81bc17e23e56","Type":"ContainerStarted","Data":"cd97e6af203a701968b0c2349db9294699fbbd9206a593083097068a8f588ff7"} Dec 10 23:41:05 crc kubenswrapper[4791]: I1210 23:41:05.723677 4791 generic.go:334] "Generic (PLEG): container finished" podID="fc193084-a565-419a-a1a4-81bc17e23e56" containerID="cd97e6af203a701968b0c2349db9294699fbbd9206a593083097068a8f588ff7" exitCode=0 Dec 10 23:41:05 crc kubenswrapper[4791]: I1210 23:41:05.723768 4791 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-jfsc6" event={"ID":"fc193084-a565-419a-a1a4-81bc17e23e56","Type":"ContainerDied","Data":"cd97e6af203a701968b0c2349db9294699fbbd9206a593083097068a8f588ff7"} Dec 10 23:41:06 crc kubenswrapper[4791]: I1210 23:41:06.735228 4791 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-jfsc6" event={"ID":"fc193084-a565-419a-a1a4-81bc17e23e56","Type":"ContainerStarted","Data":"26b7529ccd162d30eafe7b2d4342a9a2a246b96a2e26124a27081f19d4ad1ab1"} Dec 10 23:41:06 crc kubenswrapper[4791]: I1210 23:41:06.758911 4791 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-operators-jfsc6" podStartSLOduration=3.158552624 podStartE2EDuration="5.75887335s" podCreationTimestamp="2025-12-10 23:41:01 +0000 UTC" firstStartedPulling="2025-12-10 23:41:03.689678465 +0000 UTC m=+3098.119296098" lastFinishedPulling="2025-12-10 23:41:06.289999171 +0000 UTC m=+3100.719616824" observedRunningTime="2025-12-10 23:41:06.751794466 +0000 UTC m=+3101.181412089" watchObservedRunningTime="2025-12-10 23:41:06.75887335 +0000 UTC m=+3101.188490963" Dec 10 23:41:08 crc kubenswrapper[4791]: I1210 23:41:08.885485 4791 scope.go:117] "RemoveContainer" containerID="89d61e80933d85eeb2188e63c1702b39d9223c0393e3c83d39da9537a851be64" Dec 10 23:41:08 crc kubenswrapper[4791]: E1210 23:41:08.886517 4791 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-5rb5l_openshift-machine-config-operator(ba35653c-6e06-4cee-a4d6-137764090d18)\"" pod="openshift-machine-config-operator/machine-config-daemon-5rb5l" podUID="ba35653c-6e06-4cee-a4d6-137764090d18" Dec 10 23:41:12 crc kubenswrapper[4791]: I1210 23:41:12.223184 4791 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/community-operators-cdqfj"] Dec 10 23:41:12 crc kubenswrapper[4791]: I1210 23:41:12.226774 4791 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-cdqfj" Dec 10 23:41:12 crc kubenswrapper[4791]: I1210 23:41:12.238825 4791 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-cdqfj"] Dec 10 23:41:12 crc kubenswrapper[4791]: I1210 23:41:12.280681 4791 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-operators-jfsc6" Dec 10 23:41:12 crc kubenswrapper[4791]: I1210 23:41:12.280763 4791 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-operators-jfsc6" Dec 10 23:41:12 crc kubenswrapper[4791]: I1210 23:41:12.335762 4791 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/8827669c-a262-47ef-ae54-698c4941f379-catalog-content\") pod \"community-operators-cdqfj\" (UID: \"8827669c-a262-47ef-ae54-698c4941f379\") " pod="openshift-marketplace/community-operators-cdqfj" Dec 10 23:41:12 crc kubenswrapper[4791]: I1210 23:41:12.336354 4791 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/8827669c-a262-47ef-ae54-698c4941f379-utilities\") pod \"community-operators-cdqfj\" (UID: \"8827669c-a262-47ef-ae54-698c4941f379\") " pod="openshift-marketplace/community-operators-cdqfj" Dec 10 23:41:12 crc kubenswrapper[4791]: I1210 23:41:12.336530 4791 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-h9xdt\" (UniqueName: \"kubernetes.io/projected/8827669c-a262-47ef-ae54-698c4941f379-kube-api-access-h9xdt\") pod \"community-operators-cdqfj\" (UID: \"8827669c-a262-47ef-ae54-698c4941f379\") " pod="openshift-marketplace/community-operators-cdqfj" Dec 10 23:41:12 crc kubenswrapper[4791]: I1210 23:41:12.438508 4791 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/8827669c-a262-47ef-ae54-698c4941f379-catalog-content\") pod \"community-operators-cdqfj\" (UID: \"8827669c-a262-47ef-ae54-698c4941f379\") " pod="openshift-marketplace/community-operators-cdqfj" Dec 10 23:41:12 crc kubenswrapper[4791]: I1210 23:41:12.438582 4791 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/8827669c-a262-47ef-ae54-698c4941f379-utilities\") pod \"community-operators-cdqfj\" (UID: \"8827669c-a262-47ef-ae54-698c4941f379\") " pod="openshift-marketplace/community-operators-cdqfj" Dec 10 23:41:12 crc kubenswrapper[4791]: I1210 23:41:12.438717 4791 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-h9xdt\" (UniqueName: \"kubernetes.io/projected/8827669c-a262-47ef-ae54-698c4941f379-kube-api-access-h9xdt\") pod \"community-operators-cdqfj\" (UID: \"8827669c-a262-47ef-ae54-698c4941f379\") " pod="openshift-marketplace/community-operators-cdqfj" Dec 10 23:41:12 crc kubenswrapper[4791]: I1210 23:41:12.439045 4791 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/8827669c-a262-47ef-ae54-698c4941f379-catalog-content\") pod \"community-operators-cdqfj\" (UID: \"8827669c-a262-47ef-ae54-698c4941f379\") " pod="openshift-marketplace/community-operators-cdqfj" Dec 10 23:41:12 crc kubenswrapper[4791]: I1210 23:41:12.439443 4791 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/8827669c-a262-47ef-ae54-698c4941f379-utilities\") pod \"community-operators-cdqfj\" (UID: \"8827669c-a262-47ef-ae54-698c4941f379\") " pod="openshift-marketplace/community-operators-cdqfj" Dec 10 23:41:12 crc kubenswrapper[4791]: I1210 23:41:12.464303 4791 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-h9xdt\" (UniqueName: \"kubernetes.io/projected/8827669c-a262-47ef-ae54-698c4941f379-kube-api-access-h9xdt\") pod \"community-operators-cdqfj\" (UID: \"8827669c-a262-47ef-ae54-698c4941f379\") " pod="openshift-marketplace/community-operators-cdqfj" Dec 10 23:41:12 crc kubenswrapper[4791]: I1210 23:41:12.562934 4791 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-cdqfj" Dec 10 23:41:13 crc kubenswrapper[4791]: I1210 23:41:13.084650 4791 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-cdqfj"] Dec 10 23:41:13 crc kubenswrapper[4791]: I1210 23:41:13.334070 4791 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-marketplace/redhat-operators-jfsc6" podUID="fc193084-a565-419a-a1a4-81bc17e23e56" containerName="registry-server" probeResult="failure" output=< Dec 10 23:41:13 crc kubenswrapper[4791]: timeout: failed to connect service ":50051" within 1s Dec 10 23:41:13 crc kubenswrapper[4791]: > Dec 10 23:41:13 crc kubenswrapper[4791]: I1210 23:41:13.818646 4791 generic.go:334] "Generic (PLEG): container finished" podID="8827669c-a262-47ef-ae54-698c4941f379" containerID="eceab658a42d30618db05736b00cc3ecc9bbbcf74ad09bce5c8b135198851981" exitCode=0 Dec 10 23:41:13 crc kubenswrapper[4791]: I1210 23:41:13.818711 4791 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-cdqfj" event={"ID":"8827669c-a262-47ef-ae54-698c4941f379","Type":"ContainerDied","Data":"eceab658a42d30618db05736b00cc3ecc9bbbcf74ad09bce5c8b135198851981"} Dec 10 23:41:13 crc kubenswrapper[4791]: I1210 23:41:13.819010 4791 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-cdqfj" event={"ID":"8827669c-a262-47ef-ae54-698c4941f379","Type":"ContainerStarted","Data":"ae2bae5f5deb072c4af681c0fbee0bd287fdc22431ec3a4ddb3e6f04a67e2349"} Dec 10 23:41:15 crc kubenswrapper[4791]: I1210 23:41:15.840364 4791 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-cdqfj" event={"ID":"8827669c-a262-47ef-ae54-698c4941f379","Type":"ContainerStarted","Data":"3db8e22a943feb63266b0427a3164d3eb8fd7b6aa1f75ed8a94ee964ee48efd4"} Dec 10 23:41:16 crc kubenswrapper[4791]: I1210 23:41:16.851556 4791 generic.go:334] "Generic (PLEG): container finished" podID="8827669c-a262-47ef-ae54-698c4941f379" containerID="3db8e22a943feb63266b0427a3164d3eb8fd7b6aa1f75ed8a94ee964ee48efd4" exitCode=0 Dec 10 23:41:16 crc kubenswrapper[4791]: I1210 23:41:16.851706 4791 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-cdqfj" event={"ID":"8827669c-a262-47ef-ae54-698c4941f379","Type":"ContainerDied","Data":"3db8e22a943feb63266b0427a3164d3eb8fd7b6aa1f75ed8a94ee964ee48efd4"} Dec 10 23:41:17 crc kubenswrapper[4791]: I1210 23:41:17.871310 4791 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-cdqfj" event={"ID":"8827669c-a262-47ef-ae54-698c4941f379","Type":"ContainerStarted","Data":"14b47f2715b38454ba7f51112190396294d02db110f1cd0ee3056efbc1a0f1f0"} Dec 10 23:41:17 crc kubenswrapper[4791]: I1210 23:41:17.912658 4791 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/community-operators-cdqfj" podStartSLOduration=2.319499885 podStartE2EDuration="5.912638693s" podCreationTimestamp="2025-12-10 23:41:12 +0000 UTC" firstStartedPulling="2025-12-10 23:41:13.82153488 +0000 UTC m=+3108.251152503" lastFinishedPulling="2025-12-10 23:41:17.414673688 +0000 UTC m=+3111.844291311" observedRunningTime="2025-12-10 23:41:17.904210851 +0000 UTC m=+3112.333828464" watchObservedRunningTime="2025-12-10 23:41:17.912638693 +0000 UTC m=+3112.342256306" Dec 10 23:41:21 crc kubenswrapper[4791]: I1210 23:41:21.886564 4791 scope.go:117] "RemoveContainer" containerID="89d61e80933d85eeb2188e63c1702b39d9223c0393e3c83d39da9537a851be64" Dec 10 23:41:21 crc kubenswrapper[4791]: E1210 23:41:21.887452 4791 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-5rb5l_openshift-machine-config-operator(ba35653c-6e06-4cee-a4d6-137764090d18)\"" pod="openshift-machine-config-operator/machine-config-daemon-5rb5l" podUID="ba35653c-6e06-4cee-a4d6-137764090d18" Dec 10 23:41:22 crc kubenswrapper[4791]: I1210 23:41:22.325477 4791 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-operators-jfsc6" Dec 10 23:41:22 crc kubenswrapper[4791]: I1210 23:41:22.375564 4791 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-operators-jfsc6" Dec 10 23:41:22 crc kubenswrapper[4791]: I1210 23:41:22.560054 4791 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-jfsc6"] Dec 10 23:41:22 crc kubenswrapper[4791]: I1210 23:41:22.564309 4791 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/community-operators-cdqfj" Dec 10 23:41:22 crc kubenswrapper[4791]: I1210 23:41:22.564417 4791 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/community-operators-cdqfj" Dec 10 23:41:22 crc kubenswrapper[4791]: I1210 23:41:22.627299 4791 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/community-operators-cdqfj" Dec 10 23:41:22 crc kubenswrapper[4791]: I1210 23:41:22.972191 4791 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/community-operators-cdqfj" Dec 10 23:41:23 crc kubenswrapper[4791]: I1210 23:41:23.940511 4791 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-operators-jfsc6" podUID="fc193084-a565-419a-a1a4-81bc17e23e56" containerName="registry-server" containerID="cri-o://26b7529ccd162d30eafe7b2d4342a9a2a246b96a2e26124a27081f19d4ad1ab1" gracePeriod=2 Dec 10 23:41:24 crc kubenswrapper[4791]: I1210 23:41:24.533894 4791 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-jfsc6" Dec 10 23:41:24 crc kubenswrapper[4791]: I1210 23:41:24.727271 4791 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/fc193084-a565-419a-a1a4-81bc17e23e56-catalog-content\") pod \"fc193084-a565-419a-a1a4-81bc17e23e56\" (UID: \"fc193084-a565-419a-a1a4-81bc17e23e56\") " Dec 10 23:41:24 crc kubenswrapper[4791]: I1210 23:41:24.727507 4791 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/fc193084-a565-419a-a1a4-81bc17e23e56-utilities\") pod \"fc193084-a565-419a-a1a4-81bc17e23e56\" (UID: \"fc193084-a565-419a-a1a4-81bc17e23e56\") " Dec 10 23:41:24 crc kubenswrapper[4791]: I1210 23:41:24.727653 4791 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-xn9tp\" (UniqueName: \"kubernetes.io/projected/fc193084-a565-419a-a1a4-81bc17e23e56-kube-api-access-xn9tp\") pod \"fc193084-a565-419a-a1a4-81bc17e23e56\" (UID: \"fc193084-a565-419a-a1a4-81bc17e23e56\") " Dec 10 23:41:24 crc kubenswrapper[4791]: I1210 23:41:24.731165 4791 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/fc193084-a565-419a-a1a4-81bc17e23e56-utilities" (OuterVolumeSpecName: "utilities") pod "fc193084-a565-419a-a1a4-81bc17e23e56" (UID: "fc193084-a565-419a-a1a4-81bc17e23e56"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 10 23:41:24 crc kubenswrapper[4791]: I1210 23:41:24.739670 4791 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/fc193084-a565-419a-a1a4-81bc17e23e56-kube-api-access-xn9tp" (OuterVolumeSpecName: "kube-api-access-xn9tp") pod "fc193084-a565-419a-a1a4-81bc17e23e56" (UID: "fc193084-a565-419a-a1a4-81bc17e23e56"). InnerVolumeSpecName "kube-api-access-xn9tp". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 10 23:41:24 crc kubenswrapper[4791]: I1210 23:41:24.829648 4791 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/fc193084-a565-419a-a1a4-81bc17e23e56-utilities\") on node \"crc\" DevicePath \"\"" Dec 10 23:41:24 crc kubenswrapper[4791]: I1210 23:41:24.829675 4791 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-xn9tp\" (UniqueName: \"kubernetes.io/projected/fc193084-a565-419a-a1a4-81bc17e23e56-kube-api-access-xn9tp\") on node \"crc\" DevicePath \"\"" Dec 10 23:41:24 crc kubenswrapper[4791]: I1210 23:41:24.886520 4791 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/fc193084-a565-419a-a1a4-81bc17e23e56-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "fc193084-a565-419a-a1a4-81bc17e23e56" (UID: "fc193084-a565-419a-a1a4-81bc17e23e56"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 10 23:41:24 crc kubenswrapper[4791]: I1210 23:41:24.931234 4791 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/fc193084-a565-419a-a1a4-81bc17e23e56-catalog-content\") on node \"crc\" DevicePath \"\"" Dec 10 23:41:24 crc kubenswrapper[4791]: I1210 23:41:24.956034 4791 generic.go:334] "Generic (PLEG): container finished" podID="fc193084-a565-419a-a1a4-81bc17e23e56" containerID="26b7529ccd162d30eafe7b2d4342a9a2a246b96a2e26124a27081f19d4ad1ab1" exitCode=0 Dec 10 23:41:24 crc kubenswrapper[4791]: I1210 23:41:24.956101 4791 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-jfsc6" Dec 10 23:41:24 crc kubenswrapper[4791]: I1210 23:41:24.956099 4791 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-jfsc6" event={"ID":"fc193084-a565-419a-a1a4-81bc17e23e56","Type":"ContainerDied","Data":"26b7529ccd162d30eafe7b2d4342a9a2a246b96a2e26124a27081f19d4ad1ab1"} Dec 10 23:41:24 crc kubenswrapper[4791]: I1210 23:41:24.956235 4791 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-jfsc6" event={"ID":"fc193084-a565-419a-a1a4-81bc17e23e56","Type":"ContainerDied","Data":"6ca7e06725cf59ba9741d718c12457a5e1071fb53065b85fece43a74470c3672"} Dec 10 23:41:24 crc kubenswrapper[4791]: I1210 23:41:24.956266 4791 scope.go:117] "RemoveContainer" containerID="26b7529ccd162d30eafe7b2d4342a9a2a246b96a2e26124a27081f19d4ad1ab1" Dec 10 23:41:24 crc kubenswrapper[4791]: I1210 23:41:24.973078 4791 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-cdqfj"] Dec 10 23:41:24 crc kubenswrapper[4791]: I1210 23:41:24.973801 4791 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/community-operators-cdqfj" podUID="8827669c-a262-47ef-ae54-698c4941f379" containerName="registry-server" containerID="cri-o://14b47f2715b38454ba7f51112190396294d02db110f1cd0ee3056efbc1a0f1f0" gracePeriod=2 Dec 10 23:41:24 crc kubenswrapper[4791]: I1210 23:41:24.988386 4791 scope.go:117] "RemoveContainer" containerID="cd97e6af203a701968b0c2349db9294699fbbd9206a593083097068a8f588ff7" Dec 10 23:41:25 crc kubenswrapper[4791]: I1210 23:41:25.002163 4791 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-jfsc6"] Dec 10 23:41:25 crc kubenswrapper[4791]: I1210 23:41:25.029711 4791 scope.go:117] "RemoveContainer" containerID="4e460fded1d3c9998007abb7998b5dfeda0669b4b6574cf05861539cce4c6797" Dec 10 23:41:25 crc kubenswrapper[4791]: I1210 23:41:25.033397 4791 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-operators-jfsc6"] Dec 10 23:41:25 crc kubenswrapper[4791]: I1210 23:41:25.178070 4791 scope.go:117] "RemoveContainer" containerID="26b7529ccd162d30eafe7b2d4342a9a2a246b96a2e26124a27081f19d4ad1ab1" Dec 10 23:41:25 crc kubenswrapper[4791]: E1210 23:41:25.179518 4791 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"26b7529ccd162d30eafe7b2d4342a9a2a246b96a2e26124a27081f19d4ad1ab1\": container with ID starting with 26b7529ccd162d30eafe7b2d4342a9a2a246b96a2e26124a27081f19d4ad1ab1 not found: ID does not exist" containerID="26b7529ccd162d30eafe7b2d4342a9a2a246b96a2e26124a27081f19d4ad1ab1" Dec 10 23:41:25 crc kubenswrapper[4791]: I1210 23:41:25.179583 4791 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"26b7529ccd162d30eafe7b2d4342a9a2a246b96a2e26124a27081f19d4ad1ab1"} err="failed to get container status \"26b7529ccd162d30eafe7b2d4342a9a2a246b96a2e26124a27081f19d4ad1ab1\": rpc error: code = NotFound desc = could not find container \"26b7529ccd162d30eafe7b2d4342a9a2a246b96a2e26124a27081f19d4ad1ab1\": container with ID starting with 26b7529ccd162d30eafe7b2d4342a9a2a246b96a2e26124a27081f19d4ad1ab1 not found: ID does not exist" Dec 10 23:41:25 crc kubenswrapper[4791]: I1210 23:41:25.179619 4791 scope.go:117] "RemoveContainer" containerID="cd97e6af203a701968b0c2349db9294699fbbd9206a593083097068a8f588ff7" Dec 10 23:41:25 crc kubenswrapper[4791]: E1210 23:41:25.180053 4791 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"cd97e6af203a701968b0c2349db9294699fbbd9206a593083097068a8f588ff7\": container with ID starting with cd97e6af203a701968b0c2349db9294699fbbd9206a593083097068a8f588ff7 not found: ID does not exist" containerID="cd97e6af203a701968b0c2349db9294699fbbd9206a593083097068a8f588ff7" Dec 10 23:41:25 crc kubenswrapper[4791]: I1210 23:41:25.180094 4791 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"cd97e6af203a701968b0c2349db9294699fbbd9206a593083097068a8f588ff7"} err="failed to get container status \"cd97e6af203a701968b0c2349db9294699fbbd9206a593083097068a8f588ff7\": rpc error: code = NotFound desc = could not find container \"cd97e6af203a701968b0c2349db9294699fbbd9206a593083097068a8f588ff7\": container with ID starting with cd97e6af203a701968b0c2349db9294699fbbd9206a593083097068a8f588ff7 not found: ID does not exist" Dec 10 23:41:25 crc kubenswrapper[4791]: I1210 23:41:25.180135 4791 scope.go:117] "RemoveContainer" containerID="4e460fded1d3c9998007abb7998b5dfeda0669b4b6574cf05861539cce4c6797" Dec 10 23:41:25 crc kubenswrapper[4791]: E1210 23:41:25.180508 4791 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"4e460fded1d3c9998007abb7998b5dfeda0669b4b6574cf05861539cce4c6797\": container with ID starting with 4e460fded1d3c9998007abb7998b5dfeda0669b4b6574cf05861539cce4c6797 not found: ID does not exist" containerID="4e460fded1d3c9998007abb7998b5dfeda0669b4b6574cf05861539cce4c6797" Dec 10 23:41:25 crc kubenswrapper[4791]: I1210 23:41:25.180537 4791 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"4e460fded1d3c9998007abb7998b5dfeda0669b4b6574cf05861539cce4c6797"} err="failed to get container status \"4e460fded1d3c9998007abb7998b5dfeda0669b4b6574cf05861539cce4c6797\": rpc error: code = NotFound desc = could not find container \"4e460fded1d3c9998007abb7998b5dfeda0669b4b6574cf05861539cce4c6797\": container with ID starting with 4e460fded1d3c9998007abb7998b5dfeda0669b4b6574cf05861539cce4c6797 not found: ID does not exist" Dec 10 23:41:25 crc kubenswrapper[4791]: I1210 23:41:25.513426 4791 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-cdqfj" Dec 10 23:41:25 crc kubenswrapper[4791]: I1210 23:41:25.645495 4791 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-h9xdt\" (UniqueName: \"kubernetes.io/projected/8827669c-a262-47ef-ae54-698c4941f379-kube-api-access-h9xdt\") pod \"8827669c-a262-47ef-ae54-698c4941f379\" (UID: \"8827669c-a262-47ef-ae54-698c4941f379\") " Dec 10 23:41:25 crc kubenswrapper[4791]: I1210 23:41:25.645630 4791 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/8827669c-a262-47ef-ae54-698c4941f379-catalog-content\") pod \"8827669c-a262-47ef-ae54-698c4941f379\" (UID: \"8827669c-a262-47ef-ae54-698c4941f379\") " Dec 10 23:41:25 crc kubenswrapper[4791]: I1210 23:41:25.647682 4791 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/8827669c-a262-47ef-ae54-698c4941f379-utilities\") pod \"8827669c-a262-47ef-ae54-698c4941f379\" (UID: \"8827669c-a262-47ef-ae54-698c4941f379\") " Dec 10 23:41:25 crc kubenswrapper[4791]: I1210 23:41:25.648609 4791 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/8827669c-a262-47ef-ae54-698c4941f379-utilities" (OuterVolumeSpecName: "utilities") pod "8827669c-a262-47ef-ae54-698c4941f379" (UID: "8827669c-a262-47ef-ae54-698c4941f379"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 10 23:41:25 crc kubenswrapper[4791]: I1210 23:41:25.650179 4791 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/8827669c-a262-47ef-ae54-698c4941f379-kube-api-access-h9xdt" (OuterVolumeSpecName: "kube-api-access-h9xdt") pod "8827669c-a262-47ef-ae54-698c4941f379" (UID: "8827669c-a262-47ef-ae54-698c4941f379"). InnerVolumeSpecName "kube-api-access-h9xdt". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 10 23:41:25 crc kubenswrapper[4791]: I1210 23:41:25.719595 4791 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/8827669c-a262-47ef-ae54-698c4941f379-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "8827669c-a262-47ef-ae54-698c4941f379" (UID: "8827669c-a262-47ef-ae54-698c4941f379"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 10 23:41:25 crc kubenswrapper[4791]: I1210 23:41:25.750634 4791 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-h9xdt\" (UniqueName: \"kubernetes.io/projected/8827669c-a262-47ef-ae54-698c4941f379-kube-api-access-h9xdt\") on node \"crc\" DevicePath \"\"" Dec 10 23:41:25 crc kubenswrapper[4791]: I1210 23:41:25.750933 4791 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/8827669c-a262-47ef-ae54-698c4941f379-catalog-content\") on node \"crc\" DevicePath \"\"" Dec 10 23:41:25 crc kubenswrapper[4791]: I1210 23:41:25.751071 4791 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/8827669c-a262-47ef-ae54-698c4941f379-utilities\") on node \"crc\" DevicePath \"\"" Dec 10 23:41:25 crc kubenswrapper[4791]: I1210 23:41:25.900909 4791 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="fc193084-a565-419a-a1a4-81bc17e23e56" path="/var/lib/kubelet/pods/fc193084-a565-419a-a1a4-81bc17e23e56/volumes" Dec 10 23:41:25 crc kubenswrapper[4791]: I1210 23:41:25.970999 4791 generic.go:334] "Generic (PLEG): container finished" podID="8827669c-a262-47ef-ae54-698c4941f379" containerID="14b47f2715b38454ba7f51112190396294d02db110f1cd0ee3056efbc1a0f1f0" exitCode=0 Dec 10 23:41:25 crc kubenswrapper[4791]: I1210 23:41:25.971051 4791 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-cdqfj" event={"ID":"8827669c-a262-47ef-ae54-698c4941f379","Type":"ContainerDied","Data":"14b47f2715b38454ba7f51112190396294d02db110f1cd0ee3056efbc1a0f1f0"} Dec 10 23:41:25 crc kubenswrapper[4791]: I1210 23:41:25.971108 4791 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-cdqfj" Dec 10 23:41:25 crc kubenswrapper[4791]: I1210 23:41:25.971134 4791 scope.go:117] "RemoveContainer" containerID="14b47f2715b38454ba7f51112190396294d02db110f1cd0ee3056efbc1a0f1f0" Dec 10 23:41:25 crc kubenswrapper[4791]: I1210 23:41:25.971115 4791 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-cdqfj" event={"ID":"8827669c-a262-47ef-ae54-698c4941f379","Type":"ContainerDied","Data":"ae2bae5f5deb072c4af681c0fbee0bd287fdc22431ec3a4ddb3e6f04a67e2349"} Dec 10 23:41:26 crc kubenswrapper[4791]: I1210 23:41:26.001215 4791 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-cdqfj"] Dec 10 23:41:26 crc kubenswrapper[4791]: I1210 23:41:26.009797 4791 scope.go:117] "RemoveContainer" containerID="3db8e22a943feb63266b0427a3164d3eb8fd7b6aa1f75ed8a94ee964ee48efd4" Dec 10 23:41:26 crc kubenswrapper[4791]: I1210 23:41:26.012433 4791 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/community-operators-cdqfj"] Dec 10 23:41:26 crc kubenswrapper[4791]: I1210 23:41:26.068237 4791 scope.go:117] "RemoveContainer" containerID="eceab658a42d30618db05736b00cc3ecc9bbbcf74ad09bce5c8b135198851981" Dec 10 23:41:26 crc kubenswrapper[4791]: I1210 23:41:26.092665 4791 scope.go:117] "RemoveContainer" containerID="14b47f2715b38454ba7f51112190396294d02db110f1cd0ee3056efbc1a0f1f0" Dec 10 23:41:26 crc kubenswrapper[4791]: E1210 23:41:26.096476 4791 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"14b47f2715b38454ba7f51112190396294d02db110f1cd0ee3056efbc1a0f1f0\": container with ID starting with 14b47f2715b38454ba7f51112190396294d02db110f1cd0ee3056efbc1a0f1f0 not found: ID does not exist" containerID="14b47f2715b38454ba7f51112190396294d02db110f1cd0ee3056efbc1a0f1f0" Dec 10 23:41:26 crc kubenswrapper[4791]: I1210 23:41:26.096524 4791 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"14b47f2715b38454ba7f51112190396294d02db110f1cd0ee3056efbc1a0f1f0"} err="failed to get container status \"14b47f2715b38454ba7f51112190396294d02db110f1cd0ee3056efbc1a0f1f0\": rpc error: code = NotFound desc = could not find container \"14b47f2715b38454ba7f51112190396294d02db110f1cd0ee3056efbc1a0f1f0\": container with ID starting with 14b47f2715b38454ba7f51112190396294d02db110f1cd0ee3056efbc1a0f1f0 not found: ID does not exist" Dec 10 23:41:26 crc kubenswrapper[4791]: I1210 23:41:26.096564 4791 scope.go:117] "RemoveContainer" containerID="3db8e22a943feb63266b0427a3164d3eb8fd7b6aa1f75ed8a94ee964ee48efd4" Dec 10 23:41:26 crc kubenswrapper[4791]: E1210 23:41:26.097627 4791 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"3db8e22a943feb63266b0427a3164d3eb8fd7b6aa1f75ed8a94ee964ee48efd4\": container with ID starting with 3db8e22a943feb63266b0427a3164d3eb8fd7b6aa1f75ed8a94ee964ee48efd4 not found: ID does not exist" containerID="3db8e22a943feb63266b0427a3164d3eb8fd7b6aa1f75ed8a94ee964ee48efd4" Dec 10 23:41:26 crc kubenswrapper[4791]: I1210 23:41:26.097723 4791 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"3db8e22a943feb63266b0427a3164d3eb8fd7b6aa1f75ed8a94ee964ee48efd4"} err="failed to get container status \"3db8e22a943feb63266b0427a3164d3eb8fd7b6aa1f75ed8a94ee964ee48efd4\": rpc error: code = NotFound desc = could not find container \"3db8e22a943feb63266b0427a3164d3eb8fd7b6aa1f75ed8a94ee964ee48efd4\": container with ID starting with 3db8e22a943feb63266b0427a3164d3eb8fd7b6aa1f75ed8a94ee964ee48efd4 not found: ID does not exist" Dec 10 23:41:26 crc kubenswrapper[4791]: I1210 23:41:26.097746 4791 scope.go:117] "RemoveContainer" containerID="eceab658a42d30618db05736b00cc3ecc9bbbcf74ad09bce5c8b135198851981" Dec 10 23:41:26 crc kubenswrapper[4791]: E1210 23:41:26.098168 4791 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"eceab658a42d30618db05736b00cc3ecc9bbbcf74ad09bce5c8b135198851981\": container with ID starting with eceab658a42d30618db05736b00cc3ecc9bbbcf74ad09bce5c8b135198851981 not found: ID does not exist" containerID="eceab658a42d30618db05736b00cc3ecc9bbbcf74ad09bce5c8b135198851981" Dec 10 23:41:26 crc kubenswrapper[4791]: I1210 23:41:26.098254 4791 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"eceab658a42d30618db05736b00cc3ecc9bbbcf74ad09bce5c8b135198851981"} err="failed to get container status \"eceab658a42d30618db05736b00cc3ecc9bbbcf74ad09bce5c8b135198851981\": rpc error: code = NotFound desc = could not find container \"eceab658a42d30618db05736b00cc3ecc9bbbcf74ad09bce5c8b135198851981\": container with ID starting with eceab658a42d30618db05736b00cc3ecc9bbbcf74ad09bce5c8b135198851981 not found: ID does not exist" Dec 10 23:41:27 crc kubenswrapper[4791]: I1210 23:41:27.896962 4791 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="8827669c-a262-47ef-ae54-698c4941f379" path="/var/lib/kubelet/pods/8827669c-a262-47ef-ae54-698c4941f379/volumes" Dec 10 23:41:31 crc kubenswrapper[4791]: E1210 23:41:31.952083 4791 cadvisor_stats_provider.go:516] "Partial failure issuing cadvisor.ContainerInfoV2" err="partial failures: [\"/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod8827669c_a262_47ef_ae54_698c4941f379.slice/crio-ae2bae5f5deb072c4af681c0fbee0bd287fdc22431ec3a4ddb3e6f04a67e2349\": RecentStats: unable to find data in memory cache], [\"/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod8827669c_a262_47ef_ae54_698c4941f379.slice\": RecentStats: unable to find data in memory cache]" Dec 10 23:41:32 crc kubenswrapper[4791]: I1210 23:41:32.884779 4791 scope.go:117] "RemoveContainer" containerID="89d61e80933d85eeb2188e63c1702b39d9223c0393e3c83d39da9537a851be64" Dec 10 23:41:32 crc kubenswrapper[4791]: E1210 23:41:32.885503 4791 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-5rb5l_openshift-machine-config-operator(ba35653c-6e06-4cee-a4d6-137764090d18)\"" pod="openshift-machine-config-operator/machine-config-daemon-5rb5l" podUID="ba35653c-6e06-4cee-a4d6-137764090d18" Dec 10 23:41:42 crc kubenswrapper[4791]: E1210 23:41:42.184912 4791 cadvisor_stats_provider.go:516] "Partial failure issuing cadvisor.ContainerInfoV2" err="partial failures: [\"/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod8827669c_a262_47ef_ae54_698c4941f379.slice\": RecentStats: unable to find data in memory cache], [\"/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod8827669c_a262_47ef_ae54_698c4941f379.slice/crio-ae2bae5f5deb072c4af681c0fbee0bd287fdc22431ec3a4ddb3e6f04a67e2349\": RecentStats: unable to find data in memory cache]" Dec 10 23:41:45 crc kubenswrapper[4791]: I1210 23:41:45.894754 4791 scope.go:117] "RemoveContainer" containerID="89d61e80933d85eeb2188e63c1702b39d9223c0393e3c83d39da9537a851be64" Dec 10 23:41:45 crc kubenswrapper[4791]: E1210 23:41:45.895648 4791 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-5rb5l_openshift-machine-config-operator(ba35653c-6e06-4cee-a4d6-137764090d18)\"" pod="openshift-machine-config-operator/machine-config-daemon-5rb5l" podUID="ba35653c-6e06-4cee-a4d6-137764090d18" Dec 10 23:41:52 crc kubenswrapper[4791]: E1210 23:41:52.482170 4791 cadvisor_stats_provider.go:516] "Partial failure issuing cadvisor.ContainerInfoV2" err="partial failures: [\"/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod8827669c_a262_47ef_ae54_698c4941f379.slice/crio-ae2bae5f5deb072c4af681c0fbee0bd287fdc22431ec3a4ddb3e6f04a67e2349\": RecentStats: unable to find data in memory cache], [\"/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod8827669c_a262_47ef_ae54_698c4941f379.slice\": RecentStats: unable to find data in memory cache]" Dec 10 23:41:56 crc kubenswrapper[4791]: I1210 23:41:56.885533 4791 scope.go:117] "RemoveContainer" containerID="89d61e80933d85eeb2188e63c1702b39d9223c0393e3c83d39da9537a851be64" Dec 10 23:41:56 crc kubenswrapper[4791]: E1210 23:41:56.886308 4791 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-5rb5l_openshift-machine-config-operator(ba35653c-6e06-4cee-a4d6-137764090d18)\"" pod="openshift-machine-config-operator/machine-config-daemon-5rb5l" podUID="ba35653c-6e06-4cee-a4d6-137764090d18" Dec 10 23:42:02 crc kubenswrapper[4791]: E1210 23:42:02.803270 4791 cadvisor_stats_provider.go:516] "Partial failure issuing cadvisor.ContainerInfoV2" err="partial failures: [\"/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod8827669c_a262_47ef_ae54_698c4941f379.slice/crio-ae2bae5f5deb072c4af681c0fbee0bd287fdc22431ec3a4ddb3e6f04a67e2349\": RecentStats: unable to find data in memory cache], [\"/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod8827669c_a262_47ef_ae54_698c4941f379.slice\": RecentStats: unable to find data in memory cache]" Dec 10 23:42:09 crc kubenswrapper[4791]: I1210 23:42:09.885329 4791 scope.go:117] "RemoveContainer" containerID="89d61e80933d85eeb2188e63c1702b39d9223c0393e3c83d39da9537a851be64" Dec 10 23:42:09 crc kubenswrapper[4791]: E1210 23:42:09.886286 4791 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-5rb5l_openshift-machine-config-operator(ba35653c-6e06-4cee-a4d6-137764090d18)\"" pod="openshift-machine-config-operator/machine-config-daemon-5rb5l" podUID="ba35653c-6e06-4cee-a4d6-137764090d18" Dec 10 23:42:11 crc kubenswrapper[4791]: I1210 23:42:11.300860 4791 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/certified-operators-dnkgw"] Dec 10 23:42:11 crc kubenswrapper[4791]: E1210 23:42:11.301797 4791 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="fc193084-a565-419a-a1a4-81bc17e23e56" containerName="extract-utilities" Dec 10 23:42:11 crc kubenswrapper[4791]: I1210 23:42:11.301825 4791 state_mem.go:107] "Deleted CPUSet assignment" podUID="fc193084-a565-419a-a1a4-81bc17e23e56" containerName="extract-utilities" Dec 10 23:42:11 crc kubenswrapper[4791]: E1210 23:42:11.301866 4791 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="fc193084-a565-419a-a1a4-81bc17e23e56" containerName="registry-server" Dec 10 23:42:11 crc kubenswrapper[4791]: I1210 23:42:11.301878 4791 state_mem.go:107] "Deleted CPUSet assignment" podUID="fc193084-a565-419a-a1a4-81bc17e23e56" containerName="registry-server" Dec 10 23:42:11 crc kubenswrapper[4791]: E1210 23:42:11.301902 4791 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="fc193084-a565-419a-a1a4-81bc17e23e56" containerName="extract-content" Dec 10 23:42:11 crc kubenswrapper[4791]: I1210 23:42:11.301916 4791 state_mem.go:107] "Deleted CPUSet assignment" podUID="fc193084-a565-419a-a1a4-81bc17e23e56" containerName="extract-content" Dec 10 23:42:11 crc kubenswrapper[4791]: E1210 23:42:11.301952 4791 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="8827669c-a262-47ef-ae54-698c4941f379" containerName="registry-server" Dec 10 23:42:11 crc kubenswrapper[4791]: I1210 23:42:11.301964 4791 state_mem.go:107] "Deleted CPUSet assignment" podUID="8827669c-a262-47ef-ae54-698c4941f379" containerName="registry-server" Dec 10 23:42:11 crc kubenswrapper[4791]: E1210 23:42:11.301990 4791 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="8827669c-a262-47ef-ae54-698c4941f379" containerName="extract-utilities" Dec 10 23:42:11 crc kubenswrapper[4791]: I1210 23:42:11.302002 4791 state_mem.go:107] "Deleted CPUSet assignment" podUID="8827669c-a262-47ef-ae54-698c4941f379" containerName="extract-utilities" Dec 10 23:42:11 crc kubenswrapper[4791]: E1210 23:42:11.302031 4791 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="8827669c-a262-47ef-ae54-698c4941f379" containerName="extract-content" Dec 10 23:42:11 crc kubenswrapper[4791]: I1210 23:42:11.302043 4791 state_mem.go:107] "Deleted CPUSet assignment" podUID="8827669c-a262-47ef-ae54-698c4941f379" containerName="extract-content" Dec 10 23:42:11 crc kubenswrapper[4791]: I1210 23:42:11.302462 4791 memory_manager.go:354] "RemoveStaleState removing state" podUID="8827669c-a262-47ef-ae54-698c4941f379" containerName="registry-server" Dec 10 23:42:11 crc kubenswrapper[4791]: I1210 23:42:11.302519 4791 memory_manager.go:354] "RemoveStaleState removing state" podUID="fc193084-a565-419a-a1a4-81bc17e23e56" containerName="registry-server" Dec 10 23:42:11 crc kubenswrapper[4791]: I1210 23:42:11.305069 4791 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-dnkgw" Dec 10 23:42:11 crc kubenswrapper[4791]: I1210 23:42:11.314402 4791 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-dnkgw"] Dec 10 23:42:11 crc kubenswrapper[4791]: I1210 23:42:11.323200 4791 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-wfqrv\" (UniqueName: \"kubernetes.io/projected/1c5e5471-de5a-4d1c-a983-11c4864d885d-kube-api-access-wfqrv\") pod \"certified-operators-dnkgw\" (UID: \"1c5e5471-de5a-4d1c-a983-11c4864d885d\") " pod="openshift-marketplace/certified-operators-dnkgw" Dec 10 23:42:11 crc kubenswrapper[4791]: I1210 23:42:11.323311 4791 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/1c5e5471-de5a-4d1c-a983-11c4864d885d-catalog-content\") pod \"certified-operators-dnkgw\" (UID: \"1c5e5471-de5a-4d1c-a983-11c4864d885d\") " pod="openshift-marketplace/certified-operators-dnkgw" Dec 10 23:42:11 crc kubenswrapper[4791]: I1210 23:42:11.323403 4791 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/1c5e5471-de5a-4d1c-a983-11c4864d885d-utilities\") pod \"certified-operators-dnkgw\" (UID: \"1c5e5471-de5a-4d1c-a983-11c4864d885d\") " pod="openshift-marketplace/certified-operators-dnkgw" Dec 10 23:42:11 crc kubenswrapper[4791]: I1210 23:42:11.425858 4791 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-wfqrv\" (UniqueName: \"kubernetes.io/projected/1c5e5471-de5a-4d1c-a983-11c4864d885d-kube-api-access-wfqrv\") pod \"certified-operators-dnkgw\" (UID: \"1c5e5471-de5a-4d1c-a983-11c4864d885d\") " pod="openshift-marketplace/certified-operators-dnkgw" Dec 10 23:42:11 crc kubenswrapper[4791]: I1210 23:42:11.425993 4791 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/1c5e5471-de5a-4d1c-a983-11c4864d885d-catalog-content\") pod \"certified-operators-dnkgw\" (UID: \"1c5e5471-de5a-4d1c-a983-11c4864d885d\") " pod="openshift-marketplace/certified-operators-dnkgw" Dec 10 23:42:11 crc kubenswrapper[4791]: I1210 23:42:11.426085 4791 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/1c5e5471-de5a-4d1c-a983-11c4864d885d-utilities\") pod \"certified-operators-dnkgw\" (UID: \"1c5e5471-de5a-4d1c-a983-11c4864d885d\") " pod="openshift-marketplace/certified-operators-dnkgw" Dec 10 23:42:11 crc kubenswrapper[4791]: I1210 23:42:11.427493 4791 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/1c5e5471-de5a-4d1c-a983-11c4864d885d-utilities\") pod \"certified-operators-dnkgw\" (UID: \"1c5e5471-de5a-4d1c-a983-11c4864d885d\") " pod="openshift-marketplace/certified-operators-dnkgw" Dec 10 23:42:11 crc kubenswrapper[4791]: I1210 23:42:11.427503 4791 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/1c5e5471-de5a-4d1c-a983-11c4864d885d-catalog-content\") pod \"certified-operators-dnkgw\" (UID: \"1c5e5471-de5a-4d1c-a983-11c4864d885d\") " pod="openshift-marketplace/certified-operators-dnkgw" Dec 10 23:42:11 crc kubenswrapper[4791]: I1210 23:42:11.456526 4791 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-wfqrv\" (UniqueName: \"kubernetes.io/projected/1c5e5471-de5a-4d1c-a983-11c4864d885d-kube-api-access-wfqrv\") pod \"certified-operators-dnkgw\" (UID: \"1c5e5471-de5a-4d1c-a983-11c4864d885d\") " pod="openshift-marketplace/certified-operators-dnkgw" Dec 10 23:42:11 crc kubenswrapper[4791]: I1210 23:42:11.631012 4791 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-dnkgw" Dec 10 23:42:12 crc kubenswrapper[4791]: I1210 23:42:12.199986 4791 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-dnkgw"] Dec 10 23:42:12 crc kubenswrapper[4791]: W1210 23:42:12.205325 4791 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod1c5e5471_de5a_4d1c_a983_11c4864d885d.slice/crio-eb441cbd100dc547882448b687344e51a71261fd6eecc40e7373cc1418c04f98 WatchSource:0}: Error finding container eb441cbd100dc547882448b687344e51a71261fd6eecc40e7373cc1418c04f98: Status 404 returned error can't find the container with id eb441cbd100dc547882448b687344e51a71261fd6eecc40e7373cc1418c04f98 Dec 10 23:42:12 crc kubenswrapper[4791]: I1210 23:42:12.475986 4791 generic.go:334] "Generic (PLEG): container finished" podID="1c5e5471-de5a-4d1c-a983-11c4864d885d" containerID="89a473e4cd415e3c739b514bce03a07f0dbb28b4f309080d2093c8daa34aca2f" exitCode=0 Dec 10 23:42:12 crc kubenswrapper[4791]: I1210 23:42:12.476022 4791 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-dnkgw" event={"ID":"1c5e5471-de5a-4d1c-a983-11c4864d885d","Type":"ContainerDied","Data":"89a473e4cd415e3c739b514bce03a07f0dbb28b4f309080d2093c8daa34aca2f"} Dec 10 23:42:12 crc kubenswrapper[4791]: I1210 23:42:12.476056 4791 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-dnkgw" event={"ID":"1c5e5471-de5a-4d1c-a983-11c4864d885d","Type":"ContainerStarted","Data":"eb441cbd100dc547882448b687344e51a71261fd6eecc40e7373cc1418c04f98"} Dec 10 23:42:13 crc kubenswrapper[4791]: E1210 23:42:13.025387 4791 cadvisor_stats_provider.go:516] "Partial failure issuing cadvisor.ContainerInfoV2" err="partial failures: [\"/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod8827669c_a262_47ef_ae54_698c4941f379.slice/crio-ae2bae5f5deb072c4af681c0fbee0bd287fdc22431ec3a4ddb3e6f04a67e2349\": RecentStats: unable to find data in memory cache], [\"/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod8827669c_a262_47ef_ae54_698c4941f379.slice\": RecentStats: unable to find data in memory cache]" Dec 10 23:42:13 crc kubenswrapper[4791]: I1210 23:42:13.486469 4791 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-dnkgw" event={"ID":"1c5e5471-de5a-4d1c-a983-11c4864d885d","Type":"ContainerStarted","Data":"ebbee9909a1714802b2d92a30f776808b4dc4673ec0ba712e00ed36c2c5ff550"} Dec 10 23:42:14 crc kubenswrapper[4791]: I1210 23:42:14.514293 4791 generic.go:334] "Generic (PLEG): container finished" podID="1c5e5471-de5a-4d1c-a983-11c4864d885d" containerID="ebbee9909a1714802b2d92a30f776808b4dc4673ec0ba712e00ed36c2c5ff550" exitCode=0 Dec 10 23:42:14 crc kubenswrapper[4791]: I1210 23:42:14.514392 4791 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-dnkgw" event={"ID":"1c5e5471-de5a-4d1c-a983-11c4864d885d","Type":"ContainerDied","Data":"ebbee9909a1714802b2d92a30f776808b4dc4673ec0ba712e00ed36c2c5ff550"} Dec 10 23:42:16 crc kubenswrapper[4791]: I1210 23:42:16.543768 4791 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-dnkgw" event={"ID":"1c5e5471-de5a-4d1c-a983-11c4864d885d","Type":"ContainerStarted","Data":"82005e1f83f8b46234beba92a0e5ec0b649aa08886b22a8443ab3211e3bf7fba"} Dec 10 23:42:16 crc kubenswrapper[4791]: I1210 23:42:16.580221 4791 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/certified-operators-dnkgw" podStartSLOduration=2.336316087 podStartE2EDuration="5.580195258s" podCreationTimestamp="2025-12-10 23:42:11 +0000 UTC" firstStartedPulling="2025-12-10 23:42:12.477688136 +0000 UTC m=+3166.907305749" lastFinishedPulling="2025-12-10 23:42:15.721567307 +0000 UTC m=+3170.151184920" observedRunningTime="2025-12-10 23:42:16.57366406 +0000 UTC m=+3171.003281673" watchObservedRunningTime="2025-12-10 23:42:16.580195258 +0000 UTC m=+3171.009812881" Dec 10 23:42:21 crc kubenswrapper[4791]: I1210 23:42:21.631199 4791 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/certified-operators-dnkgw" Dec 10 23:42:21 crc kubenswrapper[4791]: I1210 23:42:21.631918 4791 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/certified-operators-dnkgw" Dec 10 23:42:21 crc kubenswrapper[4791]: I1210 23:42:21.719529 4791 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/certified-operators-dnkgw" Dec 10 23:42:21 crc kubenswrapper[4791]: I1210 23:42:21.885516 4791 scope.go:117] "RemoveContainer" containerID="89d61e80933d85eeb2188e63c1702b39d9223c0393e3c83d39da9537a851be64" Dec 10 23:42:21 crc kubenswrapper[4791]: E1210 23:42:21.885990 4791 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-5rb5l_openshift-machine-config-operator(ba35653c-6e06-4cee-a4d6-137764090d18)\"" pod="openshift-machine-config-operator/machine-config-daemon-5rb5l" podUID="ba35653c-6e06-4cee-a4d6-137764090d18" Dec 10 23:42:22 crc kubenswrapper[4791]: I1210 23:42:22.663541 4791 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/certified-operators-dnkgw" Dec 10 23:42:22 crc kubenswrapper[4791]: I1210 23:42:22.733668 4791 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-dnkgw"] Dec 10 23:42:23 crc kubenswrapper[4791]: E1210 23:42:23.281709 4791 cadvisor_stats_provider.go:516] "Partial failure issuing cadvisor.ContainerInfoV2" err="partial failures: [\"/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod8827669c_a262_47ef_ae54_698c4941f379.slice\": RecentStats: unable to find data in memory cache], [\"/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod8827669c_a262_47ef_ae54_698c4941f379.slice/crio-ae2bae5f5deb072c4af681c0fbee0bd287fdc22431ec3a4ddb3e6f04a67e2349\": RecentStats: unable to find data in memory cache]" Dec 10 23:42:24 crc kubenswrapper[4791]: I1210 23:42:24.624640 4791 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/certified-operators-dnkgw" podUID="1c5e5471-de5a-4d1c-a983-11c4864d885d" containerName="registry-server" containerID="cri-o://82005e1f83f8b46234beba92a0e5ec0b649aa08886b22a8443ab3211e3bf7fba" gracePeriod=2 Dec 10 23:42:25 crc kubenswrapper[4791]: I1210 23:42:25.113456 4791 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-dnkgw" Dec 10 23:42:25 crc kubenswrapper[4791]: I1210 23:42:25.235023 4791 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/1c5e5471-de5a-4d1c-a983-11c4864d885d-catalog-content\") pod \"1c5e5471-de5a-4d1c-a983-11c4864d885d\" (UID: \"1c5e5471-de5a-4d1c-a983-11c4864d885d\") " Dec 10 23:42:25 crc kubenswrapper[4791]: I1210 23:42:25.238700 4791 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-wfqrv\" (UniqueName: \"kubernetes.io/projected/1c5e5471-de5a-4d1c-a983-11c4864d885d-kube-api-access-wfqrv\") pod \"1c5e5471-de5a-4d1c-a983-11c4864d885d\" (UID: \"1c5e5471-de5a-4d1c-a983-11c4864d885d\") " Dec 10 23:42:25 crc kubenswrapper[4791]: I1210 23:42:25.238803 4791 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/1c5e5471-de5a-4d1c-a983-11c4864d885d-utilities\") pod \"1c5e5471-de5a-4d1c-a983-11c4864d885d\" (UID: \"1c5e5471-de5a-4d1c-a983-11c4864d885d\") " Dec 10 23:42:25 crc kubenswrapper[4791]: I1210 23:42:25.241613 4791 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/1c5e5471-de5a-4d1c-a983-11c4864d885d-utilities" (OuterVolumeSpecName: "utilities") pod "1c5e5471-de5a-4d1c-a983-11c4864d885d" (UID: "1c5e5471-de5a-4d1c-a983-11c4864d885d"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 10 23:42:25 crc kubenswrapper[4791]: I1210 23:42:25.245598 4791 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/1c5e5471-de5a-4d1c-a983-11c4864d885d-kube-api-access-wfqrv" (OuterVolumeSpecName: "kube-api-access-wfqrv") pod "1c5e5471-de5a-4d1c-a983-11c4864d885d" (UID: "1c5e5471-de5a-4d1c-a983-11c4864d885d"). InnerVolumeSpecName "kube-api-access-wfqrv". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 10 23:42:25 crc kubenswrapper[4791]: I1210 23:42:25.304617 4791 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/1c5e5471-de5a-4d1c-a983-11c4864d885d-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "1c5e5471-de5a-4d1c-a983-11c4864d885d" (UID: "1c5e5471-de5a-4d1c-a983-11c4864d885d"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 10 23:42:25 crc kubenswrapper[4791]: I1210 23:42:25.342395 4791 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/1c5e5471-de5a-4d1c-a983-11c4864d885d-catalog-content\") on node \"crc\" DevicePath \"\"" Dec 10 23:42:25 crc kubenswrapper[4791]: I1210 23:42:25.342429 4791 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-wfqrv\" (UniqueName: \"kubernetes.io/projected/1c5e5471-de5a-4d1c-a983-11c4864d885d-kube-api-access-wfqrv\") on node \"crc\" DevicePath \"\"" Dec 10 23:42:25 crc kubenswrapper[4791]: I1210 23:42:25.342441 4791 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/1c5e5471-de5a-4d1c-a983-11c4864d885d-utilities\") on node \"crc\" DevicePath \"\"" Dec 10 23:42:25 crc kubenswrapper[4791]: I1210 23:42:25.642651 4791 generic.go:334] "Generic (PLEG): container finished" podID="1c5e5471-de5a-4d1c-a983-11c4864d885d" containerID="82005e1f83f8b46234beba92a0e5ec0b649aa08886b22a8443ab3211e3bf7fba" exitCode=0 Dec 10 23:42:25 crc kubenswrapper[4791]: I1210 23:42:25.642704 4791 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-dnkgw" event={"ID":"1c5e5471-de5a-4d1c-a983-11c4864d885d","Type":"ContainerDied","Data":"82005e1f83f8b46234beba92a0e5ec0b649aa08886b22a8443ab3211e3bf7fba"} Dec 10 23:42:25 crc kubenswrapper[4791]: I1210 23:42:25.642735 4791 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-dnkgw" event={"ID":"1c5e5471-de5a-4d1c-a983-11c4864d885d","Type":"ContainerDied","Data":"eb441cbd100dc547882448b687344e51a71261fd6eecc40e7373cc1418c04f98"} Dec 10 23:42:25 crc kubenswrapper[4791]: I1210 23:42:25.642748 4791 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-dnkgw" Dec 10 23:42:25 crc kubenswrapper[4791]: I1210 23:42:25.642763 4791 scope.go:117] "RemoveContainer" containerID="82005e1f83f8b46234beba92a0e5ec0b649aa08886b22a8443ab3211e3bf7fba" Dec 10 23:42:25 crc kubenswrapper[4791]: I1210 23:42:25.662402 4791 scope.go:117] "RemoveContainer" containerID="ebbee9909a1714802b2d92a30f776808b4dc4673ec0ba712e00ed36c2c5ff550" Dec 10 23:42:25 crc kubenswrapper[4791]: I1210 23:42:25.694191 4791 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-dnkgw"] Dec 10 23:42:25 crc kubenswrapper[4791]: I1210 23:42:25.702043 4791 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/certified-operators-dnkgw"] Dec 10 23:42:25 crc kubenswrapper[4791]: I1210 23:42:25.706759 4791 scope.go:117] "RemoveContainer" containerID="89a473e4cd415e3c739b514bce03a07f0dbb28b4f309080d2093c8daa34aca2f" Dec 10 23:42:25 crc kubenswrapper[4791]: I1210 23:42:25.740529 4791 scope.go:117] "RemoveContainer" containerID="82005e1f83f8b46234beba92a0e5ec0b649aa08886b22a8443ab3211e3bf7fba" Dec 10 23:42:25 crc kubenswrapper[4791]: E1210 23:42:25.743424 4791 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"82005e1f83f8b46234beba92a0e5ec0b649aa08886b22a8443ab3211e3bf7fba\": container with ID starting with 82005e1f83f8b46234beba92a0e5ec0b649aa08886b22a8443ab3211e3bf7fba not found: ID does not exist" containerID="82005e1f83f8b46234beba92a0e5ec0b649aa08886b22a8443ab3211e3bf7fba" Dec 10 23:42:25 crc kubenswrapper[4791]: I1210 23:42:25.743469 4791 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"82005e1f83f8b46234beba92a0e5ec0b649aa08886b22a8443ab3211e3bf7fba"} err="failed to get container status \"82005e1f83f8b46234beba92a0e5ec0b649aa08886b22a8443ab3211e3bf7fba\": rpc error: code = NotFound desc = could not find container \"82005e1f83f8b46234beba92a0e5ec0b649aa08886b22a8443ab3211e3bf7fba\": container with ID starting with 82005e1f83f8b46234beba92a0e5ec0b649aa08886b22a8443ab3211e3bf7fba not found: ID does not exist" Dec 10 23:42:25 crc kubenswrapper[4791]: I1210 23:42:25.743497 4791 scope.go:117] "RemoveContainer" containerID="ebbee9909a1714802b2d92a30f776808b4dc4673ec0ba712e00ed36c2c5ff550" Dec 10 23:42:25 crc kubenswrapper[4791]: E1210 23:42:25.743842 4791 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"ebbee9909a1714802b2d92a30f776808b4dc4673ec0ba712e00ed36c2c5ff550\": container with ID starting with ebbee9909a1714802b2d92a30f776808b4dc4673ec0ba712e00ed36c2c5ff550 not found: ID does not exist" containerID="ebbee9909a1714802b2d92a30f776808b4dc4673ec0ba712e00ed36c2c5ff550" Dec 10 23:42:25 crc kubenswrapper[4791]: I1210 23:42:25.743905 4791 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"ebbee9909a1714802b2d92a30f776808b4dc4673ec0ba712e00ed36c2c5ff550"} err="failed to get container status \"ebbee9909a1714802b2d92a30f776808b4dc4673ec0ba712e00ed36c2c5ff550\": rpc error: code = NotFound desc = could not find container \"ebbee9909a1714802b2d92a30f776808b4dc4673ec0ba712e00ed36c2c5ff550\": container with ID starting with ebbee9909a1714802b2d92a30f776808b4dc4673ec0ba712e00ed36c2c5ff550 not found: ID does not exist" Dec 10 23:42:25 crc kubenswrapper[4791]: I1210 23:42:25.743928 4791 scope.go:117] "RemoveContainer" containerID="89a473e4cd415e3c739b514bce03a07f0dbb28b4f309080d2093c8daa34aca2f" Dec 10 23:42:25 crc kubenswrapper[4791]: E1210 23:42:25.744269 4791 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"89a473e4cd415e3c739b514bce03a07f0dbb28b4f309080d2093c8daa34aca2f\": container with ID starting with 89a473e4cd415e3c739b514bce03a07f0dbb28b4f309080d2093c8daa34aca2f not found: ID does not exist" containerID="89a473e4cd415e3c739b514bce03a07f0dbb28b4f309080d2093c8daa34aca2f" Dec 10 23:42:25 crc kubenswrapper[4791]: I1210 23:42:25.744292 4791 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"89a473e4cd415e3c739b514bce03a07f0dbb28b4f309080d2093c8daa34aca2f"} err="failed to get container status \"89a473e4cd415e3c739b514bce03a07f0dbb28b4f309080d2093c8daa34aca2f\": rpc error: code = NotFound desc = could not find container \"89a473e4cd415e3c739b514bce03a07f0dbb28b4f309080d2093c8daa34aca2f\": container with ID starting with 89a473e4cd415e3c739b514bce03a07f0dbb28b4f309080d2093c8daa34aca2f not found: ID does not exist" Dec 10 23:42:25 crc kubenswrapper[4791]: I1210 23:42:25.897149 4791 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="1c5e5471-de5a-4d1c-a983-11c4864d885d" path="/var/lib/kubelet/pods/1c5e5471-de5a-4d1c-a983-11c4864d885d/volumes" Dec 10 23:42:33 crc kubenswrapper[4791]: I1210 23:42:33.885523 4791 scope.go:117] "RemoveContainer" containerID="89d61e80933d85eeb2188e63c1702b39d9223c0393e3c83d39da9537a851be64" Dec 10 23:42:33 crc kubenswrapper[4791]: E1210 23:42:33.886707 4791 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-5rb5l_openshift-machine-config-operator(ba35653c-6e06-4cee-a4d6-137764090d18)\"" pod="openshift-machine-config-operator/machine-config-daemon-5rb5l" podUID="ba35653c-6e06-4cee-a4d6-137764090d18" Dec 10 23:42:48 crc kubenswrapper[4791]: I1210 23:42:48.884803 4791 scope.go:117] "RemoveContainer" containerID="89d61e80933d85eeb2188e63c1702b39d9223c0393e3c83d39da9537a851be64" Dec 10 23:42:48 crc kubenswrapper[4791]: E1210 23:42:48.885553 4791 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-5rb5l_openshift-machine-config-operator(ba35653c-6e06-4cee-a4d6-137764090d18)\"" pod="openshift-machine-config-operator/machine-config-daemon-5rb5l" podUID="ba35653c-6e06-4cee-a4d6-137764090d18" Dec 10 23:42:59 crc kubenswrapper[4791]: I1210 23:42:59.885369 4791 scope.go:117] "RemoveContainer" containerID="89d61e80933d85eeb2188e63c1702b39d9223c0393e3c83d39da9537a851be64" Dec 10 23:42:59 crc kubenswrapper[4791]: E1210 23:42:59.886237 4791 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-5rb5l_openshift-machine-config-operator(ba35653c-6e06-4cee-a4d6-137764090d18)\"" pod="openshift-machine-config-operator/machine-config-daemon-5rb5l" podUID="ba35653c-6e06-4cee-a4d6-137764090d18" Dec 10 23:43:11 crc kubenswrapper[4791]: I1210 23:43:11.886071 4791 scope.go:117] "RemoveContainer" containerID="89d61e80933d85eeb2188e63c1702b39d9223c0393e3c83d39da9537a851be64" Dec 10 23:43:11 crc kubenswrapper[4791]: E1210 23:43:11.887180 4791 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-5rb5l_openshift-machine-config-operator(ba35653c-6e06-4cee-a4d6-137764090d18)\"" pod="openshift-machine-config-operator/machine-config-daemon-5rb5l" podUID="ba35653c-6e06-4cee-a4d6-137764090d18" Dec 10 23:43:22 crc kubenswrapper[4791]: I1210 23:43:22.885252 4791 scope.go:117] "RemoveContainer" containerID="89d61e80933d85eeb2188e63c1702b39d9223c0393e3c83d39da9537a851be64" Dec 10 23:43:22 crc kubenswrapper[4791]: E1210 23:43:22.885876 4791 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-5rb5l_openshift-machine-config-operator(ba35653c-6e06-4cee-a4d6-137764090d18)\"" pod="openshift-machine-config-operator/machine-config-daemon-5rb5l" podUID="ba35653c-6e06-4cee-a4d6-137764090d18" Dec 10 23:43:34 crc kubenswrapper[4791]: I1210 23:43:34.885421 4791 scope.go:117] "RemoveContainer" containerID="89d61e80933d85eeb2188e63c1702b39d9223c0393e3c83d39da9537a851be64" Dec 10 23:43:34 crc kubenswrapper[4791]: E1210 23:43:34.886842 4791 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-5rb5l_openshift-machine-config-operator(ba35653c-6e06-4cee-a4d6-137764090d18)\"" pod="openshift-machine-config-operator/machine-config-daemon-5rb5l" podUID="ba35653c-6e06-4cee-a4d6-137764090d18" Dec 10 23:43:49 crc kubenswrapper[4791]: I1210 23:43:49.885213 4791 scope.go:117] "RemoveContainer" containerID="89d61e80933d85eeb2188e63c1702b39d9223c0393e3c83d39da9537a851be64" Dec 10 23:43:49 crc kubenswrapper[4791]: E1210 23:43:49.886677 4791 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-5rb5l_openshift-machine-config-operator(ba35653c-6e06-4cee-a4d6-137764090d18)\"" pod="openshift-machine-config-operator/machine-config-daemon-5rb5l" podUID="ba35653c-6e06-4cee-a4d6-137764090d18" Dec 10 23:44:04 crc kubenswrapper[4791]: I1210 23:44:04.884712 4791 scope.go:117] "RemoveContainer" containerID="89d61e80933d85eeb2188e63c1702b39d9223c0393e3c83d39da9537a851be64" Dec 10 23:44:05 crc kubenswrapper[4791]: I1210 23:44:05.642754 4791 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-5rb5l" event={"ID":"ba35653c-6e06-4cee-a4d6-137764090d18","Type":"ContainerStarted","Data":"b11a39d149fd20d0528f36ebfc87a9f2befcafcaa18f22a73d7824b34c5c7890"} Dec 10 23:45:00 crc kubenswrapper[4791]: I1210 23:45:00.152970 4791 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29423505-hg9bv"] Dec 10 23:45:00 crc kubenswrapper[4791]: E1210 23:45:00.154063 4791 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="1c5e5471-de5a-4d1c-a983-11c4864d885d" containerName="extract-content" Dec 10 23:45:00 crc kubenswrapper[4791]: I1210 23:45:00.154081 4791 state_mem.go:107] "Deleted CPUSet assignment" podUID="1c5e5471-de5a-4d1c-a983-11c4864d885d" containerName="extract-content" Dec 10 23:45:00 crc kubenswrapper[4791]: E1210 23:45:00.154101 4791 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="1c5e5471-de5a-4d1c-a983-11c4864d885d" containerName="registry-server" Dec 10 23:45:00 crc kubenswrapper[4791]: I1210 23:45:00.154109 4791 state_mem.go:107] "Deleted CPUSet assignment" podUID="1c5e5471-de5a-4d1c-a983-11c4864d885d" containerName="registry-server" Dec 10 23:45:00 crc kubenswrapper[4791]: E1210 23:45:00.154136 4791 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="1c5e5471-de5a-4d1c-a983-11c4864d885d" containerName="extract-utilities" Dec 10 23:45:00 crc kubenswrapper[4791]: I1210 23:45:00.154144 4791 state_mem.go:107] "Deleted CPUSet assignment" podUID="1c5e5471-de5a-4d1c-a983-11c4864d885d" containerName="extract-utilities" Dec 10 23:45:00 crc kubenswrapper[4791]: I1210 23:45:00.154413 4791 memory_manager.go:354] "RemoveStaleState removing state" podUID="1c5e5471-de5a-4d1c-a983-11c4864d885d" containerName="registry-server" Dec 10 23:45:00 crc kubenswrapper[4791]: I1210 23:45:00.155223 4791 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29423505-hg9bv" Dec 10 23:45:00 crc kubenswrapper[4791]: I1210 23:45:00.199062 4791 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"collect-profiles-dockercfg-kzf4t" Dec 10 23:45:00 crc kubenswrapper[4791]: I1210 23:45:00.199326 4791 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-operator-lifecycle-manager"/"collect-profiles-config" Dec 10 23:45:00 crc kubenswrapper[4791]: I1210 23:45:00.200406 4791 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/c41bf7d8-8cf8-4fd2-81d9-3cd14cee6fda-config-volume\") pod \"collect-profiles-29423505-hg9bv\" (UID: \"c41bf7d8-8cf8-4fd2-81d9-3cd14cee6fda\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29423505-hg9bv" Dec 10 23:45:00 crc kubenswrapper[4791]: I1210 23:45:00.200498 4791 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-t9vpc\" (UniqueName: \"kubernetes.io/projected/c41bf7d8-8cf8-4fd2-81d9-3cd14cee6fda-kube-api-access-t9vpc\") pod \"collect-profiles-29423505-hg9bv\" (UID: \"c41bf7d8-8cf8-4fd2-81d9-3cd14cee6fda\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29423505-hg9bv" Dec 10 23:45:00 crc kubenswrapper[4791]: I1210 23:45:00.200550 4791 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/c41bf7d8-8cf8-4fd2-81d9-3cd14cee6fda-secret-volume\") pod \"collect-profiles-29423505-hg9bv\" (UID: \"c41bf7d8-8cf8-4fd2-81d9-3cd14cee6fda\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29423505-hg9bv" Dec 10 23:45:00 crc kubenswrapper[4791]: I1210 23:45:00.209060 4791 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29423505-hg9bv"] Dec 10 23:45:00 crc kubenswrapper[4791]: I1210 23:45:00.302554 4791 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/c41bf7d8-8cf8-4fd2-81d9-3cd14cee6fda-config-volume\") pod \"collect-profiles-29423505-hg9bv\" (UID: \"c41bf7d8-8cf8-4fd2-81d9-3cd14cee6fda\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29423505-hg9bv" Dec 10 23:45:00 crc kubenswrapper[4791]: I1210 23:45:00.302618 4791 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-t9vpc\" (UniqueName: \"kubernetes.io/projected/c41bf7d8-8cf8-4fd2-81d9-3cd14cee6fda-kube-api-access-t9vpc\") pod \"collect-profiles-29423505-hg9bv\" (UID: \"c41bf7d8-8cf8-4fd2-81d9-3cd14cee6fda\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29423505-hg9bv" Dec 10 23:45:00 crc kubenswrapper[4791]: I1210 23:45:00.302665 4791 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/c41bf7d8-8cf8-4fd2-81d9-3cd14cee6fda-secret-volume\") pod \"collect-profiles-29423505-hg9bv\" (UID: \"c41bf7d8-8cf8-4fd2-81d9-3cd14cee6fda\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29423505-hg9bv" Dec 10 23:45:00 crc kubenswrapper[4791]: I1210 23:45:00.303519 4791 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/c41bf7d8-8cf8-4fd2-81d9-3cd14cee6fda-config-volume\") pod \"collect-profiles-29423505-hg9bv\" (UID: \"c41bf7d8-8cf8-4fd2-81d9-3cd14cee6fda\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29423505-hg9bv" Dec 10 23:45:00 crc kubenswrapper[4791]: I1210 23:45:00.309053 4791 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/c41bf7d8-8cf8-4fd2-81d9-3cd14cee6fda-secret-volume\") pod \"collect-profiles-29423505-hg9bv\" (UID: \"c41bf7d8-8cf8-4fd2-81d9-3cd14cee6fda\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29423505-hg9bv" Dec 10 23:45:00 crc kubenswrapper[4791]: I1210 23:45:00.320916 4791 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-t9vpc\" (UniqueName: \"kubernetes.io/projected/c41bf7d8-8cf8-4fd2-81d9-3cd14cee6fda-kube-api-access-t9vpc\") pod \"collect-profiles-29423505-hg9bv\" (UID: \"c41bf7d8-8cf8-4fd2-81d9-3cd14cee6fda\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29423505-hg9bv" Dec 10 23:45:00 crc kubenswrapper[4791]: I1210 23:45:00.520691 4791 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29423505-hg9bv" Dec 10 23:45:01 crc kubenswrapper[4791]: I1210 23:45:01.024367 4791 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29423505-hg9bv"] Dec 10 23:45:01 crc kubenswrapper[4791]: I1210 23:45:01.277509 4791 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29423505-hg9bv" event={"ID":"c41bf7d8-8cf8-4fd2-81d9-3cd14cee6fda","Type":"ContainerStarted","Data":"70759bc397837a28be63c40a126e9bda155606f709ec47f6e8358f0628cb1e48"} Dec 10 23:45:01 crc kubenswrapper[4791]: I1210 23:45:01.279039 4791 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29423505-hg9bv" event={"ID":"c41bf7d8-8cf8-4fd2-81d9-3cd14cee6fda","Type":"ContainerStarted","Data":"93b73be4b14538c24e0bb9ec6388c718d7cdbd36b2d949193bc071f106690bd4"} Dec 10 23:45:01 crc kubenswrapper[4791]: I1210 23:45:01.299857 4791 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-operator-lifecycle-manager/collect-profiles-29423505-hg9bv" podStartSLOduration=1.299835075 podStartE2EDuration="1.299835075s" podCreationTimestamp="2025-12-10 23:45:00 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-10 23:45:01.293712982 +0000 UTC m=+3335.723330595" watchObservedRunningTime="2025-12-10 23:45:01.299835075 +0000 UTC m=+3335.729452688" Dec 10 23:45:02 crc kubenswrapper[4791]: I1210 23:45:02.286588 4791 generic.go:334] "Generic (PLEG): container finished" podID="c41bf7d8-8cf8-4fd2-81d9-3cd14cee6fda" containerID="70759bc397837a28be63c40a126e9bda155606f709ec47f6e8358f0628cb1e48" exitCode=0 Dec 10 23:45:02 crc kubenswrapper[4791]: I1210 23:45:02.286707 4791 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29423505-hg9bv" event={"ID":"c41bf7d8-8cf8-4fd2-81d9-3cd14cee6fda","Type":"ContainerDied","Data":"70759bc397837a28be63c40a126e9bda155606f709ec47f6e8358f0628cb1e48"} Dec 10 23:45:03 crc kubenswrapper[4791]: I1210 23:45:03.831312 4791 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29423505-hg9bv" Dec 10 23:45:03 crc kubenswrapper[4791]: I1210 23:45:03.988364 4791 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/c41bf7d8-8cf8-4fd2-81d9-3cd14cee6fda-secret-volume\") pod \"c41bf7d8-8cf8-4fd2-81d9-3cd14cee6fda\" (UID: \"c41bf7d8-8cf8-4fd2-81d9-3cd14cee6fda\") " Dec 10 23:45:03 crc kubenswrapper[4791]: I1210 23:45:03.988465 4791 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/c41bf7d8-8cf8-4fd2-81d9-3cd14cee6fda-config-volume\") pod \"c41bf7d8-8cf8-4fd2-81d9-3cd14cee6fda\" (UID: \"c41bf7d8-8cf8-4fd2-81d9-3cd14cee6fda\") " Dec 10 23:45:03 crc kubenswrapper[4791]: I1210 23:45:03.988553 4791 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-t9vpc\" (UniqueName: \"kubernetes.io/projected/c41bf7d8-8cf8-4fd2-81d9-3cd14cee6fda-kube-api-access-t9vpc\") pod \"c41bf7d8-8cf8-4fd2-81d9-3cd14cee6fda\" (UID: \"c41bf7d8-8cf8-4fd2-81d9-3cd14cee6fda\") " Dec 10 23:45:03 crc kubenswrapper[4791]: I1210 23:45:03.989018 4791 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/c41bf7d8-8cf8-4fd2-81d9-3cd14cee6fda-config-volume" (OuterVolumeSpecName: "config-volume") pod "c41bf7d8-8cf8-4fd2-81d9-3cd14cee6fda" (UID: "c41bf7d8-8cf8-4fd2-81d9-3cd14cee6fda"). InnerVolumeSpecName "config-volume". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 10 23:45:03 crc kubenswrapper[4791]: I1210 23:45:03.989326 4791 reconciler_common.go:293] "Volume detached for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/c41bf7d8-8cf8-4fd2-81d9-3cd14cee6fda-config-volume\") on node \"crc\" DevicePath \"\"" Dec 10 23:45:03 crc kubenswrapper[4791]: I1210 23:45:03.993502 4791 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/c41bf7d8-8cf8-4fd2-81d9-3cd14cee6fda-secret-volume" (OuterVolumeSpecName: "secret-volume") pod "c41bf7d8-8cf8-4fd2-81d9-3cd14cee6fda" (UID: "c41bf7d8-8cf8-4fd2-81d9-3cd14cee6fda"). InnerVolumeSpecName "secret-volume". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 10 23:45:04 crc kubenswrapper[4791]: I1210 23:45:04.011002 4791 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/c41bf7d8-8cf8-4fd2-81d9-3cd14cee6fda-kube-api-access-t9vpc" (OuterVolumeSpecName: "kube-api-access-t9vpc") pod "c41bf7d8-8cf8-4fd2-81d9-3cd14cee6fda" (UID: "c41bf7d8-8cf8-4fd2-81d9-3cd14cee6fda"). InnerVolumeSpecName "kube-api-access-t9vpc". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 10 23:45:04 crc kubenswrapper[4791]: I1210 23:45:04.091756 4791 reconciler_common.go:293] "Volume detached for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/c41bf7d8-8cf8-4fd2-81d9-3cd14cee6fda-secret-volume\") on node \"crc\" DevicePath \"\"" Dec 10 23:45:04 crc kubenswrapper[4791]: I1210 23:45:04.091805 4791 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-t9vpc\" (UniqueName: \"kubernetes.io/projected/c41bf7d8-8cf8-4fd2-81d9-3cd14cee6fda-kube-api-access-t9vpc\") on node \"crc\" DevicePath \"\"" Dec 10 23:45:04 crc kubenswrapper[4791]: I1210 23:45:04.308274 4791 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29423505-hg9bv" event={"ID":"c41bf7d8-8cf8-4fd2-81d9-3cd14cee6fda","Type":"ContainerDied","Data":"93b73be4b14538c24e0bb9ec6388c718d7cdbd36b2d949193bc071f106690bd4"} Dec 10 23:45:04 crc kubenswrapper[4791]: I1210 23:45:04.308324 4791 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="93b73be4b14538c24e0bb9ec6388c718d7cdbd36b2d949193bc071f106690bd4" Dec 10 23:45:04 crc kubenswrapper[4791]: I1210 23:45:04.308376 4791 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29423505-hg9bv" Dec 10 23:45:04 crc kubenswrapper[4791]: I1210 23:45:04.372727 4791 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29423460-rp94k"] Dec 10 23:45:04 crc kubenswrapper[4791]: I1210 23:45:04.380410 4791 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29423460-rp94k"] Dec 10 23:45:05 crc kubenswrapper[4791]: I1210 23:45:05.904249 4791 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="262dcaa0-a67b-421d-ad84-95eaaa3d6a7f" path="/var/lib/kubelet/pods/262dcaa0-a67b-421d-ad84-95eaaa3d6a7f/volumes" Dec 10 23:45:54 crc kubenswrapper[4791]: I1210 23:45:54.670597 4791 scope.go:117] "RemoveContainer" containerID="0ff11197939d88bcbbaa9c417821a4a555a4f94ceb603408b0d6283461991b2c" Dec 10 23:46:25 crc kubenswrapper[4791]: I1210 23:46:25.038166 4791 patch_prober.go:28] interesting pod/machine-config-daemon-5rb5l container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 10 23:46:25 crc kubenswrapper[4791]: I1210 23:46:25.039096 4791 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-5rb5l" podUID="ba35653c-6e06-4cee-a4d6-137764090d18" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 10 23:46:55 crc kubenswrapper[4791]: I1210 23:46:55.037498 4791 patch_prober.go:28] interesting pod/machine-config-daemon-5rb5l container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 10 23:46:55 crc kubenswrapper[4791]: I1210 23:46:55.037891 4791 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-5rb5l" podUID="ba35653c-6e06-4cee-a4d6-137764090d18" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 10 23:47:04 crc kubenswrapper[4791]: I1210 23:47:04.549103 4791 generic.go:334] "Generic (PLEG): container finished" podID="cf11ce8f-d433-49ff-9390-064a2f4ec60d" containerID="f73cf0bd24d3f0b1d51a66e7bef480518a1fbb1721aadb250c779b5578569e42" exitCode=0 Dec 10 23:47:04 crc kubenswrapper[4791]: I1210 23:47:04.549242 4791 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/tempest-tests-tempest" event={"ID":"cf11ce8f-d433-49ff-9390-064a2f4ec60d","Type":"ContainerDied","Data":"f73cf0bd24d3f0b1d51a66e7bef480518a1fbb1721aadb250c779b5578569e42"} Dec 10 23:47:06 crc kubenswrapper[4791]: I1210 23:47:06.016183 4791 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/tempest-tests-tempest" Dec 10 23:47:06 crc kubenswrapper[4791]: I1210 23:47:06.211607 4791 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"test-operator-logs\" (UniqueName: \"kubernetes.io/local-volume/local-storage05-crc\") pod \"cf11ce8f-d433-49ff-9390-064a2f4ec60d\" (UID: \"cf11ce8f-d433-49ff-9390-064a2f4ec60d\") " Dec 10 23:47:06 crc kubenswrapper[4791]: I1210 23:47:06.211671 4791 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"openstack-config-secret\" (UniqueName: \"kubernetes.io/secret/cf11ce8f-d433-49ff-9390-064a2f4ec60d-openstack-config-secret\") pod \"cf11ce8f-d433-49ff-9390-064a2f4ec60d\" (UID: \"cf11ce8f-d433-49ff-9390-064a2f4ec60d\") " Dec 10 23:47:06 crc kubenswrapper[4791]: I1210 23:47:06.211700 4791 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"test-operator-ephemeral-workdir\" (UniqueName: \"kubernetes.io/empty-dir/cf11ce8f-d433-49ff-9390-064a2f4ec60d-test-operator-ephemeral-workdir\") pod \"cf11ce8f-d433-49ff-9390-064a2f4ec60d\" (UID: \"cf11ce8f-d433-49ff-9390-064a2f4ec60d\") " Dec 10 23:47:06 crc kubenswrapper[4791]: I1210 23:47:06.211720 4791 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-5q7hd\" (UniqueName: \"kubernetes.io/projected/cf11ce8f-d433-49ff-9390-064a2f4ec60d-kube-api-access-5q7hd\") pod \"cf11ce8f-d433-49ff-9390-064a2f4ec60d\" (UID: \"cf11ce8f-d433-49ff-9390-064a2f4ec60d\") " Dec 10 23:47:06 crc kubenswrapper[4791]: I1210 23:47:06.211777 4791 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/cf11ce8f-d433-49ff-9390-064a2f4ec60d-ssh-key\") pod \"cf11ce8f-d433-49ff-9390-064a2f4ec60d\" (UID: \"cf11ce8f-d433-49ff-9390-064a2f4ec60d\") " Dec 10 23:47:06 crc kubenswrapper[4791]: I1210 23:47:06.211895 4791 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"openstack-config\" (UniqueName: \"kubernetes.io/configmap/cf11ce8f-d433-49ff-9390-064a2f4ec60d-openstack-config\") pod \"cf11ce8f-d433-49ff-9390-064a2f4ec60d\" (UID: \"cf11ce8f-d433-49ff-9390-064a2f4ec60d\") " Dec 10 23:47:06 crc kubenswrapper[4791]: I1210 23:47:06.211939 4791 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ca-certs\" (UniqueName: \"kubernetes.io/secret/cf11ce8f-d433-49ff-9390-064a2f4ec60d-ca-certs\") pod \"cf11ce8f-d433-49ff-9390-064a2f4ec60d\" (UID: \"cf11ce8f-d433-49ff-9390-064a2f4ec60d\") " Dec 10 23:47:06 crc kubenswrapper[4791]: I1210 23:47:06.211963 4791 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/cf11ce8f-d433-49ff-9390-064a2f4ec60d-config-data\") pod \"cf11ce8f-d433-49ff-9390-064a2f4ec60d\" (UID: \"cf11ce8f-d433-49ff-9390-064a2f4ec60d\") " Dec 10 23:47:06 crc kubenswrapper[4791]: I1210 23:47:06.211991 4791 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"test-operator-ephemeral-temporary\" (UniqueName: \"kubernetes.io/empty-dir/cf11ce8f-d433-49ff-9390-064a2f4ec60d-test-operator-ephemeral-temporary\") pod \"cf11ce8f-d433-49ff-9390-064a2f4ec60d\" (UID: \"cf11ce8f-d433-49ff-9390-064a2f4ec60d\") " Dec 10 23:47:06 crc kubenswrapper[4791]: I1210 23:47:06.214037 4791 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/cf11ce8f-d433-49ff-9390-064a2f4ec60d-config-data" (OuterVolumeSpecName: "config-data") pod "cf11ce8f-d433-49ff-9390-064a2f4ec60d" (UID: "cf11ce8f-d433-49ff-9390-064a2f4ec60d"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 10 23:47:06 crc kubenswrapper[4791]: I1210 23:47:06.214051 4791 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/cf11ce8f-d433-49ff-9390-064a2f4ec60d-test-operator-ephemeral-temporary" (OuterVolumeSpecName: "test-operator-ephemeral-temporary") pod "cf11ce8f-d433-49ff-9390-064a2f4ec60d" (UID: "cf11ce8f-d433-49ff-9390-064a2f4ec60d"). InnerVolumeSpecName "test-operator-ephemeral-temporary". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 10 23:47:06 crc kubenswrapper[4791]: I1210 23:47:06.218012 4791 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/local-volume/local-storage05-crc" (OuterVolumeSpecName: "test-operator-logs") pod "cf11ce8f-d433-49ff-9390-064a2f4ec60d" (UID: "cf11ce8f-d433-49ff-9390-064a2f4ec60d"). InnerVolumeSpecName "local-storage05-crc". PluginName "kubernetes.io/local-volume", VolumeGidValue "" Dec 10 23:47:06 crc kubenswrapper[4791]: I1210 23:47:06.218289 4791 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/cf11ce8f-d433-49ff-9390-064a2f4ec60d-kube-api-access-5q7hd" (OuterVolumeSpecName: "kube-api-access-5q7hd") pod "cf11ce8f-d433-49ff-9390-064a2f4ec60d" (UID: "cf11ce8f-d433-49ff-9390-064a2f4ec60d"). InnerVolumeSpecName "kube-api-access-5q7hd". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 10 23:47:06 crc kubenswrapper[4791]: I1210 23:47:06.219565 4791 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/cf11ce8f-d433-49ff-9390-064a2f4ec60d-test-operator-ephemeral-workdir" (OuterVolumeSpecName: "test-operator-ephemeral-workdir") pod "cf11ce8f-d433-49ff-9390-064a2f4ec60d" (UID: "cf11ce8f-d433-49ff-9390-064a2f4ec60d"). InnerVolumeSpecName "test-operator-ephemeral-workdir". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 10 23:47:06 crc kubenswrapper[4791]: I1210 23:47:06.245214 4791 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/cf11ce8f-d433-49ff-9390-064a2f4ec60d-openstack-config-secret" (OuterVolumeSpecName: "openstack-config-secret") pod "cf11ce8f-d433-49ff-9390-064a2f4ec60d" (UID: "cf11ce8f-d433-49ff-9390-064a2f4ec60d"). InnerVolumeSpecName "openstack-config-secret". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 10 23:47:06 crc kubenswrapper[4791]: I1210 23:47:06.249116 4791 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/cf11ce8f-d433-49ff-9390-064a2f4ec60d-ca-certs" (OuterVolumeSpecName: "ca-certs") pod "cf11ce8f-d433-49ff-9390-064a2f4ec60d" (UID: "cf11ce8f-d433-49ff-9390-064a2f4ec60d"). InnerVolumeSpecName "ca-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 10 23:47:06 crc kubenswrapper[4791]: I1210 23:47:06.259671 4791 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/cf11ce8f-d433-49ff-9390-064a2f4ec60d-openstack-config" (OuterVolumeSpecName: "openstack-config") pod "cf11ce8f-d433-49ff-9390-064a2f4ec60d" (UID: "cf11ce8f-d433-49ff-9390-064a2f4ec60d"). InnerVolumeSpecName "openstack-config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 10 23:47:06 crc kubenswrapper[4791]: I1210 23:47:06.269294 4791 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/cf11ce8f-d433-49ff-9390-064a2f4ec60d-ssh-key" (OuterVolumeSpecName: "ssh-key") pod "cf11ce8f-d433-49ff-9390-064a2f4ec60d" (UID: "cf11ce8f-d433-49ff-9390-064a2f4ec60d"). InnerVolumeSpecName "ssh-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 10 23:47:06 crc kubenswrapper[4791]: I1210 23:47:06.316759 4791 reconciler_common.go:293] "Volume detached for volume \"openstack-config\" (UniqueName: \"kubernetes.io/configmap/cf11ce8f-d433-49ff-9390-064a2f4ec60d-openstack-config\") on node \"crc\" DevicePath \"\"" Dec 10 23:47:06 crc kubenswrapper[4791]: I1210 23:47:06.316865 4791 reconciler_common.go:293] "Volume detached for volume \"ca-certs\" (UniqueName: \"kubernetes.io/secret/cf11ce8f-d433-49ff-9390-064a2f4ec60d-ca-certs\") on node \"crc\" DevicePath \"\"" Dec 10 23:47:06 crc kubenswrapper[4791]: I1210 23:47:06.316884 4791 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/cf11ce8f-d433-49ff-9390-064a2f4ec60d-config-data\") on node \"crc\" DevicePath \"\"" Dec 10 23:47:06 crc kubenswrapper[4791]: I1210 23:47:06.316907 4791 reconciler_common.go:293] "Volume detached for volume \"test-operator-ephemeral-temporary\" (UniqueName: \"kubernetes.io/empty-dir/cf11ce8f-d433-49ff-9390-064a2f4ec60d-test-operator-ephemeral-temporary\") on node \"crc\" DevicePath \"\"" Dec 10 23:47:06 crc kubenswrapper[4791]: I1210 23:47:06.316963 4791 reconciler_common.go:286] "operationExecutor.UnmountDevice started for volume \"local-storage05-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage05-crc\") on node \"crc\" " Dec 10 23:47:06 crc kubenswrapper[4791]: I1210 23:47:06.316984 4791 reconciler_common.go:293] "Volume detached for volume \"openstack-config-secret\" (UniqueName: \"kubernetes.io/secret/cf11ce8f-d433-49ff-9390-064a2f4ec60d-openstack-config-secret\") on node \"crc\" DevicePath \"\"" Dec 10 23:47:06 crc kubenswrapper[4791]: I1210 23:47:06.317006 4791 reconciler_common.go:293] "Volume detached for volume \"test-operator-ephemeral-workdir\" (UniqueName: \"kubernetes.io/empty-dir/cf11ce8f-d433-49ff-9390-064a2f4ec60d-test-operator-ephemeral-workdir\") on node \"crc\" DevicePath \"\"" Dec 10 23:47:06 crc kubenswrapper[4791]: I1210 23:47:06.317026 4791 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-5q7hd\" (UniqueName: \"kubernetes.io/projected/cf11ce8f-d433-49ff-9390-064a2f4ec60d-kube-api-access-5q7hd\") on node \"crc\" DevicePath \"\"" Dec 10 23:47:06 crc kubenswrapper[4791]: I1210 23:47:06.317045 4791 reconciler_common.go:293] "Volume detached for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/cf11ce8f-d433-49ff-9390-064a2f4ec60d-ssh-key\") on node \"crc\" DevicePath \"\"" Dec 10 23:47:06 crc kubenswrapper[4791]: I1210 23:47:06.352136 4791 operation_generator.go:917] UnmountDevice succeeded for volume "local-storage05-crc" (UniqueName: "kubernetes.io/local-volume/local-storage05-crc") on node "crc" Dec 10 23:47:06 crc kubenswrapper[4791]: I1210 23:47:06.419389 4791 reconciler_common.go:293] "Volume detached for volume \"local-storage05-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage05-crc\") on node \"crc\" DevicePath \"\"" Dec 10 23:47:06 crc kubenswrapper[4791]: I1210 23:47:06.571424 4791 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/tempest-tests-tempest" event={"ID":"cf11ce8f-d433-49ff-9390-064a2f4ec60d","Type":"ContainerDied","Data":"9f1caab4a7c10473b84a7da85950f86eeabba357974aa29c56bc77427421353d"} Dec 10 23:47:06 crc kubenswrapper[4791]: I1210 23:47:06.571464 4791 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/tempest-tests-tempest" Dec 10 23:47:06 crc kubenswrapper[4791]: I1210 23:47:06.571471 4791 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="9f1caab4a7c10473b84a7da85950f86eeabba357974aa29c56bc77427421353d" Dec 10 23:47:16 crc kubenswrapper[4791]: I1210 23:47:16.853206 4791 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/test-operator-logs-pod-tempest-tempest-tests-tempest"] Dec 10 23:47:16 crc kubenswrapper[4791]: E1210 23:47:16.854643 4791 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="cf11ce8f-d433-49ff-9390-064a2f4ec60d" containerName="tempest-tests-tempest-tests-runner" Dec 10 23:47:16 crc kubenswrapper[4791]: I1210 23:47:16.854671 4791 state_mem.go:107] "Deleted CPUSet assignment" podUID="cf11ce8f-d433-49ff-9390-064a2f4ec60d" containerName="tempest-tests-tempest-tests-runner" Dec 10 23:47:16 crc kubenswrapper[4791]: E1210 23:47:16.854709 4791 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="c41bf7d8-8cf8-4fd2-81d9-3cd14cee6fda" containerName="collect-profiles" Dec 10 23:47:16 crc kubenswrapper[4791]: I1210 23:47:16.854722 4791 state_mem.go:107] "Deleted CPUSet assignment" podUID="c41bf7d8-8cf8-4fd2-81d9-3cd14cee6fda" containerName="collect-profiles" Dec 10 23:47:16 crc kubenswrapper[4791]: I1210 23:47:16.855100 4791 memory_manager.go:354] "RemoveStaleState removing state" podUID="cf11ce8f-d433-49ff-9390-064a2f4ec60d" containerName="tempest-tests-tempest-tests-runner" Dec 10 23:47:16 crc kubenswrapper[4791]: I1210 23:47:16.855140 4791 memory_manager.go:354] "RemoveStaleState removing state" podUID="c41bf7d8-8cf8-4fd2-81d9-3cd14cee6fda" containerName="collect-profiles" Dec 10 23:47:16 crc kubenswrapper[4791]: I1210 23:47:16.856218 4791 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/test-operator-logs-pod-tempest-tempest-tests-tempest" Dec 10 23:47:16 crc kubenswrapper[4791]: I1210 23:47:16.860049 4791 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"default-dockercfg-2z5p4" Dec 10 23:47:16 crc kubenswrapper[4791]: I1210 23:47:16.882432 4791 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/test-operator-logs-pod-tempest-tempest-tests-tempest"] Dec 10 23:47:16 crc kubenswrapper[4791]: I1210 23:47:16.962876 4791 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-n4rs8\" (UniqueName: \"kubernetes.io/projected/00c94f56-3118-486c-a325-9493d5bfd911-kube-api-access-n4rs8\") pod \"test-operator-logs-pod-tempest-tempest-tests-tempest\" (UID: \"00c94f56-3118-486c-a325-9493d5bfd911\") " pod="openstack/test-operator-logs-pod-tempest-tempest-tests-tempest" Dec 10 23:47:16 crc kubenswrapper[4791]: I1210 23:47:16.963094 4791 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage05-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage05-crc\") pod \"test-operator-logs-pod-tempest-tempest-tests-tempest\" (UID: \"00c94f56-3118-486c-a325-9493d5bfd911\") " pod="openstack/test-operator-logs-pod-tempest-tempest-tests-tempest" Dec 10 23:47:17 crc kubenswrapper[4791]: I1210 23:47:17.064555 4791 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-n4rs8\" (UniqueName: \"kubernetes.io/projected/00c94f56-3118-486c-a325-9493d5bfd911-kube-api-access-n4rs8\") pod \"test-operator-logs-pod-tempest-tempest-tests-tempest\" (UID: \"00c94f56-3118-486c-a325-9493d5bfd911\") " pod="openstack/test-operator-logs-pod-tempest-tempest-tests-tempest" Dec 10 23:47:17 crc kubenswrapper[4791]: I1210 23:47:17.064688 4791 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage05-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage05-crc\") pod \"test-operator-logs-pod-tempest-tempest-tests-tempest\" (UID: \"00c94f56-3118-486c-a325-9493d5bfd911\") " pod="openstack/test-operator-logs-pod-tempest-tempest-tests-tempest" Dec 10 23:47:17 crc kubenswrapper[4791]: I1210 23:47:17.065118 4791 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage05-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage05-crc\") pod \"test-operator-logs-pod-tempest-tempest-tests-tempest\" (UID: \"00c94f56-3118-486c-a325-9493d5bfd911\") device mount path \"/mnt/openstack/pv05\"" pod="openstack/test-operator-logs-pod-tempest-tempest-tests-tempest" Dec 10 23:47:17 crc kubenswrapper[4791]: I1210 23:47:17.105971 4791 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage05-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage05-crc\") pod \"test-operator-logs-pod-tempest-tempest-tests-tempest\" (UID: \"00c94f56-3118-486c-a325-9493d5bfd911\") " pod="openstack/test-operator-logs-pod-tempest-tempest-tests-tempest" Dec 10 23:47:17 crc kubenswrapper[4791]: I1210 23:47:17.113504 4791 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-n4rs8\" (UniqueName: \"kubernetes.io/projected/00c94f56-3118-486c-a325-9493d5bfd911-kube-api-access-n4rs8\") pod \"test-operator-logs-pod-tempest-tempest-tests-tempest\" (UID: \"00c94f56-3118-486c-a325-9493d5bfd911\") " pod="openstack/test-operator-logs-pod-tempest-tempest-tests-tempest" Dec 10 23:47:17 crc kubenswrapper[4791]: I1210 23:47:17.190010 4791 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/test-operator-logs-pod-tempest-tempest-tests-tempest" Dec 10 23:47:17 crc kubenswrapper[4791]: I1210 23:47:17.736334 4791 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/test-operator-logs-pod-tempest-tempest-tests-tempest"] Dec 10 23:47:17 crc kubenswrapper[4791]: I1210 23:47:17.747000 4791 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Dec 10 23:47:17 crc kubenswrapper[4791]: I1210 23:47:17.773266 4791 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/test-operator-logs-pod-tempest-tempest-tests-tempest" event={"ID":"00c94f56-3118-486c-a325-9493d5bfd911","Type":"ContainerStarted","Data":"fd5e72515623408229388b726cb111fcfde3a6287655b04c8371d18e8110f214"} Dec 10 23:47:19 crc kubenswrapper[4791]: I1210 23:47:19.796908 4791 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/test-operator-logs-pod-tempest-tempest-tests-tempest" event={"ID":"00c94f56-3118-486c-a325-9493d5bfd911","Type":"ContainerStarted","Data":"0bc0a26c912d9089a84f179e7a6c84c19567ec14bbbe724e21b39a0beb3288c9"} Dec 10 23:47:19 crc kubenswrapper[4791]: I1210 23:47:19.812463 4791 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/test-operator-logs-pod-tempest-tempest-tests-tempest" podStartSLOduration=2.548753264 podStartE2EDuration="3.812432281s" podCreationTimestamp="2025-12-10 23:47:16 +0000 UTC" firstStartedPulling="2025-12-10 23:47:17.746710496 +0000 UTC m=+3472.176328119" lastFinishedPulling="2025-12-10 23:47:19.010389503 +0000 UTC m=+3473.440007136" observedRunningTime="2025-12-10 23:47:19.811874566 +0000 UTC m=+3474.241492219" watchObservedRunningTime="2025-12-10 23:47:19.812432281 +0000 UTC m=+3474.242049934" Dec 10 23:47:25 crc kubenswrapper[4791]: I1210 23:47:25.038292 4791 patch_prober.go:28] interesting pod/machine-config-daemon-5rb5l container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 10 23:47:25 crc kubenswrapper[4791]: I1210 23:47:25.039691 4791 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-5rb5l" podUID="ba35653c-6e06-4cee-a4d6-137764090d18" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 10 23:47:25 crc kubenswrapper[4791]: I1210 23:47:25.039777 4791 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-5rb5l" Dec 10 23:47:25 crc kubenswrapper[4791]: I1210 23:47:25.041050 4791 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"b11a39d149fd20d0528f36ebfc87a9f2befcafcaa18f22a73d7824b34c5c7890"} pod="openshift-machine-config-operator/machine-config-daemon-5rb5l" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Dec 10 23:47:25 crc kubenswrapper[4791]: I1210 23:47:25.041178 4791 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-5rb5l" podUID="ba35653c-6e06-4cee-a4d6-137764090d18" containerName="machine-config-daemon" containerID="cri-o://b11a39d149fd20d0528f36ebfc87a9f2befcafcaa18f22a73d7824b34c5c7890" gracePeriod=600 Dec 10 23:47:25 crc kubenswrapper[4791]: I1210 23:47:25.866692 4791 generic.go:334] "Generic (PLEG): container finished" podID="ba35653c-6e06-4cee-a4d6-137764090d18" containerID="b11a39d149fd20d0528f36ebfc87a9f2befcafcaa18f22a73d7824b34c5c7890" exitCode=0 Dec 10 23:47:25 crc kubenswrapper[4791]: I1210 23:47:25.866772 4791 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-5rb5l" event={"ID":"ba35653c-6e06-4cee-a4d6-137764090d18","Type":"ContainerDied","Data":"b11a39d149fd20d0528f36ebfc87a9f2befcafcaa18f22a73d7824b34c5c7890"} Dec 10 23:47:25 crc kubenswrapper[4791]: I1210 23:47:25.867479 4791 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-5rb5l" event={"ID":"ba35653c-6e06-4cee-a4d6-137764090d18","Type":"ContainerStarted","Data":"f44d2e51d8375135a5705f7cf71dde0bd0c6901d20f892b1346dc2eac144cbc0"} Dec 10 23:47:25 crc kubenswrapper[4791]: I1210 23:47:25.867530 4791 scope.go:117] "RemoveContainer" containerID="89d61e80933d85eeb2188e63c1702b39d9223c0393e3c83d39da9537a851be64" Dec 10 23:47:41 crc kubenswrapper[4791]: I1210 23:47:41.300137 4791 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-must-gather-6t57w/must-gather-drn2z"] Dec 10 23:47:41 crc kubenswrapper[4791]: I1210 23:47:41.302609 4791 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-6t57w/must-gather-drn2z" Dec 10 23:47:41 crc kubenswrapper[4791]: I1210 23:47:41.304848 4791 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-must-gather-6t57w"/"kube-root-ca.crt" Dec 10 23:47:41 crc kubenswrapper[4791]: I1210 23:47:41.305113 4791 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-must-gather-6t57w"/"openshift-service-ca.crt" Dec 10 23:47:41 crc kubenswrapper[4791]: I1210 23:47:41.305633 4791 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-must-gather-6t57w"/"default-dockercfg-p9ffh" Dec 10 23:47:41 crc kubenswrapper[4791]: I1210 23:47:41.309320 4791 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-must-gather-6t57w/must-gather-drn2z"] Dec 10 23:47:41 crc kubenswrapper[4791]: I1210 23:47:41.361775 4791 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"must-gather-output\" (UniqueName: \"kubernetes.io/empty-dir/2171ff1e-63b0-41ff-8019-5a72a0c59f5f-must-gather-output\") pod \"must-gather-drn2z\" (UID: \"2171ff1e-63b0-41ff-8019-5a72a0c59f5f\") " pod="openshift-must-gather-6t57w/must-gather-drn2z" Dec 10 23:47:41 crc kubenswrapper[4791]: I1210 23:47:41.361843 4791 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-29zpv\" (UniqueName: \"kubernetes.io/projected/2171ff1e-63b0-41ff-8019-5a72a0c59f5f-kube-api-access-29zpv\") pod \"must-gather-drn2z\" (UID: \"2171ff1e-63b0-41ff-8019-5a72a0c59f5f\") " pod="openshift-must-gather-6t57w/must-gather-drn2z" Dec 10 23:47:41 crc kubenswrapper[4791]: I1210 23:47:41.463067 4791 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-29zpv\" (UniqueName: \"kubernetes.io/projected/2171ff1e-63b0-41ff-8019-5a72a0c59f5f-kube-api-access-29zpv\") pod \"must-gather-drn2z\" (UID: \"2171ff1e-63b0-41ff-8019-5a72a0c59f5f\") " pod="openshift-must-gather-6t57w/must-gather-drn2z" Dec 10 23:47:41 crc kubenswrapper[4791]: I1210 23:47:41.463301 4791 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"must-gather-output\" (UniqueName: \"kubernetes.io/empty-dir/2171ff1e-63b0-41ff-8019-5a72a0c59f5f-must-gather-output\") pod \"must-gather-drn2z\" (UID: \"2171ff1e-63b0-41ff-8019-5a72a0c59f5f\") " pod="openshift-must-gather-6t57w/must-gather-drn2z" Dec 10 23:47:41 crc kubenswrapper[4791]: I1210 23:47:41.463698 4791 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"must-gather-output\" (UniqueName: \"kubernetes.io/empty-dir/2171ff1e-63b0-41ff-8019-5a72a0c59f5f-must-gather-output\") pod \"must-gather-drn2z\" (UID: \"2171ff1e-63b0-41ff-8019-5a72a0c59f5f\") " pod="openshift-must-gather-6t57w/must-gather-drn2z" Dec 10 23:47:41 crc kubenswrapper[4791]: I1210 23:47:41.504544 4791 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-29zpv\" (UniqueName: \"kubernetes.io/projected/2171ff1e-63b0-41ff-8019-5a72a0c59f5f-kube-api-access-29zpv\") pod \"must-gather-drn2z\" (UID: \"2171ff1e-63b0-41ff-8019-5a72a0c59f5f\") " pod="openshift-must-gather-6t57w/must-gather-drn2z" Dec 10 23:47:41 crc kubenswrapper[4791]: I1210 23:47:41.622894 4791 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-6t57w/must-gather-drn2z" Dec 10 23:47:42 crc kubenswrapper[4791]: I1210 23:47:42.125916 4791 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-must-gather-6t57w/must-gather-drn2z"] Dec 10 23:47:43 crc kubenswrapper[4791]: I1210 23:47:43.058577 4791 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-6t57w/must-gather-drn2z" event={"ID":"2171ff1e-63b0-41ff-8019-5a72a0c59f5f","Type":"ContainerStarted","Data":"b192885bcb7a6095d4caeb8ce43999436a8c24b2b8a23c196030367368c6c64c"} Dec 10 23:47:49 crc kubenswrapper[4791]: I1210 23:47:49.123417 4791 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-6t57w/must-gather-drn2z" event={"ID":"2171ff1e-63b0-41ff-8019-5a72a0c59f5f","Type":"ContainerStarted","Data":"706ee31dc0ed7f4f1c4e136cb4a5a2a8f0f4228c2e114dd3b83a3906e1db6991"} Dec 10 23:47:49 crc kubenswrapper[4791]: I1210 23:47:49.123943 4791 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-6t57w/must-gather-drn2z" event={"ID":"2171ff1e-63b0-41ff-8019-5a72a0c59f5f","Type":"ContainerStarted","Data":"3f8fb9ccc563247b2f211de395f436764efe1eba5da3c41369225bf5feb95308"} Dec 10 23:47:49 crc kubenswrapper[4791]: I1210 23:47:49.148796 4791 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-must-gather-6t57w/must-gather-drn2z" podStartSLOduration=1.992921317 podStartE2EDuration="8.148778359s" podCreationTimestamp="2025-12-10 23:47:41 +0000 UTC" firstStartedPulling="2025-12-10 23:47:42.134398913 +0000 UTC m=+3496.564016536" lastFinishedPulling="2025-12-10 23:47:48.290255965 +0000 UTC m=+3502.719873578" observedRunningTime="2025-12-10 23:47:49.143078228 +0000 UTC m=+3503.572695851" watchObservedRunningTime="2025-12-10 23:47:49.148778359 +0000 UTC m=+3503.578395972" Dec 10 23:47:52 crc kubenswrapper[4791]: I1210 23:47:52.294097 4791 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-must-gather-6t57w/crc-debug-xgtff"] Dec 10 23:47:52 crc kubenswrapper[4791]: I1210 23:47:52.296099 4791 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-6t57w/crc-debug-xgtff" Dec 10 23:47:52 crc kubenswrapper[4791]: I1210 23:47:52.397482 4791 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host\" (UniqueName: \"kubernetes.io/host-path/a9688305-09b7-45c9-8efe-e8015f0a96cb-host\") pod \"crc-debug-xgtff\" (UID: \"a9688305-09b7-45c9-8efe-e8015f0a96cb\") " pod="openshift-must-gather-6t57w/crc-debug-xgtff" Dec 10 23:47:52 crc kubenswrapper[4791]: I1210 23:47:52.397798 4791 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-84flx\" (UniqueName: \"kubernetes.io/projected/a9688305-09b7-45c9-8efe-e8015f0a96cb-kube-api-access-84flx\") pod \"crc-debug-xgtff\" (UID: \"a9688305-09b7-45c9-8efe-e8015f0a96cb\") " pod="openshift-must-gather-6t57w/crc-debug-xgtff" Dec 10 23:47:52 crc kubenswrapper[4791]: I1210 23:47:52.500053 4791 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host\" (UniqueName: \"kubernetes.io/host-path/a9688305-09b7-45c9-8efe-e8015f0a96cb-host\") pod \"crc-debug-xgtff\" (UID: \"a9688305-09b7-45c9-8efe-e8015f0a96cb\") " pod="openshift-must-gather-6t57w/crc-debug-xgtff" Dec 10 23:47:52 crc kubenswrapper[4791]: I1210 23:47:52.500159 4791 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-84flx\" (UniqueName: \"kubernetes.io/projected/a9688305-09b7-45c9-8efe-e8015f0a96cb-kube-api-access-84flx\") pod \"crc-debug-xgtff\" (UID: \"a9688305-09b7-45c9-8efe-e8015f0a96cb\") " pod="openshift-must-gather-6t57w/crc-debug-xgtff" Dec 10 23:47:52 crc kubenswrapper[4791]: I1210 23:47:52.500185 4791 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host\" (UniqueName: \"kubernetes.io/host-path/a9688305-09b7-45c9-8efe-e8015f0a96cb-host\") pod \"crc-debug-xgtff\" (UID: \"a9688305-09b7-45c9-8efe-e8015f0a96cb\") " pod="openshift-must-gather-6t57w/crc-debug-xgtff" Dec 10 23:47:52 crc kubenswrapper[4791]: I1210 23:47:52.526203 4791 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-84flx\" (UniqueName: \"kubernetes.io/projected/a9688305-09b7-45c9-8efe-e8015f0a96cb-kube-api-access-84flx\") pod \"crc-debug-xgtff\" (UID: \"a9688305-09b7-45c9-8efe-e8015f0a96cb\") " pod="openshift-must-gather-6t57w/crc-debug-xgtff" Dec 10 23:47:52 crc kubenswrapper[4791]: I1210 23:47:52.617959 4791 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-6t57w/crc-debug-xgtff" Dec 10 23:47:52 crc kubenswrapper[4791]: W1210 23:47:52.647131 4791 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-poda9688305_09b7_45c9_8efe_e8015f0a96cb.slice/crio-4680b92ad578445c465fc64919286dc215c0530a3d49f218ea3cb3ab67a5299b WatchSource:0}: Error finding container 4680b92ad578445c465fc64919286dc215c0530a3d49f218ea3cb3ab67a5299b: Status 404 returned error can't find the container with id 4680b92ad578445c465fc64919286dc215c0530a3d49f218ea3cb3ab67a5299b Dec 10 23:47:53 crc kubenswrapper[4791]: I1210 23:47:53.169697 4791 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-6t57w/crc-debug-xgtff" event={"ID":"a9688305-09b7-45c9-8efe-e8015f0a96cb","Type":"ContainerStarted","Data":"4680b92ad578445c465fc64919286dc215c0530a3d49f218ea3cb3ab67a5299b"} Dec 10 23:48:05 crc kubenswrapper[4791]: I1210 23:48:05.282424 4791 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-6t57w/crc-debug-xgtff" event={"ID":"a9688305-09b7-45c9-8efe-e8015f0a96cb","Type":"ContainerStarted","Data":"8e71c17af47465b1c6ac87a038183dd522ea5fb166a468e84343ee6f185d31de"} Dec 10 23:48:05 crc kubenswrapper[4791]: I1210 23:48:05.298265 4791 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-must-gather-6t57w/crc-debug-xgtff" podStartSLOduration=1.007302792 podStartE2EDuration="13.298244989s" podCreationTimestamp="2025-12-10 23:47:52 +0000 UTC" firstStartedPulling="2025-12-10 23:47:52.649040485 +0000 UTC m=+3507.078658098" lastFinishedPulling="2025-12-10 23:48:04.939982692 +0000 UTC m=+3519.369600295" observedRunningTime="2025-12-10 23:48:05.293537606 +0000 UTC m=+3519.723155219" watchObservedRunningTime="2025-12-10 23:48:05.298244989 +0000 UTC m=+3519.727862602" Dec 10 23:48:48 crc kubenswrapper[4791]: I1210 23:48:48.739119 4791 generic.go:334] "Generic (PLEG): container finished" podID="a9688305-09b7-45c9-8efe-e8015f0a96cb" containerID="8e71c17af47465b1c6ac87a038183dd522ea5fb166a468e84343ee6f185d31de" exitCode=0 Dec 10 23:48:48 crc kubenswrapper[4791]: I1210 23:48:48.740389 4791 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-6t57w/crc-debug-xgtff" event={"ID":"a9688305-09b7-45c9-8efe-e8015f0a96cb","Type":"ContainerDied","Data":"8e71c17af47465b1c6ac87a038183dd522ea5fb166a468e84343ee6f185d31de"} Dec 10 23:48:49 crc kubenswrapper[4791]: I1210 23:48:49.891685 4791 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-6t57w/crc-debug-xgtff" Dec 10 23:48:49 crc kubenswrapper[4791]: I1210 23:48:49.944903 4791 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-must-gather-6t57w/crc-debug-xgtff"] Dec 10 23:48:49 crc kubenswrapper[4791]: I1210 23:48:49.956036 4791 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-must-gather-6t57w/crc-debug-xgtff"] Dec 10 23:48:50 crc kubenswrapper[4791]: I1210 23:48:50.010791 4791 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-84flx\" (UniqueName: \"kubernetes.io/projected/a9688305-09b7-45c9-8efe-e8015f0a96cb-kube-api-access-84flx\") pod \"a9688305-09b7-45c9-8efe-e8015f0a96cb\" (UID: \"a9688305-09b7-45c9-8efe-e8015f0a96cb\") " Dec 10 23:48:50 crc kubenswrapper[4791]: I1210 23:48:50.011115 4791 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"host\" (UniqueName: \"kubernetes.io/host-path/a9688305-09b7-45c9-8efe-e8015f0a96cb-host\") pod \"a9688305-09b7-45c9-8efe-e8015f0a96cb\" (UID: \"a9688305-09b7-45c9-8efe-e8015f0a96cb\") " Dec 10 23:48:50 crc kubenswrapper[4791]: I1210 23:48:50.011387 4791 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/a9688305-09b7-45c9-8efe-e8015f0a96cb-host" (OuterVolumeSpecName: "host") pod "a9688305-09b7-45c9-8efe-e8015f0a96cb" (UID: "a9688305-09b7-45c9-8efe-e8015f0a96cb"). InnerVolumeSpecName "host". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 10 23:48:50 crc kubenswrapper[4791]: I1210 23:48:50.012608 4791 reconciler_common.go:293] "Volume detached for volume \"host\" (UniqueName: \"kubernetes.io/host-path/a9688305-09b7-45c9-8efe-e8015f0a96cb-host\") on node \"crc\" DevicePath \"\"" Dec 10 23:48:50 crc kubenswrapper[4791]: I1210 23:48:50.018878 4791 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/a9688305-09b7-45c9-8efe-e8015f0a96cb-kube-api-access-84flx" (OuterVolumeSpecName: "kube-api-access-84flx") pod "a9688305-09b7-45c9-8efe-e8015f0a96cb" (UID: "a9688305-09b7-45c9-8efe-e8015f0a96cb"). InnerVolumeSpecName "kube-api-access-84flx". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 10 23:48:50 crc kubenswrapper[4791]: I1210 23:48:50.114290 4791 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-84flx\" (UniqueName: \"kubernetes.io/projected/a9688305-09b7-45c9-8efe-e8015f0a96cb-kube-api-access-84flx\") on node \"crc\" DevicePath \"\"" Dec 10 23:48:50 crc kubenswrapper[4791]: I1210 23:48:50.564117 4791 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-marketplace-qllwx"] Dec 10 23:48:50 crc kubenswrapper[4791]: E1210 23:48:50.564936 4791 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="a9688305-09b7-45c9-8efe-e8015f0a96cb" containerName="container-00" Dec 10 23:48:50 crc kubenswrapper[4791]: I1210 23:48:50.564960 4791 state_mem.go:107] "Deleted CPUSet assignment" podUID="a9688305-09b7-45c9-8efe-e8015f0a96cb" containerName="container-00" Dec 10 23:48:50 crc kubenswrapper[4791]: I1210 23:48:50.565239 4791 memory_manager.go:354] "RemoveStaleState removing state" podUID="a9688305-09b7-45c9-8efe-e8015f0a96cb" containerName="container-00" Dec 10 23:48:50 crc kubenswrapper[4791]: I1210 23:48:50.567231 4791 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-qllwx" Dec 10 23:48:50 crc kubenswrapper[4791]: I1210 23:48:50.582201 4791 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-qllwx"] Dec 10 23:48:50 crc kubenswrapper[4791]: I1210 23:48:50.626537 4791 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-6frkx\" (UniqueName: \"kubernetes.io/projected/7b92afeb-c305-478c-9489-4673510df2f8-kube-api-access-6frkx\") pod \"redhat-marketplace-qllwx\" (UID: \"7b92afeb-c305-478c-9489-4673510df2f8\") " pod="openshift-marketplace/redhat-marketplace-qllwx" Dec 10 23:48:50 crc kubenswrapper[4791]: I1210 23:48:50.626625 4791 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/7b92afeb-c305-478c-9489-4673510df2f8-utilities\") pod \"redhat-marketplace-qllwx\" (UID: \"7b92afeb-c305-478c-9489-4673510df2f8\") " pod="openshift-marketplace/redhat-marketplace-qllwx" Dec 10 23:48:50 crc kubenswrapper[4791]: I1210 23:48:50.626687 4791 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/7b92afeb-c305-478c-9489-4673510df2f8-catalog-content\") pod \"redhat-marketplace-qllwx\" (UID: \"7b92afeb-c305-478c-9489-4673510df2f8\") " pod="openshift-marketplace/redhat-marketplace-qllwx" Dec 10 23:48:50 crc kubenswrapper[4791]: I1210 23:48:50.728226 4791 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-6frkx\" (UniqueName: \"kubernetes.io/projected/7b92afeb-c305-478c-9489-4673510df2f8-kube-api-access-6frkx\") pod \"redhat-marketplace-qllwx\" (UID: \"7b92afeb-c305-478c-9489-4673510df2f8\") " pod="openshift-marketplace/redhat-marketplace-qllwx" Dec 10 23:48:50 crc kubenswrapper[4791]: I1210 23:48:50.728295 4791 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/7b92afeb-c305-478c-9489-4673510df2f8-utilities\") pod \"redhat-marketplace-qllwx\" (UID: \"7b92afeb-c305-478c-9489-4673510df2f8\") " pod="openshift-marketplace/redhat-marketplace-qllwx" Dec 10 23:48:50 crc kubenswrapper[4791]: I1210 23:48:50.728353 4791 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/7b92afeb-c305-478c-9489-4673510df2f8-catalog-content\") pod \"redhat-marketplace-qllwx\" (UID: \"7b92afeb-c305-478c-9489-4673510df2f8\") " pod="openshift-marketplace/redhat-marketplace-qllwx" Dec 10 23:48:50 crc kubenswrapper[4791]: I1210 23:48:50.728826 4791 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/7b92afeb-c305-478c-9489-4673510df2f8-utilities\") pod \"redhat-marketplace-qllwx\" (UID: \"7b92afeb-c305-478c-9489-4673510df2f8\") " pod="openshift-marketplace/redhat-marketplace-qllwx" Dec 10 23:48:50 crc kubenswrapper[4791]: I1210 23:48:50.728949 4791 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/7b92afeb-c305-478c-9489-4673510df2f8-catalog-content\") pod \"redhat-marketplace-qllwx\" (UID: \"7b92afeb-c305-478c-9489-4673510df2f8\") " pod="openshift-marketplace/redhat-marketplace-qllwx" Dec 10 23:48:50 crc kubenswrapper[4791]: I1210 23:48:50.747312 4791 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-6frkx\" (UniqueName: \"kubernetes.io/projected/7b92afeb-c305-478c-9489-4673510df2f8-kube-api-access-6frkx\") pod \"redhat-marketplace-qllwx\" (UID: \"7b92afeb-c305-478c-9489-4673510df2f8\") " pod="openshift-marketplace/redhat-marketplace-qllwx" Dec 10 23:48:50 crc kubenswrapper[4791]: I1210 23:48:50.764679 4791 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="4680b92ad578445c465fc64919286dc215c0530a3d49f218ea3cb3ab67a5299b" Dec 10 23:48:50 crc kubenswrapper[4791]: I1210 23:48:50.764762 4791 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-6t57w/crc-debug-xgtff" Dec 10 23:48:50 crc kubenswrapper[4791]: I1210 23:48:50.902761 4791 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-qllwx" Dec 10 23:48:51 crc kubenswrapper[4791]: I1210 23:48:51.193934 4791 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-must-gather-6t57w/crc-debug-gs847"] Dec 10 23:48:51 crc kubenswrapper[4791]: I1210 23:48:51.195374 4791 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-6t57w/crc-debug-gs847" Dec 10 23:48:51 crc kubenswrapper[4791]: I1210 23:48:51.340147 4791 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-pvdp4\" (UniqueName: \"kubernetes.io/projected/701644a7-09e5-4253-a133-c607172e147e-kube-api-access-pvdp4\") pod \"crc-debug-gs847\" (UID: \"701644a7-09e5-4253-a133-c607172e147e\") " pod="openshift-must-gather-6t57w/crc-debug-gs847" Dec 10 23:48:51 crc kubenswrapper[4791]: I1210 23:48:51.340417 4791 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host\" (UniqueName: \"kubernetes.io/host-path/701644a7-09e5-4253-a133-c607172e147e-host\") pod \"crc-debug-gs847\" (UID: \"701644a7-09e5-4253-a133-c607172e147e\") " pod="openshift-must-gather-6t57w/crc-debug-gs847" Dec 10 23:48:51 crc kubenswrapper[4791]: I1210 23:48:51.394502 4791 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-qllwx"] Dec 10 23:48:51 crc kubenswrapper[4791]: I1210 23:48:51.442788 4791 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-pvdp4\" (UniqueName: \"kubernetes.io/projected/701644a7-09e5-4253-a133-c607172e147e-kube-api-access-pvdp4\") pod \"crc-debug-gs847\" (UID: \"701644a7-09e5-4253-a133-c607172e147e\") " pod="openshift-must-gather-6t57w/crc-debug-gs847" Dec 10 23:48:51 crc kubenswrapper[4791]: I1210 23:48:51.442922 4791 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host\" (UniqueName: \"kubernetes.io/host-path/701644a7-09e5-4253-a133-c607172e147e-host\") pod \"crc-debug-gs847\" (UID: \"701644a7-09e5-4253-a133-c607172e147e\") " pod="openshift-must-gather-6t57w/crc-debug-gs847" Dec 10 23:48:51 crc kubenswrapper[4791]: I1210 23:48:51.443050 4791 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host\" (UniqueName: \"kubernetes.io/host-path/701644a7-09e5-4253-a133-c607172e147e-host\") pod \"crc-debug-gs847\" (UID: \"701644a7-09e5-4253-a133-c607172e147e\") " pod="openshift-must-gather-6t57w/crc-debug-gs847" Dec 10 23:48:51 crc kubenswrapper[4791]: I1210 23:48:51.463037 4791 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-pvdp4\" (UniqueName: \"kubernetes.io/projected/701644a7-09e5-4253-a133-c607172e147e-kube-api-access-pvdp4\") pod \"crc-debug-gs847\" (UID: \"701644a7-09e5-4253-a133-c607172e147e\") " pod="openshift-must-gather-6t57w/crc-debug-gs847" Dec 10 23:48:51 crc kubenswrapper[4791]: I1210 23:48:51.517170 4791 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-6t57w/crc-debug-gs847" Dec 10 23:48:51 crc kubenswrapper[4791]: W1210 23:48:51.546811 4791 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod701644a7_09e5_4253_a133_c607172e147e.slice/crio-d1f9b5e73563b7bd6ac3a81f4bf76c1b09e93a2ef28ab79927a33f4098649cac WatchSource:0}: Error finding container d1f9b5e73563b7bd6ac3a81f4bf76c1b09e93a2ef28ab79927a33f4098649cac: Status 404 returned error can't find the container with id d1f9b5e73563b7bd6ac3a81f4bf76c1b09e93a2ef28ab79927a33f4098649cac Dec 10 23:48:51 crc kubenswrapper[4791]: I1210 23:48:51.778370 4791 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-qllwx" event={"ID":"7b92afeb-c305-478c-9489-4673510df2f8","Type":"ContainerStarted","Data":"29f7568cbeff56ce1613ba9705330dbf4b99569b045ec0cad217c94c21432b8b"} Dec 10 23:48:51 crc kubenswrapper[4791]: I1210 23:48:51.780903 4791 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-6t57w/crc-debug-gs847" event={"ID":"701644a7-09e5-4253-a133-c607172e147e","Type":"ContainerStarted","Data":"d1f9b5e73563b7bd6ac3a81f4bf76c1b09e93a2ef28ab79927a33f4098649cac"} Dec 10 23:48:51 crc kubenswrapper[4791]: I1210 23:48:51.897672 4791 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="a9688305-09b7-45c9-8efe-e8015f0a96cb" path="/var/lib/kubelet/pods/a9688305-09b7-45c9-8efe-e8015f0a96cb/volumes" Dec 10 23:48:52 crc kubenswrapper[4791]: I1210 23:48:52.794707 4791 generic.go:334] "Generic (PLEG): container finished" podID="701644a7-09e5-4253-a133-c607172e147e" containerID="3a06af291cb227022261a4ab29bbe121f0a396429f2de38662668e6cd79b155d" exitCode=0 Dec 10 23:48:52 crc kubenswrapper[4791]: I1210 23:48:52.794839 4791 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-6t57w/crc-debug-gs847" event={"ID":"701644a7-09e5-4253-a133-c607172e147e","Type":"ContainerDied","Data":"3a06af291cb227022261a4ab29bbe121f0a396429f2de38662668e6cd79b155d"} Dec 10 23:48:52 crc kubenswrapper[4791]: I1210 23:48:52.798905 4791 generic.go:334] "Generic (PLEG): container finished" podID="7b92afeb-c305-478c-9489-4673510df2f8" containerID="1dba14f5aa6207910f1dc1fe34733c54fb89324c490f834bcef91797bd2f214a" exitCode=0 Dec 10 23:48:52 crc kubenswrapper[4791]: I1210 23:48:52.798943 4791 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-qllwx" event={"ID":"7b92afeb-c305-478c-9489-4673510df2f8","Type":"ContainerDied","Data":"1dba14f5aa6207910f1dc1fe34733c54fb89324c490f834bcef91797bd2f214a"} Dec 10 23:48:53 crc kubenswrapper[4791]: I1210 23:48:53.398515 4791 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-must-gather-6t57w/crc-debug-gs847"] Dec 10 23:48:53 crc kubenswrapper[4791]: I1210 23:48:53.409155 4791 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-must-gather-6t57w/crc-debug-gs847"] Dec 10 23:48:53 crc kubenswrapper[4791]: I1210 23:48:53.810597 4791 generic.go:334] "Generic (PLEG): container finished" podID="7b92afeb-c305-478c-9489-4673510df2f8" containerID="d689a2f97d9e9d1e0820889731eb2d2f70a6c08f196b7919d9fec267f214c923" exitCode=0 Dec 10 23:48:53 crc kubenswrapper[4791]: I1210 23:48:53.810641 4791 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-qllwx" event={"ID":"7b92afeb-c305-478c-9489-4673510df2f8","Type":"ContainerDied","Data":"d689a2f97d9e9d1e0820889731eb2d2f70a6c08f196b7919d9fec267f214c923"} Dec 10 23:48:53 crc kubenswrapper[4791]: I1210 23:48:53.895409 4791 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-6t57w/crc-debug-gs847" Dec 10 23:48:53 crc kubenswrapper[4791]: I1210 23:48:53.990324 4791 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"host\" (UniqueName: \"kubernetes.io/host-path/701644a7-09e5-4253-a133-c607172e147e-host\") pod \"701644a7-09e5-4253-a133-c607172e147e\" (UID: \"701644a7-09e5-4253-a133-c607172e147e\") " Dec 10 23:48:53 crc kubenswrapper[4791]: I1210 23:48:53.990470 4791 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-pvdp4\" (UniqueName: \"kubernetes.io/projected/701644a7-09e5-4253-a133-c607172e147e-kube-api-access-pvdp4\") pod \"701644a7-09e5-4253-a133-c607172e147e\" (UID: \"701644a7-09e5-4253-a133-c607172e147e\") " Dec 10 23:48:53 crc kubenswrapper[4791]: I1210 23:48:53.991073 4791 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/701644a7-09e5-4253-a133-c607172e147e-host" (OuterVolumeSpecName: "host") pod "701644a7-09e5-4253-a133-c607172e147e" (UID: "701644a7-09e5-4253-a133-c607172e147e"). InnerVolumeSpecName "host". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 10 23:48:54 crc kubenswrapper[4791]: I1210 23:48:54.000587 4791 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/701644a7-09e5-4253-a133-c607172e147e-kube-api-access-pvdp4" (OuterVolumeSpecName: "kube-api-access-pvdp4") pod "701644a7-09e5-4253-a133-c607172e147e" (UID: "701644a7-09e5-4253-a133-c607172e147e"). InnerVolumeSpecName "kube-api-access-pvdp4". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 10 23:48:54 crc kubenswrapper[4791]: I1210 23:48:54.093148 4791 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-pvdp4\" (UniqueName: \"kubernetes.io/projected/701644a7-09e5-4253-a133-c607172e147e-kube-api-access-pvdp4\") on node \"crc\" DevicePath \"\"" Dec 10 23:48:54 crc kubenswrapper[4791]: I1210 23:48:54.093215 4791 reconciler_common.go:293] "Volume detached for volume \"host\" (UniqueName: \"kubernetes.io/host-path/701644a7-09e5-4253-a133-c607172e147e-host\") on node \"crc\" DevicePath \"\"" Dec 10 23:48:54 crc kubenswrapper[4791]: I1210 23:48:54.601888 4791 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-must-gather-6t57w/crc-debug-c65z5"] Dec 10 23:48:54 crc kubenswrapper[4791]: E1210 23:48:54.602873 4791 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="701644a7-09e5-4253-a133-c607172e147e" containerName="container-00" Dec 10 23:48:54 crc kubenswrapper[4791]: I1210 23:48:54.602904 4791 state_mem.go:107] "Deleted CPUSet assignment" podUID="701644a7-09e5-4253-a133-c607172e147e" containerName="container-00" Dec 10 23:48:54 crc kubenswrapper[4791]: I1210 23:48:54.603196 4791 memory_manager.go:354] "RemoveStaleState removing state" podUID="701644a7-09e5-4253-a133-c607172e147e" containerName="container-00" Dec 10 23:48:54 crc kubenswrapper[4791]: I1210 23:48:54.604030 4791 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-6t57w/crc-debug-c65z5" Dec 10 23:48:54 crc kubenswrapper[4791]: I1210 23:48:54.705276 4791 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-jqsqp\" (UniqueName: \"kubernetes.io/projected/8906f851-2b6f-4387-872f-a9a6c4267b74-kube-api-access-jqsqp\") pod \"crc-debug-c65z5\" (UID: \"8906f851-2b6f-4387-872f-a9a6c4267b74\") " pod="openshift-must-gather-6t57w/crc-debug-c65z5" Dec 10 23:48:54 crc kubenswrapper[4791]: I1210 23:48:54.705411 4791 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host\" (UniqueName: \"kubernetes.io/host-path/8906f851-2b6f-4387-872f-a9a6c4267b74-host\") pod \"crc-debug-c65z5\" (UID: \"8906f851-2b6f-4387-872f-a9a6c4267b74\") " pod="openshift-must-gather-6t57w/crc-debug-c65z5" Dec 10 23:48:54 crc kubenswrapper[4791]: I1210 23:48:54.807255 4791 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host\" (UniqueName: \"kubernetes.io/host-path/8906f851-2b6f-4387-872f-a9a6c4267b74-host\") pod \"crc-debug-c65z5\" (UID: \"8906f851-2b6f-4387-872f-a9a6c4267b74\") " pod="openshift-must-gather-6t57w/crc-debug-c65z5" Dec 10 23:48:54 crc kubenswrapper[4791]: I1210 23:48:54.807429 4791 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host\" (UniqueName: \"kubernetes.io/host-path/8906f851-2b6f-4387-872f-a9a6c4267b74-host\") pod \"crc-debug-c65z5\" (UID: \"8906f851-2b6f-4387-872f-a9a6c4267b74\") " pod="openshift-must-gather-6t57w/crc-debug-c65z5" Dec 10 23:48:54 crc kubenswrapper[4791]: I1210 23:48:54.807725 4791 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-jqsqp\" (UniqueName: \"kubernetes.io/projected/8906f851-2b6f-4387-872f-a9a6c4267b74-kube-api-access-jqsqp\") pod \"crc-debug-c65z5\" (UID: \"8906f851-2b6f-4387-872f-a9a6c4267b74\") " pod="openshift-must-gather-6t57w/crc-debug-c65z5" Dec 10 23:48:54 crc kubenswrapper[4791]: I1210 23:48:54.820086 4791 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-qllwx" event={"ID":"7b92afeb-c305-478c-9489-4673510df2f8","Type":"ContainerStarted","Data":"8d8fa96559f238e2154d85f1fbc720a0ec9da9f532946e9c0c2a85b131ab38d0"} Dec 10 23:48:54 crc kubenswrapper[4791]: I1210 23:48:54.821968 4791 scope.go:117] "RemoveContainer" containerID="3a06af291cb227022261a4ab29bbe121f0a396429f2de38662668e6cd79b155d" Dec 10 23:48:54 crc kubenswrapper[4791]: I1210 23:48:54.822054 4791 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-6t57w/crc-debug-gs847" Dec 10 23:48:54 crc kubenswrapper[4791]: I1210 23:48:54.845467 4791 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-jqsqp\" (UniqueName: \"kubernetes.io/projected/8906f851-2b6f-4387-872f-a9a6c4267b74-kube-api-access-jqsqp\") pod \"crc-debug-c65z5\" (UID: \"8906f851-2b6f-4387-872f-a9a6c4267b74\") " pod="openshift-must-gather-6t57w/crc-debug-c65z5" Dec 10 23:48:54 crc kubenswrapper[4791]: I1210 23:48:54.849671 4791 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-marketplace-qllwx" podStartSLOduration=3.369309194 podStartE2EDuration="4.849643898s" podCreationTimestamp="2025-12-10 23:48:50 +0000 UTC" firstStartedPulling="2025-12-10 23:48:52.802316822 +0000 UTC m=+3567.231934475" lastFinishedPulling="2025-12-10 23:48:54.282651536 +0000 UTC m=+3568.712269179" observedRunningTime="2025-12-10 23:48:54.844645377 +0000 UTC m=+3569.274262990" watchObservedRunningTime="2025-12-10 23:48:54.849643898 +0000 UTC m=+3569.279261551" Dec 10 23:48:54 crc kubenswrapper[4791]: I1210 23:48:54.923137 4791 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-6t57w/crc-debug-c65z5" Dec 10 23:48:54 crc kubenswrapper[4791]: W1210 23:48:54.949758 4791 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod8906f851_2b6f_4387_872f_a9a6c4267b74.slice/crio-9fa825b534945e1979d3f79cd195d311c5fcd4ab947b5118692feaed7f7f8cea WatchSource:0}: Error finding container 9fa825b534945e1979d3f79cd195d311c5fcd4ab947b5118692feaed7f7f8cea: Status 404 returned error can't find the container with id 9fa825b534945e1979d3f79cd195d311c5fcd4ab947b5118692feaed7f7f8cea Dec 10 23:48:55 crc kubenswrapper[4791]: I1210 23:48:55.831941 4791 generic.go:334] "Generic (PLEG): container finished" podID="8906f851-2b6f-4387-872f-a9a6c4267b74" containerID="5ba1ecb0bd4b0571936b904f67fa42236a90b08a68f51337c69397ae126964dd" exitCode=0 Dec 10 23:48:55 crc kubenswrapper[4791]: I1210 23:48:55.832013 4791 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-6t57w/crc-debug-c65z5" event={"ID":"8906f851-2b6f-4387-872f-a9a6c4267b74","Type":"ContainerDied","Data":"5ba1ecb0bd4b0571936b904f67fa42236a90b08a68f51337c69397ae126964dd"} Dec 10 23:48:55 crc kubenswrapper[4791]: I1210 23:48:55.832395 4791 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-6t57w/crc-debug-c65z5" event={"ID":"8906f851-2b6f-4387-872f-a9a6c4267b74","Type":"ContainerStarted","Data":"9fa825b534945e1979d3f79cd195d311c5fcd4ab947b5118692feaed7f7f8cea"} Dec 10 23:48:55 crc kubenswrapper[4791]: I1210 23:48:55.876349 4791 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-must-gather-6t57w/crc-debug-c65z5"] Dec 10 23:48:55 crc kubenswrapper[4791]: I1210 23:48:55.899993 4791 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="701644a7-09e5-4253-a133-c607172e147e" path="/var/lib/kubelet/pods/701644a7-09e5-4253-a133-c607172e147e/volumes" Dec 10 23:48:55 crc kubenswrapper[4791]: I1210 23:48:55.900821 4791 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-must-gather-6t57w/crc-debug-c65z5"] Dec 10 23:48:56 crc kubenswrapper[4791]: I1210 23:48:56.980541 4791 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-6t57w/crc-debug-c65z5" Dec 10 23:48:57 crc kubenswrapper[4791]: I1210 23:48:57.053031 4791 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"host\" (UniqueName: \"kubernetes.io/host-path/8906f851-2b6f-4387-872f-a9a6c4267b74-host\") pod \"8906f851-2b6f-4387-872f-a9a6c4267b74\" (UID: \"8906f851-2b6f-4387-872f-a9a6c4267b74\") " Dec 10 23:48:57 crc kubenswrapper[4791]: I1210 23:48:57.053102 4791 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-jqsqp\" (UniqueName: \"kubernetes.io/projected/8906f851-2b6f-4387-872f-a9a6c4267b74-kube-api-access-jqsqp\") pod \"8906f851-2b6f-4387-872f-a9a6c4267b74\" (UID: \"8906f851-2b6f-4387-872f-a9a6c4267b74\") " Dec 10 23:48:57 crc kubenswrapper[4791]: I1210 23:48:57.053173 4791 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/8906f851-2b6f-4387-872f-a9a6c4267b74-host" (OuterVolumeSpecName: "host") pod "8906f851-2b6f-4387-872f-a9a6c4267b74" (UID: "8906f851-2b6f-4387-872f-a9a6c4267b74"). InnerVolumeSpecName "host". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 10 23:48:57 crc kubenswrapper[4791]: I1210 23:48:57.053757 4791 reconciler_common.go:293] "Volume detached for volume \"host\" (UniqueName: \"kubernetes.io/host-path/8906f851-2b6f-4387-872f-a9a6c4267b74-host\") on node \"crc\" DevicePath \"\"" Dec 10 23:48:57 crc kubenswrapper[4791]: I1210 23:48:57.065578 4791 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/8906f851-2b6f-4387-872f-a9a6c4267b74-kube-api-access-jqsqp" (OuterVolumeSpecName: "kube-api-access-jqsqp") pod "8906f851-2b6f-4387-872f-a9a6c4267b74" (UID: "8906f851-2b6f-4387-872f-a9a6c4267b74"). InnerVolumeSpecName "kube-api-access-jqsqp". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 10 23:48:57 crc kubenswrapper[4791]: I1210 23:48:57.155308 4791 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-jqsqp\" (UniqueName: \"kubernetes.io/projected/8906f851-2b6f-4387-872f-a9a6c4267b74-kube-api-access-jqsqp\") on node \"crc\" DevicePath \"\"" Dec 10 23:48:57 crc kubenswrapper[4791]: I1210 23:48:57.856495 4791 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="9fa825b534945e1979d3f79cd195d311c5fcd4ab947b5118692feaed7f7f8cea" Dec 10 23:48:57 crc kubenswrapper[4791]: I1210 23:48:57.856567 4791 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-6t57w/crc-debug-c65z5" Dec 10 23:48:57 crc kubenswrapper[4791]: I1210 23:48:57.896960 4791 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="8906f851-2b6f-4387-872f-a9a6c4267b74" path="/var/lib/kubelet/pods/8906f851-2b6f-4387-872f-a9a6c4267b74/volumes" Dec 10 23:49:00 crc kubenswrapper[4791]: I1210 23:49:00.903447 4791 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-marketplace-qllwx" Dec 10 23:49:00 crc kubenswrapper[4791]: I1210 23:49:00.903777 4791 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-marketplace-qllwx" Dec 10 23:49:00 crc kubenswrapper[4791]: I1210 23:49:00.962329 4791 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-marketplace-qllwx" Dec 10 23:49:01 crc kubenswrapper[4791]: I1210 23:49:01.958827 4791 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-marketplace-qllwx" Dec 10 23:49:02 crc kubenswrapper[4791]: I1210 23:49:02.008156 4791 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-qllwx"] Dec 10 23:49:03 crc kubenswrapper[4791]: I1210 23:49:03.911569 4791 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-marketplace-qllwx" podUID="7b92afeb-c305-478c-9489-4673510df2f8" containerName="registry-server" containerID="cri-o://8d8fa96559f238e2154d85f1fbc720a0ec9da9f532946e9c0c2a85b131ab38d0" gracePeriod=2 Dec 10 23:49:04 crc kubenswrapper[4791]: I1210 23:49:04.391681 4791 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-qllwx" Dec 10 23:49:04 crc kubenswrapper[4791]: I1210 23:49:04.486713 4791 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/7b92afeb-c305-478c-9489-4673510df2f8-catalog-content\") pod \"7b92afeb-c305-478c-9489-4673510df2f8\" (UID: \"7b92afeb-c305-478c-9489-4673510df2f8\") " Dec 10 23:49:04 crc kubenswrapper[4791]: I1210 23:49:04.486775 4791 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-6frkx\" (UniqueName: \"kubernetes.io/projected/7b92afeb-c305-478c-9489-4673510df2f8-kube-api-access-6frkx\") pod \"7b92afeb-c305-478c-9489-4673510df2f8\" (UID: \"7b92afeb-c305-478c-9489-4673510df2f8\") " Dec 10 23:49:04 crc kubenswrapper[4791]: I1210 23:49:04.486804 4791 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/7b92afeb-c305-478c-9489-4673510df2f8-utilities\") pod \"7b92afeb-c305-478c-9489-4673510df2f8\" (UID: \"7b92afeb-c305-478c-9489-4673510df2f8\") " Dec 10 23:49:04 crc kubenswrapper[4791]: I1210 23:49:04.487617 4791 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/7b92afeb-c305-478c-9489-4673510df2f8-utilities" (OuterVolumeSpecName: "utilities") pod "7b92afeb-c305-478c-9489-4673510df2f8" (UID: "7b92afeb-c305-478c-9489-4673510df2f8"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 10 23:49:04 crc kubenswrapper[4791]: I1210 23:49:04.506320 4791 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/7b92afeb-c305-478c-9489-4673510df2f8-kube-api-access-6frkx" (OuterVolumeSpecName: "kube-api-access-6frkx") pod "7b92afeb-c305-478c-9489-4673510df2f8" (UID: "7b92afeb-c305-478c-9489-4673510df2f8"). InnerVolumeSpecName "kube-api-access-6frkx". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 10 23:49:04 crc kubenswrapper[4791]: I1210 23:49:04.508127 4791 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/7b92afeb-c305-478c-9489-4673510df2f8-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "7b92afeb-c305-478c-9489-4673510df2f8" (UID: "7b92afeb-c305-478c-9489-4673510df2f8"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 10 23:49:04 crc kubenswrapper[4791]: I1210 23:49:04.588640 4791 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/7b92afeb-c305-478c-9489-4673510df2f8-catalog-content\") on node \"crc\" DevicePath \"\"" Dec 10 23:49:04 crc kubenswrapper[4791]: I1210 23:49:04.588672 4791 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-6frkx\" (UniqueName: \"kubernetes.io/projected/7b92afeb-c305-478c-9489-4673510df2f8-kube-api-access-6frkx\") on node \"crc\" DevicePath \"\"" Dec 10 23:49:04 crc kubenswrapper[4791]: I1210 23:49:04.588684 4791 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/7b92afeb-c305-478c-9489-4673510df2f8-utilities\") on node \"crc\" DevicePath \"\"" Dec 10 23:49:04 crc kubenswrapper[4791]: I1210 23:49:04.922501 4791 generic.go:334] "Generic (PLEG): container finished" podID="7b92afeb-c305-478c-9489-4673510df2f8" containerID="8d8fa96559f238e2154d85f1fbc720a0ec9da9f532946e9c0c2a85b131ab38d0" exitCode=0 Dec 10 23:49:04 crc kubenswrapper[4791]: I1210 23:49:04.922614 4791 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-qllwx" Dec 10 23:49:04 crc kubenswrapper[4791]: I1210 23:49:04.922629 4791 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-qllwx" event={"ID":"7b92afeb-c305-478c-9489-4673510df2f8","Type":"ContainerDied","Data":"8d8fa96559f238e2154d85f1fbc720a0ec9da9f532946e9c0c2a85b131ab38d0"} Dec 10 23:49:04 crc kubenswrapper[4791]: I1210 23:49:04.923719 4791 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-qllwx" event={"ID":"7b92afeb-c305-478c-9489-4673510df2f8","Type":"ContainerDied","Data":"29f7568cbeff56ce1613ba9705330dbf4b99569b045ec0cad217c94c21432b8b"} Dec 10 23:49:04 crc kubenswrapper[4791]: I1210 23:49:04.923756 4791 scope.go:117] "RemoveContainer" containerID="8d8fa96559f238e2154d85f1fbc720a0ec9da9f532946e9c0c2a85b131ab38d0" Dec 10 23:49:04 crc kubenswrapper[4791]: I1210 23:49:04.944164 4791 scope.go:117] "RemoveContainer" containerID="d689a2f97d9e9d1e0820889731eb2d2f70a6c08f196b7919d9fec267f214c923" Dec 10 23:49:04 crc kubenswrapper[4791]: I1210 23:49:04.970550 4791 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-qllwx"] Dec 10 23:49:04 crc kubenswrapper[4791]: I1210 23:49:04.982409 4791 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-marketplace-qllwx"] Dec 10 23:49:04 crc kubenswrapper[4791]: I1210 23:49:04.998154 4791 scope.go:117] "RemoveContainer" containerID="1dba14f5aa6207910f1dc1fe34733c54fb89324c490f834bcef91797bd2f214a" Dec 10 23:49:05 crc kubenswrapper[4791]: I1210 23:49:05.020849 4791 scope.go:117] "RemoveContainer" containerID="8d8fa96559f238e2154d85f1fbc720a0ec9da9f532946e9c0c2a85b131ab38d0" Dec 10 23:49:05 crc kubenswrapper[4791]: E1210 23:49:05.021326 4791 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"8d8fa96559f238e2154d85f1fbc720a0ec9da9f532946e9c0c2a85b131ab38d0\": container with ID starting with 8d8fa96559f238e2154d85f1fbc720a0ec9da9f532946e9c0c2a85b131ab38d0 not found: ID does not exist" containerID="8d8fa96559f238e2154d85f1fbc720a0ec9da9f532946e9c0c2a85b131ab38d0" Dec 10 23:49:05 crc kubenswrapper[4791]: I1210 23:49:05.021385 4791 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"8d8fa96559f238e2154d85f1fbc720a0ec9da9f532946e9c0c2a85b131ab38d0"} err="failed to get container status \"8d8fa96559f238e2154d85f1fbc720a0ec9da9f532946e9c0c2a85b131ab38d0\": rpc error: code = NotFound desc = could not find container \"8d8fa96559f238e2154d85f1fbc720a0ec9da9f532946e9c0c2a85b131ab38d0\": container with ID starting with 8d8fa96559f238e2154d85f1fbc720a0ec9da9f532946e9c0c2a85b131ab38d0 not found: ID does not exist" Dec 10 23:49:05 crc kubenswrapper[4791]: I1210 23:49:05.021416 4791 scope.go:117] "RemoveContainer" containerID="d689a2f97d9e9d1e0820889731eb2d2f70a6c08f196b7919d9fec267f214c923" Dec 10 23:49:05 crc kubenswrapper[4791]: E1210 23:49:05.021821 4791 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"d689a2f97d9e9d1e0820889731eb2d2f70a6c08f196b7919d9fec267f214c923\": container with ID starting with d689a2f97d9e9d1e0820889731eb2d2f70a6c08f196b7919d9fec267f214c923 not found: ID does not exist" containerID="d689a2f97d9e9d1e0820889731eb2d2f70a6c08f196b7919d9fec267f214c923" Dec 10 23:49:05 crc kubenswrapper[4791]: I1210 23:49:05.021848 4791 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"d689a2f97d9e9d1e0820889731eb2d2f70a6c08f196b7919d9fec267f214c923"} err="failed to get container status \"d689a2f97d9e9d1e0820889731eb2d2f70a6c08f196b7919d9fec267f214c923\": rpc error: code = NotFound desc = could not find container \"d689a2f97d9e9d1e0820889731eb2d2f70a6c08f196b7919d9fec267f214c923\": container with ID starting with d689a2f97d9e9d1e0820889731eb2d2f70a6c08f196b7919d9fec267f214c923 not found: ID does not exist" Dec 10 23:49:05 crc kubenswrapper[4791]: I1210 23:49:05.021866 4791 scope.go:117] "RemoveContainer" containerID="1dba14f5aa6207910f1dc1fe34733c54fb89324c490f834bcef91797bd2f214a" Dec 10 23:49:05 crc kubenswrapper[4791]: E1210 23:49:05.022161 4791 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"1dba14f5aa6207910f1dc1fe34733c54fb89324c490f834bcef91797bd2f214a\": container with ID starting with 1dba14f5aa6207910f1dc1fe34733c54fb89324c490f834bcef91797bd2f214a not found: ID does not exist" containerID="1dba14f5aa6207910f1dc1fe34733c54fb89324c490f834bcef91797bd2f214a" Dec 10 23:49:05 crc kubenswrapper[4791]: I1210 23:49:05.022205 4791 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"1dba14f5aa6207910f1dc1fe34733c54fb89324c490f834bcef91797bd2f214a"} err="failed to get container status \"1dba14f5aa6207910f1dc1fe34733c54fb89324c490f834bcef91797bd2f214a\": rpc error: code = NotFound desc = could not find container \"1dba14f5aa6207910f1dc1fe34733c54fb89324c490f834bcef91797bd2f214a\": container with ID starting with 1dba14f5aa6207910f1dc1fe34733c54fb89324c490f834bcef91797bd2f214a not found: ID does not exist" Dec 10 23:49:05 crc kubenswrapper[4791]: I1210 23:49:05.901738 4791 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="7b92afeb-c305-478c-9489-4673510df2f8" path="/var/lib/kubelet/pods/7b92afeb-c305-478c-9489-4673510df2f8/volumes" Dec 10 23:49:11 crc kubenswrapper[4791]: I1210 23:49:11.091742 4791 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_barbican-api-68fb764d56-4mfz4_eb482cb0-7ac1-4097-8799-4a5d78b6b5e5/barbican-api/0.log" Dec 10 23:49:11 crc kubenswrapper[4791]: I1210 23:49:11.187407 4791 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_barbican-api-68fb764d56-4mfz4_eb482cb0-7ac1-4097-8799-4a5d78b6b5e5/barbican-api-log/0.log" Dec 10 23:49:11 crc kubenswrapper[4791]: I1210 23:49:11.270666 4791 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_barbican-keystone-listener-67867746d6-mvxp4_4d7d941e-8494-46ea-990c-fe83db67dd2a/barbican-keystone-listener/0.log" Dec 10 23:49:11 crc kubenswrapper[4791]: I1210 23:49:11.357908 4791 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_barbican-keystone-listener-67867746d6-mvxp4_4d7d941e-8494-46ea-990c-fe83db67dd2a/barbican-keystone-listener-log/0.log" Dec 10 23:49:11 crc kubenswrapper[4791]: I1210 23:49:11.498284 4791 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_barbican-worker-77fb99879-xzsrk_8a302b7c-c2df-4023-95da-4ef8e86dbc0b/barbican-worker/0.log" Dec 10 23:49:11 crc kubenswrapper[4791]: I1210 23:49:11.532940 4791 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_barbican-worker-77fb99879-xzsrk_8a302b7c-c2df-4023-95da-4ef8e86dbc0b/barbican-worker-log/0.log" Dec 10 23:49:11 crc kubenswrapper[4791]: I1210 23:49:11.635911 4791 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_bootstrap-edpm-deployment-openstack-edpm-ipam-fpng5_07a0187d-1677-4da9-9a3b-8da49e8f6819/bootstrap-edpm-deployment-openstack-edpm-ipam/0.log" Dec 10 23:49:11 crc kubenswrapper[4791]: I1210 23:49:11.804191 4791 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ceilometer-0_c85398ce-a5d4-414a-8811-775a4ce71193/ceilometer-central-agent/0.log" Dec 10 23:49:11 crc kubenswrapper[4791]: I1210 23:49:11.822464 4791 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ceilometer-0_c85398ce-a5d4-414a-8811-775a4ce71193/ceilometer-notification-agent/0.log" Dec 10 23:49:11 crc kubenswrapper[4791]: I1210 23:49:11.846120 4791 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ceilometer-0_c85398ce-a5d4-414a-8811-775a4ce71193/proxy-httpd/0.log" Dec 10 23:49:12 crc kubenswrapper[4791]: I1210 23:49:12.165621 4791 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_cinder-api-0_f807e752-6cf5-4ad6-bc21-f22777b62c83/cinder-api-log/0.log" Dec 10 23:49:12 crc kubenswrapper[4791]: I1210 23:49:12.172536 4791 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ceilometer-0_c85398ce-a5d4-414a-8811-775a4ce71193/sg-core/0.log" Dec 10 23:49:12 crc kubenswrapper[4791]: I1210 23:49:12.281684 4791 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_cinder-api-0_f807e752-6cf5-4ad6-bc21-f22777b62c83/cinder-api/0.log" Dec 10 23:49:12 crc kubenswrapper[4791]: I1210 23:49:12.409174 4791 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_cinder-scheduler-0_e703a39e-cce5-4b36-b017-611836d0fa18/cinder-scheduler/0.log" Dec 10 23:49:12 crc kubenswrapper[4791]: I1210 23:49:12.432088 4791 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_cinder-scheduler-0_e703a39e-cce5-4b36-b017-611836d0fa18/probe/0.log" Dec 10 23:49:12 crc kubenswrapper[4791]: I1210 23:49:12.610519 4791 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_configure-os-edpm-deployment-openstack-edpm-ipam-cb6z5_cbb47bf3-4d3a-4e7e-bc54-c8a3a2ec98c6/configure-os-edpm-deployment-openstack-edpm-ipam/0.log" Dec 10 23:49:12 crc kubenswrapper[4791]: I1210 23:49:12.611977 4791 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_configure-network-edpm-deployment-openstack-edpm-ipam-wf5sq_852ca005-3b2f-407e-ae86-9b1c5a5fe182/configure-network-edpm-deployment-openstack-edpm-ipam/0.log" Dec 10 23:49:12 crc kubenswrapper[4791]: I1210 23:49:12.791120 4791 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_dnsmasq-dns-cb6ffcf87-wcxkk_07473608-7cf2-4bcb-b7bb-046165c73afe/init/0.log" Dec 10 23:49:12 crc kubenswrapper[4791]: I1210 23:49:12.986946 4791 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_dnsmasq-dns-cb6ffcf87-wcxkk_07473608-7cf2-4bcb-b7bb-046165c73afe/init/0.log" Dec 10 23:49:13 crc kubenswrapper[4791]: I1210 23:49:13.005312 4791 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_dnsmasq-dns-cb6ffcf87-wcxkk_07473608-7cf2-4bcb-b7bb-046165c73afe/dnsmasq-dns/0.log" Dec 10 23:49:13 crc kubenswrapper[4791]: I1210 23:49:13.051566 4791 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_download-cache-edpm-deployment-openstack-edpm-ipam-nx9lt_36b35496-ca5d-49ac-bb9f-cb697fa27af6/download-cache-edpm-deployment-openstack-edpm-ipam/0.log" Dec 10 23:49:13 crc kubenswrapper[4791]: I1210 23:49:13.194935 4791 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_glance-default-external-api-0_9655fe55-b388-4296-b9e4-198ea42c70cd/glance-httpd/0.log" Dec 10 23:49:13 crc kubenswrapper[4791]: I1210 23:49:13.282128 4791 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_glance-default-external-api-0_9655fe55-b388-4296-b9e4-198ea42c70cd/glance-log/0.log" Dec 10 23:49:13 crc kubenswrapper[4791]: I1210 23:49:13.431229 4791 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_glance-default-internal-api-0_d3317b5c-5c3b-4138-8f93-1bd9a159f8dd/glance-httpd/0.log" Dec 10 23:49:13 crc kubenswrapper[4791]: I1210 23:49:13.442452 4791 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_glance-default-internal-api-0_d3317b5c-5c3b-4138-8f93-1bd9a159f8dd/glance-log/0.log" Dec 10 23:49:13 crc kubenswrapper[4791]: I1210 23:49:13.695043 4791 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_horizon-6999c5bf5b-7cr9t_0fab3b41-ed57-4536-ade0-f5f6ccde40e2/horizon/0.log" Dec 10 23:49:13 crc kubenswrapper[4791]: I1210 23:49:13.788722 4791 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_install-certs-edpm-deployment-openstack-edpm-ipam-n9w4z_7eb11eed-a935-4ce6-b348-2c7c8f6ef16a/install-certs-edpm-deployment-openstack-edpm-ipam/0.log" Dec 10 23:49:13 crc kubenswrapper[4791]: I1210 23:49:13.917299 4791 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_horizon-6999c5bf5b-7cr9t_0fab3b41-ed57-4536-ade0-f5f6ccde40e2/horizon-log/0.log" Dec 10 23:49:14 crc kubenswrapper[4791]: I1210 23:49:14.055234 4791 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_install-os-edpm-deployment-openstack-edpm-ipam-c6nvg_19add645-abcb-46e3-a103-e39305c0be1c/install-os-edpm-deployment-openstack-edpm-ipam/0.log" Dec 10 23:49:14 crc kubenswrapper[4791]: I1210 23:49:14.226681 4791 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_kube-state-metrics-0_a5c489de-39fe-42e1-963b-10a99cb531b6/kube-state-metrics/0.log" Dec 10 23:49:14 crc kubenswrapper[4791]: I1210 23:49:14.260209 4791 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_keystone-54b7f6dc79-9klwj_5a5460d8-5939-40d3-8453-a487b3c4b58f/keystone-api/0.log" Dec 10 23:49:14 crc kubenswrapper[4791]: I1210 23:49:14.400273 4791 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_libvirt-edpm-deployment-openstack-edpm-ipam-9czng_b180a2cd-d91a-4203-b0f8-7de5c1be8226/libvirt-edpm-deployment-openstack-edpm-ipam/0.log" Dec 10 23:49:14 crc kubenswrapper[4791]: I1210 23:49:14.767027 4791 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_neutron-856d8bb659-x5kwg_60652b01-72de-481d-b2ea-8c349e9aee52/neutron-httpd/0.log" Dec 10 23:49:14 crc kubenswrapper[4791]: I1210 23:49:14.796255 4791 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_neutron-856d8bb659-x5kwg_60652b01-72de-481d-b2ea-8c349e9aee52/neutron-api/0.log" Dec 10 23:49:15 crc kubenswrapper[4791]: I1210 23:49:15.034514 4791 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_neutron-metadata-edpm-deployment-openstack-edpm-ipam-nnggl_9b204a02-650a-4f54-b257-5791c8399bf6/neutron-metadata-edpm-deployment-openstack-edpm-ipam/0.log" Dec 10 23:49:15 crc kubenswrapper[4791]: I1210 23:49:15.418821 4791 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_nova-api-0_70ddc653-9583-4f3c-a0f2-a2bfa952e7a4/nova-api-log/0.log" Dec 10 23:49:15 crc kubenswrapper[4791]: I1210 23:49:15.577616 4791 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_nova-api-0_70ddc653-9583-4f3c-a0f2-a2bfa952e7a4/nova-api-api/0.log" Dec 10 23:49:15 crc kubenswrapper[4791]: I1210 23:49:15.700293 4791 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_nova-cell1-conductor-0_2dff7039-7fed-41ac-8703-8e6023146318/nova-cell1-conductor-conductor/0.log" Dec 10 23:49:15 crc kubenswrapper[4791]: I1210 23:49:15.701250 4791 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_nova-cell0-conductor-0_1082e5c5-2e93-4950-93b4-c66ee224c711/nova-cell0-conductor-conductor/0.log" Dec 10 23:49:15 crc kubenswrapper[4791]: I1210 23:49:15.969009 4791 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_nova-cell1-novncproxy-0_4c145576-0efc-440b-bc4d-5468995305b4/nova-cell1-novncproxy-novncproxy/0.log" Dec 10 23:49:15 crc kubenswrapper[4791]: I1210 23:49:15.985755 4791 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_nova-edpm-deployment-openstack-edpm-ipam-5krlr_f9fc3f8e-9635-48a4-8ec3-20034a1bc7ad/nova-edpm-deployment-openstack-edpm-ipam/0.log" Dec 10 23:49:16 crc kubenswrapper[4791]: I1210 23:49:16.287739 4791 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_nova-metadata-0_8ee8d247-355c-4e17-965c-0b3c958a546c/nova-metadata-log/0.log" Dec 10 23:49:16 crc kubenswrapper[4791]: I1210 23:49:16.440850 4791 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_nova-scheduler-0_b1d1b38e-2897-4c52-a4e4-d994dfffe805/nova-scheduler-scheduler/0.log" Dec 10 23:49:16 crc kubenswrapper[4791]: I1210 23:49:16.531910 4791 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_openstack-cell1-galera-0_f04fa829-3ebe-445d-a4e5-f7592ac682f3/mysql-bootstrap/0.log" Dec 10 23:49:16 crc kubenswrapper[4791]: I1210 23:49:16.729087 4791 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_openstack-cell1-galera-0_f04fa829-3ebe-445d-a4e5-f7592ac682f3/mysql-bootstrap/0.log" Dec 10 23:49:16 crc kubenswrapper[4791]: I1210 23:49:16.736213 4791 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_openstack-cell1-galera-0_f04fa829-3ebe-445d-a4e5-f7592ac682f3/galera/0.log" Dec 10 23:49:16 crc kubenswrapper[4791]: I1210 23:49:16.951627 4791 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_openstack-galera-0_aed2b986-c040-4191-864b-47f29c5c8537/mysql-bootstrap/0.log" Dec 10 23:49:17 crc kubenswrapper[4791]: I1210 23:49:17.195694 4791 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_openstack-galera-0_aed2b986-c040-4191-864b-47f29c5c8537/mysql-bootstrap/0.log" Dec 10 23:49:17 crc kubenswrapper[4791]: I1210 23:49:17.252596 4791 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_openstack-galera-0_aed2b986-c040-4191-864b-47f29c5c8537/galera/0.log" Dec 10 23:49:17 crc kubenswrapper[4791]: I1210 23:49:17.273032 4791 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_nova-metadata-0_8ee8d247-355c-4e17-965c-0b3c958a546c/nova-metadata-metadata/0.log" Dec 10 23:49:17 crc kubenswrapper[4791]: I1210 23:49:17.393851 4791 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_openstackclient_2d25921f-5a9e-4e3b-9c51-f07f95aa02a1/openstackclient/0.log" Dec 10 23:49:17 crc kubenswrapper[4791]: I1210 23:49:17.443229 4791 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovn-controller-metrics-djrh2_53ceaa05-0398-4cf9-a787-874adab015aa/openstack-network-exporter/0.log" Dec 10 23:49:17 crc kubenswrapper[4791]: I1210 23:49:17.645372 4791 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovn-controller-ovs-phsfg_ac28ec72-8d7b-4576-b2ba-c93731d50267/ovsdb-server-init/0.log" Dec 10 23:49:17 crc kubenswrapper[4791]: I1210 23:49:17.822272 4791 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovn-controller-ovs-phsfg_ac28ec72-8d7b-4576-b2ba-c93731d50267/ovsdb-server/0.log" Dec 10 23:49:17 crc kubenswrapper[4791]: I1210 23:49:17.917171 4791 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovn-controller-ovs-phsfg_ac28ec72-8d7b-4576-b2ba-c93731d50267/ovsdb-server-init/0.log" Dec 10 23:49:17 crc kubenswrapper[4791]: I1210 23:49:17.918772 4791 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovn-controller-ovs-phsfg_ac28ec72-8d7b-4576-b2ba-c93731d50267/ovs-vswitchd/0.log" Dec 10 23:49:18 crc kubenswrapper[4791]: I1210 23:49:18.055271 4791 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovn-controller-xb29h_780fd8a7-cfb0-4958-afb3-1d060398d3ea/ovn-controller/0.log" Dec 10 23:49:18 crc kubenswrapper[4791]: I1210 23:49:18.226388 4791 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovn-edpm-deployment-openstack-edpm-ipam-ql7sj_9b2b641e-b123-4328-9151-f4c95e1405f2/ovn-edpm-deployment-openstack-edpm-ipam/0.log" Dec 10 23:49:18 crc kubenswrapper[4791]: I1210 23:49:18.262738 4791 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovn-northd-0_b73f3e7f-f699-460a-8d40-6abc63f13a1c/openstack-network-exporter/0.log" Dec 10 23:49:18 crc kubenswrapper[4791]: I1210 23:49:18.287866 4791 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovn-northd-0_b73f3e7f-f699-460a-8d40-6abc63f13a1c/ovn-northd/0.log" Dec 10 23:49:18 crc kubenswrapper[4791]: I1210 23:49:18.513643 4791 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovsdbserver-nb-0_a4bfff72-3a66-40f0-a06a-432d3cb4fff4/ovsdbserver-nb/0.log" Dec 10 23:49:18 crc kubenswrapper[4791]: I1210 23:49:18.533381 4791 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovsdbserver-nb-0_a4bfff72-3a66-40f0-a06a-432d3cb4fff4/openstack-network-exporter/0.log" Dec 10 23:49:18 crc kubenswrapper[4791]: I1210 23:49:18.850112 4791 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovsdbserver-sb-0_61769416-0728-4475-b965-dedb50d4455d/openstack-network-exporter/0.log" Dec 10 23:49:18 crc kubenswrapper[4791]: I1210 23:49:18.900478 4791 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovsdbserver-sb-0_61769416-0728-4475-b965-dedb50d4455d/ovsdbserver-sb/0.log" Dec 10 23:49:19 crc kubenswrapper[4791]: I1210 23:49:19.011980 4791 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_placement-6487596ddd-btmd7_0babf4ec-03b9-48de-875d-a27a8b7ed119/placement-api/0.log" Dec 10 23:49:19 crc kubenswrapper[4791]: I1210 23:49:19.134472 4791 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_placement-6487596ddd-btmd7_0babf4ec-03b9-48de-875d-a27a8b7ed119/placement-log/0.log" Dec 10 23:49:19 crc kubenswrapper[4791]: I1210 23:49:19.160972 4791 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_rabbitmq-cell1-server-0_3add67e0-b67a-438b-b83f-4c9f6733b5cb/setup-container/0.log" Dec 10 23:49:19 crc kubenswrapper[4791]: I1210 23:49:19.339684 4791 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_rabbitmq-cell1-server-0_3add67e0-b67a-438b-b83f-4c9f6733b5cb/setup-container/0.log" Dec 10 23:49:19 crc kubenswrapper[4791]: I1210 23:49:19.377930 4791 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_rabbitmq-cell1-server-0_3add67e0-b67a-438b-b83f-4c9f6733b5cb/rabbitmq/0.log" Dec 10 23:49:19 crc kubenswrapper[4791]: I1210 23:49:19.440581 4791 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_rabbitmq-server-0_5f51c37d-2f6f-4ef0-a08c-77216e7db57b/setup-container/0.log" Dec 10 23:49:19 crc kubenswrapper[4791]: I1210 23:49:19.651590 4791 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_rabbitmq-server-0_5f51c37d-2f6f-4ef0-a08c-77216e7db57b/setup-container/0.log" Dec 10 23:49:19 crc kubenswrapper[4791]: I1210 23:49:19.666651 4791 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_rabbitmq-server-0_5f51c37d-2f6f-4ef0-a08c-77216e7db57b/rabbitmq/0.log" Dec 10 23:49:19 crc kubenswrapper[4791]: I1210 23:49:19.734398 4791 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_reboot-os-edpm-deployment-openstack-edpm-ipam-fbsgr_72f75339-ac62-4e10-a6a2-634ac54461cd/reboot-os-edpm-deployment-openstack-edpm-ipam/0.log" Dec 10 23:49:19 crc kubenswrapper[4791]: I1210 23:49:19.973143 4791 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_redhat-edpm-deployment-openstack-edpm-ipam-2zj6q_e77c2ca5-1bf1-4739-b1c3-b0913fcc9d99/redhat-edpm-deployment-openstack-edpm-ipam/0.log" Dec 10 23:49:19 crc kubenswrapper[4791]: I1210 23:49:19.996670 4791 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_repo-setup-edpm-deployment-openstack-edpm-ipam-xb2ph_eff77dbf-4e46-4119-9db0-eb32a1d9d11c/repo-setup-edpm-deployment-openstack-edpm-ipam/0.log" Dec 10 23:49:20 crc kubenswrapper[4791]: I1210 23:49:20.191401 4791 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ssh-known-hosts-edpm-deployment-8zwc6_12dfad91-cb70-4237-be65-427b3e919a05/ssh-known-hosts-edpm-deployment/0.log" Dec 10 23:49:20 crc kubenswrapper[4791]: I1210 23:49:20.199235 4791 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_run-os-edpm-deployment-openstack-edpm-ipam-5w7qc_62988878-9637-47ad-9877-fd81f32199d2/run-os-edpm-deployment-openstack-edpm-ipam/0.log" Dec 10 23:49:20 crc kubenswrapper[4791]: I1210 23:49:20.567963 4791 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-proxy-c7bbc5ff-rlprn_644890b3-8cae-45bc-b3eb-e0e6bfa8d957/proxy-server/0.log" Dec 10 23:49:20 crc kubenswrapper[4791]: I1210 23:49:20.614052 4791 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-proxy-c7bbc5ff-rlprn_644890b3-8cae-45bc-b3eb-e0e6bfa8d957/proxy-httpd/0.log" Dec 10 23:49:20 crc kubenswrapper[4791]: I1210 23:49:20.732611 4791 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-ring-rebalance-bqplg_16bd9b65-844e-4537-8d5b-c0a73666c2b2/swift-ring-rebalance/0.log" Dec 10 23:49:20 crc kubenswrapper[4791]: I1210 23:49:20.761916 4791 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_69a6c950-8a3b-4a7c-b284-ebd20157eb20/account-auditor/0.log" Dec 10 23:49:20 crc kubenswrapper[4791]: I1210 23:49:20.850751 4791 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_69a6c950-8a3b-4a7c-b284-ebd20157eb20/account-reaper/0.log" Dec 10 23:49:21 crc kubenswrapper[4791]: I1210 23:49:21.014977 4791 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_69a6c950-8a3b-4a7c-b284-ebd20157eb20/account-replicator/0.log" Dec 10 23:49:21 crc kubenswrapper[4791]: I1210 23:49:21.025530 4791 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_69a6c950-8a3b-4a7c-b284-ebd20157eb20/account-server/0.log" Dec 10 23:49:21 crc kubenswrapper[4791]: I1210 23:49:21.043244 4791 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_69a6c950-8a3b-4a7c-b284-ebd20157eb20/container-auditor/0.log" Dec 10 23:49:21 crc kubenswrapper[4791]: I1210 23:49:21.108201 4791 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_69a6c950-8a3b-4a7c-b284-ebd20157eb20/container-replicator/0.log" Dec 10 23:49:21 crc kubenswrapper[4791]: I1210 23:49:21.218694 4791 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_69a6c950-8a3b-4a7c-b284-ebd20157eb20/container-server/0.log" Dec 10 23:49:21 crc kubenswrapper[4791]: I1210 23:49:21.255953 4791 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_69a6c950-8a3b-4a7c-b284-ebd20157eb20/container-updater/0.log" Dec 10 23:49:21 crc kubenswrapper[4791]: I1210 23:49:21.346553 4791 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_69a6c950-8a3b-4a7c-b284-ebd20157eb20/object-auditor/0.log" Dec 10 23:49:21 crc kubenswrapper[4791]: I1210 23:49:21.356144 4791 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_69a6c950-8a3b-4a7c-b284-ebd20157eb20/object-expirer/0.log" Dec 10 23:49:21 crc kubenswrapper[4791]: I1210 23:49:21.436937 4791 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_69a6c950-8a3b-4a7c-b284-ebd20157eb20/object-replicator/0.log" Dec 10 23:49:21 crc kubenswrapper[4791]: I1210 23:49:21.456603 4791 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_69a6c950-8a3b-4a7c-b284-ebd20157eb20/object-server/0.log" Dec 10 23:49:21 crc kubenswrapper[4791]: I1210 23:49:21.529234 4791 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_69a6c950-8a3b-4a7c-b284-ebd20157eb20/rsync/0.log" Dec 10 23:49:21 crc kubenswrapper[4791]: I1210 23:49:21.555433 4791 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_69a6c950-8a3b-4a7c-b284-ebd20157eb20/object-updater/0.log" Dec 10 23:49:21 crc kubenswrapper[4791]: I1210 23:49:21.700726 4791 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_69a6c950-8a3b-4a7c-b284-ebd20157eb20/swift-recon-cron/0.log" Dec 10 23:49:21 crc kubenswrapper[4791]: I1210 23:49:21.756004 4791 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_telemetry-edpm-deployment-openstack-edpm-ipam-kxtwt_282a792a-a4a9-4e6d-b7ae-f5cedbe55787/telemetry-edpm-deployment-openstack-edpm-ipam/0.log" Dec 10 23:49:21 crc kubenswrapper[4791]: I1210 23:49:21.928519 4791 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_tempest-tests-tempest_cf11ce8f-d433-49ff-9390-064a2f4ec60d/tempest-tests-tempest-tests-runner/0.log" Dec 10 23:49:21 crc kubenswrapper[4791]: I1210 23:49:21.988369 4791 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_test-operator-logs-pod-tempest-tempest-tests-tempest_00c94f56-3118-486c-a325-9493d5bfd911/test-operator-logs-container/0.log" Dec 10 23:49:22 crc kubenswrapper[4791]: I1210 23:49:22.152080 4791 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_validate-network-edpm-deployment-openstack-edpm-ipam-gtc8x_73316991-2ef6-4257-8118-55378571b7d7/validate-network-edpm-deployment-openstack-edpm-ipam/0.log" Dec 10 23:49:25 crc kubenswrapper[4791]: I1210 23:49:25.037611 4791 patch_prober.go:28] interesting pod/machine-config-daemon-5rb5l container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 10 23:49:25 crc kubenswrapper[4791]: I1210 23:49:25.038174 4791 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-5rb5l" podUID="ba35653c-6e06-4cee-a4d6-137764090d18" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 10 23:49:30 crc kubenswrapper[4791]: I1210 23:49:30.280980 4791 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_memcached-0_49839437-3696-4ac8-98cf-2cc16048f21a/memcached/0.log" Dec 10 23:49:46 crc kubenswrapper[4791]: I1210 23:49:46.881502 4791 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_50d664eb6e76f12a148f821b8e085abb03568bf5cb6cd29276e1382c05lclqs_c76a21e1-575e-42e2-8372-e79986d26dbb/util/0.log" Dec 10 23:49:47 crc kubenswrapper[4791]: I1210 23:49:47.030747 4791 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_50d664eb6e76f12a148f821b8e085abb03568bf5cb6cd29276e1382c05lclqs_c76a21e1-575e-42e2-8372-e79986d26dbb/util/0.log" Dec 10 23:49:47 crc kubenswrapper[4791]: I1210 23:49:47.059145 4791 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_50d664eb6e76f12a148f821b8e085abb03568bf5cb6cd29276e1382c05lclqs_c76a21e1-575e-42e2-8372-e79986d26dbb/pull/0.log" Dec 10 23:49:47 crc kubenswrapper[4791]: I1210 23:49:47.086858 4791 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_50d664eb6e76f12a148f821b8e085abb03568bf5cb6cd29276e1382c05lclqs_c76a21e1-575e-42e2-8372-e79986d26dbb/pull/0.log" Dec 10 23:49:47 crc kubenswrapper[4791]: I1210 23:49:47.230899 4791 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_50d664eb6e76f12a148f821b8e085abb03568bf5cb6cd29276e1382c05lclqs_c76a21e1-575e-42e2-8372-e79986d26dbb/util/0.log" Dec 10 23:49:47 crc kubenswrapper[4791]: I1210 23:49:47.256462 4791 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_50d664eb6e76f12a148f821b8e085abb03568bf5cb6cd29276e1382c05lclqs_c76a21e1-575e-42e2-8372-e79986d26dbb/extract/0.log" Dec 10 23:49:47 crc kubenswrapper[4791]: I1210 23:49:47.260653 4791 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_50d664eb6e76f12a148f821b8e085abb03568bf5cb6cd29276e1382c05lclqs_c76a21e1-575e-42e2-8372-e79986d26dbb/pull/0.log" Dec 10 23:49:47 crc kubenswrapper[4791]: I1210 23:49:47.390614 4791 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_barbican-operator-controller-manager-7d9dfd778-4ndx5_dc042f4f-d336-484e-8c9d-60658c82f84a/kube-rbac-proxy/0.log" Dec 10 23:49:47 crc kubenswrapper[4791]: I1210 23:49:47.445857 4791 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_cinder-operator-controller-manager-6c677c69b-6jmn2_48cd2bae-61e0-446c-ac2a-48e70bff5187/kube-rbac-proxy/0.log" Dec 10 23:49:47 crc kubenswrapper[4791]: I1210 23:49:47.493742 4791 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_barbican-operator-controller-manager-7d9dfd778-4ndx5_dc042f4f-d336-484e-8c9d-60658c82f84a/manager/0.log" Dec 10 23:49:47 crc kubenswrapper[4791]: I1210 23:49:47.661415 4791 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_cinder-operator-controller-manager-6c677c69b-6jmn2_48cd2bae-61e0-446c-ac2a-48e70bff5187/manager/0.log" Dec 10 23:49:47 crc kubenswrapper[4791]: I1210 23:49:47.681639 4791 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_designate-operator-controller-manager-697fb699cf-vg956_19bf0bfe-4e8b-47a4-a919-4227b3e47bb3/kube-rbac-proxy/0.log" Dec 10 23:49:47 crc kubenswrapper[4791]: I1210 23:49:47.700294 4791 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_designate-operator-controller-manager-697fb699cf-vg956_19bf0bfe-4e8b-47a4-a919-4227b3e47bb3/manager/0.log" Dec 10 23:49:47 crc kubenswrapper[4791]: I1210 23:49:47.844688 4791 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_glance-operator-controller-manager-5697bb5779-cvbqc_f0b16aa2-e60e-4f0e-9679-9afc0a5ae027/kube-rbac-proxy/0.log" Dec 10 23:49:47 crc kubenswrapper[4791]: I1210 23:49:47.931475 4791 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_glance-operator-controller-manager-5697bb5779-cvbqc_f0b16aa2-e60e-4f0e-9679-9afc0a5ae027/manager/0.log" Dec 10 23:49:48 crc kubenswrapper[4791]: I1210 23:49:48.014800 4791 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_heat-operator-controller-manager-5f64f6f8bb-fx67c_d181199e-5cd2-408b-b167-059b34491a15/kube-rbac-proxy/0.log" Dec 10 23:49:48 crc kubenswrapper[4791]: I1210 23:49:48.039228 4791 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_heat-operator-controller-manager-5f64f6f8bb-fx67c_d181199e-5cd2-408b-b167-059b34491a15/manager/0.log" Dec 10 23:49:48 crc kubenswrapper[4791]: I1210 23:49:48.117144 4791 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_horizon-operator-controller-manager-68c6d99b8f-2lq47_e3fdc812-fbd3-4d35-8555-d55e9390ebd0/kube-rbac-proxy/0.log" Dec 10 23:49:48 crc kubenswrapper[4791]: I1210 23:49:48.222006 4791 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_horizon-operator-controller-manager-68c6d99b8f-2lq47_e3fdc812-fbd3-4d35-8555-d55e9390ebd0/manager/0.log" Dec 10 23:49:48 crc kubenswrapper[4791]: I1210 23:49:48.317801 4791 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_infra-operator-controller-manager-78d48bff9d-mmczn_416b455d-b397-4aad-baf1-88e880619eb5/kube-rbac-proxy/0.log" Dec 10 23:49:48 crc kubenswrapper[4791]: I1210 23:49:48.508635 4791 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_ironic-operator-controller-manager-967d97867-r6jdv_cb5fb221-76b0-48b4-ad27-7d0c51581ec4/kube-rbac-proxy/0.log" Dec 10 23:49:48 crc kubenswrapper[4791]: I1210 23:49:48.519442 4791 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_infra-operator-controller-manager-78d48bff9d-mmczn_416b455d-b397-4aad-baf1-88e880619eb5/manager/0.log" Dec 10 23:49:48 crc kubenswrapper[4791]: I1210 23:49:48.588062 4791 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_ironic-operator-controller-manager-967d97867-r6jdv_cb5fb221-76b0-48b4-ad27-7d0c51581ec4/manager/0.log" Dec 10 23:49:48 crc kubenswrapper[4791]: I1210 23:49:48.658950 4791 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_keystone-operator-controller-manager-7765d96ddf-2wsxl_311d4e7c-a0c7-47f6-8deb-7262068e8beb/kube-rbac-proxy/0.log" Dec 10 23:49:48 crc kubenswrapper[4791]: I1210 23:49:48.747333 4791 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_keystone-operator-controller-manager-7765d96ddf-2wsxl_311d4e7c-a0c7-47f6-8deb-7262068e8beb/manager/0.log" Dec 10 23:49:48 crc kubenswrapper[4791]: I1210 23:49:48.824031 4791 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_manila-operator-controller-manager-5b5fd79c9c-tsbtp_12b75f38-fb6a-4020-95b9-29c7dd872849/manager/0.log" Dec 10 23:49:48 crc kubenswrapper[4791]: I1210 23:49:48.875630 4791 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_manila-operator-controller-manager-5b5fd79c9c-tsbtp_12b75f38-fb6a-4020-95b9-29c7dd872849/kube-rbac-proxy/0.log" Dec 10 23:49:48 crc kubenswrapper[4791]: I1210 23:49:48.994466 4791 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_mariadb-operator-controller-manager-79c8c4686c-2wmm4_590c13cc-efa0-431a-9f2f-65fb97c1bd3c/kube-rbac-proxy/0.log" Dec 10 23:49:49 crc kubenswrapper[4791]: I1210 23:49:49.058742 4791 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_mariadb-operator-controller-manager-79c8c4686c-2wmm4_590c13cc-efa0-431a-9f2f-65fb97c1bd3c/manager/0.log" Dec 10 23:49:49 crc kubenswrapper[4791]: I1210 23:49:49.185105 4791 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_neutron-operator-controller-manager-5fdfd5b6b5-g8f76_86bbb7b2-5f55-46a2-b426-013ba183977f/kube-rbac-proxy/0.log" Dec 10 23:49:49 crc kubenswrapper[4791]: I1210 23:49:49.186263 4791 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_neutron-operator-controller-manager-5fdfd5b6b5-g8f76_86bbb7b2-5f55-46a2-b426-013ba183977f/manager/0.log" Dec 10 23:49:49 crc kubenswrapper[4791]: I1210 23:49:49.265627 4791 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_nova-operator-controller-manager-697bc559fc-2lkcl_3b2af4ca-48b4-463f-b0aa-c5cfd9099d67/kube-rbac-proxy/0.log" Dec 10 23:49:49 crc kubenswrapper[4791]: I1210 23:49:49.411785 4791 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_octavia-operator-controller-manager-998648c74-sjp6q_23a95e98-8ae3-4ac4-945d-0cae5af5d1ac/kube-rbac-proxy/0.log" Dec 10 23:49:49 crc kubenswrapper[4791]: I1210 23:49:49.425455 4791 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_nova-operator-controller-manager-697bc559fc-2lkcl_3b2af4ca-48b4-463f-b0aa-c5cfd9099d67/manager/0.log" Dec 10 23:49:49 crc kubenswrapper[4791]: I1210 23:49:49.482929 4791 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_octavia-operator-controller-manager-998648c74-sjp6q_23a95e98-8ae3-4ac4-945d-0cae5af5d1ac/manager/0.log" Dec 10 23:49:49 crc kubenswrapper[4791]: I1210 23:49:49.581844 4791 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_openstack-baremetal-operator-controller-manager-84b575879fv4t28_d5907af9-fe3f-4da0-ba70-993d36b25746/kube-rbac-proxy/0.log" Dec 10 23:49:49 crc kubenswrapper[4791]: I1210 23:49:49.621454 4791 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_openstack-baremetal-operator-controller-manager-84b575879fv4t28_d5907af9-fe3f-4da0-ba70-993d36b25746/manager/0.log" Dec 10 23:49:49 crc kubenswrapper[4791]: I1210 23:49:49.953982 4791 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_openstack-operator-controller-operator-6cc5d54456-w47nt_35509dee-f21b-4ab9-95ae-b93d2a4d93d2/operator/0.log" Dec 10 23:49:50 crc kubenswrapper[4791]: I1210 23:49:50.078819 4791 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_openstack-operator-index-zg7wn_f96dcde6-8c38-41a4-b1bd-b1c3007ff451/registry-server/0.log" Dec 10 23:49:50 crc kubenswrapper[4791]: I1210 23:49:50.189211 4791 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_ovn-operator-controller-manager-b6456fdb6-6rddm_87c40827-309c-4c68-ae46-bbe449a9e66c/kube-rbac-proxy/0.log" Dec 10 23:49:50 crc kubenswrapper[4791]: I1210 23:49:50.256565 4791 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_ovn-operator-controller-manager-b6456fdb6-6rddm_87c40827-309c-4c68-ae46-bbe449a9e66c/manager/0.log" Dec 10 23:49:50 crc kubenswrapper[4791]: I1210 23:49:50.428906 4791 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_placement-operator-controller-manager-78f8948974-bhgvx_b67ef71c-85c5-41ca-bf93-d1ca8a2fd007/kube-rbac-proxy/0.log" Dec 10 23:49:50 crc kubenswrapper[4791]: I1210 23:49:50.503265 4791 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_placement-operator-controller-manager-78f8948974-bhgvx_b67ef71c-85c5-41ca-bf93-d1ca8a2fd007/manager/0.log" Dec 10 23:49:50 crc kubenswrapper[4791]: I1210 23:49:50.685931 4791 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_rabbitmq-cluster-operator-manager-668c99d594-2bgc5_b3ae5e53-9077-4dd0-996d-2417f28b4736/operator/0.log" Dec 10 23:49:50 crc kubenswrapper[4791]: I1210 23:49:50.687791 4791 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_swift-operator-controller-manager-9d58d64bc-8vwnp_4cf9497a-4193-4d34-a68a-48085af6e8b5/kube-rbac-proxy/0.log" Dec 10 23:49:50 crc kubenswrapper[4791]: I1210 23:49:50.895316 4791 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_openstack-operator-controller-manager-686dfd865c-cft5f_e2c4ecea-2852-4e39-93fb-aee7cbe31aa4/manager/0.log" Dec 10 23:49:50 crc kubenswrapper[4791]: I1210 23:49:50.927250 4791 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_telemetry-operator-controller-manager-58d5ff84df-t6tnb_2418f025-d37e-4316-b804-9ab53c326c82/manager/0.log" Dec 10 23:49:50 crc kubenswrapper[4791]: I1210 23:49:50.955545 4791 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_telemetry-operator-controller-manager-58d5ff84df-t6tnb_2418f025-d37e-4316-b804-9ab53c326c82/kube-rbac-proxy/0.log" Dec 10 23:49:50 crc kubenswrapper[4791]: I1210 23:49:50.968010 4791 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_swift-operator-controller-manager-9d58d64bc-8vwnp_4cf9497a-4193-4d34-a68a-48085af6e8b5/manager/0.log" Dec 10 23:49:51 crc kubenswrapper[4791]: I1210 23:49:51.088580 4791 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_test-operator-controller-manager-5854674fcc-rw7ct_bda40eaa-446c-475e-8f2f-76f76caa0b85/kube-rbac-proxy/0.log" Dec 10 23:49:51 crc kubenswrapper[4791]: I1210 23:49:51.147651 4791 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_test-operator-controller-manager-5854674fcc-rw7ct_bda40eaa-446c-475e-8f2f-76f76caa0b85/manager/0.log" Dec 10 23:49:51 crc kubenswrapper[4791]: I1210 23:49:51.167246 4791 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_watcher-operator-controller-manager-75944c9b7-7cdb7_21e66f28-5ed9-475e-aa7d-a105ad1f2f9e/kube-rbac-proxy/0.log" Dec 10 23:49:51 crc kubenswrapper[4791]: I1210 23:49:51.267884 4791 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_watcher-operator-controller-manager-75944c9b7-7cdb7_21e66f28-5ed9-475e-aa7d-a105ad1f2f9e/manager/0.log" Dec 10 23:49:55 crc kubenswrapper[4791]: I1210 23:49:55.037850 4791 patch_prober.go:28] interesting pod/machine-config-daemon-5rb5l container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 10 23:49:55 crc kubenswrapper[4791]: I1210 23:49:55.038741 4791 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-5rb5l" podUID="ba35653c-6e06-4cee-a4d6-137764090d18" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 10 23:50:09 crc kubenswrapper[4791]: I1210 23:50:09.394184 4791 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-machine-api_control-plane-machine-set-operator-78cbb6b69f-ms8jw_c6c773c2-29e4-477d-91fb-f63052272f06/control-plane-machine-set-operator/0.log" Dec 10 23:50:09 crc kubenswrapper[4791]: I1210 23:50:09.556864 4791 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-machine-api_machine-api-operator-5694c8668f-cpbhz_cd65e91e-5b02-4f78-829e-93b4596ff8f6/kube-rbac-proxy/0.log" Dec 10 23:50:09 crc kubenswrapper[4791]: I1210 23:50:09.647358 4791 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-machine-api_machine-api-operator-5694c8668f-cpbhz_cd65e91e-5b02-4f78-829e-93b4596ff8f6/machine-api-operator/0.log" Dec 10 23:50:21 crc kubenswrapper[4791]: I1210 23:50:21.879493 4791 log.go:25] "Finished parsing log file" path="/var/log/pods/cert-manager_cert-manager-5b446d88c5-w9jv9_5f070f43-21a8-43bf-80db-9038ed2dd9f5/cert-manager-controller/0.log" Dec 10 23:50:22 crc kubenswrapper[4791]: I1210 23:50:22.078318 4791 log.go:25] "Finished parsing log file" path="/var/log/pods/cert-manager_cert-manager-webhook-5655c58dd6-xzg6b_33717f3f-c6f5-4be9-a7c5-d6a7ae599dc9/cert-manager-webhook/0.log" Dec 10 23:50:22 crc kubenswrapper[4791]: I1210 23:50:22.107871 4791 log.go:25] "Finished parsing log file" path="/var/log/pods/cert-manager_cert-manager-cainjector-7f985d654d-ntt86_18fded69-b7d4-42c9-80de-1c78e6389edf/cert-manager-cainjector/0.log" Dec 10 23:50:25 crc kubenswrapper[4791]: I1210 23:50:25.037695 4791 patch_prober.go:28] interesting pod/machine-config-daemon-5rb5l container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 10 23:50:25 crc kubenswrapper[4791]: I1210 23:50:25.038003 4791 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-5rb5l" podUID="ba35653c-6e06-4cee-a4d6-137764090d18" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 10 23:50:25 crc kubenswrapper[4791]: I1210 23:50:25.038042 4791 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-5rb5l" Dec 10 23:50:25 crc kubenswrapper[4791]: I1210 23:50:25.040286 4791 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"f44d2e51d8375135a5705f7cf71dde0bd0c6901d20f892b1346dc2eac144cbc0"} pod="openshift-machine-config-operator/machine-config-daemon-5rb5l" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Dec 10 23:50:25 crc kubenswrapper[4791]: I1210 23:50:25.040374 4791 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-5rb5l" podUID="ba35653c-6e06-4cee-a4d6-137764090d18" containerName="machine-config-daemon" containerID="cri-o://f44d2e51d8375135a5705f7cf71dde0bd0c6901d20f892b1346dc2eac144cbc0" gracePeriod=600 Dec 10 23:50:25 crc kubenswrapper[4791]: E1210 23:50:25.180689 4791 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-5rb5l_openshift-machine-config-operator(ba35653c-6e06-4cee-a4d6-137764090d18)\"" pod="openshift-machine-config-operator/machine-config-daemon-5rb5l" podUID="ba35653c-6e06-4cee-a4d6-137764090d18" Dec 10 23:50:25 crc kubenswrapper[4791]: I1210 23:50:25.762217 4791 generic.go:334] "Generic (PLEG): container finished" podID="ba35653c-6e06-4cee-a4d6-137764090d18" containerID="f44d2e51d8375135a5705f7cf71dde0bd0c6901d20f892b1346dc2eac144cbc0" exitCode=0 Dec 10 23:50:25 crc kubenswrapper[4791]: I1210 23:50:25.762267 4791 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-5rb5l" event={"ID":"ba35653c-6e06-4cee-a4d6-137764090d18","Type":"ContainerDied","Data":"f44d2e51d8375135a5705f7cf71dde0bd0c6901d20f892b1346dc2eac144cbc0"} Dec 10 23:50:25 crc kubenswrapper[4791]: I1210 23:50:25.762307 4791 scope.go:117] "RemoveContainer" containerID="b11a39d149fd20d0528f36ebfc87a9f2befcafcaa18f22a73d7824b34c5c7890" Dec 10 23:50:25 crc kubenswrapper[4791]: I1210 23:50:25.763258 4791 scope.go:117] "RemoveContainer" containerID="f44d2e51d8375135a5705f7cf71dde0bd0c6901d20f892b1346dc2eac144cbc0" Dec 10 23:50:25 crc kubenswrapper[4791]: E1210 23:50:25.763900 4791 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-5rb5l_openshift-machine-config-operator(ba35653c-6e06-4cee-a4d6-137764090d18)\"" pod="openshift-machine-config-operator/machine-config-daemon-5rb5l" podUID="ba35653c-6e06-4cee-a4d6-137764090d18" Dec 10 23:50:34 crc kubenswrapper[4791]: I1210 23:50:34.013131 4791 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-nmstate_nmstate-console-plugin-7fbb5f6569-xxb9s_14000c4a-8a7d-4ade-86e6-819e08eaf42a/nmstate-console-plugin/0.log" Dec 10 23:50:34 crc kubenswrapper[4791]: I1210 23:50:34.165737 4791 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-nmstate_nmstate-handler-2j8xz_fd7f0c56-2613-4ba7-b0a4-4ff765493973/nmstate-handler/0.log" Dec 10 23:50:34 crc kubenswrapper[4791]: I1210 23:50:34.195521 4791 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-nmstate_nmstate-metrics-7f946cbc9-mph8d_e85292bb-7cee-4b20-8c97-b1d1902e6001/kube-rbac-proxy/0.log" Dec 10 23:50:34 crc kubenswrapper[4791]: I1210 23:50:34.221551 4791 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-nmstate_nmstate-metrics-7f946cbc9-mph8d_e85292bb-7cee-4b20-8c97-b1d1902e6001/nmstate-metrics/0.log" Dec 10 23:50:34 crc kubenswrapper[4791]: I1210 23:50:34.324306 4791 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-nmstate_nmstate-operator-5b5b58f5c8-9qtjh_f6780aea-2f04-4dc3-96c6-8e4e7ef4284e/nmstate-operator/0.log" Dec 10 23:50:34 crc kubenswrapper[4791]: I1210 23:50:34.399418 4791 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-nmstate_nmstate-webhook-5f6d4c5ccb-x65fz_99524090-326f-47fd-98d9-bf7e8352cc4f/nmstate-webhook/0.log" Dec 10 23:50:38 crc kubenswrapper[4791]: I1210 23:50:38.885072 4791 scope.go:117] "RemoveContainer" containerID="f44d2e51d8375135a5705f7cf71dde0bd0c6901d20f892b1346dc2eac144cbc0" Dec 10 23:50:38 crc kubenswrapper[4791]: E1210 23:50:38.888194 4791 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-5rb5l_openshift-machine-config-operator(ba35653c-6e06-4cee-a4d6-137764090d18)\"" pod="openshift-machine-config-operator/machine-config-daemon-5rb5l" podUID="ba35653c-6e06-4cee-a4d6-137764090d18" Dec 10 23:50:49 crc kubenswrapper[4791]: I1210 23:50:49.273345 4791 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_controller-f8648f98b-q9cdb_e52e894c-94e4-4958-b261-c5c373c29040/kube-rbac-proxy/0.log" Dec 10 23:50:49 crc kubenswrapper[4791]: I1210 23:50:49.350065 4791 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_controller-f8648f98b-q9cdb_e52e894c-94e4-4958-b261-c5c373c29040/controller/0.log" Dec 10 23:50:49 crc kubenswrapper[4791]: I1210 23:50:49.453368 4791 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-gqlpm_a68714bc-7787-46c8-ba8f-21f7af6d9d67/cp-frr-files/0.log" Dec 10 23:50:49 crc kubenswrapper[4791]: I1210 23:50:49.619864 4791 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-gqlpm_a68714bc-7787-46c8-ba8f-21f7af6d9d67/cp-reloader/0.log" Dec 10 23:50:49 crc kubenswrapper[4791]: I1210 23:50:49.662301 4791 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-gqlpm_a68714bc-7787-46c8-ba8f-21f7af6d9d67/cp-metrics/0.log" Dec 10 23:50:49 crc kubenswrapper[4791]: I1210 23:50:49.662970 4791 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-gqlpm_a68714bc-7787-46c8-ba8f-21f7af6d9d67/cp-frr-files/0.log" Dec 10 23:50:49 crc kubenswrapper[4791]: I1210 23:50:49.687795 4791 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-gqlpm_a68714bc-7787-46c8-ba8f-21f7af6d9d67/cp-reloader/0.log" Dec 10 23:50:49 crc kubenswrapper[4791]: I1210 23:50:49.845307 4791 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-gqlpm_a68714bc-7787-46c8-ba8f-21f7af6d9d67/cp-metrics/0.log" Dec 10 23:50:49 crc kubenswrapper[4791]: I1210 23:50:49.854378 4791 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-gqlpm_a68714bc-7787-46c8-ba8f-21f7af6d9d67/cp-frr-files/0.log" Dec 10 23:50:49 crc kubenswrapper[4791]: I1210 23:50:49.859257 4791 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-gqlpm_a68714bc-7787-46c8-ba8f-21f7af6d9d67/cp-reloader/0.log" Dec 10 23:50:49 crc kubenswrapper[4791]: I1210 23:50:49.872667 4791 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-gqlpm_a68714bc-7787-46c8-ba8f-21f7af6d9d67/cp-metrics/0.log" Dec 10 23:50:49 crc kubenswrapper[4791]: I1210 23:50:49.995806 4791 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-gqlpm_a68714bc-7787-46c8-ba8f-21f7af6d9d67/cp-reloader/0.log" Dec 10 23:50:50 crc kubenswrapper[4791]: I1210 23:50:50.030623 4791 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-gqlpm_a68714bc-7787-46c8-ba8f-21f7af6d9d67/cp-frr-files/0.log" Dec 10 23:50:50 crc kubenswrapper[4791]: I1210 23:50:50.032170 4791 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-gqlpm_a68714bc-7787-46c8-ba8f-21f7af6d9d67/cp-metrics/0.log" Dec 10 23:50:50 crc kubenswrapper[4791]: I1210 23:50:50.043652 4791 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-gqlpm_a68714bc-7787-46c8-ba8f-21f7af6d9d67/controller/0.log" Dec 10 23:50:50 crc kubenswrapper[4791]: I1210 23:50:50.213980 4791 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-gqlpm_a68714bc-7787-46c8-ba8f-21f7af6d9d67/frr-metrics/0.log" Dec 10 23:50:50 crc kubenswrapper[4791]: I1210 23:50:50.218831 4791 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-gqlpm_a68714bc-7787-46c8-ba8f-21f7af6d9d67/kube-rbac-proxy/0.log" Dec 10 23:50:50 crc kubenswrapper[4791]: I1210 23:50:50.219135 4791 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-gqlpm_a68714bc-7787-46c8-ba8f-21f7af6d9d67/kube-rbac-proxy-frr/0.log" Dec 10 23:50:50 crc kubenswrapper[4791]: I1210 23:50:50.403957 4791 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-webhook-server-7fcb986d4-whw8m_d1dc382e-929c-44ce-b5c3-73410e626d0c/frr-k8s-webhook-server/0.log" Dec 10 23:50:50 crc kubenswrapper[4791]: I1210 23:50:50.483703 4791 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-gqlpm_a68714bc-7787-46c8-ba8f-21f7af6d9d67/reloader/0.log" Dec 10 23:50:50 crc kubenswrapper[4791]: I1210 23:50:50.675796 4791 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_metallb-operator-controller-manager-867ddc9894-mhxh7_72a07cc5-20c1-4f60-b6d7-84dfe1f78be2/manager/0.log" Dec 10 23:50:50 crc kubenswrapper[4791]: I1210 23:50:50.928134 4791 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_metallb-operator-webhook-server-568c9f7b75-dtmkk_f1ea99a7-677c-4d8b-89c6-3b24466632c2/webhook-server/0.log" Dec 10 23:50:51 crc kubenswrapper[4791]: I1210 23:50:51.061714 4791 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_speaker-jfggv_10e04a8d-1cac-4c7a-9b14-be8db393651a/kube-rbac-proxy/0.log" Dec 10 23:50:51 crc kubenswrapper[4791]: I1210 23:50:51.559848 4791 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-gqlpm_a68714bc-7787-46c8-ba8f-21f7af6d9d67/frr/0.log" Dec 10 23:50:51 crc kubenswrapper[4791]: I1210 23:50:51.718124 4791 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_speaker-jfggv_10e04a8d-1cac-4c7a-9b14-be8db393651a/speaker/0.log" Dec 10 23:50:53 crc kubenswrapper[4791]: I1210 23:50:53.884872 4791 scope.go:117] "RemoveContainer" containerID="f44d2e51d8375135a5705f7cf71dde0bd0c6901d20f892b1346dc2eac144cbc0" Dec 10 23:50:53 crc kubenswrapper[4791]: E1210 23:50:53.885529 4791 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-5rb5l_openshift-machine-config-operator(ba35653c-6e06-4cee-a4d6-137764090d18)\"" pod="openshift-machine-config-operator/machine-config-daemon-5rb5l" podUID="ba35653c-6e06-4cee-a4d6-137764090d18" Dec 10 23:51:04 crc kubenswrapper[4791]: I1210 23:51:04.107985 4791 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_5064f9f8917b246f69f5d7fc025e7e6c34236c02bca31167615d38212fppnkc_4ebf6fba-353e-4d14-9e9e-f2be85033920/util/0.log" Dec 10 23:51:04 crc kubenswrapper[4791]: I1210 23:51:04.257794 4791 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_5064f9f8917b246f69f5d7fc025e7e6c34236c02bca31167615d38212fppnkc_4ebf6fba-353e-4d14-9e9e-f2be85033920/util/0.log" Dec 10 23:51:04 crc kubenswrapper[4791]: I1210 23:51:04.277140 4791 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_5064f9f8917b246f69f5d7fc025e7e6c34236c02bca31167615d38212fppnkc_4ebf6fba-353e-4d14-9e9e-f2be85033920/pull/0.log" Dec 10 23:51:04 crc kubenswrapper[4791]: I1210 23:51:04.333410 4791 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_5064f9f8917b246f69f5d7fc025e7e6c34236c02bca31167615d38212fppnkc_4ebf6fba-353e-4d14-9e9e-f2be85033920/pull/0.log" Dec 10 23:51:04 crc kubenswrapper[4791]: I1210 23:51:04.487885 4791 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_5064f9f8917b246f69f5d7fc025e7e6c34236c02bca31167615d38212fppnkc_4ebf6fba-353e-4d14-9e9e-f2be85033920/util/0.log" Dec 10 23:51:04 crc kubenswrapper[4791]: I1210 23:51:04.489824 4791 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_5064f9f8917b246f69f5d7fc025e7e6c34236c02bca31167615d38212fppnkc_4ebf6fba-353e-4d14-9e9e-f2be85033920/extract/0.log" Dec 10 23:51:04 crc kubenswrapper[4791]: I1210 23:51:04.493857 4791 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_5064f9f8917b246f69f5d7fc025e7e6c34236c02bca31167615d38212fppnkc_4ebf6fba-353e-4d14-9e9e-f2be85033920/pull/0.log" Dec 10 23:51:04 crc kubenswrapper[4791]: I1210 23:51:04.717549 4791 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_af69698b82fdf008f5ff9e195cf808a654240e16b13dcd924b74994f83xxh8j_d1f6824d-b614-4d99-a538-2917ad1c4bb3/util/0.log" Dec 10 23:51:04 crc kubenswrapper[4791]: I1210 23:51:04.874114 4791 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_af69698b82fdf008f5ff9e195cf808a654240e16b13dcd924b74994f83xxh8j_d1f6824d-b614-4d99-a538-2917ad1c4bb3/util/0.log" Dec 10 23:51:04 crc kubenswrapper[4791]: I1210 23:51:04.874767 4791 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_af69698b82fdf008f5ff9e195cf808a654240e16b13dcd924b74994f83xxh8j_d1f6824d-b614-4d99-a538-2917ad1c4bb3/pull/0.log" Dec 10 23:51:04 crc kubenswrapper[4791]: I1210 23:51:04.885692 4791 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_af69698b82fdf008f5ff9e195cf808a654240e16b13dcd924b74994f83xxh8j_d1f6824d-b614-4d99-a538-2917ad1c4bb3/pull/0.log" Dec 10 23:51:05 crc kubenswrapper[4791]: I1210 23:51:05.041614 4791 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_af69698b82fdf008f5ff9e195cf808a654240e16b13dcd924b74994f83xxh8j_d1f6824d-b614-4d99-a538-2917ad1c4bb3/util/0.log" Dec 10 23:51:05 crc kubenswrapper[4791]: I1210 23:51:05.047517 4791 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_af69698b82fdf008f5ff9e195cf808a654240e16b13dcd924b74994f83xxh8j_d1f6824d-b614-4d99-a538-2917ad1c4bb3/pull/0.log" Dec 10 23:51:05 crc kubenswrapper[4791]: I1210 23:51:05.065415 4791 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_af69698b82fdf008f5ff9e195cf808a654240e16b13dcd924b74994f83xxh8j_d1f6824d-b614-4d99-a538-2917ad1c4bb3/extract/0.log" Dec 10 23:51:05 crc kubenswrapper[4791]: I1210 23:51:05.192327 4791 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_certified-operators-658k2_3c0bf306-a118-4131-9888-17cb46b832ac/extract-utilities/0.log" Dec 10 23:51:05 crc kubenswrapper[4791]: I1210 23:51:05.393410 4791 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_certified-operators-658k2_3c0bf306-a118-4131-9888-17cb46b832ac/extract-utilities/0.log" Dec 10 23:51:05 crc kubenswrapper[4791]: I1210 23:51:05.400832 4791 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_certified-operators-658k2_3c0bf306-a118-4131-9888-17cb46b832ac/extract-content/0.log" Dec 10 23:51:05 crc kubenswrapper[4791]: I1210 23:51:05.447036 4791 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_certified-operators-658k2_3c0bf306-a118-4131-9888-17cb46b832ac/extract-content/0.log" Dec 10 23:51:05 crc kubenswrapper[4791]: I1210 23:51:05.605083 4791 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_certified-operators-658k2_3c0bf306-a118-4131-9888-17cb46b832ac/extract-utilities/0.log" Dec 10 23:51:05 crc kubenswrapper[4791]: I1210 23:51:05.611924 4791 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_certified-operators-658k2_3c0bf306-a118-4131-9888-17cb46b832ac/extract-content/0.log" Dec 10 23:51:05 crc kubenswrapper[4791]: I1210 23:51:05.739064 4791 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_community-operators-x582q_d02b36c2-1b51-4466-af31-2e69357acc9c/extract-utilities/0.log" Dec 10 23:51:05 crc kubenswrapper[4791]: I1210 23:51:05.866882 4791 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_certified-operators-658k2_3c0bf306-a118-4131-9888-17cb46b832ac/registry-server/0.log" Dec 10 23:51:05 crc kubenswrapper[4791]: I1210 23:51:05.945071 4791 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_community-operators-x582q_d02b36c2-1b51-4466-af31-2e69357acc9c/extract-content/0.log" Dec 10 23:51:05 crc kubenswrapper[4791]: I1210 23:51:05.965492 4791 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_community-operators-x582q_d02b36c2-1b51-4466-af31-2e69357acc9c/extract-utilities/0.log" Dec 10 23:51:05 crc kubenswrapper[4791]: I1210 23:51:05.994246 4791 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_community-operators-x582q_d02b36c2-1b51-4466-af31-2e69357acc9c/extract-content/0.log" Dec 10 23:51:06 crc kubenswrapper[4791]: I1210 23:51:06.263427 4791 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_community-operators-x582q_d02b36c2-1b51-4466-af31-2e69357acc9c/extract-utilities/0.log" Dec 10 23:51:06 crc kubenswrapper[4791]: I1210 23:51:06.265290 4791 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_community-operators-x582q_d02b36c2-1b51-4466-af31-2e69357acc9c/extract-content/0.log" Dec 10 23:51:06 crc kubenswrapper[4791]: I1210 23:51:06.461462 4791 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_marketplace-operator-79b997595-8b9dm_c2479901-785c-4a0a-b494-90675f64b71d/marketplace-operator/0.log" Dec 10 23:51:06 crc kubenswrapper[4791]: I1210 23:51:06.546572 4791 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-marketplace-fcnzr_44085f7f-f0e9-4dec-a816-a687a0dd06ec/extract-utilities/0.log" Dec 10 23:51:06 crc kubenswrapper[4791]: I1210 23:51:06.792724 4791 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-marketplace-fcnzr_44085f7f-f0e9-4dec-a816-a687a0dd06ec/extract-content/0.log" Dec 10 23:51:06 crc kubenswrapper[4791]: I1210 23:51:06.801509 4791 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-marketplace-fcnzr_44085f7f-f0e9-4dec-a816-a687a0dd06ec/extract-utilities/0.log" Dec 10 23:51:06 crc kubenswrapper[4791]: I1210 23:51:06.813574 4791 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_community-operators-x582q_d02b36c2-1b51-4466-af31-2e69357acc9c/registry-server/0.log" Dec 10 23:51:06 crc kubenswrapper[4791]: I1210 23:51:06.828547 4791 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-marketplace-fcnzr_44085f7f-f0e9-4dec-a816-a687a0dd06ec/extract-content/0.log" Dec 10 23:51:07 crc kubenswrapper[4791]: I1210 23:51:07.031679 4791 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-marketplace-fcnzr_44085f7f-f0e9-4dec-a816-a687a0dd06ec/extract-content/0.log" Dec 10 23:51:07 crc kubenswrapper[4791]: I1210 23:51:07.116834 4791 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-marketplace-fcnzr_44085f7f-f0e9-4dec-a816-a687a0dd06ec/extract-utilities/0.log" Dec 10 23:51:07 crc kubenswrapper[4791]: I1210 23:51:07.168266 4791 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-marketplace-fcnzr_44085f7f-f0e9-4dec-a816-a687a0dd06ec/registry-server/0.log" Dec 10 23:51:07 crc kubenswrapper[4791]: I1210 23:51:07.194056 4791 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-operators-f8n6m_a68872ea-89d3-42a0-b2da-877e2d37218b/extract-utilities/0.log" Dec 10 23:51:07 crc kubenswrapper[4791]: I1210 23:51:07.433994 4791 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-operators-f8n6m_a68872ea-89d3-42a0-b2da-877e2d37218b/extract-utilities/0.log" Dec 10 23:51:07 crc kubenswrapper[4791]: I1210 23:51:07.462097 4791 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-operators-f8n6m_a68872ea-89d3-42a0-b2da-877e2d37218b/extract-content/0.log" Dec 10 23:51:07 crc kubenswrapper[4791]: I1210 23:51:07.478053 4791 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-operators-f8n6m_a68872ea-89d3-42a0-b2da-877e2d37218b/extract-content/0.log" Dec 10 23:51:07 crc kubenswrapper[4791]: I1210 23:51:07.629071 4791 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-operators-f8n6m_a68872ea-89d3-42a0-b2da-877e2d37218b/extract-utilities/0.log" Dec 10 23:51:07 crc kubenswrapper[4791]: I1210 23:51:07.683439 4791 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-operators-f8n6m_a68872ea-89d3-42a0-b2da-877e2d37218b/extract-content/0.log" Dec 10 23:51:08 crc kubenswrapper[4791]: I1210 23:51:08.184445 4791 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-operators-f8n6m_a68872ea-89d3-42a0-b2da-877e2d37218b/registry-server/0.log" Dec 10 23:51:08 crc kubenswrapper[4791]: I1210 23:51:08.885360 4791 scope.go:117] "RemoveContainer" containerID="f44d2e51d8375135a5705f7cf71dde0bd0c6901d20f892b1346dc2eac144cbc0" Dec 10 23:51:08 crc kubenswrapper[4791]: E1210 23:51:08.885697 4791 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-5rb5l_openshift-machine-config-operator(ba35653c-6e06-4cee-a4d6-137764090d18)\"" pod="openshift-machine-config-operator/machine-config-daemon-5rb5l" podUID="ba35653c-6e06-4cee-a4d6-137764090d18" Dec 10 23:51:09 crc kubenswrapper[4791]: I1210 23:51:09.621609 4791 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-operators-n6bwl"] Dec 10 23:51:09 crc kubenswrapper[4791]: E1210 23:51:09.626549 4791 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="7b92afeb-c305-478c-9489-4673510df2f8" containerName="extract-content" Dec 10 23:51:09 crc kubenswrapper[4791]: I1210 23:51:09.626593 4791 state_mem.go:107] "Deleted CPUSet assignment" podUID="7b92afeb-c305-478c-9489-4673510df2f8" containerName="extract-content" Dec 10 23:51:09 crc kubenswrapper[4791]: E1210 23:51:09.626638 4791 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="7b92afeb-c305-478c-9489-4673510df2f8" containerName="registry-server" Dec 10 23:51:09 crc kubenswrapper[4791]: I1210 23:51:09.626650 4791 state_mem.go:107] "Deleted CPUSet assignment" podUID="7b92afeb-c305-478c-9489-4673510df2f8" containerName="registry-server" Dec 10 23:51:09 crc kubenswrapper[4791]: E1210 23:51:09.626697 4791 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="7b92afeb-c305-478c-9489-4673510df2f8" containerName="extract-utilities" Dec 10 23:51:09 crc kubenswrapper[4791]: I1210 23:51:09.626707 4791 state_mem.go:107] "Deleted CPUSet assignment" podUID="7b92afeb-c305-478c-9489-4673510df2f8" containerName="extract-utilities" Dec 10 23:51:09 crc kubenswrapper[4791]: E1210 23:51:09.626725 4791 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="8906f851-2b6f-4387-872f-a9a6c4267b74" containerName="container-00" Dec 10 23:51:09 crc kubenswrapper[4791]: I1210 23:51:09.626733 4791 state_mem.go:107] "Deleted CPUSet assignment" podUID="8906f851-2b6f-4387-872f-a9a6c4267b74" containerName="container-00" Dec 10 23:51:09 crc kubenswrapper[4791]: I1210 23:51:09.627077 4791 memory_manager.go:354] "RemoveStaleState removing state" podUID="7b92afeb-c305-478c-9489-4673510df2f8" containerName="registry-server" Dec 10 23:51:09 crc kubenswrapper[4791]: I1210 23:51:09.627119 4791 memory_manager.go:354] "RemoveStaleState removing state" podUID="8906f851-2b6f-4387-872f-a9a6c4267b74" containerName="container-00" Dec 10 23:51:09 crc kubenswrapper[4791]: I1210 23:51:09.628586 4791 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-n6bwl" Dec 10 23:51:09 crc kubenswrapper[4791]: I1210 23:51:09.640741 4791 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-n6bwl"] Dec 10 23:51:09 crc kubenswrapper[4791]: I1210 23:51:09.743506 4791 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/6ab7384a-c2be-4ef5-8491-321e4828c1d1-catalog-content\") pod \"redhat-operators-n6bwl\" (UID: \"6ab7384a-c2be-4ef5-8491-321e4828c1d1\") " pod="openshift-marketplace/redhat-operators-n6bwl" Dec 10 23:51:09 crc kubenswrapper[4791]: I1210 23:51:09.743612 4791 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-fskm6\" (UniqueName: \"kubernetes.io/projected/6ab7384a-c2be-4ef5-8491-321e4828c1d1-kube-api-access-fskm6\") pod \"redhat-operators-n6bwl\" (UID: \"6ab7384a-c2be-4ef5-8491-321e4828c1d1\") " pod="openshift-marketplace/redhat-operators-n6bwl" Dec 10 23:51:09 crc kubenswrapper[4791]: I1210 23:51:09.743653 4791 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/6ab7384a-c2be-4ef5-8491-321e4828c1d1-utilities\") pod \"redhat-operators-n6bwl\" (UID: \"6ab7384a-c2be-4ef5-8491-321e4828c1d1\") " pod="openshift-marketplace/redhat-operators-n6bwl" Dec 10 23:51:09 crc kubenswrapper[4791]: I1210 23:51:09.845238 4791 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/6ab7384a-c2be-4ef5-8491-321e4828c1d1-catalog-content\") pod \"redhat-operators-n6bwl\" (UID: \"6ab7384a-c2be-4ef5-8491-321e4828c1d1\") " pod="openshift-marketplace/redhat-operators-n6bwl" Dec 10 23:51:09 crc kubenswrapper[4791]: I1210 23:51:09.845452 4791 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-fskm6\" (UniqueName: \"kubernetes.io/projected/6ab7384a-c2be-4ef5-8491-321e4828c1d1-kube-api-access-fskm6\") pod \"redhat-operators-n6bwl\" (UID: \"6ab7384a-c2be-4ef5-8491-321e4828c1d1\") " pod="openshift-marketplace/redhat-operators-n6bwl" Dec 10 23:51:09 crc kubenswrapper[4791]: I1210 23:51:09.845503 4791 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/6ab7384a-c2be-4ef5-8491-321e4828c1d1-utilities\") pod \"redhat-operators-n6bwl\" (UID: \"6ab7384a-c2be-4ef5-8491-321e4828c1d1\") " pod="openshift-marketplace/redhat-operators-n6bwl" Dec 10 23:51:09 crc kubenswrapper[4791]: I1210 23:51:09.846178 4791 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/6ab7384a-c2be-4ef5-8491-321e4828c1d1-utilities\") pod \"redhat-operators-n6bwl\" (UID: \"6ab7384a-c2be-4ef5-8491-321e4828c1d1\") " pod="openshift-marketplace/redhat-operators-n6bwl" Dec 10 23:51:09 crc kubenswrapper[4791]: I1210 23:51:09.846538 4791 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/6ab7384a-c2be-4ef5-8491-321e4828c1d1-catalog-content\") pod \"redhat-operators-n6bwl\" (UID: \"6ab7384a-c2be-4ef5-8491-321e4828c1d1\") " pod="openshift-marketplace/redhat-operators-n6bwl" Dec 10 23:51:09 crc kubenswrapper[4791]: I1210 23:51:09.869309 4791 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-fskm6\" (UniqueName: \"kubernetes.io/projected/6ab7384a-c2be-4ef5-8491-321e4828c1d1-kube-api-access-fskm6\") pod \"redhat-operators-n6bwl\" (UID: \"6ab7384a-c2be-4ef5-8491-321e4828c1d1\") " pod="openshift-marketplace/redhat-operators-n6bwl" Dec 10 23:51:09 crc kubenswrapper[4791]: I1210 23:51:09.954831 4791 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-n6bwl" Dec 10 23:51:10 crc kubenswrapper[4791]: I1210 23:51:10.470071 4791 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-n6bwl"] Dec 10 23:51:11 crc kubenswrapper[4791]: I1210 23:51:11.298902 4791 generic.go:334] "Generic (PLEG): container finished" podID="6ab7384a-c2be-4ef5-8491-321e4828c1d1" containerID="7bcbd2ab9cd9fe08bd8f0f2a1edcb537e64e94d2a23a4b19cdd2ecfc95c62d10" exitCode=0 Dec 10 23:51:11 crc kubenswrapper[4791]: I1210 23:51:11.299382 4791 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-n6bwl" event={"ID":"6ab7384a-c2be-4ef5-8491-321e4828c1d1","Type":"ContainerDied","Data":"7bcbd2ab9cd9fe08bd8f0f2a1edcb537e64e94d2a23a4b19cdd2ecfc95c62d10"} Dec 10 23:51:11 crc kubenswrapper[4791]: I1210 23:51:11.299409 4791 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-n6bwl" event={"ID":"6ab7384a-c2be-4ef5-8491-321e4828c1d1","Type":"ContainerStarted","Data":"b41c6dcabec5bbbcbd29f6f7c75ba582ea675b15a2b41b1bc8a5b4bb11a36979"} Dec 10 23:51:12 crc kubenswrapper[4791]: I1210 23:51:12.309411 4791 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-n6bwl" event={"ID":"6ab7384a-c2be-4ef5-8491-321e4828c1d1","Type":"ContainerStarted","Data":"a403aa6cfb198893f7ada45749e72bcf86c660e56c0a322b69d095c7e8a54b2a"} Dec 10 23:51:13 crc kubenswrapper[4791]: I1210 23:51:13.323053 4791 generic.go:334] "Generic (PLEG): container finished" podID="6ab7384a-c2be-4ef5-8491-321e4828c1d1" containerID="a403aa6cfb198893f7ada45749e72bcf86c660e56c0a322b69d095c7e8a54b2a" exitCode=0 Dec 10 23:51:13 crc kubenswrapper[4791]: I1210 23:51:13.323126 4791 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-n6bwl" event={"ID":"6ab7384a-c2be-4ef5-8491-321e4828c1d1","Type":"ContainerDied","Data":"a403aa6cfb198893f7ada45749e72bcf86c660e56c0a322b69d095c7e8a54b2a"} Dec 10 23:51:14 crc kubenswrapper[4791]: I1210 23:51:14.334968 4791 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-n6bwl" event={"ID":"6ab7384a-c2be-4ef5-8491-321e4828c1d1","Type":"ContainerStarted","Data":"bf4a2032a92f8ad2eea1a917c8b1af5fc7ccc42394c85cff9bf53de25744e550"} Dec 10 23:51:14 crc kubenswrapper[4791]: I1210 23:51:14.356439 4791 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-operators-n6bwl" podStartSLOduration=2.869250795 podStartE2EDuration="5.356417003s" podCreationTimestamp="2025-12-10 23:51:09 +0000 UTC" firstStartedPulling="2025-12-10 23:51:11.301500342 +0000 UTC m=+3705.731117955" lastFinishedPulling="2025-12-10 23:51:13.78866655 +0000 UTC m=+3708.218284163" observedRunningTime="2025-12-10 23:51:14.350402473 +0000 UTC m=+3708.780020086" watchObservedRunningTime="2025-12-10 23:51:14.356417003 +0000 UTC m=+3708.786034616" Dec 10 23:51:19 crc kubenswrapper[4791]: I1210 23:51:19.884434 4791 scope.go:117] "RemoveContainer" containerID="f44d2e51d8375135a5705f7cf71dde0bd0c6901d20f892b1346dc2eac144cbc0" Dec 10 23:51:19 crc kubenswrapper[4791]: E1210 23:51:19.885141 4791 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-5rb5l_openshift-machine-config-operator(ba35653c-6e06-4cee-a4d6-137764090d18)\"" pod="openshift-machine-config-operator/machine-config-daemon-5rb5l" podUID="ba35653c-6e06-4cee-a4d6-137764090d18" Dec 10 23:51:19 crc kubenswrapper[4791]: I1210 23:51:19.955684 4791 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-operators-n6bwl" Dec 10 23:51:19 crc kubenswrapper[4791]: I1210 23:51:19.955731 4791 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-operators-n6bwl" Dec 10 23:51:21 crc kubenswrapper[4791]: I1210 23:51:21.012112 4791 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-marketplace/redhat-operators-n6bwl" podUID="6ab7384a-c2be-4ef5-8491-321e4828c1d1" containerName="registry-server" probeResult="failure" output=< Dec 10 23:51:21 crc kubenswrapper[4791]: timeout: failed to connect service ":50051" within 1s Dec 10 23:51:21 crc kubenswrapper[4791]: > Dec 10 23:51:30 crc kubenswrapper[4791]: I1210 23:51:30.021022 4791 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-operators-n6bwl" Dec 10 23:51:30 crc kubenswrapper[4791]: I1210 23:51:30.086125 4791 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-operators-n6bwl" Dec 10 23:51:30 crc kubenswrapper[4791]: I1210 23:51:30.258656 4791 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-n6bwl"] Dec 10 23:51:31 crc kubenswrapper[4791]: I1210 23:51:31.488912 4791 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-operators-n6bwl" podUID="6ab7384a-c2be-4ef5-8491-321e4828c1d1" containerName="registry-server" containerID="cri-o://bf4a2032a92f8ad2eea1a917c8b1af5fc7ccc42394c85cff9bf53de25744e550" gracePeriod=2 Dec 10 23:51:31 crc kubenswrapper[4791]: I1210 23:51:31.889119 4791 scope.go:117] "RemoveContainer" containerID="f44d2e51d8375135a5705f7cf71dde0bd0c6901d20f892b1346dc2eac144cbc0" Dec 10 23:51:31 crc kubenswrapper[4791]: E1210 23:51:31.889714 4791 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-5rb5l_openshift-machine-config-operator(ba35653c-6e06-4cee-a4d6-137764090d18)\"" pod="openshift-machine-config-operator/machine-config-daemon-5rb5l" podUID="ba35653c-6e06-4cee-a4d6-137764090d18" Dec 10 23:51:32 crc kubenswrapper[4791]: I1210 23:51:32.101981 4791 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-n6bwl" Dec 10 23:51:32 crc kubenswrapper[4791]: I1210 23:51:32.216876 4791 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/6ab7384a-c2be-4ef5-8491-321e4828c1d1-utilities\") pod \"6ab7384a-c2be-4ef5-8491-321e4828c1d1\" (UID: \"6ab7384a-c2be-4ef5-8491-321e4828c1d1\") " Dec 10 23:51:32 crc kubenswrapper[4791]: I1210 23:51:32.217129 4791 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/6ab7384a-c2be-4ef5-8491-321e4828c1d1-catalog-content\") pod \"6ab7384a-c2be-4ef5-8491-321e4828c1d1\" (UID: \"6ab7384a-c2be-4ef5-8491-321e4828c1d1\") " Dec 10 23:51:32 crc kubenswrapper[4791]: I1210 23:51:32.217198 4791 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-fskm6\" (UniqueName: \"kubernetes.io/projected/6ab7384a-c2be-4ef5-8491-321e4828c1d1-kube-api-access-fskm6\") pod \"6ab7384a-c2be-4ef5-8491-321e4828c1d1\" (UID: \"6ab7384a-c2be-4ef5-8491-321e4828c1d1\") " Dec 10 23:51:32 crc kubenswrapper[4791]: I1210 23:51:32.218085 4791 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/6ab7384a-c2be-4ef5-8491-321e4828c1d1-utilities" (OuterVolumeSpecName: "utilities") pod "6ab7384a-c2be-4ef5-8491-321e4828c1d1" (UID: "6ab7384a-c2be-4ef5-8491-321e4828c1d1"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 10 23:51:32 crc kubenswrapper[4791]: I1210 23:51:32.242682 4791 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/6ab7384a-c2be-4ef5-8491-321e4828c1d1-kube-api-access-fskm6" (OuterVolumeSpecName: "kube-api-access-fskm6") pod "6ab7384a-c2be-4ef5-8491-321e4828c1d1" (UID: "6ab7384a-c2be-4ef5-8491-321e4828c1d1"). InnerVolumeSpecName "kube-api-access-fskm6". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 10 23:51:32 crc kubenswrapper[4791]: I1210 23:51:32.319867 4791 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-fskm6\" (UniqueName: \"kubernetes.io/projected/6ab7384a-c2be-4ef5-8491-321e4828c1d1-kube-api-access-fskm6\") on node \"crc\" DevicePath \"\"" Dec 10 23:51:32 crc kubenswrapper[4791]: I1210 23:51:32.319898 4791 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/6ab7384a-c2be-4ef5-8491-321e4828c1d1-utilities\") on node \"crc\" DevicePath \"\"" Dec 10 23:51:32 crc kubenswrapper[4791]: I1210 23:51:32.324700 4791 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/6ab7384a-c2be-4ef5-8491-321e4828c1d1-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "6ab7384a-c2be-4ef5-8491-321e4828c1d1" (UID: "6ab7384a-c2be-4ef5-8491-321e4828c1d1"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 10 23:51:32 crc kubenswrapper[4791]: I1210 23:51:32.421580 4791 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/6ab7384a-c2be-4ef5-8491-321e4828c1d1-catalog-content\") on node \"crc\" DevicePath \"\"" Dec 10 23:51:32 crc kubenswrapper[4791]: I1210 23:51:32.499637 4791 generic.go:334] "Generic (PLEG): container finished" podID="6ab7384a-c2be-4ef5-8491-321e4828c1d1" containerID="bf4a2032a92f8ad2eea1a917c8b1af5fc7ccc42394c85cff9bf53de25744e550" exitCode=0 Dec 10 23:51:32 crc kubenswrapper[4791]: I1210 23:51:32.499678 4791 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-n6bwl" event={"ID":"6ab7384a-c2be-4ef5-8491-321e4828c1d1","Type":"ContainerDied","Data":"bf4a2032a92f8ad2eea1a917c8b1af5fc7ccc42394c85cff9bf53de25744e550"} Dec 10 23:51:32 crc kubenswrapper[4791]: I1210 23:51:32.499692 4791 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-n6bwl" Dec 10 23:51:32 crc kubenswrapper[4791]: I1210 23:51:32.499702 4791 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-n6bwl" event={"ID":"6ab7384a-c2be-4ef5-8491-321e4828c1d1","Type":"ContainerDied","Data":"b41c6dcabec5bbbcbd29f6f7c75ba582ea675b15a2b41b1bc8a5b4bb11a36979"} Dec 10 23:51:32 crc kubenswrapper[4791]: I1210 23:51:32.499721 4791 scope.go:117] "RemoveContainer" containerID="bf4a2032a92f8ad2eea1a917c8b1af5fc7ccc42394c85cff9bf53de25744e550" Dec 10 23:51:32 crc kubenswrapper[4791]: I1210 23:51:32.571159 4791 scope.go:117] "RemoveContainer" containerID="a403aa6cfb198893f7ada45749e72bcf86c660e56c0a322b69d095c7e8a54b2a" Dec 10 23:51:32 crc kubenswrapper[4791]: I1210 23:51:32.574611 4791 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-n6bwl"] Dec 10 23:51:32 crc kubenswrapper[4791]: I1210 23:51:32.586537 4791 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-operators-n6bwl"] Dec 10 23:51:32 crc kubenswrapper[4791]: I1210 23:51:32.647526 4791 scope.go:117] "RemoveContainer" containerID="7bcbd2ab9cd9fe08bd8f0f2a1edcb537e64e94d2a23a4b19cdd2ecfc95c62d10" Dec 10 23:51:32 crc kubenswrapper[4791]: I1210 23:51:32.708510 4791 scope.go:117] "RemoveContainer" containerID="bf4a2032a92f8ad2eea1a917c8b1af5fc7ccc42394c85cff9bf53de25744e550" Dec 10 23:51:32 crc kubenswrapper[4791]: E1210 23:51:32.713622 4791 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"bf4a2032a92f8ad2eea1a917c8b1af5fc7ccc42394c85cff9bf53de25744e550\": container with ID starting with bf4a2032a92f8ad2eea1a917c8b1af5fc7ccc42394c85cff9bf53de25744e550 not found: ID does not exist" containerID="bf4a2032a92f8ad2eea1a917c8b1af5fc7ccc42394c85cff9bf53de25744e550" Dec 10 23:51:32 crc kubenswrapper[4791]: I1210 23:51:32.713853 4791 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"bf4a2032a92f8ad2eea1a917c8b1af5fc7ccc42394c85cff9bf53de25744e550"} err="failed to get container status \"bf4a2032a92f8ad2eea1a917c8b1af5fc7ccc42394c85cff9bf53de25744e550\": rpc error: code = NotFound desc = could not find container \"bf4a2032a92f8ad2eea1a917c8b1af5fc7ccc42394c85cff9bf53de25744e550\": container with ID starting with bf4a2032a92f8ad2eea1a917c8b1af5fc7ccc42394c85cff9bf53de25744e550 not found: ID does not exist" Dec 10 23:51:32 crc kubenswrapper[4791]: I1210 23:51:32.714006 4791 scope.go:117] "RemoveContainer" containerID="a403aa6cfb198893f7ada45749e72bcf86c660e56c0a322b69d095c7e8a54b2a" Dec 10 23:51:32 crc kubenswrapper[4791]: E1210 23:51:32.715594 4791 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"a403aa6cfb198893f7ada45749e72bcf86c660e56c0a322b69d095c7e8a54b2a\": container with ID starting with a403aa6cfb198893f7ada45749e72bcf86c660e56c0a322b69d095c7e8a54b2a not found: ID does not exist" containerID="a403aa6cfb198893f7ada45749e72bcf86c660e56c0a322b69d095c7e8a54b2a" Dec 10 23:51:32 crc kubenswrapper[4791]: I1210 23:51:32.715639 4791 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"a403aa6cfb198893f7ada45749e72bcf86c660e56c0a322b69d095c7e8a54b2a"} err="failed to get container status \"a403aa6cfb198893f7ada45749e72bcf86c660e56c0a322b69d095c7e8a54b2a\": rpc error: code = NotFound desc = could not find container \"a403aa6cfb198893f7ada45749e72bcf86c660e56c0a322b69d095c7e8a54b2a\": container with ID starting with a403aa6cfb198893f7ada45749e72bcf86c660e56c0a322b69d095c7e8a54b2a not found: ID does not exist" Dec 10 23:51:32 crc kubenswrapper[4791]: I1210 23:51:32.715671 4791 scope.go:117] "RemoveContainer" containerID="7bcbd2ab9cd9fe08bd8f0f2a1edcb537e64e94d2a23a4b19cdd2ecfc95c62d10" Dec 10 23:51:32 crc kubenswrapper[4791]: E1210 23:51:32.716481 4791 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"7bcbd2ab9cd9fe08bd8f0f2a1edcb537e64e94d2a23a4b19cdd2ecfc95c62d10\": container with ID starting with 7bcbd2ab9cd9fe08bd8f0f2a1edcb537e64e94d2a23a4b19cdd2ecfc95c62d10 not found: ID does not exist" containerID="7bcbd2ab9cd9fe08bd8f0f2a1edcb537e64e94d2a23a4b19cdd2ecfc95c62d10" Dec 10 23:51:32 crc kubenswrapper[4791]: I1210 23:51:32.716508 4791 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"7bcbd2ab9cd9fe08bd8f0f2a1edcb537e64e94d2a23a4b19cdd2ecfc95c62d10"} err="failed to get container status \"7bcbd2ab9cd9fe08bd8f0f2a1edcb537e64e94d2a23a4b19cdd2ecfc95c62d10\": rpc error: code = NotFound desc = could not find container \"7bcbd2ab9cd9fe08bd8f0f2a1edcb537e64e94d2a23a4b19cdd2ecfc95c62d10\": container with ID starting with 7bcbd2ab9cd9fe08bd8f0f2a1edcb537e64e94d2a23a4b19cdd2ecfc95c62d10 not found: ID does not exist" Dec 10 23:51:33 crc kubenswrapper[4791]: I1210 23:51:33.896115 4791 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="6ab7384a-c2be-4ef5-8491-321e4828c1d1" path="/var/lib/kubelet/pods/6ab7384a-c2be-4ef5-8491-321e4828c1d1/volumes" Dec 10 23:51:43 crc kubenswrapper[4791]: E1210 23:51:43.931172 4791 upgradeaware.go:427] Error proxying data from client to backend: readfrom tcp 38.102.83.30:51874->38.102.83.30:41701: write tcp 38.102.83.30:51874->38.102.83.30:41701: write: broken pipe Dec 10 23:51:45 crc kubenswrapper[4791]: I1210 23:51:45.892237 4791 scope.go:117] "RemoveContainer" containerID="f44d2e51d8375135a5705f7cf71dde0bd0c6901d20f892b1346dc2eac144cbc0" Dec 10 23:51:45 crc kubenswrapper[4791]: E1210 23:51:45.892993 4791 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-5rb5l_openshift-machine-config-operator(ba35653c-6e06-4cee-a4d6-137764090d18)\"" pod="openshift-machine-config-operator/machine-config-daemon-5rb5l" podUID="ba35653c-6e06-4cee-a4d6-137764090d18" Dec 10 23:51:56 crc kubenswrapper[4791]: I1210 23:51:56.885291 4791 scope.go:117] "RemoveContainer" containerID="f44d2e51d8375135a5705f7cf71dde0bd0c6901d20f892b1346dc2eac144cbc0" Dec 10 23:51:56 crc kubenswrapper[4791]: E1210 23:51:56.887968 4791 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-5rb5l_openshift-machine-config-operator(ba35653c-6e06-4cee-a4d6-137764090d18)\"" pod="openshift-machine-config-operator/machine-config-daemon-5rb5l" podUID="ba35653c-6e06-4cee-a4d6-137764090d18" Dec 10 23:52:02 crc kubenswrapper[4791]: I1210 23:52:02.787822 4791 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/community-operators-xdnq8"] Dec 10 23:52:02 crc kubenswrapper[4791]: E1210 23:52:02.789041 4791 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="6ab7384a-c2be-4ef5-8491-321e4828c1d1" containerName="extract-content" Dec 10 23:52:02 crc kubenswrapper[4791]: I1210 23:52:02.789065 4791 state_mem.go:107] "Deleted CPUSet assignment" podUID="6ab7384a-c2be-4ef5-8491-321e4828c1d1" containerName="extract-content" Dec 10 23:52:02 crc kubenswrapper[4791]: E1210 23:52:02.789122 4791 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="6ab7384a-c2be-4ef5-8491-321e4828c1d1" containerName="registry-server" Dec 10 23:52:02 crc kubenswrapper[4791]: I1210 23:52:02.789133 4791 state_mem.go:107] "Deleted CPUSet assignment" podUID="6ab7384a-c2be-4ef5-8491-321e4828c1d1" containerName="registry-server" Dec 10 23:52:02 crc kubenswrapper[4791]: E1210 23:52:02.789157 4791 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="6ab7384a-c2be-4ef5-8491-321e4828c1d1" containerName="extract-utilities" Dec 10 23:52:02 crc kubenswrapper[4791]: I1210 23:52:02.789171 4791 state_mem.go:107] "Deleted CPUSet assignment" podUID="6ab7384a-c2be-4ef5-8491-321e4828c1d1" containerName="extract-utilities" Dec 10 23:52:02 crc kubenswrapper[4791]: I1210 23:52:02.789486 4791 memory_manager.go:354] "RemoveStaleState removing state" podUID="6ab7384a-c2be-4ef5-8491-321e4828c1d1" containerName="registry-server" Dec 10 23:52:02 crc kubenswrapper[4791]: I1210 23:52:02.792561 4791 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-xdnq8" Dec 10 23:52:02 crc kubenswrapper[4791]: I1210 23:52:02.816287 4791 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-xdnq8"] Dec 10 23:52:02 crc kubenswrapper[4791]: I1210 23:52:02.995403 4791 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/dfc8cceb-9cfd-44c4-b148-7a33b2914124-catalog-content\") pod \"community-operators-xdnq8\" (UID: \"dfc8cceb-9cfd-44c4-b148-7a33b2914124\") " pod="openshift-marketplace/community-operators-xdnq8" Dec 10 23:52:02 crc kubenswrapper[4791]: I1210 23:52:02.995463 4791 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-7pqdg\" (UniqueName: \"kubernetes.io/projected/dfc8cceb-9cfd-44c4-b148-7a33b2914124-kube-api-access-7pqdg\") pod \"community-operators-xdnq8\" (UID: \"dfc8cceb-9cfd-44c4-b148-7a33b2914124\") " pod="openshift-marketplace/community-operators-xdnq8" Dec 10 23:52:02 crc kubenswrapper[4791]: I1210 23:52:02.995516 4791 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/dfc8cceb-9cfd-44c4-b148-7a33b2914124-utilities\") pod \"community-operators-xdnq8\" (UID: \"dfc8cceb-9cfd-44c4-b148-7a33b2914124\") " pod="openshift-marketplace/community-operators-xdnq8" Dec 10 23:52:03 crc kubenswrapper[4791]: I1210 23:52:03.097464 4791 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/dfc8cceb-9cfd-44c4-b148-7a33b2914124-catalog-content\") pod \"community-operators-xdnq8\" (UID: \"dfc8cceb-9cfd-44c4-b148-7a33b2914124\") " pod="openshift-marketplace/community-operators-xdnq8" Dec 10 23:52:03 crc kubenswrapper[4791]: I1210 23:52:03.097521 4791 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-7pqdg\" (UniqueName: \"kubernetes.io/projected/dfc8cceb-9cfd-44c4-b148-7a33b2914124-kube-api-access-7pqdg\") pod \"community-operators-xdnq8\" (UID: \"dfc8cceb-9cfd-44c4-b148-7a33b2914124\") " pod="openshift-marketplace/community-operators-xdnq8" Dec 10 23:52:03 crc kubenswrapper[4791]: I1210 23:52:03.097566 4791 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/dfc8cceb-9cfd-44c4-b148-7a33b2914124-utilities\") pod \"community-operators-xdnq8\" (UID: \"dfc8cceb-9cfd-44c4-b148-7a33b2914124\") " pod="openshift-marketplace/community-operators-xdnq8" Dec 10 23:52:03 crc kubenswrapper[4791]: I1210 23:52:03.098139 4791 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/dfc8cceb-9cfd-44c4-b148-7a33b2914124-utilities\") pod \"community-operators-xdnq8\" (UID: \"dfc8cceb-9cfd-44c4-b148-7a33b2914124\") " pod="openshift-marketplace/community-operators-xdnq8" Dec 10 23:52:03 crc kubenswrapper[4791]: I1210 23:52:03.098467 4791 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/dfc8cceb-9cfd-44c4-b148-7a33b2914124-catalog-content\") pod \"community-operators-xdnq8\" (UID: \"dfc8cceb-9cfd-44c4-b148-7a33b2914124\") " pod="openshift-marketplace/community-operators-xdnq8" Dec 10 23:52:03 crc kubenswrapper[4791]: I1210 23:52:03.148187 4791 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-7pqdg\" (UniqueName: \"kubernetes.io/projected/dfc8cceb-9cfd-44c4-b148-7a33b2914124-kube-api-access-7pqdg\") pod \"community-operators-xdnq8\" (UID: \"dfc8cceb-9cfd-44c4-b148-7a33b2914124\") " pod="openshift-marketplace/community-operators-xdnq8" Dec 10 23:52:03 crc kubenswrapper[4791]: I1210 23:52:03.417831 4791 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-xdnq8" Dec 10 23:52:03 crc kubenswrapper[4791]: I1210 23:52:03.925869 4791 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-xdnq8"] Dec 10 23:52:03 crc kubenswrapper[4791]: W1210 23:52:03.937371 4791 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-poddfc8cceb_9cfd_44c4_b148_7a33b2914124.slice/crio-ee25d728e84283e5cc07e3ca6a53403df06a9f4fcee5d769252d55fbc09e2b04 WatchSource:0}: Error finding container ee25d728e84283e5cc07e3ca6a53403df06a9f4fcee5d769252d55fbc09e2b04: Status 404 returned error can't find the container with id ee25d728e84283e5cc07e3ca6a53403df06a9f4fcee5d769252d55fbc09e2b04 Dec 10 23:52:03 crc kubenswrapper[4791]: I1210 23:52:03.971747 4791 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-xdnq8" event={"ID":"dfc8cceb-9cfd-44c4-b148-7a33b2914124","Type":"ContainerStarted","Data":"ee25d728e84283e5cc07e3ca6a53403df06a9f4fcee5d769252d55fbc09e2b04"} Dec 10 23:52:04 crc kubenswrapper[4791]: I1210 23:52:04.983555 4791 generic.go:334] "Generic (PLEG): container finished" podID="dfc8cceb-9cfd-44c4-b148-7a33b2914124" containerID="0ef50d7dd037f03a84fe00cc7904f1a36059b5713bccf931dcb00e24f1fe1c1c" exitCode=0 Dec 10 23:52:04 crc kubenswrapper[4791]: I1210 23:52:04.983938 4791 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-xdnq8" event={"ID":"dfc8cceb-9cfd-44c4-b148-7a33b2914124","Type":"ContainerDied","Data":"0ef50d7dd037f03a84fe00cc7904f1a36059b5713bccf931dcb00e24f1fe1c1c"} Dec 10 23:52:05 crc kubenswrapper[4791]: I1210 23:52:05.998484 4791 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-xdnq8" event={"ID":"dfc8cceb-9cfd-44c4-b148-7a33b2914124","Type":"ContainerStarted","Data":"b0b8c22337fbb54556078de9bdfbaf07d88eb96e2a8bb7c25324d23f4b12e639"} Dec 10 23:52:07 crc kubenswrapper[4791]: I1210 23:52:07.015613 4791 generic.go:334] "Generic (PLEG): container finished" podID="dfc8cceb-9cfd-44c4-b148-7a33b2914124" containerID="b0b8c22337fbb54556078de9bdfbaf07d88eb96e2a8bb7c25324d23f4b12e639" exitCode=0 Dec 10 23:52:07 crc kubenswrapper[4791]: I1210 23:52:07.015826 4791 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-xdnq8" event={"ID":"dfc8cceb-9cfd-44c4-b148-7a33b2914124","Type":"ContainerDied","Data":"b0b8c22337fbb54556078de9bdfbaf07d88eb96e2a8bb7c25324d23f4b12e639"} Dec 10 23:52:08 crc kubenswrapper[4791]: I1210 23:52:08.025960 4791 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-xdnq8" event={"ID":"dfc8cceb-9cfd-44c4-b148-7a33b2914124","Type":"ContainerStarted","Data":"2f683abb13c8a9dea97149673e99e6fa0ba61abf8ffb0f2489d5e5f64b168fa0"} Dec 10 23:52:08 crc kubenswrapper[4791]: I1210 23:52:08.067127 4791 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/community-operators-xdnq8" podStartSLOduration=3.407614656 podStartE2EDuration="6.067111307s" podCreationTimestamp="2025-12-10 23:52:02 +0000 UTC" firstStartedPulling="2025-12-10 23:52:04.986877873 +0000 UTC m=+3759.416495506" lastFinishedPulling="2025-12-10 23:52:07.646374544 +0000 UTC m=+3762.075992157" observedRunningTime="2025-12-10 23:52:08.059180353 +0000 UTC m=+3762.488797986" watchObservedRunningTime="2025-12-10 23:52:08.067111307 +0000 UTC m=+3762.496728910" Dec 10 23:52:09 crc kubenswrapper[4791]: I1210 23:52:09.884963 4791 scope.go:117] "RemoveContainer" containerID="f44d2e51d8375135a5705f7cf71dde0bd0c6901d20f892b1346dc2eac144cbc0" Dec 10 23:52:09 crc kubenswrapper[4791]: E1210 23:52:09.886042 4791 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-5rb5l_openshift-machine-config-operator(ba35653c-6e06-4cee-a4d6-137764090d18)\"" pod="openshift-machine-config-operator/machine-config-daemon-5rb5l" podUID="ba35653c-6e06-4cee-a4d6-137764090d18" Dec 10 23:52:13 crc kubenswrapper[4791]: I1210 23:52:13.418657 4791 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/community-operators-xdnq8" Dec 10 23:52:13 crc kubenswrapper[4791]: I1210 23:52:13.420444 4791 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/community-operators-xdnq8" Dec 10 23:52:13 crc kubenswrapper[4791]: I1210 23:52:13.516032 4791 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/community-operators-xdnq8" Dec 10 23:52:14 crc kubenswrapper[4791]: I1210 23:52:14.175311 4791 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/community-operators-xdnq8" Dec 10 23:52:14 crc kubenswrapper[4791]: I1210 23:52:14.240139 4791 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-xdnq8"] Dec 10 23:52:16 crc kubenswrapper[4791]: I1210 23:52:16.102722 4791 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/community-operators-xdnq8" podUID="dfc8cceb-9cfd-44c4-b148-7a33b2914124" containerName="registry-server" containerID="cri-o://2f683abb13c8a9dea97149673e99e6fa0ba61abf8ffb0f2489d5e5f64b168fa0" gracePeriod=2 Dec 10 23:52:16 crc kubenswrapper[4791]: I1210 23:52:16.616968 4791 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-xdnq8" Dec 10 23:52:16 crc kubenswrapper[4791]: I1210 23:52:16.763868 4791 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/dfc8cceb-9cfd-44c4-b148-7a33b2914124-catalog-content\") pod \"dfc8cceb-9cfd-44c4-b148-7a33b2914124\" (UID: \"dfc8cceb-9cfd-44c4-b148-7a33b2914124\") " Dec 10 23:52:16 crc kubenswrapper[4791]: I1210 23:52:16.764071 4791 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-7pqdg\" (UniqueName: \"kubernetes.io/projected/dfc8cceb-9cfd-44c4-b148-7a33b2914124-kube-api-access-7pqdg\") pod \"dfc8cceb-9cfd-44c4-b148-7a33b2914124\" (UID: \"dfc8cceb-9cfd-44c4-b148-7a33b2914124\") " Dec 10 23:52:16 crc kubenswrapper[4791]: I1210 23:52:16.764111 4791 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/dfc8cceb-9cfd-44c4-b148-7a33b2914124-utilities\") pod \"dfc8cceb-9cfd-44c4-b148-7a33b2914124\" (UID: \"dfc8cceb-9cfd-44c4-b148-7a33b2914124\") " Dec 10 23:52:16 crc kubenswrapper[4791]: I1210 23:52:16.765236 4791 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/dfc8cceb-9cfd-44c4-b148-7a33b2914124-utilities" (OuterVolumeSpecName: "utilities") pod "dfc8cceb-9cfd-44c4-b148-7a33b2914124" (UID: "dfc8cceb-9cfd-44c4-b148-7a33b2914124"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 10 23:52:16 crc kubenswrapper[4791]: I1210 23:52:16.775665 4791 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/dfc8cceb-9cfd-44c4-b148-7a33b2914124-kube-api-access-7pqdg" (OuterVolumeSpecName: "kube-api-access-7pqdg") pod "dfc8cceb-9cfd-44c4-b148-7a33b2914124" (UID: "dfc8cceb-9cfd-44c4-b148-7a33b2914124"). InnerVolumeSpecName "kube-api-access-7pqdg". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 10 23:52:16 crc kubenswrapper[4791]: I1210 23:52:16.813244 4791 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/dfc8cceb-9cfd-44c4-b148-7a33b2914124-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "dfc8cceb-9cfd-44c4-b148-7a33b2914124" (UID: "dfc8cceb-9cfd-44c4-b148-7a33b2914124"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 10 23:52:16 crc kubenswrapper[4791]: I1210 23:52:16.865829 4791 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-7pqdg\" (UniqueName: \"kubernetes.io/projected/dfc8cceb-9cfd-44c4-b148-7a33b2914124-kube-api-access-7pqdg\") on node \"crc\" DevicePath \"\"" Dec 10 23:52:16 crc kubenswrapper[4791]: I1210 23:52:16.865862 4791 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/dfc8cceb-9cfd-44c4-b148-7a33b2914124-utilities\") on node \"crc\" DevicePath \"\"" Dec 10 23:52:16 crc kubenswrapper[4791]: I1210 23:52:16.865873 4791 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/dfc8cceb-9cfd-44c4-b148-7a33b2914124-catalog-content\") on node \"crc\" DevicePath \"\"" Dec 10 23:52:17 crc kubenswrapper[4791]: I1210 23:52:17.112165 4791 generic.go:334] "Generic (PLEG): container finished" podID="dfc8cceb-9cfd-44c4-b148-7a33b2914124" containerID="2f683abb13c8a9dea97149673e99e6fa0ba61abf8ffb0f2489d5e5f64b168fa0" exitCode=0 Dec 10 23:52:17 crc kubenswrapper[4791]: I1210 23:52:17.112205 4791 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-xdnq8" event={"ID":"dfc8cceb-9cfd-44c4-b148-7a33b2914124","Type":"ContainerDied","Data":"2f683abb13c8a9dea97149673e99e6fa0ba61abf8ffb0f2489d5e5f64b168fa0"} Dec 10 23:52:17 crc kubenswrapper[4791]: I1210 23:52:17.112230 4791 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-xdnq8" event={"ID":"dfc8cceb-9cfd-44c4-b148-7a33b2914124","Type":"ContainerDied","Data":"ee25d728e84283e5cc07e3ca6a53403df06a9f4fcee5d769252d55fbc09e2b04"} Dec 10 23:52:17 crc kubenswrapper[4791]: I1210 23:52:17.112249 4791 scope.go:117] "RemoveContainer" containerID="2f683abb13c8a9dea97149673e99e6fa0ba61abf8ffb0f2489d5e5f64b168fa0" Dec 10 23:52:17 crc kubenswrapper[4791]: I1210 23:52:17.112380 4791 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-xdnq8" Dec 10 23:52:17 crc kubenswrapper[4791]: I1210 23:52:17.134810 4791 scope.go:117] "RemoveContainer" containerID="b0b8c22337fbb54556078de9bdfbaf07d88eb96e2a8bb7c25324d23f4b12e639" Dec 10 23:52:17 crc kubenswrapper[4791]: I1210 23:52:17.178843 4791 scope.go:117] "RemoveContainer" containerID="0ef50d7dd037f03a84fe00cc7904f1a36059b5713bccf931dcb00e24f1fe1c1c" Dec 10 23:52:17 crc kubenswrapper[4791]: I1210 23:52:17.181187 4791 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-xdnq8"] Dec 10 23:52:17 crc kubenswrapper[4791]: I1210 23:52:17.193920 4791 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/community-operators-xdnq8"] Dec 10 23:52:17 crc kubenswrapper[4791]: I1210 23:52:17.236196 4791 scope.go:117] "RemoveContainer" containerID="2f683abb13c8a9dea97149673e99e6fa0ba61abf8ffb0f2489d5e5f64b168fa0" Dec 10 23:52:17 crc kubenswrapper[4791]: E1210 23:52:17.236813 4791 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"2f683abb13c8a9dea97149673e99e6fa0ba61abf8ffb0f2489d5e5f64b168fa0\": container with ID starting with 2f683abb13c8a9dea97149673e99e6fa0ba61abf8ffb0f2489d5e5f64b168fa0 not found: ID does not exist" containerID="2f683abb13c8a9dea97149673e99e6fa0ba61abf8ffb0f2489d5e5f64b168fa0" Dec 10 23:52:17 crc kubenswrapper[4791]: I1210 23:52:17.236891 4791 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"2f683abb13c8a9dea97149673e99e6fa0ba61abf8ffb0f2489d5e5f64b168fa0"} err="failed to get container status \"2f683abb13c8a9dea97149673e99e6fa0ba61abf8ffb0f2489d5e5f64b168fa0\": rpc error: code = NotFound desc = could not find container \"2f683abb13c8a9dea97149673e99e6fa0ba61abf8ffb0f2489d5e5f64b168fa0\": container with ID starting with 2f683abb13c8a9dea97149673e99e6fa0ba61abf8ffb0f2489d5e5f64b168fa0 not found: ID does not exist" Dec 10 23:52:17 crc kubenswrapper[4791]: I1210 23:52:17.236929 4791 scope.go:117] "RemoveContainer" containerID="b0b8c22337fbb54556078de9bdfbaf07d88eb96e2a8bb7c25324d23f4b12e639" Dec 10 23:52:17 crc kubenswrapper[4791]: E1210 23:52:17.237283 4791 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"b0b8c22337fbb54556078de9bdfbaf07d88eb96e2a8bb7c25324d23f4b12e639\": container with ID starting with b0b8c22337fbb54556078de9bdfbaf07d88eb96e2a8bb7c25324d23f4b12e639 not found: ID does not exist" containerID="b0b8c22337fbb54556078de9bdfbaf07d88eb96e2a8bb7c25324d23f4b12e639" Dec 10 23:52:17 crc kubenswrapper[4791]: I1210 23:52:17.237318 4791 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"b0b8c22337fbb54556078de9bdfbaf07d88eb96e2a8bb7c25324d23f4b12e639"} err="failed to get container status \"b0b8c22337fbb54556078de9bdfbaf07d88eb96e2a8bb7c25324d23f4b12e639\": rpc error: code = NotFound desc = could not find container \"b0b8c22337fbb54556078de9bdfbaf07d88eb96e2a8bb7c25324d23f4b12e639\": container with ID starting with b0b8c22337fbb54556078de9bdfbaf07d88eb96e2a8bb7c25324d23f4b12e639 not found: ID does not exist" Dec 10 23:52:17 crc kubenswrapper[4791]: I1210 23:52:17.237335 4791 scope.go:117] "RemoveContainer" containerID="0ef50d7dd037f03a84fe00cc7904f1a36059b5713bccf931dcb00e24f1fe1c1c" Dec 10 23:52:17 crc kubenswrapper[4791]: E1210 23:52:17.237608 4791 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"0ef50d7dd037f03a84fe00cc7904f1a36059b5713bccf931dcb00e24f1fe1c1c\": container with ID starting with 0ef50d7dd037f03a84fe00cc7904f1a36059b5713bccf931dcb00e24f1fe1c1c not found: ID does not exist" containerID="0ef50d7dd037f03a84fe00cc7904f1a36059b5713bccf931dcb00e24f1fe1c1c" Dec 10 23:52:17 crc kubenswrapper[4791]: I1210 23:52:17.237638 4791 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"0ef50d7dd037f03a84fe00cc7904f1a36059b5713bccf931dcb00e24f1fe1c1c"} err="failed to get container status \"0ef50d7dd037f03a84fe00cc7904f1a36059b5713bccf931dcb00e24f1fe1c1c\": rpc error: code = NotFound desc = could not find container \"0ef50d7dd037f03a84fe00cc7904f1a36059b5713bccf931dcb00e24f1fe1c1c\": container with ID starting with 0ef50d7dd037f03a84fe00cc7904f1a36059b5713bccf931dcb00e24f1fe1c1c not found: ID does not exist" Dec 10 23:52:17 crc kubenswrapper[4791]: I1210 23:52:17.906022 4791 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="dfc8cceb-9cfd-44c4-b148-7a33b2914124" path="/var/lib/kubelet/pods/dfc8cceb-9cfd-44c4-b148-7a33b2914124/volumes" Dec 10 23:52:23 crc kubenswrapper[4791]: I1210 23:52:23.885277 4791 scope.go:117] "RemoveContainer" containerID="f44d2e51d8375135a5705f7cf71dde0bd0c6901d20f892b1346dc2eac144cbc0" Dec 10 23:52:23 crc kubenswrapper[4791]: E1210 23:52:23.886582 4791 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-5rb5l_openshift-machine-config-operator(ba35653c-6e06-4cee-a4d6-137764090d18)\"" pod="openshift-machine-config-operator/machine-config-daemon-5rb5l" podUID="ba35653c-6e06-4cee-a4d6-137764090d18" Dec 10 23:52:36 crc kubenswrapper[4791]: I1210 23:52:36.884721 4791 scope.go:117] "RemoveContainer" containerID="f44d2e51d8375135a5705f7cf71dde0bd0c6901d20f892b1346dc2eac144cbc0" Dec 10 23:52:36 crc kubenswrapper[4791]: E1210 23:52:36.885736 4791 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-5rb5l_openshift-machine-config-operator(ba35653c-6e06-4cee-a4d6-137764090d18)\"" pod="openshift-machine-config-operator/machine-config-daemon-5rb5l" podUID="ba35653c-6e06-4cee-a4d6-137764090d18" Dec 10 23:52:45 crc kubenswrapper[4791]: I1210 23:52:45.945232 4791 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/certified-operators-cmz9s"] Dec 10 23:52:45 crc kubenswrapper[4791]: E1210 23:52:45.946325 4791 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="dfc8cceb-9cfd-44c4-b148-7a33b2914124" containerName="registry-server" Dec 10 23:52:45 crc kubenswrapper[4791]: I1210 23:52:45.946365 4791 state_mem.go:107] "Deleted CPUSet assignment" podUID="dfc8cceb-9cfd-44c4-b148-7a33b2914124" containerName="registry-server" Dec 10 23:52:45 crc kubenswrapper[4791]: E1210 23:52:45.946386 4791 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="dfc8cceb-9cfd-44c4-b148-7a33b2914124" containerName="extract-utilities" Dec 10 23:52:45 crc kubenswrapper[4791]: I1210 23:52:45.946397 4791 state_mem.go:107] "Deleted CPUSet assignment" podUID="dfc8cceb-9cfd-44c4-b148-7a33b2914124" containerName="extract-utilities" Dec 10 23:52:45 crc kubenswrapper[4791]: E1210 23:52:45.946433 4791 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="dfc8cceb-9cfd-44c4-b148-7a33b2914124" containerName="extract-content" Dec 10 23:52:45 crc kubenswrapper[4791]: I1210 23:52:45.946443 4791 state_mem.go:107] "Deleted CPUSet assignment" podUID="dfc8cceb-9cfd-44c4-b148-7a33b2914124" containerName="extract-content" Dec 10 23:52:45 crc kubenswrapper[4791]: I1210 23:52:45.946759 4791 memory_manager.go:354] "RemoveStaleState removing state" podUID="dfc8cceb-9cfd-44c4-b148-7a33b2914124" containerName="registry-server" Dec 10 23:52:45 crc kubenswrapper[4791]: I1210 23:52:45.948935 4791 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-cmz9s" Dec 10 23:52:45 crc kubenswrapper[4791]: I1210 23:52:45.971571 4791 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-cmz9s"] Dec 10 23:52:46 crc kubenswrapper[4791]: I1210 23:52:46.092240 4791 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/2b323bf3-c265-4df1-969d-eb4d52f13929-catalog-content\") pod \"certified-operators-cmz9s\" (UID: \"2b323bf3-c265-4df1-969d-eb4d52f13929\") " pod="openshift-marketplace/certified-operators-cmz9s" Dec 10 23:52:46 crc kubenswrapper[4791]: I1210 23:52:46.092364 4791 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/2b323bf3-c265-4df1-969d-eb4d52f13929-utilities\") pod \"certified-operators-cmz9s\" (UID: \"2b323bf3-c265-4df1-969d-eb4d52f13929\") " pod="openshift-marketplace/certified-operators-cmz9s" Dec 10 23:52:46 crc kubenswrapper[4791]: I1210 23:52:46.092563 4791 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-48mgr\" (UniqueName: \"kubernetes.io/projected/2b323bf3-c265-4df1-969d-eb4d52f13929-kube-api-access-48mgr\") pod \"certified-operators-cmz9s\" (UID: \"2b323bf3-c265-4df1-969d-eb4d52f13929\") " pod="openshift-marketplace/certified-operators-cmz9s" Dec 10 23:52:46 crc kubenswrapper[4791]: I1210 23:52:46.195142 4791 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/2b323bf3-c265-4df1-969d-eb4d52f13929-catalog-content\") pod \"certified-operators-cmz9s\" (UID: \"2b323bf3-c265-4df1-969d-eb4d52f13929\") " pod="openshift-marketplace/certified-operators-cmz9s" Dec 10 23:52:46 crc kubenswrapper[4791]: I1210 23:52:46.195485 4791 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/2b323bf3-c265-4df1-969d-eb4d52f13929-utilities\") pod \"certified-operators-cmz9s\" (UID: \"2b323bf3-c265-4df1-969d-eb4d52f13929\") " pod="openshift-marketplace/certified-operators-cmz9s" Dec 10 23:52:46 crc kubenswrapper[4791]: I1210 23:52:46.195641 4791 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-48mgr\" (UniqueName: \"kubernetes.io/projected/2b323bf3-c265-4df1-969d-eb4d52f13929-kube-api-access-48mgr\") pod \"certified-operators-cmz9s\" (UID: \"2b323bf3-c265-4df1-969d-eb4d52f13929\") " pod="openshift-marketplace/certified-operators-cmz9s" Dec 10 23:52:46 crc kubenswrapper[4791]: I1210 23:52:46.196108 4791 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/2b323bf3-c265-4df1-969d-eb4d52f13929-catalog-content\") pod \"certified-operators-cmz9s\" (UID: \"2b323bf3-c265-4df1-969d-eb4d52f13929\") " pod="openshift-marketplace/certified-operators-cmz9s" Dec 10 23:52:46 crc kubenswrapper[4791]: I1210 23:52:46.196174 4791 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/2b323bf3-c265-4df1-969d-eb4d52f13929-utilities\") pod \"certified-operators-cmz9s\" (UID: \"2b323bf3-c265-4df1-969d-eb4d52f13929\") " pod="openshift-marketplace/certified-operators-cmz9s" Dec 10 23:52:46 crc kubenswrapper[4791]: I1210 23:52:46.224407 4791 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-48mgr\" (UniqueName: \"kubernetes.io/projected/2b323bf3-c265-4df1-969d-eb4d52f13929-kube-api-access-48mgr\") pod \"certified-operators-cmz9s\" (UID: \"2b323bf3-c265-4df1-969d-eb4d52f13929\") " pod="openshift-marketplace/certified-operators-cmz9s" Dec 10 23:52:46 crc kubenswrapper[4791]: I1210 23:52:46.273794 4791 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-cmz9s" Dec 10 23:52:46 crc kubenswrapper[4791]: I1210 23:52:46.809446 4791 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-cmz9s"] Dec 10 23:52:47 crc kubenswrapper[4791]: I1210 23:52:47.442878 4791 generic.go:334] "Generic (PLEG): container finished" podID="2b323bf3-c265-4df1-969d-eb4d52f13929" containerID="0cb59a6503fe4e15136b829c6137627c27e067ca73e687aaf97ea67db494819d" exitCode=0 Dec 10 23:52:47 crc kubenswrapper[4791]: I1210 23:52:47.443007 4791 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-cmz9s" event={"ID":"2b323bf3-c265-4df1-969d-eb4d52f13929","Type":"ContainerDied","Data":"0cb59a6503fe4e15136b829c6137627c27e067ca73e687aaf97ea67db494819d"} Dec 10 23:52:47 crc kubenswrapper[4791]: I1210 23:52:47.443214 4791 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-cmz9s" event={"ID":"2b323bf3-c265-4df1-969d-eb4d52f13929","Type":"ContainerStarted","Data":"7c3fb8b5a151b2e07ff9fe0e1c793f118b6e076ec21d6d4654c2511552a3d15c"} Dec 10 23:52:47 crc kubenswrapper[4791]: I1210 23:52:47.445555 4791 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Dec 10 23:52:47 crc kubenswrapper[4791]: I1210 23:52:47.890656 4791 scope.go:117] "RemoveContainer" containerID="f44d2e51d8375135a5705f7cf71dde0bd0c6901d20f892b1346dc2eac144cbc0" Dec 10 23:52:47 crc kubenswrapper[4791]: E1210 23:52:47.891295 4791 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-5rb5l_openshift-machine-config-operator(ba35653c-6e06-4cee-a4d6-137764090d18)\"" pod="openshift-machine-config-operator/machine-config-daemon-5rb5l" podUID="ba35653c-6e06-4cee-a4d6-137764090d18" Dec 10 23:52:54 crc kubenswrapper[4791]: I1210 23:52:54.515990 4791 generic.go:334] "Generic (PLEG): container finished" podID="2171ff1e-63b0-41ff-8019-5a72a0c59f5f" containerID="3f8fb9ccc563247b2f211de395f436764efe1eba5da3c41369225bf5feb95308" exitCode=0 Dec 10 23:52:54 crc kubenswrapper[4791]: I1210 23:52:54.516127 4791 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-6t57w/must-gather-drn2z" event={"ID":"2171ff1e-63b0-41ff-8019-5a72a0c59f5f","Type":"ContainerDied","Data":"3f8fb9ccc563247b2f211de395f436764efe1eba5da3c41369225bf5feb95308"} Dec 10 23:52:54 crc kubenswrapper[4791]: I1210 23:52:54.517211 4791 scope.go:117] "RemoveContainer" containerID="3f8fb9ccc563247b2f211de395f436764efe1eba5da3c41369225bf5feb95308" Dec 10 23:52:54 crc kubenswrapper[4791]: I1210 23:52:54.519289 4791 generic.go:334] "Generic (PLEG): container finished" podID="2b323bf3-c265-4df1-969d-eb4d52f13929" containerID="6d84430b787598105fdb57ce76f24b0237a26a578ec32111ab30a2c7700a9a88" exitCode=0 Dec 10 23:52:54 crc kubenswrapper[4791]: I1210 23:52:54.519321 4791 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-cmz9s" event={"ID":"2b323bf3-c265-4df1-969d-eb4d52f13929","Type":"ContainerDied","Data":"6d84430b787598105fdb57ce76f24b0237a26a578ec32111ab30a2c7700a9a88"} Dec 10 23:52:54 crc kubenswrapper[4791]: I1210 23:52:54.899917 4791 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-must-gather-6t57w_must-gather-drn2z_2171ff1e-63b0-41ff-8019-5a72a0c59f5f/gather/0.log" Dec 10 23:52:55 crc kubenswrapper[4791]: I1210 23:52:55.530957 4791 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-cmz9s" event={"ID":"2b323bf3-c265-4df1-969d-eb4d52f13929","Type":"ContainerStarted","Data":"0df53d68040369e4c480959d8bf536538fd03b31b6f7e08043be69459ca56f1b"} Dec 10 23:52:55 crc kubenswrapper[4791]: I1210 23:52:55.555579 4791 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/certified-operators-cmz9s" podStartSLOduration=3.023604067 podStartE2EDuration="10.555553538s" podCreationTimestamp="2025-12-10 23:52:45 +0000 UTC" firstStartedPulling="2025-12-10 23:52:47.444969187 +0000 UTC m=+3801.874586840" lastFinishedPulling="2025-12-10 23:52:54.976918698 +0000 UTC m=+3809.406536311" observedRunningTime="2025-12-10 23:52:55.547326786 +0000 UTC m=+3809.976944409" watchObservedRunningTime="2025-12-10 23:52:55.555553538 +0000 UTC m=+3809.985171151" Dec 10 23:52:56 crc kubenswrapper[4791]: I1210 23:52:56.274329 4791 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/certified-operators-cmz9s" Dec 10 23:52:56 crc kubenswrapper[4791]: I1210 23:52:56.274730 4791 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/certified-operators-cmz9s" Dec 10 23:52:57 crc kubenswrapper[4791]: I1210 23:52:57.322893 4791 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-marketplace/certified-operators-cmz9s" podUID="2b323bf3-c265-4df1-969d-eb4d52f13929" containerName="registry-server" probeResult="failure" output=< Dec 10 23:52:57 crc kubenswrapper[4791]: timeout: failed to connect service ":50051" within 1s Dec 10 23:52:57 crc kubenswrapper[4791]: > Dec 10 23:53:01 crc kubenswrapper[4791]: I1210 23:53:01.885571 4791 scope.go:117] "RemoveContainer" containerID="f44d2e51d8375135a5705f7cf71dde0bd0c6901d20f892b1346dc2eac144cbc0" Dec 10 23:53:01 crc kubenswrapper[4791]: E1210 23:53:01.886355 4791 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-5rb5l_openshift-machine-config-operator(ba35653c-6e06-4cee-a4d6-137764090d18)\"" pod="openshift-machine-config-operator/machine-config-daemon-5rb5l" podUID="ba35653c-6e06-4cee-a4d6-137764090d18" Dec 10 23:53:02 crc kubenswrapper[4791]: I1210 23:53:02.181629 4791 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-must-gather-6t57w/must-gather-drn2z"] Dec 10 23:53:02 crc kubenswrapper[4791]: I1210 23:53:02.182231 4791 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-must-gather-6t57w/must-gather-drn2z" podUID="2171ff1e-63b0-41ff-8019-5a72a0c59f5f" containerName="copy" containerID="cri-o://706ee31dc0ed7f4f1c4e136cb4a5a2a8f0f4228c2e114dd3b83a3906e1db6991" gracePeriod=2 Dec 10 23:53:02 crc kubenswrapper[4791]: I1210 23:53:02.194663 4791 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-must-gather-6t57w/must-gather-drn2z"] Dec 10 23:53:02 crc kubenswrapper[4791]: I1210 23:53:02.614425 4791 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-must-gather-6t57w_must-gather-drn2z_2171ff1e-63b0-41ff-8019-5a72a0c59f5f/copy/0.log" Dec 10 23:53:02 crc kubenswrapper[4791]: I1210 23:53:02.614942 4791 generic.go:334] "Generic (PLEG): container finished" podID="2171ff1e-63b0-41ff-8019-5a72a0c59f5f" containerID="706ee31dc0ed7f4f1c4e136cb4a5a2a8f0f4228c2e114dd3b83a3906e1db6991" exitCode=143 Dec 10 23:53:02 crc kubenswrapper[4791]: I1210 23:53:02.614984 4791 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="b192885bcb7a6095d4caeb8ce43999436a8c24b2b8a23c196030367368c6c64c" Dec 10 23:53:02 crc kubenswrapper[4791]: I1210 23:53:02.646938 4791 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-must-gather-6t57w_must-gather-drn2z_2171ff1e-63b0-41ff-8019-5a72a0c59f5f/copy/0.log" Dec 10 23:53:02 crc kubenswrapper[4791]: I1210 23:53:02.647660 4791 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-6t57w/must-gather-drn2z" Dec 10 23:53:02 crc kubenswrapper[4791]: I1210 23:53:02.684837 4791 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-29zpv\" (UniqueName: \"kubernetes.io/projected/2171ff1e-63b0-41ff-8019-5a72a0c59f5f-kube-api-access-29zpv\") pod \"2171ff1e-63b0-41ff-8019-5a72a0c59f5f\" (UID: \"2171ff1e-63b0-41ff-8019-5a72a0c59f5f\") " Dec 10 23:53:02 crc kubenswrapper[4791]: I1210 23:53:02.684943 4791 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"must-gather-output\" (UniqueName: \"kubernetes.io/empty-dir/2171ff1e-63b0-41ff-8019-5a72a0c59f5f-must-gather-output\") pod \"2171ff1e-63b0-41ff-8019-5a72a0c59f5f\" (UID: \"2171ff1e-63b0-41ff-8019-5a72a0c59f5f\") " Dec 10 23:53:02 crc kubenswrapper[4791]: I1210 23:53:02.690757 4791 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/2171ff1e-63b0-41ff-8019-5a72a0c59f5f-kube-api-access-29zpv" (OuterVolumeSpecName: "kube-api-access-29zpv") pod "2171ff1e-63b0-41ff-8019-5a72a0c59f5f" (UID: "2171ff1e-63b0-41ff-8019-5a72a0c59f5f"). InnerVolumeSpecName "kube-api-access-29zpv". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 10 23:53:02 crc kubenswrapper[4791]: I1210 23:53:02.786127 4791 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-29zpv\" (UniqueName: \"kubernetes.io/projected/2171ff1e-63b0-41ff-8019-5a72a0c59f5f-kube-api-access-29zpv\") on node \"crc\" DevicePath \"\"" Dec 10 23:53:02 crc kubenswrapper[4791]: I1210 23:53:02.835286 4791 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/2171ff1e-63b0-41ff-8019-5a72a0c59f5f-must-gather-output" (OuterVolumeSpecName: "must-gather-output") pod "2171ff1e-63b0-41ff-8019-5a72a0c59f5f" (UID: "2171ff1e-63b0-41ff-8019-5a72a0c59f5f"). InnerVolumeSpecName "must-gather-output". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 10 23:53:02 crc kubenswrapper[4791]: I1210 23:53:02.886943 4791 reconciler_common.go:293] "Volume detached for volume \"must-gather-output\" (UniqueName: \"kubernetes.io/empty-dir/2171ff1e-63b0-41ff-8019-5a72a0c59f5f-must-gather-output\") on node \"crc\" DevicePath \"\"" Dec 10 23:53:03 crc kubenswrapper[4791]: I1210 23:53:03.625990 4791 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-6t57w/must-gather-drn2z" Dec 10 23:53:03 crc kubenswrapper[4791]: I1210 23:53:03.901244 4791 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="2171ff1e-63b0-41ff-8019-5a72a0c59f5f" path="/var/lib/kubelet/pods/2171ff1e-63b0-41ff-8019-5a72a0c59f5f/volumes" Dec 10 23:53:06 crc kubenswrapper[4791]: I1210 23:53:06.350192 4791 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/certified-operators-cmz9s" Dec 10 23:53:06 crc kubenswrapper[4791]: I1210 23:53:06.421298 4791 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/certified-operators-cmz9s" Dec 10 23:53:06 crc kubenswrapper[4791]: I1210 23:53:06.505539 4791 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-cmz9s"] Dec 10 23:53:06 crc kubenswrapper[4791]: I1210 23:53:06.599863 4791 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-658k2"] Dec 10 23:53:06 crc kubenswrapper[4791]: I1210 23:53:06.600157 4791 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/certified-operators-658k2" podUID="3c0bf306-a118-4131-9888-17cb46b832ac" containerName="registry-server" containerID="cri-o://e125b0c8405a37578f72d03da423f5dd68e703bcae56bf123ed904bf0a200b0c" gracePeriod=2 Dec 10 23:53:07 crc kubenswrapper[4791]: I1210 23:53:07.147862 4791 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-658k2" Dec 10 23:53:07 crc kubenswrapper[4791]: I1210 23:53:07.173282 4791 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/3c0bf306-a118-4131-9888-17cb46b832ac-catalog-content\") pod \"3c0bf306-a118-4131-9888-17cb46b832ac\" (UID: \"3c0bf306-a118-4131-9888-17cb46b832ac\") " Dec 10 23:53:07 crc kubenswrapper[4791]: I1210 23:53:07.173680 4791 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/3c0bf306-a118-4131-9888-17cb46b832ac-utilities\") pod \"3c0bf306-a118-4131-9888-17cb46b832ac\" (UID: \"3c0bf306-a118-4131-9888-17cb46b832ac\") " Dec 10 23:53:07 crc kubenswrapper[4791]: I1210 23:53:07.173733 4791 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-6rlqj\" (UniqueName: \"kubernetes.io/projected/3c0bf306-a118-4131-9888-17cb46b832ac-kube-api-access-6rlqj\") pod \"3c0bf306-a118-4131-9888-17cb46b832ac\" (UID: \"3c0bf306-a118-4131-9888-17cb46b832ac\") " Dec 10 23:53:07 crc kubenswrapper[4791]: I1210 23:53:07.174898 4791 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/3c0bf306-a118-4131-9888-17cb46b832ac-utilities" (OuterVolumeSpecName: "utilities") pod "3c0bf306-a118-4131-9888-17cb46b832ac" (UID: "3c0bf306-a118-4131-9888-17cb46b832ac"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 10 23:53:07 crc kubenswrapper[4791]: I1210 23:53:07.179989 4791 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/3c0bf306-a118-4131-9888-17cb46b832ac-kube-api-access-6rlqj" (OuterVolumeSpecName: "kube-api-access-6rlqj") pod "3c0bf306-a118-4131-9888-17cb46b832ac" (UID: "3c0bf306-a118-4131-9888-17cb46b832ac"). InnerVolumeSpecName "kube-api-access-6rlqj". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 10 23:53:07 crc kubenswrapper[4791]: I1210 23:53:07.226159 4791 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/3c0bf306-a118-4131-9888-17cb46b832ac-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "3c0bf306-a118-4131-9888-17cb46b832ac" (UID: "3c0bf306-a118-4131-9888-17cb46b832ac"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 10 23:53:07 crc kubenswrapper[4791]: I1210 23:53:07.275878 4791 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/3c0bf306-a118-4131-9888-17cb46b832ac-utilities\") on node \"crc\" DevicePath \"\"" Dec 10 23:53:07 crc kubenswrapper[4791]: I1210 23:53:07.275910 4791 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-6rlqj\" (UniqueName: \"kubernetes.io/projected/3c0bf306-a118-4131-9888-17cb46b832ac-kube-api-access-6rlqj\") on node \"crc\" DevicePath \"\"" Dec 10 23:53:07 crc kubenswrapper[4791]: I1210 23:53:07.275921 4791 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/3c0bf306-a118-4131-9888-17cb46b832ac-catalog-content\") on node \"crc\" DevicePath \"\"" Dec 10 23:53:07 crc kubenswrapper[4791]: I1210 23:53:07.661425 4791 generic.go:334] "Generic (PLEG): container finished" podID="3c0bf306-a118-4131-9888-17cb46b832ac" containerID="e125b0c8405a37578f72d03da423f5dd68e703bcae56bf123ed904bf0a200b0c" exitCode=0 Dec 10 23:53:07 crc kubenswrapper[4791]: I1210 23:53:07.661486 4791 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-658k2" event={"ID":"3c0bf306-a118-4131-9888-17cb46b832ac","Type":"ContainerDied","Data":"e125b0c8405a37578f72d03da423f5dd68e703bcae56bf123ed904bf0a200b0c"} Dec 10 23:53:07 crc kubenswrapper[4791]: I1210 23:53:07.661516 4791 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-658k2" Dec 10 23:53:07 crc kubenswrapper[4791]: I1210 23:53:07.661834 4791 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-658k2" event={"ID":"3c0bf306-a118-4131-9888-17cb46b832ac","Type":"ContainerDied","Data":"e29ecbe1e6a8f8b33f8c56e642091c475c866ea90221ba1783aa6774e56e4a3f"} Dec 10 23:53:07 crc kubenswrapper[4791]: I1210 23:53:07.661865 4791 scope.go:117] "RemoveContainer" containerID="e125b0c8405a37578f72d03da423f5dd68e703bcae56bf123ed904bf0a200b0c" Dec 10 23:53:07 crc kubenswrapper[4791]: I1210 23:53:07.696005 4791 scope.go:117] "RemoveContainer" containerID="c0a0a94f1f423d22c3b351631f5690046b276000a4a13392f21a648fb53856a2" Dec 10 23:53:07 crc kubenswrapper[4791]: I1210 23:53:07.716981 4791 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-658k2"] Dec 10 23:53:07 crc kubenswrapper[4791]: I1210 23:53:07.726266 4791 scope.go:117] "RemoveContainer" containerID="d69cecbe0cf9ac8005b5fa04c7ec40cdfdc06981ffa2e667ffb1e36e6954161a" Dec 10 23:53:07 crc kubenswrapper[4791]: I1210 23:53:07.732030 4791 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/certified-operators-658k2"] Dec 10 23:53:07 crc kubenswrapper[4791]: I1210 23:53:07.783996 4791 scope.go:117] "RemoveContainer" containerID="e125b0c8405a37578f72d03da423f5dd68e703bcae56bf123ed904bf0a200b0c" Dec 10 23:53:07 crc kubenswrapper[4791]: E1210 23:53:07.784470 4791 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"e125b0c8405a37578f72d03da423f5dd68e703bcae56bf123ed904bf0a200b0c\": container with ID starting with e125b0c8405a37578f72d03da423f5dd68e703bcae56bf123ed904bf0a200b0c not found: ID does not exist" containerID="e125b0c8405a37578f72d03da423f5dd68e703bcae56bf123ed904bf0a200b0c" Dec 10 23:53:07 crc kubenswrapper[4791]: I1210 23:53:07.784520 4791 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"e125b0c8405a37578f72d03da423f5dd68e703bcae56bf123ed904bf0a200b0c"} err="failed to get container status \"e125b0c8405a37578f72d03da423f5dd68e703bcae56bf123ed904bf0a200b0c\": rpc error: code = NotFound desc = could not find container \"e125b0c8405a37578f72d03da423f5dd68e703bcae56bf123ed904bf0a200b0c\": container with ID starting with e125b0c8405a37578f72d03da423f5dd68e703bcae56bf123ed904bf0a200b0c not found: ID does not exist" Dec 10 23:53:07 crc kubenswrapper[4791]: I1210 23:53:07.784550 4791 scope.go:117] "RemoveContainer" containerID="c0a0a94f1f423d22c3b351631f5690046b276000a4a13392f21a648fb53856a2" Dec 10 23:53:07 crc kubenswrapper[4791]: E1210 23:53:07.784822 4791 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"c0a0a94f1f423d22c3b351631f5690046b276000a4a13392f21a648fb53856a2\": container with ID starting with c0a0a94f1f423d22c3b351631f5690046b276000a4a13392f21a648fb53856a2 not found: ID does not exist" containerID="c0a0a94f1f423d22c3b351631f5690046b276000a4a13392f21a648fb53856a2" Dec 10 23:53:07 crc kubenswrapper[4791]: I1210 23:53:07.784846 4791 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"c0a0a94f1f423d22c3b351631f5690046b276000a4a13392f21a648fb53856a2"} err="failed to get container status \"c0a0a94f1f423d22c3b351631f5690046b276000a4a13392f21a648fb53856a2\": rpc error: code = NotFound desc = could not find container \"c0a0a94f1f423d22c3b351631f5690046b276000a4a13392f21a648fb53856a2\": container with ID starting with c0a0a94f1f423d22c3b351631f5690046b276000a4a13392f21a648fb53856a2 not found: ID does not exist" Dec 10 23:53:07 crc kubenswrapper[4791]: I1210 23:53:07.784860 4791 scope.go:117] "RemoveContainer" containerID="d69cecbe0cf9ac8005b5fa04c7ec40cdfdc06981ffa2e667ffb1e36e6954161a" Dec 10 23:53:07 crc kubenswrapper[4791]: E1210 23:53:07.785146 4791 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"d69cecbe0cf9ac8005b5fa04c7ec40cdfdc06981ffa2e667ffb1e36e6954161a\": container with ID starting with d69cecbe0cf9ac8005b5fa04c7ec40cdfdc06981ffa2e667ffb1e36e6954161a not found: ID does not exist" containerID="d69cecbe0cf9ac8005b5fa04c7ec40cdfdc06981ffa2e667ffb1e36e6954161a" Dec 10 23:53:07 crc kubenswrapper[4791]: I1210 23:53:07.785207 4791 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"d69cecbe0cf9ac8005b5fa04c7ec40cdfdc06981ffa2e667ffb1e36e6954161a"} err="failed to get container status \"d69cecbe0cf9ac8005b5fa04c7ec40cdfdc06981ffa2e667ffb1e36e6954161a\": rpc error: code = NotFound desc = could not find container \"d69cecbe0cf9ac8005b5fa04c7ec40cdfdc06981ffa2e667ffb1e36e6954161a\": container with ID starting with d69cecbe0cf9ac8005b5fa04c7ec40cdfdc06981ffa2e667ffb1e36e6954161a not found: ID does not exist" Dec 10 23:53:07 crc kubenswrapper[4791]: I1210 23:53:07.895664 4791 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="3c0bf306-a118-4131-9888-17cb46b832ac" path="/var/lib/kubelet/pods/3c0bf306-a118-4131-9888-17cb46b832ac/volumes" Dec 10 23:53:12 crc kubenswrapper[4791]: I1210 23:53:12.885814 4791 scope.go:117] "RemoveContainer" containerID="f44d2e51d8375135a5705f7cf71dde0bd0c6901d20f892b1346dc2eac144cbc0" Dec 10 23:53:12 crc kubenswrapper[4791]: E1210 23:53:12.886680 4791 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-5rb5l_openshift-machine-config-operator(ba35653c-6e06-4cee-a4d6-137764090d18)\"" pod="openshift-machine-config-operator/machine-config-daemon-5rb5l" podUID="ba35653c-6e06-4cee-a4d6-137764090d18" Dec 10 23:53:23 crc kubenswrapper[4791]: I1210 23:53:23.885024 4791 scope.go:117] "RemoveContainer" containerID="f44d2e51d8375135a5705f7cf71dde0bd0c6901d20f892b1346dc2eac144cbc0" Dec 10 23:53:23 crc kubenswrapper[4791]: E1210 23:53:23.886061 4791 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-5rb5l_openshift-machine-config-operator(ba35653c-6e06-4cee-a4d6-137764090d18)\"" pod="openshift-machine-config-operator/machine-config-daemon-5rb5l" podUID="ba35653c-6e06-4cee-a4d6-137764090d18" Dec 10 23:53:37 crc kubenswrapper[4791]: I1210 23:53:37.885673 4791 scope.go:117] "RemoveContainer" containerID="f44d2e51d8375135a5705f7cf71dde0bd0c6901d20f892b1346dc2eac144cbc0" Dec 10 23:53:37 crc kubenswrapper[4791]: E1210 23:53:37.886842 4791 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-5rb5l_openshift-machine-config-operator(ba35653c-6e06-4cee-a4d6-137764090d18)\"" pod="openshift-machine-config-operator/machine-config-daemon-5rb5l" podUID="ba35653c-6e06-4cee-a4d6-137764090d18" Dec 10 23:53:50 crc kubenswrapper[4791]: I1210 23:53:50.885159 4791 scope.go:117] "RemoveContainer" containerID="f44d2e51d8375135a5705f7cf71dde0bd0c6901d20f892b1346dc2eac144cbc0" Dec 10 23:53:50 crc kubenswrapper[4791]: E1210 23:53:50.886272 4791 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-5rb5l_openshift-machine-config-operator(ba35653c-6e06-4cee-a4d6-137764090d18)\"" pod="openshift-machine-config-operator/machine-config-daemon-5rb5l" podUID="ba35653c-6e06-4cee-a4d6-137764090d18" Dec 10 23:53:54 crc kubenswrapper[4791]: I1210 23:53:54.978364 4791 scope.go:117] "RemoveContainer" containerID="706ee31dc0ed7f4f1c4e136cb4a5a2a8f0f4228c2e114dd3b83a3906e1db6991" Dec 10 23:53:55 crc kubenswrapper[4791]: I1210 23:53:55.012696 4791 scope.go:117] "RemoveContainer" containerID="3f8fb9ccc563247b2f211de395f436764efe1eba5da3c41369225bf5feb95308" Dec 10 23:54:01 crc kubenswrapper[4791]: I1210 23:54:01.884736 4791 scope.go:117] "RemoveContainer" containerID="f44d2e51d8375135a5705f7cf71dde0bd0c6901d20f892b1346dc2eac144cbc0" Dec 10 23:54:01 crc kubenswrapper[4791]: E1210 23:54:01.885467 4791 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-5rb5l_openshift-machine-config-operator(ba35653c-6e06-4cee-a4d6-137764090d18)\"" pod="openshift-machine-config-operator/machine-config-daemon-5rb5l" podUID="ba35653c-6e06-4cee-a4d6-137764090d18" Dec 10 23:54:12 crc kubenswrapper[4791]: I1210 23:54:12.885477 4791 scope.go:117] "RemoveContainer" containerID="f44d2e51d8375135a5705f7cf71dde0bd0c6901d20f892b1346dc2eac144cbc0" Dec 10 23:54:12 crc kubenswrapper[4791]: E1210 23:54:12.886494 4791 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-5rb5l_openshift-machine-config-operator(ba35653c-6e06-4cee-a4d6-137764090d18)\"" pod="openshift-machine-config-operator/machine-config-daemon-5rb5l" podUID="ba35653c-6e06-4cee-a4d6-137764090d18" Dec 10 23:54:26 crc kubenswrapper[4791]: I1210 23:54:26.886173 4791 scope.go:117] "RemoveContainer" containerID="f44d2e51d8375135a5705f7cf71dde0bd0c6901d20f892b1346dc2eac144cbc0" Dec 10 23:54:26 crc kubenswrapper[4791]: E1210 23:54:26.887860 4791 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-5rb5l_openshift-machine-config-operator(ba35653c-6e06-4cee-a4d6-137764090d18)\"" pod="openshift-machine-config-operator/machine-config-daemon-5rb5l" podUID="ba35653c-6e06-4cee-a4d6-137764090d18" Dec 10 23:54:40 crc kubenswrapper[4791]: I1210 23:54:40.885083 4791 scope.go:117] "RemoveContainer" containerID="f44d2e51d8375135a5705f7cf71dde0bd0c6901d20f892b1346dc2eac144cbc0" Dec 10 23:54:40 crc kubenswrapper[4791]: E1210 23:54:40.885770 4791 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-5rb5l_openshift-machine-config-operator(ba35653c-6e06-4cee-a4d6-137764090d18)\"" pod="openshift-machine-config-operator/machine-config-daemon-5rb5l" podUID="ba35653c-6e06-4cee-a4d6-137764090d18" Dec 10 23:54:53 crc kubenswrapper[4791]: I1210 23:54:53.885127 4791 scope.go:117] "RemoveContainer" containerID="f44d2e51d8375135a5705f7cf71dde0bd0c6901d20f892b1346dc2eac144cbc0" Dec 10 23:54:53 crc kubenswrapper[4791]: E1210 23:54:53.886060 4791 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-5rb5l_openshift-machine-config-operator(ba35653c-6e06-4cee-a4d6-137764090d18)\"" pod="openshift-machine-config-operator/machine-config-daemon-5rb5l" podUID="ba35653c-6e06-4cee-a4d6-137764090d18" Dec 10 23:54:55 crc kubenswrapper[4791]: I1210 23:54:55.139881 4791 scope.go:117] "RemoveContainer" containerID="8e71c17af47465b1c6ac87a038183dd522ea5fb166a468e84343ee6f185d31de" Dec 10 23:54:55 crc kubenswrapper[4791]: I1210 23:54:55.161182 4791 scope.go:117] "RemoveContainer" containerID="5ba1ecb0bd4b0571936b904f67fa42236a90b08a68f51337c69397ae126964dd" Dec 10 23:55:05 crc kubenswrapper[4791]: I1210 23:55:05.905663 4791 scope.go:117] "RemoveContainer" containerID="f44d2e51d8375135a5705f7cf71dde0bd0c6901d20f892b1346dc2eac144cbc0" Dec 10 23:55:05 crc kubenswrapper[4791]: E1210 23:55:05.906918 4791 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-5rb5l_openshift-machine-config-operator(ba35653c-6e06-4cee-a4d6-137764090d18)\"" pod="openshift-machine-config-operator/machine-config-daemon-5rb5l" podUID="ba35653c-6e06-4cee-a4d6-137764090d18" Dec 10 23:55:18 crc kubenswrapper[4791]: I1210 23:55:18.884988 4791 scope.go:117] "RemoveContainer" containerID="f44d2e51d8375135a5705f7cf71dde0bd0c6901d20f892b1346dc2eac144cbc0" Dec 10 23:55:18 crc kubenswrapper[4791]: E1210 23:55:18.886152 4791 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-5rb5l_openshift-machine-config-operator(ba35653c-6e06-4cee-a4d6-137764090d18)\"" pod="openshift-machine-config-operator/machine-config-daemon-5rb5l" podUID="ba35653c-6e06-4cee-a4d6-137764090d18" Dec 10 23:55:33 crc kubenswrapper[4791]: I1210 23:55:33.885056 4791 scope.go:117] "RemoveContainer" containerID="f44d2e51d8375135a5705f7cf71dde0bd0c6901d20f892b1346dc2eac144cbc0" Dec 10 23:55:34 crc kubenswrapper[4791]: I1210 23:55:34.321576 4791 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-5rb5l" event={"ID":"ba35653c-6e06-4cee-a4d6-137764090d18","Type":"ContainerStarted","Data":"1a35e34c3589dfe3497551cd10aa8a1e1fa1f05668706b1469b66da53eca8dc3"} Dec 10 23:55:46 crc kubenswrapper[4791]: I1210 23:55:46.572508 4791 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-must-gather-2dlqt/must-gather-rqrn8"] Dec 10 23:55:46 crc kubenswrapper[4791]: E1210 23:55:46.573384 4791 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="3c0bf306-a118-4131-9888-17cb46b832ac" containerName="extract-content" Dec 10 23:55:46 crc kubenswrapper[4791]: I1210 23:55:46.573397 4791 state_mem.go:107] "Deleted CPUSet assignment" podUID="3c0bf306-a118-4131-9888-17cb46b832ac" containerName="extract-content" Dec 10 23:55:46 crc kubenswrapper[4791]: E1210 23:55:46.573424 4791 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="3c0bf306-a118-4131-9888-17cb46b832ac" containerName="registry-server" Dec 10 23:55:46 crc kubenswrapper[4791]: I1210 23:55:46.573430 4791 state_mem.go:107] "Deleted CPUSet assignment" podUID="3c0bf306-a118-4131-9888-17cb46b832ac" containerName="registry-server" Dec 10 23:55:46 crc kubenswrapper[4791]: E1210 23:55:46.573441 4791 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="2171ff1e-63b0-41ff-8019-5a72a0c59f5f" containerName="copy" Dec 10 23:55:46 crc kubenswrapper[4791]: I1210 23:55:46.573447 4791 state_mem.go:107] "Deleted CPUSet assignment" podUID="2171ff1e-63b0-41ff-8019-5a72a0c59f5f" containerName="copy" Dec 10 23:55:46 crc kubenswrapper[4791]: E1210 23:55:46.573460 4791 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="2171ff1e-63b0-41ff-8019-5a72a0c59f5f" containerName="gather" Dec 10 23:55:46 crc kubenswrapper[4791]: I1210 23:55:46.573466 4791 state_mem.go:107] "Deleted CPUSet assignment" podUID="2171ff1e-63b0-41ff-8019-5a72a0c59f5f" containerName="gather" Dec 10 23:55:46 crc kubenswrapper[4791]: E1210 23:55:46.573481 4791 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="3c0bf306-a118-4131-9888-17cb46b832ac" containerName="extract-utilities" Dec 10 23:55:46 crc kubenswrapper[4791]: I1210 23:55:46.573487 4791 state_mem.go:107] "Deleted CPUSet assignment" podUID="3c0bf306-a118-4131-9888-17cb46b832ac" containerName="extract-utilities" Dec 10 23:55:46 crc kubenswrapper[4791]: I1210 23:55:46.573669 4791 memory_manager.go:354] "RemoveStaleState removing state" podUID="2171ff1e-63b0-41ff-8019-5a72a0c59f5f" containerName="gather" Dec 10 23:55:46 crc kubenswrapper[4791]: I1210 23:55:46.573679 4791 memory_manager.go:354] "RemoveStaleState removing state" podUID="2171ff1e-63b0-41ff-8019-5a72a0c59f5f" containerName="copy" Dec 10 23:55:46 crc kubenswrapper[4791]: I1210 23:55:46.573693 4791 memory_manager.go:354] "RemoveStaleState removing state" podUID="3c0bf306-a118-4131-9888-17cb46b832ac" containerName="registry-server" Dec 10 23:55:46 crc kubenswrapper[4791]: I1210 23:55:46.574679 4791 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-2dlqt/must-gather-rqrn8" Dec 10 23:55:46 crc kubenswrapper[4791]: I1210 23:55:46.577370 4791 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-must-gather-2dlqt"/"kube-root-ca.crt" Dec 10 23:55:46 crc kubenswrapper[4791]: I1210 23:55:46.577628 4791 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-must-gather-2dlqt"/"openshift-service-ca.crt" Dec 10 23:55:46 crc kubenswrapper[4791]: I1210 23:55:46.577645 4791 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-must-gather-2dlqt"/"default-dockercfg-t4khc" Dec 10 23:55:46 crc kubenswrapper[4791]: I1210 23:55:46.582916 4791 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-must-gather-2dlqt/must-gather-rqrn8"] Dec 10 23:55:46 crc kubenswrapper[4791]: I1210 23:55:46.675831 4791 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-ppqw4\" (UniqueName: \"kubernetes.io/projected/6e53fefe-b61d-4ba3-a360-4ae9f1213d1c-kube-api-access-ppqw4\") pod \"must-gather-rqrn8\" (UID: \"6e53fefe-b61d-4ba3-a360-4ae9f1213d1c\") " pod="openshift-must-gather-2dlqt/must-gather-rqrn8" Dec 10 23:55:46 crc kubenswrapper[4791]: I1210 23:55:46.675901 4791 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"must-gather-output\" (UniqueName: \"kubernetes.io/empty-dir/6e53fefe-b61d-4ba3-a360-4ae9f1213d1c-must-gather-output\") pod \"must-gather-rqrn8\" (UID: \"6e53fefe-b61d-4ba3-a360-4ae9f1213d1c\") " pod="openshift-must-gather-2dlqt/must-gather-rqrn8" Dec 10 23:55:46 crc kubenswrapper[4791]: I1210 23:55:46.777516 4791 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-ppqw4\" (UniqueName: \"kubernetes.io/projected/6e53fefe-b61d-4ba3-a360-4ae9f1213d1c-kube-api-access-ppqw4\") pod \"must-gather-rqrn8\" (UID: \"6e53fefe-b61d-4ba3-a360-4ae9f1213d1c\") " pod="openshift-must-gather-2dlqt/must-gather-rqrn8" Dec 10 23:55:46 crc kubenswrapper[4791]: I1210 23:55:46.777619 4791 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"must-gather-output\" (UniqueName: \"kubernetes.io/empty-dir/6e53fefe-b61d-4ba3-a360-4ae9f1213d1c-must-gather-output\") pod \"must-gather-rqrn8\" (UID: \"6e53fefe-b61d-4ba3-a360-4ae9f1213d1c\") " pod="openshift-must-gather-2dlqt/must-gather-rqrn8" Dec 10 23:55:46 crc kubenswrapper[4791]: I1210 23:55:46.778237 4791 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"must-gather-output\" (UniqueName: \"kubernetes.io/empty-dir/6e53fefe-b61d-4ba3-a360-4ae9f1213d1c-must-gather-output\") pod \"must-gather-rqrn8\" (UID: \"6e53fefe-b61d-4ba3-a360-4ae9f1213d1c\") " pod="openshift-must-gather-2dlqt/must-gather-rqrn8" Dec 10 23:55:46 crc kubenswrapper[4791]: I1210 23:55:46.804020 4791 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-ppqw4\" (UniqueName: \"kubernetes.io/projected/6e53fefe-b61d-4ba3-a360-4ae9f1213d1c-kube-api-access-ppqw4\") pod \"must-gather-rqrn8\" (UID: \"6e53fefe-b61d-4ba3-a360-4ae9f1213d1c\") " pod="openshift-must-gather-2dlqt/must-gather-rqrn8" Dec 10 23:55:46 crc kubenswrapper[4791]: I1210 23:55:46.892550 4791 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-2dlqt/must-gather-rqrn8" Dec 10 23:55:47 crc kubenswrapper[4791]: I1210 23:55:47.424301 4791 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-must-gather-2dlqt/must-gather-rqrn8"] Dec 10 23:55:47 crc kubenswrapper[4791]: I1210 23:55:47.482387 4791 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-2dlqt/must-gather-rqrn8" event={"ID":"6e53fefe-b61d-4ba3-a360-4ae9f1213d1c","Type":"ContainerStarted","Data":"644f8eee8f2a708c979e53548c8dcace32181b66d62fba863a7db23b9b3d34fb"} Dec 10 23:55:48 crc kubenswrapper[4791]: I1210 23:55:48.496720 4791 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-2dlqt/must-gather-rqrn8" event={"ID":"6e53fefe-b61d-4ba3-a360-4ae9f1213d1c","Type":"ContainerStarted","Data":"7fc47c42954af916ee253f740108cdb8db79306d0712fc787fe2aa1fe0310e3d"} Dec 10 23:55:48 crc kubenswrapper[4791]: I1210 23:55:48.496764 4791 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-2dlqt/must-gather-rqrn8" event={"ID":"6e53fefe-b61d-4ba3-a360-4ae9f1213d1c","Type":"ContainerStarted","Data":"fd094fcb02987fc8ae45b115d82c5662af63b3157267e481ca35913a767e24aa"} Dec 10 23:55:48 crc kubenswrapper[4791]: I1210 23:55:48.524676 4791 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-must-gather-2dlqt/must-gather-rqrn8" podStartSLOduration=2.524656997 podStartE2EDuration="2.524656997s" podCreationTimestamp="2025-12-10 23:55:46 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-10 23:55:48.516585149 +0000 UTC m=+3982.946202772" watchObservedRunningTime="2025-12-10 23:55:48.524656997 +0000 UTC m=+3982.954274610" Dec 10 23:55:51 crc kubenswrapper[4791]: I1210 23:55:51.573083 4791 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-must-gather-2dlqt/crc-debug-rkmw6"] Dec 10 23:55:51 crc kubenswrapper[4791]: I1210 23:55:51.576265 4791 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-2dlqt/crc-debug-rkmw6" Dec 10 23:55:51 crc kubenswrapper[4791]: I1210 23:55:51.750707 4791 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host\" (UniqueName: \"kubernetes.io/host-path/948ffcf1-1737-45d6-b33c-f0e05b1dbc88-host\") pod \"crc-debug-rkmw6\" (UID: \"948ffcf1-1737-45d6-b33c-f0e05b1dbc88\") " pod="openshift-must-gather-2dlqt/crc-debug-rkmw6" Dec 10 23:55:51 crc kubenswrapper[4791]: I1210 23:55:51.750966 4791 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-tvzdm\" (UniqueName: \"kubernetes.io/projected/948ffcf1-1737-45d6-b33c-f0e05b1dbc88-kube-api-access-tvzdm\") pod \"crc-debug-rkmw6\" (UID: \"948ffcf1-1737-45d6-b33c-f0e05b1dbc88\") " pod="openshift-must-gather-2dlqt/crc-debug-rkmw6" Dec 10 23:55:51 crc kubenswrapper[4791]: I1210 23:55:51.852672 4791 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-tvzdm\" (UniqueName: \"kubernetes.io/projected/948ffcf1-1737-45d6-b33c-f0e05b1dbc88-kube-api-access-tvzdm\") pod \"crc-debug-rkmw6\" (UID: \"948ffcf1-1737-45d6-b33c-f0e05b1dbc88\") " pod="openshift-must-gather-2dlqt/crc-debug-rkmw6" Dec 10 23:55:51 crc kubenswrapper[4791]: I1210 23:55:51.852842 4791 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host\" (UniqueName: \"kubernetes.io/host-path/948ffcf1-1737-45d6-b33c-f0e05b1dbc88-host\") pod \"crc-debug-rkmw6\" (UID: \"948ffcf1-1737-45d6-b33c-f0e05b1dbc88\") " pod="openshift-must-gather-2dlqt/crc-debug-rkmw6" Dec 10 23:55:51 crc kubenswrapper[4791]: I1210 23:55:51.852954 4791 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host\" (UniqueName: \"kubernetes.io/host-path/948ffcf1-1737-45d6-b33c-f0e05b1dbc88-host\") pod \"crc-debug-rkmw6\" (UID: \"948ffcf1-1737-45d6-b33c-f0e05b1dbc88\") " pod="openshift-must-gather-2dlqt/crc-debug-rkmw6" Dec 10 23:55:51 crc kubenswrapper[4791]: I1210 23:55:51.875028 4791 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-tvzdm\" (UniqueName: \"kubernetes.io/projected/948ffcf1-1737-45d6-b33c-f0e05b1dbc88-kube-api-access-tvzdm\") pod \"crc-debug-rkmw6\" (UID: \"948ffcf1-1737-45d6-b33c-f0e05b1dbc88\") " pod="openshift-must-gather-2dlqt/crc-debug-rkmw6" Dec 10 23:55:51 crc kubenswrapper[4791]: I1210 23:55:51.893929 4791 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-2dlqt/crc-debug-rkmw6" Dec 10 23:55:51 crc kubenswrapper[4791]: W1210 23:55:51.923023 4791 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod948ffcf1_1737_45d6_b33c_f0e05b1dbc88.slice/crio-65398cb617403c78f16a52cac672eb3260f02fddec9ad166540e577ecc9356c9 WatchSource:0}: Error finding container 65398cb617403c78f16a52cac672eb3260f02fddec9ad166540e577ecc9356c9: Status 404 returned error can't find the container with id 65398cb617403c78f16a52cac672eb3260f02fddec9ad166540e577ecc9356c9 Dec 10 23:55:52 crc kubenswrapper[4791]: I1210 23:55:52.565072 4791 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-2dlqt/crc-debug-rkmw6" event={"ID":"948ffcf1-1737-45d6-b33c-f0e05b1dbc88","Type":"ContainerStarted","Data":"0db4c4df2ef5868f5f3f1324b3e21890206443bb723d5f2324df65ffbad1aef6"} Dec 10 23:55:52 crc kubenswrapper[4791]: I1210 23:55:52.565653 4791 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-2dlqt/crc-debug-rkmw6" event={"ID":"948ffcf1-1737-45d6-b33c-f0e05b1dbc88","Type":"ContainerStarted","Data":"65398cb617403c78f16a52cac672eb3260f02fddec9ad166540e577ecc9356c9"} Dec 10 23:55:52 crc kubenswrapper[4791]: I1210 23:55:52.583866 4791 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-must-gather-2dlqt/crc-debug-rkmw6" podStartSLOduration=1.5838487479999999 podStartE2EDuration="1.583848748s" podCreationTimestamp="2025-12-10 23:55:51 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-10 23:55:52.578781575 +0000 UTC m=+3987.008399188" watchObservedRunningTime="2025-12-10 23:55:52.583848748 +0000 UTC m=+3987.013466361" Dec 10 23:56:28 crc kubenswrapper[4791]: I1210 23:56:28.155375 4791 generic.go:334] "Generic (PLEG): container finished" podID="948ffcf1-1737-45d6-b33c-f0e05b1dbc88" containerID="0db4c4df2ef5868f5f3f1324b3e21890206443bb723d5f2324df65ffbad1aef6" exitCode=0 Dec 10 23:56:28 crc kubenswrapper[4791]: I1210 23:56:28.155915 4791 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-2dlqt/crc-debug-rkmw6" event={"ID":"948ffcf1-1737-45d6-b33c-f0e05b1dbc88","Type":"ContainerDied","Data":"0db4c4df2ef5868f5f3f1324b3e21890206443bb723d5f2324df65ffbad1aef6"} Dec 10 23:56:29 crc kubenswrapper[4791]: I1210 23:56:29.279907 4791 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-2dlqt/crc-debug-rkmw6" Dec 10 23:56:29 crc kubenswrapper[4791]: I1210 23:56:29.314880 4791 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-must-gather-2dlqt/crc-debug-rkmw6"] Dec 10 23:56:29 crc kubenswrapper[4791]: I1210 23:56:29.319680 4791 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-tvzdm\" (UniqueName: \"kubernetes.io/projected/948ffcf1-1737-45d6-b33c-f0e05b1dbc88-kube-api-access-tvzdm\") pod \"948ffcf1-1737-45d6-b33c-f0e05b1dbc88\" (UID: \"948ffcf1-1737-45d6-b33c-f0e05b1dbc88\") " Dec 10 23:56:29 crc kubenswrapper[4791]: I1210 23:56:29.319763 4791 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"host\" (UniqueName: \"kubernetes.io/host-path/948ffcf1-1737-45d6-b33c-f0e05b1dbc88-host\") pod \"948ffcf1-1737-45d6-b33c-f0e05b1dbc88\" (UID: \"948ffcf1-1737-45d6-b33c-f0e05b1dbc88\") " Dec 10 23:56:29 crc kubenswrapper[4791]: I1210 23:56:29.319838 4791 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/948ffcf1-1737-45d6-b33c-f0e05b1dbc88-host" (OuterVolumeSpecName: "host") pod "948ffcf1-1737-45d6-b33c-f0e05b1dbc88" (UID: "948ffcf1-1737-45d6-b33c-f0e05b1dbc88"). InnerVolumeSpecName "host". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 10 23:56:29 crc kubenswrapper[4791]: I1210 23:56:29.320482 4791 reconciler_common.go:293] "Volume detached for volume \"host\" (UniqueName: \"kubernetes.io/host-path/948ffcf1-1737-45d6-b33c-f0e05b1dbc88-host\") on node \"crc\" DevicePath \"\"" Dec 10 23:56:29 crc kubenswrapper[4791]: I1210 23:56:29.324148 4791 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-must-gather-2dlqt/crc-debug-rkmw6"] Dec 10 23:56:29 crc kubenswrapper[4791]: I1210 23:56:29.325493 4791 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/948ffcf1-1737-45d6-b33c-f0e05b1dbc88-kube-api-access-tvzdm" (OuterVolumeSpecName: "kube-api-access-tvzdm") pod "948ffcf1-1737-45d6-b33c-f0e05b1dbc88" (UID: "948ffcf1-1737-45d6-b33c-f0e05b1dbc88"). InnerVolumeSpecName "kube-api-access-tvzdm". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 10 23:56:29 crc kubenswrapper[4791]: I1210 23:56:29.422252 4791 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-tvzdm\" (UniqueName: \"kubernetes.io/projected/948ffcf1-1737-45d6-b33c-f0e05b1dbc88-kube-api-access-tvzdm\") on node \"crc\" DevicePath \"\"" Dec 10 23:56:29 crc kubenswrapper[4791]: I1210 23:56:29.897198 4791 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="948ffcf1-1737-45d6-b33c-f0e05b1dbc88" path="/var/lib/kubelet/pods/948ffcf1-1737-45d6-b33c-f0e05b1dbc88/volumes" Dec 10 23:56:30 crc kubenswrapper[4791]: I1210 23:56:30.177462 4791 scope.go:117] "RemoveContainer" containerID="0db4c4df2ef5868f5f3f1324b3e21890206443bb723d5f2324df65ffbad1aef6" Dec 10 23:56:30 crc kubenswrapper[4791]: I1210 23:56:30.177884 4791 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-2dlqt/crc-debug-rkmw6" Dec 10 23:56:30 crc kubenswrapper[4791]: I1210 23:56:30.551105 4791 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-must-gather-2dlqt/crc-debug-mf9wt"] Dec 10 23:56:30 crc kubenswrapper[4791]: E1210 23:56:30.551621 4791 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="948ffcf1-1737-45d6-b33c-f0e05b1dbc88" containerName="container-00" Dec 10 23:56:30 crc kubenswrapper[4791]: I1210 23:56:30.551639 4791 state_mem.go:107] "Deleted CPUSet assignment" podUID="948ffcf1-1737-45d6-b33c-f0e05b1dbc88" containerName="container-00" Dec 10 23:56:30 crc kubenswrapper[4791]: I1210 23:56:30.551866 4791 memory_manager.go:354] "RemoveStaleState removing state" podUID="948ffcf1-1737-45d6-b33c-f0e05b1dbc88" containerName="container-00" Dec 10 23:56:30 crc kubenswrapper[4791]: I1210 23:56:30.552617 4791 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-2dlqt/crc-debug-mf9wt" Dec 10 23:56:30 crc kubenswrapper[4791]: I1210 23:56:30.641642 4791 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-mzhgr\" (UniqueName: \"kubernetes.io/projected/39b3d5f9-f8f2-426d-aa07-62aae5324c12-kube-api-access-mzhgr\") pod \"crc-debug-mf9wt\" (UID: \"39b3d5f9-f8f2-426d-aa07-62aae5324c12\") " pod="openshift-must-gather-2dlqt/crc-debug-mf9wt" Dec 10 23:56:30 crc kubenswrapper[4791]: I1210 23:56:30.642393 4791 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host\" (UniqueName: \"kubernetes.io/host-path/39b3d5f9-f8f2-426d-aa07-62aae5324c12-host\") pod \"crc-debug-mf9wt\" (UID: \"39b3d5f9-f8f2-426d-aa07-62aae5324c12\") " pod="openshift-must-gather-2dlqt/crc-debug-mf9wt" Dec 10 23:56:30 crc kubenswrapper[4791]: I1210 23:56:30.743718 4791 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-mzhgr\" (UniqueName: \"kubernetes.io/projected/39b3d5f9-f8f2-426d-aa07-62aae5324c12-kube-api-access-mzhgr\") pod \"crc-debug-mf9wt\" (UID: \"39b3d5f9-f8f2-426d-aa07-62aae5324c12\") " pod="openshift-must-gather-2dlqt/crc-debug-mf9wt" Dec 10 23:56:30 crc kubenswrapper[4791]: I1210 23:56:30.743816 4791 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host\" (UniqueName: \"kubernetes.io/host-path/39b3d5f9-f8f2-426d-aa07-62aae5324c12-host\") pod \"crc-debug-mf9wt\" (UID: \"39b3d5f9-f8f2-426d-aa07-62aae5324c12\") " pod="openshift-must-gather-2dlqt/crc-debug-mf9wt" Dec 10 23:56:30 crc kubenswrapper[4791]: I1210 23:56:30.743947 4791 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host\" (UniqueName: \"kubernetes.io/host-path/39b3d5f9-f8f2-426d-aa07-62aae5324c12-host\") pod \"crc-debug-mf9wt\" (UID: \"39b3d5f9-f8f2-426d-aa07-62aae5324c12\") " pod="openshift-must-gather-2dlqt/crc-debug-mf9wt" Dec 10 23:56:30 crc kubenswrapper[4791]: I1210 23:56:30.762040 4791 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-mzhgr\" (UniqueName: \"kubernetes.io/projected/39b3d5f9-f8f2-426d-aa07-62aae5324c12-kube-api-access-mzhgr\") pod \"crc-debug-mf9wt\" (UID: \"39b3d5f9-f8f2-426d-aa07-62aae5324c12\") " pod="openshift-must-gather-2dlqt/crc-debug-mf9wt" Dec 10 23:56:30 crc kubenswrapper[4791]: I1210 23:56:30.868266 4791 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-2dlqt/crc-debug-mf9wt" Dec 10 23:56:30 crc kubenswrapper[4791]: W1210 23:56:30.900109 4791 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod39b3d5f9_f8f2_426d_aa07_62aae5324c12.slice/crio-e53213696e6bd2fc102e34ba952b09842f0d24b1e7618a21b5afea88fee58521 WatchSource:0}: Error finding container e53213696e6bd2fc102e34ba952b09842f0d24b1e7618a21b5afea88fee58521: Status 404 returned error can't find the container with id e53213696e6bd2fc102e34ba952b09842f0d24b1e7618a21b5afea88fee58521 Dec 10 23:56:31 crc kubenswrapper[4791]: I1210 23:56:31.190238 4791 generic.go:334] "Generic (PLEG): container finished" podID="39b3d5f9-f8f2-426d-aa07-62aae5324c12" containerID="781ea9bff8b6f3c87d6f6420a0444f591a70f1b39602f3275295098ea93239bb" exitCode=0 Dec 10 23:56:31 crc kubenswrapper[4791]: I1210 23:56:31.190284 4791 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-2dlqt/crc-debug-mf9wt" event={"ID":"39b3d5f9-f8f2-426d-aa07-62aae5324c12","Type":"ContainerDied","Data":"781ea9bff8b6f3c87d6f6420a0444f591a70f1b39602f3275295098ea93239bb"} Dec 10 23:56:31 crc kubenswrapper[4791]: I1210 23:56:31.190328 4791 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-2dlqt/crc-debug-mf9wt" event={"ID":"39b3d5f9-f8f2-426d-aa07-62aae5324c12","Type":"ContainerStarted","Data":"e53213696e6bd2fc102e34ba952b09842f0d24b1e7618a21b5afea88fee58521"} Dec 10 23:56:31 crc kubenswrapper[4791]: I1210 23:56:31.605211 4791 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-must-gather-2dlqt/crc-debug-mf9wt"] Dec 10 23:56:31 crc kubenswrapper[4791]: I1210 23:56:31.614322 4791 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-must-gather-2dlqt/crc-debug-mf9wt"] Dec 10 23:56:32 crc kubenswrapper[4791]: I1210 23:56:32.329046 4791 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-2dlqt/crc-debug-mf9wt" Dec 10 23:56:32 crc kubenswrapper[4791]: I1210 23:56:32.483466 4791 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"host\" (UniqueName: \"kubernetes.io/host-path/39b3d5f9-f8f2-426d-aa07-62aae5324c12-host\") pod \"39b3d5f9-f8f2-426d-aa07-62aae5324c12\" (UID: \"39b3d5f9-f8f2-426d-aa07-62aae5324c12\") " Dec 10 23:56:32 crc kubenswrapper[4791]: I1210 23:56:32.483631 4791 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/39b3d5f9-f8f2-426d-aa07-62aae5324c12-host" (OuterVolumeSpecName: "host") pod "39b3d5f9-f8f2-426d-aa07-62aae5324c12" (UID: "39b3d5f9-f8f2-426d-aa07-62aae5324c12"). InnerVolumeSpecName "host". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 10 23:56:32 crc kubenswrapper[4791]: I1210 23:56:32.484305 4791 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-mzhgr\" (UniqueName: \"kubernetes.io/projected/39b3d5f9-f8f2-426d-aa07-62aae5324c12-kube-api-access-mzhgr\") pod \"39b3d5f9-f8f2-426d-aa07-62aae5324c12\" (UID: \"39b3d5f9-f8f2-426d-aa07-62aae5324c12\") " Dec 10 23:56:32 crc kubenswrapper[4791]: I1210 23:56:32.484999 4791 reconciler_common.go:293] "Volume detached for volume \"host\" (UniqueName: \"kubernetes.io/host-path/39b3d5f9-f8f2-426d-aa07-62aae5324c12-host\") on node \"crc\" DevicePath \"\"" Dec 10 23:56:32 crc kubenswrapper[4791]: I1210 23:56:32.493365 4791 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/39b3d5f9-f8f2-426d-aa07-62aae5324c12-kube-api-access-mzhgr" (OuterVolumeSpecName: "kube-api-access-mzhgr") pod "39b3d5f9-f8f2-426d-aa07-62aae5324c12" (UID: "39b3d5f9-f8f2-426d-aa07-62aae5324c12"). InnerVolumeSpecName "kube-api-access-mzhgr". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 10 23:56:32 crc kubenswrapper[4791]: I1210 23:56:32.586373 4791 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-mzhgr\" (UniqueName: \"kubernetes.io/projected/39b3d5f9-f8f2-426d-aa07-62aae5324c12-kube-api-access-mzhgr\") on node \"crc\" DevicePath \"\"" Dec 10 23:56:32 crc kubenswrapper[4791]: I1210 23:56:32.828711 4791 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-must-gather-2dlqt/crc-debug-9zjhh"] Dec 10 23:56:32 crc kubenswrapper[4791]: E1210 23:56:32.829162 4791 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="39b3d5f9-f8f2-426d-aa07-62aae5324c12" containerName="container-00" Dec 10 23:56:32 crc kubenswrapper[4791]: I1210 23:56:32.829178 4791 state_mem.go:107] "Deleted CPUSet assignment" podUID="39b3d5f9-f8f2-426d-aa07-62aae5324c12" containerName="container-00" Dec 10 23:56:32 crc kubenswrapper[4791]: I1210 23:56:32.829396 4791 memory_manager.go:354] "RemoveStaleState removing state" podUID="39b3d5f9-f8f2-426d-aa07-62aae5324c12" containerName="container-00" Dec 10 23:56:32 crc kubenswrapper[4791]: I1210 23:56:32.830066 4791 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-2dlqt/crc-debug-9zjhh" Dec 10 23:56:32 crc kubenswrapper[4791]: I1210 23:56:32.994676 4791 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host\" (UniqueName: \"kubernetes.io/host-path/60fb43df-96e1-49b3-a19a-835c266f55f1-host\") pod \"crc-debug-9zjhh\" (UID: \"60fb43df-96e1-49b3-a19a-835c266f55f1\") " pod="openshift-must-gather-2dlqt/crc-debug-9zjhh" Dec 10 23:56:32 crc kubenswrapper[4791]: I1210 23:56:32.994975 4791 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-q67kc\" (UniqueName: \"kubernetes.io/projected/60fb43df-96e1-49b3-a19a-835c266f55f1-kube-api-access-q67kc\") pod \"crc-debug-9zjhh\" (UID: \"60fb43df-96e1-49b3-a19a-835c266f55f1\") " pod="openshift-must-gather-2dlqt/crc-debug-9zjhh" Dec 10 23:56:33 crc kubenswrapper[4791]: I1210 23:56:33.097750 4791 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host\" (UniqueName: \"kubernetes.io/host-path/60fb43df-96e1-49b3-a19a-835c266f55f1-host\") pod \"crc-debug-9zjhh\" (UID: \"60fb43df-96e1-49b3-a19a-835c266f55f1\") " pod="openshift-must-gather-2dlqt/crc-debug-9zjhh" Dec 10 23:56:33 crc kubenswrapper[4791]: I1210 23:56:33.097876 4791 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-q67kc\" (UniqueName: \"kubernetes.io/projected/60fb43df-96e1-49b3-a19a-835c266f55f1-kube-api-access-q67kc\") pod \"crc-debug-9zjhh\" (UID: \"60fb43df-96e1-49b3-a19a-835c266f55f1\") " pod="openshift-must-gather-2dlqt/crc-debug-9zjhh" Dec 10 23:56:33 crc kubenswrapper[4791]: I1210 23:56:33.098044 4791 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host\" (UniqueName: \"kubernetes.io/host-path/60fb43df-96e1-49b3-a19a-835c266f55f1-host\") pod \"crc-debug-9zjhh\" (UID: \"60fb43df-96e1-49b3-a19a-835c266f55f1\") " pod="openshift-must-gather-2dlqt/crc-debug-9zjhh" Dec 10 23:56:33 crc kubenswrapper[4791]: I1210 23:56:33.127202 4791 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-q67kc\" (UniqueName: \"kubernetes.io/projected/60fb43df-96e1-49b3-a19a-835c266f55f1-kube-api-access-q67kc\") pod \"crc-debug-9zjhh\" (UID: \"60fb43df-96e1-49b3-a19a-835c266f55f1\") " pod="openshift-must-gather-2dlqt/crc-debug-9zjhh" Dec 10 23:56:33 crc kubenswrapper[4791]: I1210 23:56:33.160618 4791 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-2dlqt/crc-debug-9zjhh" Dec 10 23:56:33 crc kubenswrapper[4791]: I1210 23:56:33.222049 4791 scope.go:117] "RemoveContainer" containerID="781ea9bff8b6f3c87d6f6420a0444f591a70f1b39602f3275295098ea93239bb" Dec 10 23:56:33 crc kubenswrapper[4791]: I1210 23:56:33.222048 4791 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-2dlqt/crc-debug-mf9wt" Dec 10 23:56:33 crc kubenswrapper[4791]: I1210 23:56:33.224510 4791 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-2dlqt/crc-debug-9zjhh" event={"ID":"60fb43df-96e1-49b3-a19a-835c266f55f1","Type":"ContainerStarted","Data":"2051288f3801f45d5f71ffea696c6e237a3e6e1aa35b256cd9f9625f23f9b812"} Dec 10 23:56:33 crc kubenswrapper[4791]: I1210 23:56:33.898942 4791 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="39b3d5f9-f8f2-426d-aa07-62aae5324c12" path="/var/lib/kubelet/pods/39b3d5f9-f8f2-426d-aa07-62aae5324c12/volumes" Dec 10 23:56:34 crc kubenswrapper[4791]: I1210 23:56:34.236464 4791 generic.go:334] "Generic (PLEG): container finished" podID="60fb43df-96e1-49b3-a19a-835c266f55f1" containerID="0ceb51a69e48e3a27d7418ea6015fb86bb2a209fe7ec32a431442cbbcb47e1fa" exitCode=0 Dec 10 23:56:34 crc kubenswrapper[4791]: I1210 23:56:34.236517 4791 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-2dlqt/crc-debug-9zjhh" event={"ID":"60fb43df-96e1-49b3-a19a-835c266f55f1","Type":"ContainerDied","Data":"0ceb51a69e48e3a27d7418ea6015fb86bb2a209fe7ec32a431442cbbcb47e1fa"} Dec 10 23:56:34 crc kubenswrapper[4791]: I1210 23:56:34.373125 4791 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-must-gather-2dlqt/crc-debug-9zjhh"] Dec 10 23:56:34 crc kubenswrapper[4791]: I1210 23:56:34.379812 4791 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-must-gather-2dlqt/crc-debug-9zjhh"] Dec 10 23:56:35 crc kubenswrapper[4791]: I1210 23:56:35.389053 4791 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-2dlqt/crc-debug-9zjhh" Dec 10 23:56:35 crc kubenswrapper[4791]: I1210 23:56:35.576890 4791 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-q67kc\" (UniqueName: \"kubernetes.io/projected/60fb43df-96e1-49b3-a19a-835c266f55f1-kube-api-access-q67kc\") pod \"60fb43df-96e1-49b3-a19a-835c266f55f1\" (UID: \"60fb43df-96e1-49b3-a19a-835c266f55f1\") " Dec 10 23:56:35 crc kubenswrapper[4791]: I1210 23:56:35.577050 4791 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"host\" (UniqueName: \"kubernetes.io/host-path/60fb43df-96e1-49b3-a19a-835c266f55f1-host\") pod \"60fb43df-96e1-49b3-a19a-835c266f55f1\" (UID: \"60fb43df-96e1-49b3-a19a-835c266f55f1\") " Dec 10 23:56:35 crc kubenswrapper[4791]: I1210 23:56:35.577472 4791 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/60fb43df-96e1-49b3-a19a-835c266f55f1-host" (OuterVolumeSpecName: "host") pod "60fb43df-96e1-49b3-a19a-835c266f55f1" (UID: "60fb43df-96e1-49b3-a19a-835c266f55f1"). InnerVolumeSpecName "host". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 10 23:56:35 crc kubenswrapper[4791]: I1210 23:56:35.597666 4791 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/60fb43df-96e1-49b3-a19a-835c266f55f1-kube-api-access-q67kc" (OuterVolumeSpecName: "kube-api-access-q67kc") pod "60fb43df-96e1-49b3-a19a-835c266f55f1" (UID: "60fb43df-96e1-49b3-a19a-835c266f55f1"). InnerVolumeSpecName "kube-api-access-q67kc". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 10 23:56:35 crc kubenswrapper[4791]: I1210 23:56:35.678646 4791 reconciler_common.go:293] "Volume detached for volume \"host\" (UniqueName: \"kubernetes.io/host-path/60fb43df-96e1-49b3-a19a-835c266f55f1-host\") on node \"crc\" DevicePath \"\"" Dec 10 23:56:35 crc kubenswrapper[4791]: I1210 23:56:35.678682 4791 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-q67kc\" (UniqueName: \"kubernetes.io/projected/60fb43df-96e1-49b3-a19a-835c266f55f1-kube-api-access-q67kc\") on node \"crc\" DevicePath \"\"" Dec 10 23:56:35 crc kubenswrapper[4791]: I1210 23:56:35.899150 4791 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="60fb43df-96e1-49b3-a19a-835c266f55f1" path="/var/lib/kubelet/pods/60fb43df-96e1-49b3-a19a-835c266f55f1/volumes" Dec 10 23:56:36 crc kubenswrapper[4791]: I1210 23:56:36.285026 4791 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-2dlqt/crc-debug-9zjhh" Dec 10 23:56:36 crc kubenswrapper[4791]: I1210 23:56:36.284939 4791 scope.go:117] "RemoveContainer" containerID="0ceb51a69e48e3a27d7418ea6015fb86bb2a209fe7ec32a431442cbbcb47e1fa" Dec 10 23:57:01 crc kubenswrapper[4791]: I1210 23:57:01.028509 4791 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_barbican-api-68fb764d56-4mfz4_eb482cb0-7ac1-4097-8799-4a5d78b6b5e5/barbican-api/0.log" Dec 10 23:57:01 crc kubenswrapper[4791]: I1210 23:57:01.176993 4791 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_barbican-api-68fb764d56-4mfz4_eb482cb0-7ac1-4097-8799-4a5d78b6b5e5/barbican-api-log/0.log" Dec 10 23:57:01 crc kubenswrapper[4791]: I1210 23:57:01.216638 4791 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_barbican-keystone-listener-67867746d6-mvxp4_4d7d941e-8494-46ea-990c-fe83db67dd2a/barbican-keystone-listener/0.log" Dec 10 23:57:01 crc kubenswrapper[4791]: I1210 23:57:01.291657 4791 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_barbican-keystone-listener-67867746d6-mvxp4_4d7d941e-8494-46ea-990c-fe83db67dd2a/barbican-keystone-listener-log/0.log" Dec 10 23:57:01 crc kubenswrapper[4791]: I1210 23:57:01.395651 4791 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_barbican-worker-77fb99879-xzsrk_8a302b7c-c2df-4023-95da-4ef8e86dbc0b/barbican-worker-log/0.log" Dec 10 23:57:01 crc kubenswrapper[4791]: I1210 23:57:01.413868 4791 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_barbican-worker-77fb99879-xzsrk_8a302b7c-c2df-4023-95da-4ef8e86dbc0b/barbican-worker/0.log" Dec 10 23:57:02 crc kubenswrapper[4791]: I1210 23:57:02.236715 4791 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ceilometer-0_c85398ce-a5d4-414a-8811-775a4ce71193/ceilometer-central-agent/0.log" Dec 10 23:57:02 crc kubenswrapper[4791]: I1210 23:57:02.242349 4791 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_bootstrap-edpm-deployment-openstack-edpm-ipam-fpng5_07a0187d-1677-4da9-9a3b-8da49e8f6819/bootstrap-edpm-deployment-openstack-edpm-ipam/0.log" Dec 10 23:57:02 crc kubenswrapper[4791]: I1210 23:57:02.382316 4791 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ceilometer-0_c85398ce-a5d4-414a-8811-775a4ce71193/ceilometer-notification-agent/0.log" Dec 10 23:57:02 crc kubenswrapper[4791]: I1210 23:57:02.427199 4791 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ceilometer-0_c85398ce-a5d4-414a-8811-775a4ce71193/sg-core/0.log" Dec 10 23:57:02 crc kubenswrapper[4791]: I1210 23:57:02.431287 4791 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ceilometer-0_c85398ce-a5d4-414a-8811-775a4ce71193/proxy-httpd/0.log" Dec 10 23:57:02 crc kubenswrapper[4791]: I1210 23:57:02.641519 4791 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_cinder-api-0_f807e752-6cf5-4ad6-bc21-f22777b62c83/cinder-api-log/0.log" Dec 10 23:57:02 crc kubenswrapper[4791]: I1210 23:57:02.656592 4791 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_cinder-api-0_f807e752-6cf5-4ad6-bc21-f22777b62c83/cinder-api/0.log" Dec 10 23:57:02 crc kubenswrapper[4791]: I1210 23:57:02.801212 4791 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_cinder-scheduler-0_e703a39e-cce5-4b36-b017-611836d0fa18/cinder-scheduler/0.log" Dec 10 23:57:02 crc kubenswrapper[4791]: I1210 23:57:02.842486 4791 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_cinder-scheduler-0_e703a39e-cce5-4b36-b017-611836d0fa18/probe/0.log" Dec 10 23:57:02 crc kubenswrapper[4791]: I1210 23:57:02.859213 4791 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_configure-network-edpm-deployment-openstack-edpm-ipam-wf5sq_852ca005-3b2f-407e-ae86-9b1c5a5fe182/configure-network-edpm-deployment-openstack-edpm-ipam/0.log" Dec 10 23:57:03 crc kubenswrapper[4791]: I1210 23:57:03.061057 4791 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_configure-os-edpm-deployment-openstack-edpm-ipam-cb6z5_cbb47bf3-4d3a-4e7e-bc54-c8a3a2ec98c6/configure-os-edpm-deployment-openstack-edpm-ipam/0.log" Dec 10 23:57:03 crc kubenswrapper[4791]: I1210 23:57:03.142666 4791 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_dnsmasq-dns-cb6ffcf87-wcxkk_07473608-7cf2-4bcb-b7bb-046165c73afe/init/0.log" Dec 10 23:57:03 crc kubenswrapper[4791]: I1210 23:57:03.276583 4791 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_dnsmasq-dns-cb6ffcf87-wcxkk_07473608-7cf2-4bcb-b7bb-046165c73afe/init/0.log" Dec 10 23:57:03 crc kubenswrapper[4791]: I1210 23:57:03.316910 4791 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_dnsmasq-dns-cb6ffcf87-wcxkk_07473608-7cf2-4bcb-b7bb-046165c73afe/dnsmasq-dns/0.log" Dec 10 23:57:03 crc kubenswrapper[4791]: I1210 23:57:03.367232 4791 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_download-cache-edpm-deployment-openstack-edpm-ipam-nx9lt_36b35496-ca5d-49ac-bb9f-cb697fa27af6/download-cache-edpm-deployment-openstack-edpm-ipam/0.log" Dec 10 23:57:03 crc kubenswrapper[4791]: I1210 23:57:03.491406 4791 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_glance-default-external-api-0_9655fe55-b388-4296-b9e4-198ea42c70cd/glance-httpd/0.log" Dec 10 23:57:03 crc kubenswrapper[4791]: I1210 23:57:03.564101 4791 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_glance-default-external-api-0_9655fe55-b388-4296-b9e4-198ea42c70cd/glance-log/0.log" Dec 10 23:57:04 crc kubenswrapper[4791]: I1210 23:57:04.290235 4791 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_glance-default-internal-api-0_d3317b5c-5c3b-4138-8f93-1bd9a159f8dd/glance-log/0.log" Dec 10 23:57:04 crc kubenswrapper[4791]: I1210 23:57:04.300268 4791 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_glance-default-internal-api-0_d3317b5c-5c3b-4138-8f93-1bd9a159f8dd/glance-httpd/0.log" Dec 10 23:57:04 crc kubenswrapper[4791]: I1210 23:57:04.424202 4791 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_horizon-6999c5bf5b-7cr9t_0fab3b41-ed57-4536-ade0-f5f6ccde40e2/horizon/0.log" Dec 10 23:57:04 crc kubenswrapper[4791]: I1210 23:57:04.587351 4791 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_install-certs-edpm-deployment-openstack-edpm-ipam-n9w4z_7eb11eed-a935-4ce6-b348-2c7c8f6ef16a/install-certs-edpm-deployment-openstack-edpm-ipam/0.log" Dec 10 23:57:04 crc kubenswrapper[4791]: I1210 23:57:04.686664 4791 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_install-os-edpm-deployment-openstack-edpm-ipam-c6nvg_19add645-abcb-46e3-a103-e39305c0be1c/install-os-edpm-deployment-openstack-edpm-ipam/0.log" Dec 10 23:57:04 crc kubenswrapper[4791]: I1210 23:57:04.878658 4791 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_horizon-6999c5bf5b-7cr9t_0fab3b41-ed57-4536-ade0-f5f6ccde40e2/horizon-log/0.log" Dec 10 23:57:04 crc kubenswrapper[4791]: I1210 23:57:04.891231 4791 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_kube-state-metrics-0_a5c489de-39fe-42e1-963b-10a99cb531b6/kube-state-metrics/0.log" Dec 10 23:57:04 crc kubenswrapper[4791]: I1210 23:57:04.979483 4791 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_keystone-54b7f6dc79-9klwj_5a5460d8-5939-40d3-8453-a487b3c4b58f/keystone-api/0.log" Dec 10 23:57:05 crc kubenswrapper[4791]: I1210 23:57:05.132901 4791 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_libvirt-edpm-deployment-openstack-edpm-ipam-9czng_b180a2cd-d91a-4203-b0f8-7de5c1be8226/libvirt-edpm-deployment-openstack-edpm-ipam/0.log" Dec 10 23:57:05 crc kubenswrapper[4791]: I1210 23:57:05.418316 4791 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_neutron-856d8bb659-x5kwg_60652b01-72de-481d-b2ea-8c349e9aee52/neutron-httpd/0.log" Dec 10 23:57:05 crc kubenswrapper[4791]: I1210 23:57:05.439115 4791 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_neutron-856d8bb659-x5kwg_60652b01-72de-481d-b2ea-8c349e9aee52/neutron-api/0.log" Dec 10 23:57:05 crc kubenswrapper[4791]: I1210 23:57:05.448516 4791 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_neutron-metadata-edpm-deployment-openstack-edpm-ipam-nnggl_9b204a02-650a-4f54-b257-5791c8399bf6/neutron-metadata-edpm-deployment-openstack-edpm-ipam/0.log" Dec 10 23:57:06 crc kubenswrapper[4791]: I1210 23:57:06.055864 4791 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_nova-api-0_70ddc653-9583-4f3c-a0f2-a2bfa952e7a4/nova-api-log/0.log" Dec 10 23:57:06 crc kubenswrapper[4791]: I1210 23:57:06.139645 4791 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_nova-cell0-conductor-0_1082e5c5-2e93-4950-93b4-c66ee224c711/nova-cell0-conductor-conductor/0.log" Dec 10 23:57:06 crc kubenswrapper[4791]: I1210 23:57:06.384695 4791 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_nova-cell1-conductor-0_2dff7039-7fed-41ac-8703-8e6023146318/nova-cell1-conductor-conductor/0.log" Dec 10 23:57:06 crc kubenswrapper[4791]: I1210 23:57:06.518598 4791 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_nova-cell1-novncproxy-0_4c145576-0efc-440b-bc4d-5468995305b4/nova-cell1-novncproxy-novncproxy/0.log" Dec 10 23:57:06 crc kubenswrapper[4791]: I1210 23:57:06.538180 4791 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_nova-api-0_70ddc653-9583-4f3c-a0f2-a2bfa952e7a4/nova-api-api/0.log" Dec 10 23:57:06 crc kubenswrapper[4791]: I1210 23:57:06.661786 4791 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_nova-edpm-deployment-openstack-edpm-ipam-5krlr_f9fc3f8e-9635-48a4-8ec3-20034a1bc7ad/nova-edpm-deployment-openstack-edpm-ipam/0.log" Dec 10 23:57:06 crc kubenswrapper[4791]: I1210 23:57:06.857963 4791 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_nova-metadata-0_8ee8d247-355c-4e17-965c-0b3c958a546c/nova-metadata-log/0.log" Dec 10 23:57:07 crc kubenswrapper[4791]: I1210 23:57:07.109796 4791 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_openstack-cell1-galera-0_f04fa829-3ebe-445d-a4e5-f7592ac682f3/mysql-bootstrap/0.log" Dec 10 23:57:07 crc kubenswrapper[4791]: I1210 23:57:07.184568 4791 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_nova-scheduler-0_b1d1b38e-2897-4c52-a4e4-d994dfffe805/nova-scheduler-scheduler/0.log" Dec 10 23:57:07 crc kubenswrapper[4791]: I1210 23:57:07.309961 4791 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_openstack-cell1-galera-0_f04fa829-3ebe-445d-a4e5-f7592ac682f3/mysql-bootstrap/0.log" Dec 10 23:57:07 crc kubenswrapper[4791]: I1210 23:57:07.376946 4791 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_openstack-cell1-galera-0_f04fa829-3ebe-445d-a4e5-f7592ac682f3/galera/0.log" Dec 10 23:57:07 crc kubenswrapper[4791]: I1210 23:57:07.636573 4791 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_openstack-galera-0_aed2b986-c040-4191-864b-47f29c5c8537/mysql-bootstrap/0.log" Dec 10 23:57:07 crc kubenswrapper[4791]: I1210 23:57:07.906931 4791 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_openstack-galera-0_aed2b986-c040-4191-864b-47f29c5c8537/mysql-bootstrap/0.log" Dec 10 23:57:07 crc kubenswrapper[4791]: I1210 23:57:07.917453 4791 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_openstack-galera-0_aed2b986-c040-4191-864b-47f29c5c8537/galera/0.log" Dec 10 23:57:08 crc kubenswrapper[4791]: I1210 23:57:08.071199 4791 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_openstackclient_2d25921f-5a9e-4e3b-9c51-f07f95aa02a1/openstackclient/0.log" Dec 10 23:57:08 crc kubenswrapper[4791]: I1210 23:57:08.156443 4791 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_nova-metadata-0_8ee8d247-355c-4e17-965c-0b3c958a546c/nova-metadata-metadata/0.log" Dec 10 23:57:08 crc kubenswrapper[4791]: I1210 23:57:08.159098 4791 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovn-controller-metrics-djrh2_53ceaa05-0398-4cf9-a787-874adab015aa/openstack-network-exporter/0.log" Dec 10 23:57:08 crc kubenswrapper[4791]: I1210 23:57:08.321680 4791 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovn-controller-ovs-phsfg_ac28ec72-8d7b-4576-b2ba-c93731d50267/ovsdb-server-init/0.log" Dec 10 23:57:08 crc kubenswrapper[4791]: I1210 23:57:08.550752 4791 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovn-controller-ovs-phsfg_ac28ec72-8d7b-4576-b2ba-c93731d50267/ovsdb-server-init/0.log" Dec 10 23:57:08 crc kubenswrapper[4791]: I1210 23:57:08.571385 4791 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovn-controller-ovs-phsfg_ac28ec72-8d7b-4576-b2ba-c93731d50267/ovsdb-server/0.log" Dec 10 23:57:08 crc kubenswrapper[4791]: I1210 23:57:08.588889 4791 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovn-controller-ovs-phsfg_ac28ec72-8d7b-4576-b2ba-c93731d50267/ovs-vswitchd/0.log" Dec 10 23:57:08 crc kubenswrapper[4791]: I1210 23:57:08.752962 4791 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovn-controller-xb29h_780fd8a7-cfb0-4958-afb3-1d060398d3ea/ovn-controller/0.log" Dec 10 23:57:08 crc kubenswrapper[4791]: I1210 23:57:08.860320 4791 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovn-edpm-deployment-openstack-edpm-ipam-ql7sj_9b2b641e-b123-4328-9151-f4c95e1405f2/ovn-edpm-deployment-openstack-edpm-ipam/0.log" Dec 10 23:57:08 crc kubenswrapper[4791]: I1210 23:57:08.991567 4791 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovn-northd-0_b73f3e7f-f699-460a-8d40-6abc63f13a1c/openstack-network-exporter/0.log" Dec 10 23:57:09 crc kubenswrapper[4791]: I1210 23:57:09.041963 4791 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovn-northd-0_b73f3e7f-f699-460a-8d40-6abc63f13a1c/ovn-northd/0.log" Dec 10 23:57:09 crc kubenswrapper[4791]: I1210 23:57:09.062288 4791 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovsdbserver-nb-0_a4bfff72-3a66-40f0-a06a-432d3cb4fff4/openstack-network-exporter/0.log" Dec 10 23:57:09 crc kubenswrapper[4791]: I1210 23:57:09.194034 4791 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovsdbserver-nb-0_a4bfff72-3a66-40f0-a06a-432d3cb4fff4/ovsdbserver-nb/0.log" Dec 10 23:57:09 crc kubenswrapper[4791]: I1210 23:57:09.236268 4791 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovsdbserver-sb-0_61769416-0728-4475-b965-dedb50d4455d/openstack-network-exporter/0.log" Dec 10 23:57:09 crc kubenswrapper[4791]: I1210 23:57:09.320322 4791 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovsdbserver-sb-0_61769416-0728-4475-b965-dedb50d4455d/ovsdbserver-sb/0.log" Dec 10 23:57:09 crc kubenswrapper[4791]: I1210 23:57:09.547448 4791 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_placement-6487596ddd-btmd7_0babf4ec-03b9-48de-875d-a27a8b7ed119/placement-api/0.log" Dec 10 23:57:09 crc kubenswrapper[4791]: I1210 23:57:09.624038 4791 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_placement-6487596ddd-btmd7_0babf4ec-03b9-48de-875d-a27a8b7ed119/placement-log/0.log" Dec 10 23:57:09 crc kubenswrapper[4791]: I1210 23:57:09.654188 4791 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_rabbitmq-cell1-server-0_3add67e0-b67a-438b-b83f-4c9f6733b5cb/setup-container/0.log" Dec 10 23:57:09 crc kubenswrapper[4791]: I1210 23:57:09.824463 4791 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_rabbitmq-cell1-server-0_3add67e0-b67a-438b-b83f-4c9f6733b5cb/rabbitmq/0.log" Dec 10 23:57:09 crc kubenswrapper[4791]: I1210 23:57:09.831988 4791 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_rabbitmq-server-0_5f51c37d-2f6f-4ef0-a08c-77216e7db57b/setup-container/0.log" Dec 10 23:57:09 crc kubenswrapper[4791]: I1210 23:57:09.896394 4791 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_rabbitmq-cell1-server-0_3add67e0-b67a-438b-b83f-4c9f6733b5cb/setup-container/0.log" Dec 10 23:57:10 crc kubenswrapper[4791]: I1210 23:57:10.129257 4791 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_rabbitmq-server-0_5f51c37d-2f6f-4ef0-a08c-77216e7db57b/setup-container/0.log" Dec 10 23:57:10 crc kubenswrapper[4791]: I1210 23:57:10.140577 4791 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_rabbitmq-server-0_5f51c37d-2f6f-4ef0-a08c-77216e7db57b/rabbitmq/0.log" Dec 10 23:57:10 crc kubenswrapper[4791]: I1210 23:57:10.161897 4791 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_reboot-os-edpm-deployment-openstack-edpm-ipam-fbsgr_72f75339-ac62-4e10-a6a2-634ac54461cd/reboot-os-edpm-deployment-openstack-edpm-ipam/0.log" Dec 10 23:57:10 crc kubenswrapper[4791]: I1210 23:57:10.331104 4791 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_redhat-edpm-deployment-openstack-edpm-ipam-2zj6q_e77c2ca5-1bf1-4739-b1c3-b0913fcc9d99/redhat-edpm-deployment-openstack-edpm-ipam/0.log" Dec 10 23:57:10 crc kubenswrapper[4791]: I1210 23:57:10.473447 4791 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_repo-setup-edpm-deployment-openstack-edpm-ipam-xb2ph_eff77dbf-4e46-4119-9db0-eb32a1d9d11c/repo-setup-edpm-deployment-openstack-edpm-ipam/0.log" Dec 10 23:57:10 crc kubenswrapper[4791]: I1210 23:57:10.613004 4791 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_run-os-edpm-deployment-openstack-edpm-ipam-5w7qc_62988878-9637-47ad-9877-fd81f32199d2/run-os-edpm-deployment-openstack-edpm-ipam/0.log" Dec 10 23:57:10 crc kubenswrapper[4791]: I1210 23:57:10.736552 4791 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ssh-known-hosts-edpm-deployment-8zwc6_12dfad91-cb70-4237-be65-427b3e919a05/ssh-known-hosts-edpm-deployment/0.log" Dec 10 23:57:11 crc kubenswrapper[4791]: I1210 23:57:11.173871 4791 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-proxy-c7bbc5ff-rlprn_644890b3-8cae-45bc-b3eb-e0e6bfa8d957/proxy-server/0.log" Dec 10 23:57:11 crc kubenswrapper[4791]: I1210 23:57:11.346548 4791 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-ring-rebalance-bqplg_16bd9b65-844e-4537-8d5b-c0a73666c2b2/swift-ring-rebalance/0.log" Dec 10 23:57:11 crc kubenswrapper[4791]: I1210 23:57:11.366438 4791 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-proxy-c7bbc5ff-rlprn_644890b3-8cae-45bc-b3eb-e0e6bfa8d957/proxy-httpd/0.log" Dec 10 23:57:11 crc kubenswrapper[4791]: I1210 23:57:11.459425 4791 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_69a6c950-8a3b-4a7c-b284-ebd20157eb20/account-auditor/0.log" Dec 10 23:57:11 crc kubenswrapper[4791]: I1210 23:57:11.550116 4791 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_69a6c950-8a3b-4a7c-b284-ebd20157eb20/account-reaper/0.log" Dec 10 23:57:11 crc kubenswrapper[4791]: I1210 23:57:11.653705 4791 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_69a6c950-8a3b-4a7c-b284-ebd20157eb20/account-replicator/0.log" Dec 10 23:57:11 crc kubenswrapper[4791]: I1210 23:57:11.669768 4791 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_69a6c950-8a3b-4a7c-b284-ebd20157eb20/container-auditor/0.log" Dec 10 23:57:11 crc kubenswrapper[4791]: I1210 23:57:11.705083 4791 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_69a6c950-8a3b-4a7c-b284-ebd20157eb20/account-server/0.log" Dec 10 23:57:11 crc kubenswrapper[4791]: I1210 23:57:11.809227 4791 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_69a6c950-8a3b-4a7c-b284-ebd20157eb20/container-replicator/0.log" Dec 10 23:57:11 crc kubenswrapper[4791]: I1210 23:57:11.832414 4791 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_69a6c950-8a3b-4a7c-b284-ebd20157eb20/container-server/0.log" Dec 10 23:57:11 crc kubenswrapper[4791]: I1210 23:57:11.856716 4791 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_69a6c950-8a3b-4a7c-b284-ebd20157eb20/container-updater/0.log" Dec 10 23:57:11 crc kubenswrapper[4791]: I1210 23:57:11.946134 4791 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_69a6c950-8a3b-4a7c-b284-ebd20157eb20/object-auditor/0.log" Dec 10 23:57:12 crc kubenswrapper[4791]: I1210 23:57:12.018618 4791 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_69a6c950-8a3b-4a7c-b284-ebd20157eb20/object-expirer/0.log" Dec 10 23:57:12 crc kubenswrapper[4791]: I1210 23:57:12.102116 4791 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_69a6c950-8a3b-4a7c-b284-ebd20157eb20/object-replicator/0.log" Dec 10 23:57:12 crc kubenswrapper[4791]: I1210 23:57:12.108521 4791 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_69a6c950-8a3b-4a7c-b284-ebd20157eb20/object-server/0.log" Dec 10 23:57:12 crc kubenswrapper[4791]: I1210 23:57:12.203708 4791 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_69a6c950-8a3b-4a7c-b284-ebd20157eb20/object-updater/0.log" Dec 10 23:57:12 crc kubenswrapper[4791]: I1210 23:57:12.252316 4791 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_69a6c950-8a3b-4a7c-b284-ebd20157eb20/rsync/0.log" Dec 10 23:57:12 crc kubenswrapper[4791]: I1210 23:57:12.321285 4791 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_69a6c950-8a3b-4a7c-b284-ebd20157eb20/swift-recon-cron/0.log" Dec 10 23:57:12 crc kubenswrapper[4791]: I1210 23:57:12.860251 4791 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_telemetry-edpm-deployment-openstack-edpm-ipam-kxtwt_282a792a-a4a9-4e6d-b7ae-f5cedbe55787/telemetry-edpm-deployment-openstack-edpm-ipam/0.log" Dec 10 23:57:12 crc kubenswrapper[4791]: I1210 23:57:12.877297 4791 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_tempest-tests-tempest_cf11ce8f-d433-49ff-9390-064a2f4ec60d/tempest-tests-tempest-tests-runner/0.log" Dec 10 23:57:13 crc kubenswrapper[4791]: I1210 23:57:13.070295 4791 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_test-operator-logs-pod-tempest-tempest-tests-tempest_00c94f56-3118-486c-a325-9493d5bfd911/test-operator-logs-container/0.log" Dec 10 23:57:13 crc kubenswrapper[4791]: I1210 23:57:13.181139 4791 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_validate-network-edpm-deployment-openstack-edpm-ipam-gtc8x_73316991-2ef6-4257-8118-55378571b7d7/validate-network-edpm-deployment-openstack-edpm-ipam/0.log" Dec 10 23:57:23 crc kubenswrapper[4791]: I1210 23:57:23.313876 4791 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_memcached-0_49839437-3696-4ac8-98cf-2cc16048f21a/memcached/0.log" Dec 10 23:57:39 crc kubenswrapper[4791]: I1210 23:57:39.491665 4791 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_50d664eb6e76f12a148f821b8e085abb03568bf5cb6cd29276e1382c05lclqs_c76a21e1-575e-42e2-8372-e79986d26dbb/util/0.log" Dec 10 23:57:39 crc kubenswrapper[4791]: I1210 23:57:39.698834 4791 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_50d664eb6e76f12a148f821b8e085abb03568bf5cb6cd29276e1382c05lclqs_c76a21e1-575e-42e2-8372-e79986d26dbb/util/0.log" Dec 10 23:57:39 crc kubenswrapper[4791]: I1210 23:57:39.727570 4791 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_50d664eb6e76f12a148f821b8e085abb03568bf5cb6cd29276e1382c05lclqs_c76a21e1-575e-42e2-8372-e79986d26dbb/pull/0.log" Dec 10 23:57:39 crc kubenswrapper[4791]: I1210 23:57:39.746722 4791 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_50d664eb6e76f12a148f821b8e085abb03568bf5cb6cd29276e1382c05lclqs_c76a21e1-575e-42e2-8372-e79986d26dbb/pull/0.log" Dec 10 23:57:39 crc kubenswrapper[4791]: I1210 23:57:39.933527 4791 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_50d664eb6e76f12a148f821b8e085abb03568bf5cb6cd29276e1382c05lclqs_c76a21e1-575e-42e2-8372-e79986d26dbb/extract/0.log" Dec 10 23:57:39 crc kubenswrapper[4791]: I1210 23:57:39.936322 4791 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_50d664eb6e76f12a148f821b8e085abb03568bf5cb6cd29276e1382c05lclqs_c76a21e1-575e-42e2-8372-e79986d26dbb/pull/0.log" Dec 10 23:57:39 crc kubenswrapper[4791]: I1210 23:57:39.947610 4791 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_50d664eb6e76f12a148f821b8e085abb03568bf5cb6cd29276e1382c05lclqs_c76a21e1-575e-42e2-8372-e79986d26dbb/util/0.log" Dec 10 23:57:40 crc kubenswrapper[4791]: I1210 23:57:40.137158 4791 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_barbican-operator-controller-manager-7d9dfd778-4ndx5_dc042f4f-d336-484e-8c9d-60658c82f84a/kube-rbac-proxy/0.log" Dec 10 23:57:40 crc kubenswrapper[4791]: I1210 23:57:40.166058 4791 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_barbican-operator-controller-manager-7d9dfd778-4ndx5_dc042f4f-d336-484e-8c9d-60658c82f84a/manager/0.log" Dec 10 23:57:40 crc kubenswrapper[4791]: I1210 23:57:40.190732 4791 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_cinder-operator-controller-manager-6c677c69b-6jmn2_48cd2bae-61e0-446c-ac2a-48e70bff5187/kube-rbac-proxy/0.log" Dec 10 23:57:40 crc kubenswrapper[4791]: I1210 23:57:40.336367 4791 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_cinder-operator-controller-manager-6c677c69b-6jmn2_48cd2bae-61e0-446c-ac2a-48e70bff5187/manager/0.log" Dec 10 23:57:40 crc kubenswrapper[4791]: I1210 23:57:40.348586 4791 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_designate-operator-controller-manager-697fb699cf-vg956_19bf0bfe-4e8b-47a4-a919-4227b3e47bb3/kube-rbac-proxy/0.log" Dec 10 23:57:40 crc kubenswrapper[4791]: I1210 23:57:40.592315 4791 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_designate-operator-controller-manager-697fb699cf-vg956_19bf0bfe-4e8b-47a4-a919-4227b3e47bb3/manager/0.log" Dec 10 23:57:41 crc kubenswrapper[4791]: I1210 23:57:41.153902 4791 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_glance-operator-controller-manager-5697bb5779-cvbqc_f0b16aa2-e60e-4f0e-9679-9afc0a5ae027/kube-rbac-proxy/0.log" Dec 10 23:57:41 crc kubenswrapper[4791]: I1210 23:57:41.262349 4791 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_glance-operator-controller-manager-5697bb5779-cvbqc_f0b16aa2-e60e-4f0e-9679-9afc0a5ae027/manager/0.log" Dec 10 23:57:41 crc kubenswrapper[4791]: I1210 23:57:41.376410 4791 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_heat-operator-controller-manager-5f64f6f8bb-fx67c_d181199e-5cd2-408b-b167-059b34491a15/kube-rbac-proxy/0.log" Dec 10 23:57:41 crc kubenswrapper[4791]: I1210 23:57:41.421144 4791 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_horizon-operator-controller-manager-68c6d99b8f-2lq47_e3fdc812-fbd3-4d35-8555-d55e9390ebd0/kube-rbac-proxy/0.log" Dec 10 23:57:41 crc kubenswrapper[4791]: I1210 23:57:41.423687 4791 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_heat-operator-controller-manager-5f64f6f8bb-fx67c_d181199e-5cd2-408b-b167-059b34491a15/manager/0.log" Dec 10 23:57:41 crc kubenswrapper[4791]: I1210 23:57:41.550172 4791 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_horizon-operator-controller-manager-68c6d99b8f-2lq47_e3fdc812-fbd3-4d35-8555-d55e9390ebd0/manager/0.log" Dec 10 23:57:41 crc kubenswrapper[4791]: I1210 23:57:41.590218 4791 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_infra-operator-controller-manager-78d48bff9d-mmczn_416b455d-b397-4aad-baf1-88e880619eb5/kube-rbac-proxy/0.log" Dec 10 23:57:41 crc kubenswrapper[4791]: I1210 23:57:41.796527 4791 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_infra-operator-controller-manager-78d48bff9d-mmczn_416b455d-b397-4aad-baf1-88e880619eb5/manager/0.log" Dec 10 23:57:41 crc kubenswrapper[4791]: I1210 23:57:41.810981 4791 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_ironic-operator-controller-manager-967d97867-r6jdv_cb5fb221-76b0-48b4-ad27-7d0c51581ec4/kube-rbac-proxy/0.log" Dec 10 23:57:41 crc kubenswrapper[4791]: I1210 23:57:41.835697 4791 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_ironic-operator-controller-manager-967d97867-r6jdv_cb5fb221-76b0-48b4-ad27-7d0c51581ec4/manager/0.log" Dec 10 23:57:41 crc kubenswrapper[4791]: I1210 23:57:41.966989 4791 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_keystone-operator-controller-manager-7765d96ddf-2wsxl_311d4e7c-a0c7-47f6-8deb-7262068e8beb/kube-rbac-proxy/0.log" Dec 10 23:57:42 crc kubenswrapper[4791]: I1210 23:57:42.104572 4791 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_keystone-operator-controller-manager-7765d96ddf-2wsxl_311d4e7c-a0c7-47f6-8deb-7262068e8beb/manager/0.log" Dec 10 23:57:42 crc kubenswrapper[4791]: I1210 23:57:42.136784 4791 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_manila-operator-controller-manager-5b5fd79c9c-tsbtp_12b75f38-fb6a-4020-95b9-29c7dd872849/kube-rbac-proxy/0.log" Dec 10 23:57:42 crc kubenswrapper[4791]: I1210 23:57:42.178719 4791 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_manila-operator-controller-manager-5b5fd79c9c-tsbtp_12b75f38-fb6a-4020-95b9-29c7dd872849/manager/0.log" Dec 10 23:57:42 crc kubenswrapper[4791]: I1210 23:57:42.712908 4791 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_mariadb-operator-controller-manager-79c8c4686c-2wmm4_590c13cc-efa0-431a-9f2f-65fb97c1bd3c/kube-rbac-proxy/0.log" Dec 10 23:57:42 crc kubenswrapper[4791]: I1210 23:57:42.714114 4791 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_mariadb-operator-controller-manager-79c8c4686c-2wmm4_590c13cc-efa0-431a-9f2f-65fb97c1bd3c/manager/0.log" Dec 10 23:57:42 crc kubenswrapper[4791]: I1210 23:57:42.871963 4791 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_neutron-operator-controller-manager-5fdfd5b6b5-g8f76_86bbb7b2-5f55-46a2-b426-013ba183977f/kube-rbac-proxy/0.log" Dec 10 23:57:42 crc kubenswrapper[4791]: I1210 23:57:42.937556 4791 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_neutron-operator-controller-manager-5fdfd5b6b5-g8f76_86bbb7b2-5f55-46a2-b426-013ba183977f/manager/0.log" Dec 10 23:57:42 crc kubenswrapper[4791]: I1210 23:57:42.961593 4791 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_nova-operator-controller-manager-697bc559fc-2lkcl_3b2af4ca-48b4-463f-b0aa-c5cfd9099d67/kube-rbac-proxy/0.log" Dec 10 23:57:43 crc kubenswrapper[4791]: I1210 23:57:43.137267 4791 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_octavia-operator-controller-manager-998648c74-sjp6q_23a95e98-8ae3-4ac4-945d-0cae5af5d1ac/kube-rbac-proxy/0.log" Dec 10 23:57:43 crc kubenswrapper[4791]: I1210 23:57:43.227747 4791 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_octavia-operator-controller-manager-998648c74-sjp6q_23a95e98-8ae3-4ac4-945d-0cae5af5d1ac/manager/0.log" Dec 10 23:57:43 crc kubenswrapper[4791]: I1210 23:57:43.239248 4791 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_nova-operator-controller-manager-697bc559fc-2lkcl_3b2af4ca-48b4-463f-b0aa-c5cfd9099d67/manager/0.log" Dec 10 23:57:43 crc kubenswrapper[4791]: I1210 23:57:43.394479 4791 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_openstack-baremetal-operator-controller-manager-84b575879fv4t28_d5907af9-fe3f-4da0-ba70-993d36b25746/kube-rbac-proxy/0.log" Dec 10 23:57:43 crc kubenswrapper[4791]: I1210 23:57:43.405946 4791 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_openstack-baremetal-operator-controller-manager-84b575879fv4t28_d5907af9-fe3f-4da0-ba70-993d36b25746/manager/0.log" Dec 10 23:57:43 crc kubenswrapper[4791]: I1210 23:57:43.726622 4791 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_openstack-operator-index-zg7wn_f96dcde6-8c38-41a4-b1bd-b1c3007ff451/registry-server/0.log" Dec 10 23:57:43 crc kubenswrapper[4791]: I1210 23:57:43.764208 4791 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_openstack-operator-controller-operator-6cc5d54456-w47nt_35509dee-f21b-4ab9-95ae-b93d2a4d93d2/operator/0.log" Dec 10 23:57:43 crc kubenswrapper[4791]: I1210 23:57:43.809133 4791 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_ovn-operator-controller-manager-b6456fdb6-6rddm_87c40827-309c-4c68-ae46-bbe449a9e66c/kube-rbac-proxy/0.log" Dec 10 23:57:44 crc kubenswrapper[4791]: I1210 23:57:44.018951 4791 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_placement-operator-controller-manager-78f8948974-bhgvx_b67ef71c-85c5-41ca-bf93-d1ca8a2fd007/kube-rbac-proxy/0.log" Dec 10 23:57:44 crc kubenswrapper[4791]: I1210 23:57:44.070115 4791 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_placement-operator-controller-manager-78f8948974-bhgvx_b67ef71c-85c5-41ca-bf93-d1ca8a2fd007/manager/0.log" Dec 10 23:57:44 crc kubenswrapper[4791]: I1210 23:57:44.070457 4791 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_ovn-operator-controller-manager-b6456fdb6-6rddm_87c40827-309c-4c68-ae46-bbe449a9e66c/manager/0.log" Dec 10 23:57:44 crc kubenswrapper[4791]: I1210 23:57:44.288931 4791 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_rabbitmq-cluster-operator-manager-668c99d594-2bgc5_b3ae5e53-9077-4dd0-996d-2417f28b4736/operator/0.log" Dec 10 23:57:44 crc kubenswrapper[4791]: I1210 23:57:44.314200 4791 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_swift-operator-controller-manager-9d58d64bc-8vwnp_4cf9497a-4193-4d34-a68a-48085af6e8b5/kube-rbac-proxy/0.log" Dec 10 23:57:44 crc kubenswrapper[4791]: I1210 23:57:44.447311 4791 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_swift-operator-controller-manager-9d58d64bc-8vwnp_4cf9497a-4193-4d34-a68a-48085af6e8b5/manager/0.log" Dec 10 23:57:44 crc kubenswrapper[4791]: I1210 23:57:44.524737 4791 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_telemetry-operator-controller-manager-58d5ff84df-t6tnb_2418f025-d37e-4316-b804-9ab53c326c82/kube-rbac-proxy/0.log" Dec 10 23:57:44 crc kubenswrapper[4791]: I1210 23:57:44.571927 4791 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_openstack-operator-controller-manager-686dfd865c-cft5f_e2c4ecea-2852-4e39-93fb-aee7cbe31aa4/manager/0.log" Dec 10 23:57:44 crc kubenswrapper[4791]: I1210 23:57:44.647023 4791 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_telemetry-operator-controller-manager-58d5ff84df-t6tnb_2418f025-d37e-4316-b804-9ab53c326c82/manager/0.log" Dec 10 23:57:44 crc kubenswrapper[4791]: I1210 23:57:44.711622 4791 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_test-operator-controller-manager-5854674fcc-rw7ct_bda40eaa-446c-475e-8f2f-76f76caa0b85/kube-rbac-proxy/0.log" Dec 10 23:57:44 crc kubenswrapper[4791]: I1210 23:57:44.745551 4791 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_test-operator-controller-manager-5854674fcc-rw7ct_bda40eaa-446c-475e-8f2f-76f76caa0b85/manager/0.log" Dec 10 23:57:44 crc kubenswrapper[4791]: I1210 23:57:44.824973 4791 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_watcher-operator-controller-manager-75944c9b7-7cdb7_21e66f28-5ed9-475e-aa7d-a105ad1f2f9e/kube-rbac-proxy/0.log" Dec 10 23:57:44 crc kubenswrapper[4791]: I1210 23:57:44.915225 4791 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_watcher-operator-controller-manager-75944c9b7-7cdb7_21e66f28-5ed9-475e-aa7d-a105ad1f2f9e/manager/0.log" Dec 10 23:57:55 crc kubenswrapper[4791]: I1210 23:57:55.038785 4791 patch_prober.go:28] interesting pod/machine-config-daemon-5rb5l container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 10 23:57:55 crc kubenswrapper[4791]: I1210 23:57:55.039527 4791 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-5rb5l" podUID="ba35653c-6e06-4cee-a4d6-137764090d18" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 10 23:58:05 crc kubenswrapper[4791]: I1210 23:58:05.582951 4791 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-machine-api_control-plane-machine-set-operator-78cbb6b69f-ms8jw_c6c773c2-29e4-477d-91fb-f63052272f06/control-plane-machine-set-operator/0.log" Dec 10 23:58:05 crc kubenswrapper[4791]: I1210 23:58:05.717683 4791 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-machine-api_machine-api-operator-5694c8668f-cpbhz_cd65e91e-5b02-4f78-829e-93b4596ff8f6/kube-rbac-proxy/0.log" Dec 10 23:58:05 crc kubenswrapper[4791]: I1210 23:58:05.738570 4791 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-machine-api_machine-api-operator-5694c8668f-cpbhz_cd65e91e-5b02-4f78-829e-93b4596ff8f6/machine-api-operator/0.log" Dec 10 23:58:19 crc kubenswrapper[4791]: I1210 23:58:19.402898 4791 log.go:25] "Finished parsing log file" path="/var/log/pods/cert-manager_cert-manager-5b446d88c5-w9jv9_5f070f43-21a8-43bf-80db-9038ed2dd9f5/cert-manager-controller/0.log" Dec 10 23:58:19 crc kubenswrapper[4791]: I1210 23:58:19.598018 4791 log.go:25] "Finished parsing log file" path="/var/log/pods/cert-manager_cert-manager-cainjector-7f985d654d-ntt86_18fded69-b7d4-42c9-80de-1c78e6389edf/cert-manager-cainjector/0.log" Dec 10 23:58:19 crc kubenswrapper[4791]: I1210 23:58:19.674713 4791 log.go:25] "Finished parsing log file" path="/var/log/pods/cert-manager_cert-manager-webhook-5655c58dd6-xzg6b_33717f3f-c6f5-4be9-a7c5-d6a7ae599dc9/cert-manager-webhook/0.log" Dec 10 23:58:25 crc kubenswrapper[4791]: I1210 23:58:25.038061 4791 patch_prober.go:28] interesting pod/machine-config-daemon-5rb5l container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 10 23:58:25 crc kubenswrapper[4791]: I1210 23:58:25.038398 4791 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-5rb5l" podUID="ba35653c-6e06-4cee-a4d6-137764090d18" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 10 23:58:34 crc kubenswrapper[4791]: I1210 23:58:34.380474 4791 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-nmstate_nmstate-console-plugin-7fbb5f6569-xxb9s_14000c4a-8a7d-4ade-86e6-819e08eaf42a/nmstate-console-plugin/0.log" Dec 10 23:58:34 crc kubenswrapper[4791]: I1210 23:58:34.405132 4791 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-nmstate_nmstate-handler-2j8xz_fd7f0c56-2613-4ba7-b0a4-4ff765493973/nmstate-handler/0.log" Dec 10 23:58:34 crc kubenswrapper[4791]: I1210 23:58:34.597035 4791 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-nmstate_nmstate-metrics-7f946cbc9-mph8d_e85292bb-7cee-4b20-8c97-b1d1902e6001/nmstate-metrics/0.log" Dec 10 23:58:34 crc kubenswrapper[4791]: I1210 23:58:34.600613 4791 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-nmstate_nmstate-metrics-7f946cbc9-mph8d_e85292bb-7cee-4b20-8c97-b1d1902e6001/kube-rbac-proxy/0.log" Dec 10 23:58:34 crc kubenswrapper[4791]: I1210 23:58:34.751101 4791 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-nmstate_nmstate-operator-5b5b58f5c8-9qtjh_f6780aea-2f04-4dc3-96c6-8e4e7ef4284e/nmstate-operator/0.log" Dec 10 23:58:34 crc kubenswrapper[4791]: I1210 23:58:34.785333 4791 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-nmstate_nmstate-webhook-5f6d4c5ccb-x65fz_99524090-326f-47fd-98d9-bf7e8352cc4f/nmstate-webhook/0.log" Dec 10 23:58:49 crc kubenswrapper[4791]: I1210 23:58:49.741108 4791 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_controller-f8648f98b-q9cdb_e52e894c-94e4-4958-b261-c5c373c29040/kube-rbac-proxy/0.log" Dec 10 23:58:49 crc kubenswrapper[4791]: I1210 23:58:49.909273 4791 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_controller-f8648f98b-q9cdb_e52e894c-94e4-4958-b261-c5c373c29040/controller/0.log" Dec 10 23:58:49 crc kubenswrapper[4791]: I1210 23:58:49.928420 4791 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-gqlpm_a68714bc-7787-46c8-ba8f-21f7af6d9d67/cp-frr-files/0.log" Dec 10 23:58:50 crc kubenswrapper[4791]: I1210 23:58:50.161199 4791 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-gqlpm_a68714bc-7787-46c8-ba8f-21f7af6d9d67/cp-reloader/0.log" Dec 10 23:58:50 crc kubenswrapper[4791]: I1210 23:58:50.162970 4791 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-gqlpm_a68714bc-7787-46c8-ba8f-21f7af6d9d67/cp-reloader/0.log" Dec 10 23:58:50 crc kubenswrapper[4791]: I1210 23:58:50.163701 4791 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-gqlpm_a68714bc-7787-46c8-ba8f-21f7af6d9d67/cp-frr-files/0.log" Dec 10 23:58:50 crc kubenswrapper[4791]: I1210 23:58:50.201118 4791 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-gqlpm_a68714bc-7787-46c8-ba8f-21f7af6d9d67/cp-metrics/0.log" Dec 10 23:58:50 crc kubenswrapper[4791]: I1210 23:58:50.361483 4791 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-gqlpm_a68714bc-7787-46c8-ba8f-21f7af6d9d67/cp-metrics/0.log" Dec 10 23:58:50 crc kubenswrapper[4791]: I1210 23:58:50.363955 4791 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-gqlpm_a68714bc-7787-46c8-ba8f-21f7af6d9d67/cp-reloader/0.log" Dec 10 23:58:50 crc kubenswrapper[4791]: I1210 23:58:50.383962 4791 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-gqlpm_a68714bc-7787-46c8-ba8f-21f7af6d9d67/cp-metrics/0.log" Dec 10 23:58:50 crc kubenswrapper[4791]: I1210 23:58:50.415307 4791 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-gqlpm_a68714bc-7787-46c8-ba8f-21f7af6d9d67/cp-frr-files/0.log" Dec 10 23:58:50 crc kubenswrapper[4791]: I1210 23:58:50.595643 4791 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-gqlpm_a68714bc-7787-46c8-ba8f-21f7af6d9d67/controller/0.log" Dec 10 23:58:50 crc kubenswrapper[4791]: I1210 23:58:50.596455 4791 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-gqlpm_a68714bc-7787-46c8-ba8f-21f7af6d9d67/cp-metrics/0.log" Dec 10 23:58:50 crc kubenswrapper[4791]: I1210 23:58:50.600994 4791 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-gqlpm_a68714bc-7787-46c8-ba8f-21f7af6d9d67/cp-frr-files/0.log" Dec 10 23:58:50 crc kubenswrapper[4791]: I1210 23:58:50.614767 4791 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-gqlpm_a68714bc-7787-46c8-ba8f-21f7af6d9d67/cp-reloader/0.log" Dec 10 23:58:50 crc kubenswrapper[4791]: I1210 23:58:50.760484 4791 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-gqlpm_a68714bc-7787-46c8-ba8f-21f7af6d9d67/frr-metrics/0.log" Dec 10 23:58:50 crc kubenswrapper[4791]: I1210 23:58:50.790520 4791 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-gqlpm_a68714bc-7787-46c8-ba8f-21f7af6d9d67/kube-rbac-proxy/0.log" Dec 10 23:58:50 crc kubenswrapper[4791]: I1210 23:58:50.862328 4791 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-gqlpm_a68714bc-7787-46c8-ba8f-21f7af6d9d67/kube-rbac-proxy-frr/0.log" Dec 10 23:58:50 crc kubenswrapper[4791]: I1210 23:58:50.969104 4791 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-gqlpm_a68714bc-7787-46c8-ba8f-21f7af6d9d67/reloader/0.log" Dec 10 23:58:51 crc kubenswrapper[4791]: I1210 23:58:51.065660 4791 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-webhook-server-7fcb986d4-whw8m_d1dc382e-929c-44ce-b5c3-73410e626d0c/frr-k8s-webhook-server/0.log" Dec 10 23:58:51 crc kubenswrapper[4791]: I1210 23:58:51.269002 4791 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_metallb-operator-controller-manager-867ddc9894-mhxh7_72a07cc5-20c1-4f60-b6d7-84dfe1f78be2/manager/0.log" Dec 10 23:58:51 crc kubenswrapper[4791]: I1210 23:58:51.339103 4791 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_metallb-operator-webhook-server-568c9f7b75-dtmkk_f1ea99a7-677c-4d8b-89c6-3b24466632c2/webhook-server/0.log" Dec 10 23:58:51 crc kubenswrapper[4791]: I1210 23:58:51.474602 4791 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_speaker-jfggv_10e04a8d-1cac-4c7a-9b14-be8db393651a/kube-rbac-proxy/0.log" Dec 10 23:58:51 crc kubenswrapper[4791]: I1210 23:58:51.969694 4791 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_speaker-jfggv_10e04a8d-1cac-4c7a-9b14-be8db393651a/speaker/0.log" Dec 10 23:58:52 crc kubenswrapper[4791]: I1210 23:58:52.109141 4791 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-gqlpm_a68714bc-7787-46c8-ba8f-21f7af6d9d67/frr/0.log" Dec 10 23:58:55 crc kubenswrapper[4791]: I1210 23:58:55.038793 4791 patch_prober.go:28] interesting pod/machine-config-daemon-5rb5l container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 10 23:58:55 crc kubenswrapper[4791]: I1210 23:58:55.039322 4791 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-5rb5l" podUID="ba35653c-6e06-4cee-a4d6-137764090d18" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 10 23:58:55 crc kubenswrapper[4791]: I1210 23:58:55.039500 4791 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-5rb5l" Dec 10 23:58:55 crc kubenswrapper[4791]: I1210 23:58:55.040882 4791 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"1a35e34c3589dfe3497551cd10aa8a1e1fa1f05668706b1469b66da53eca8dc3"} pod="openshift-machine-config-operator/machine-config-daemon-5rb5l" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Dec 10 23:58:55 crc kubenswrapper[4791]: I1210 23:58:55.041052 4791 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-5rb5l" podUID="ba35653c-6e06-4cee-a4d6-137764090d18" containerName="machine-config-daemon" containerID="cri-o://1a35e34c3589dfe3497551cd10aa8a1e1fa1f05668706b1469b66da53eca8dc3" gracePeriod=600 Dec 10 23:58:55 crc kubenswrapper[4791]: I1210 23:58:55.363329 4791 generic.go:334] "Generic (PLEG): container finished" podID="ba35653c-6e06-4cee-a4d6-137764090d18" containerID="1a35e34c3589dfe3497551cd10aa8a1e1fa1f05668706b1469b66da53eca8dc3" exitCode=0 Dec 10 23:58:55 crc kubenswrapper[4791]: I1210 23:58:55.363373 4791 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-5rb5l" event={"ID":"ba35653c-6e06-4cee-a4d6-137764090d18","Type":"ContainerDied","Data":"1a35e34c3589dfe3497551cd10aa8a1e1fa1f05668706b1469b66da53eca8dc3"} Dec 10 23:58:55 crc kubenswrapper[4791]: I1210 23:58:55.363646 4791 scope.go:117] "RemoveContainer" containerID="f44d2e51d8375135a5705f7cf71dde0bd0c6901d20f892b1346dc2eac144cbc0" Dec 10 23:58:56 crc kubenswrapper[4791]: I1210 23:58:56.379261 4791 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-5rb5l" event={"ID":"ba35653c-6e06-4cee-a4d6-137764090d18","Type":"ContainerStarted","Data":"174b0c37e74227b1037fd59658b8e726e56afcb2c113dde44e2136ad132c104b"} Dec 10 23:58:58 crc kubenswrapper[4791]: I1210 23:58:58.489487 4791 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-marketplace-dlskm"] Dec 10 23:58:58 crc kubenswrapper[4791]: E1210 23:58:58.490635 4791 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="60fb43df-96e1-49b3-a19a-835c266f55f1" containerName="container-00" Dec 10 23:58:58 crc kubenswrapper[4791]: I1210 23:58:58.490653 4791 state_mem.go:107] "Deleted CPUSet assignment" podUID="60fb43df-96e1-49b3-a19a-835c266f55f1" containerName="container-00" Dec 10 23:58:58 crc kubenswrapper[4791]: I1210 23:58:58.490898 4791 memory_manager.go:354] "RemoveStaleState removing state" podUID="60fb43df-96e1-49b3-a19a-835c266f55f1" containerName="container-00" Dec 10 23:58:58 crc kubenswrapper[4791]: I1210 23:58:58.492583 4791 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-dlskm" Dec 10 23:58:58 crc kubenswrapper[4791]: I1210 23:58:58.519441 4791 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-dlskm"] Dec 10 23:58:58 crc kubenswrapper[4791]: I1210 23:58:58.626254 4791 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/ceb2c559-f4b0-4176-bec1-8593df45d701-utilities\") pod \"redhat-marketplace-dlskm\" (UID: \"ceb2c559-f4b0-4176-bec1-8593df45d701\") " pod="openshift-marketplace/redhat-marketplace-dlskm" Dec 10 23:58:58 crc kubenswrapper[4791]: I1210 23:58:58.626643 4791 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/ceb2c559-f4b0-4176-bec1-8593df45d701-catalog-content\") pod \"redhat-marketplace-dlskm\" (UID: \"ceb2c559-f4b0-4176-bec1-8593df45d701\") " pod="openshift-marketplace/redhat-marketplace-dlskm" Dec 10 23:58:58 crc kubenswrapper[4791]: I1210 23:58:58.626892 4791 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-vlwn5\" (UniqueName: \"kubernetes.io/projected/ceb2c559-f4b0-4176-bec1-8593df45d701-kube-api-access-vlwn5\") pod \"redhat-marketplace-dlskm\" (UID: \"ceb2c559-f4b0-4176-bec1-8593df45d701\") " pod="openshift-marketplace/redhat-marketplace-dlskm" Dec 10 23:58:58 crc kubenswrapper[4791]: I1210 23:58:58.729104 4791 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/ceb2c559-f4b0-4176-bec1-8593df45d701-utilities\") pod \"redhat-marketplace-dlskm\" (UID: \"ceb2c559-f4b0-4176-bec1-8593df45d701\") " pod="openshift-marketplace/redhat-marketplace-dlskm" Dec 10 23:58:58 crc kubenswrapper[4791]: I1210 23:58:58.729262 4791 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/ceb2c559-f4b0-4176-bec1-8593df45d701-catalog-content\") pod \"redhat-marketplace-dlskm\" (UID: \"ceb2c559-f4b0-4176-bec1-8593df45d701\") " pod="openshift-marketplace/redhat-marketplace-dlskm" Dec 10 23:58:58 crc kubenswrapper[4791]: I1210 23:58:58.729370 4791 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-vlwn5\" (UniqueName: \"kubernetes.io/projected/ceb2c559-f4b0-4176-bec1-8593df45d701-kube-api-access-vlwn5\") pod \"redhat-marketplace-dlskm\" (UID: \"ceb2c559-f4b0-4176-bec1-8593df45d701\") " pod="openshift-marketplace/redhat-marketplace-dlskm" Dec 10 23:58:58 crc kubenswrapper[4791]: I1210 23:58:58.729740 4791 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/ceb2c559-f4b0-4176-bec1-8593df45d701-catalog-content\") pod \"redhat-marketplace-dlskm\" (UID: \"ceb2c559-f4b0-4176-bec1-8593df45d701\") " pod="openshift-marketplace/redhat-marketplace-dlskm" Dec 10 23:58:58 crc kubenswrapper[4791]: I1210 23:58:58.729736 4791 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/ceb2c559-f4b0-4176-bec1-8593df45d701-utilities\") pod \"redhat-marketplace-dlskm\" (UID: \"ceb2c559-f4b0-4176-bec1-8593df45d701\") " pod="openshift-marketplace/redhat-marketplace-dlskm" Dec 10 23:58:58 crc kubenswrapper[4791]: I1210 23:58:58.748239 4791 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-vlwn5\" (UniqueName: \"kubernetes.io/projected/ceb2c559-f4b0-4176-bec1-8593df45d701-kube-api-access-vlwn5\") pod \"redhat-marketplace-dlskm\" (UID: \"ceb2c559-f4b0-4176-bec1-8593df45d701\") " pod="openshift-marketplace/redhat-marketplace-dlskm" Dec 10 23:58:58 crc kubenswrapper[4791]: I1210 23:58:58.822048 4791 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-dlskm" Dec 10 23:58:59 crc kubenswrapper[4791]: I1210 23:58:59.295818 4791 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-dlskm"] Dec 10 23:58:59 crc kubenswrapper[4791]: I1210 23:58:59.404657 4791 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-dlskm" event={"ID":"ceb2c559-f4b0-4176-bec1-8593df45d701","Type":"ContainerStarted","Data":"67ebf486ffb65360086b5b3a466beb705d506c3bf2ab40b6ce171d091138fdd8"} Dec 10 23:59:00 crc kubenswrapper[4791]: I1210 23:59:00.433998 4791 generic.go:334] "Generic (PLEG): container finished" podID="ceb2c559-f4b0-4176-bec1-8593df45d701" containerID="e552eb49d2997f23bf3a37553872ff0f4610d226bc5028c5dd98929046a25921" exitCode=0 Dec 10 23:59:00 crc kubenswrapper[4791]: I1210 23:59:00.434478 4791 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-dlskm" event={"ID":"ceb2c559-f4b0-4176-bec1-8593df45d701","Type":"ContainerDied","Data":"e552eb49d2997f23bf3a37553872ff0f4610d226bc5028c5dd98929046a25921"} Dec 10 23:59:00 crc kubenswrapper[4791]: I1210 23:59:00.440866 4791 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Dec 10 23:59:01 crc kubenswrapper[4791]: I1210 23:59:01.444830 4791 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-dlskm" event={"ID":"ceb2c559-f4b0-4176-bec1-8593df45d701","Type":"ContainerStarted","Data":"ab85b078fca3dc15e4d9db28fd30a8e088890ce8ca53e486265b3a185bf390aa"} Dec 10 23:59:02 crc kubenswrapper[4791]: I1210 23:59:02.458830 4791 generic.go:334] "Generic (PLEG): container finished" podID="ceb2c559-f4b0-4176-bec1-8593df45d701" containerID="ab85b078fca3dc15e4d9db28fd30a8e088890ce8ca53e486265b3a185bf390aa" exitCode=0 Dec 10 23:59:02 crc kubenswrapper[4791]: I1210 23:59:02.458932 4791 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-dlskm" event={"ID":"ceb2c559-f4b0-4176-bec1-8593df45d701","Type":"ContainerDied","Data":"ab85b078fca3dc15e4d9db28fd30a8e088890ce8ca53e486265b3a185bf390aa"} Dec 10 23:59:03 crc kubenswrapper[4791]: I1210 23:59:03.472107 4791 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-dlskm" event={"ID":"ceb2c559-f4b0-4176-bec1-8593df45d701","Type":"ContainerStarted","Data":"57f27dc686b512de8ce34ec4890dc880c4c6725b15baee4c75d9547527e26dcf"} Dec 10 23:59:03 crc kubenswrapper[4791]: I1210 23:59:03.498451 4791 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-marketplace-dlskm" podStartSLOduration=2.870709392 podStartE2EDuration="5.498327173s" podCreationTimestamp="2025-12-10 23:58:58 +0000 UTC" firstStartedPulling="2025-12-10 23:59:00.440547813 +0000 UTC m=+4174.870165426" lastFinishedPulling="2025-12-10 23:59:03.068165574 +0000 UTC m=+4177.497783207" observedRunningTime="2025-12-10 23:59:03.494009852 +0000 UTC m=+4177.923627465" watchObservedRunningTime="2025-12-10 23:59:03.498327173 +0000 UTC m=+4177.927944796" Dec 10 23:59:06 crc kubenswrapper[4791]: I1210 23:59:06.600142 4791 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_5064f9f8917b246f69f5d7fc025e7e6c34236c02bca31167615d38212fppnkc_4ebf6fba-353e-4d14-9e9e-f2be85033920/util/0.log" Dec 10 23:59:06 crc kubenswrapper[4791]: I1210 23:59:06.816544 4791 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_5064f9f8917b246f69f5d7fc025e7e6c34236c02bca31167615d38212fppnkc_4ebf6fba-353e-4d14-9e9e-f2be85033920/pull/0.log" Dec 10 23:59:06 crc kubenswrapper[4791]: I1210 23:59:06.850504 4791 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_5064f9f8917b246f69f5d7fc025e7e6c34236c02bca31167615d38212fppnkc_4ebf6fba-353e-4d14-9e9e-f2be85033920/pull/0.log" Dec 10 23:59:06 crc kubenswrapper[4791]: I1210 23:59:06.921300 4791 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_5064f9f8917b246f69f5d7fc025e7e6c34236c02bca31167615d38212fppnkc_4ebf6fba-353e-4d14-9e9e-f2be85033920/util/0.log" Dec 10 23:59:07 crc kubenswrapper[4791]: I1210 23:59:07.025294 4791 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_5064f9f8917b246f69f5d7fc025e7e6c34236c02bca31167615d38212fppnkc_4ebf6fba-353e-4d14-9e9e-f2be85033920/util/0.log" Dec 10 23:59:07 crc kubenswrapper[4791]: I1210 23:59:07.026957 4791 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_5064f9f8917b246f69f5d7fc025e7e6c34236c02bca31167615d38212fppnkc_4ebf6fba-353e-4d14-9e9e-f2be85033920/pull/0.log" Dec 10 23:59:07 crc kubenswrapper[4791]: I1210 23:59:07.078528 4791 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_5064f9f8917b246f69f5d7fc025e7e6c34236c02bca31167615d38212fppnkc_4ebf6fba-353e-4d14-9e9e-f2be85033920/extract/0.log" Dec 10 23:59:07 crc kubenswrapper[4791]: I1210 23:59:07.194652 4791 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_af69698b82fdf008f5ff9e195cf808a654240e16b13dcd924b74994f83xxh8j_d1f6824d-b614-4d99-a538-2917ad1c4bb3/util/0.log" Dec 10 23:59:07 crc kubenswrapper[4791]: I1210 23:59:07.370807 4791 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_af69698b82fdf008f5ff9e195cf808a654240e16b13dcd924b74994f83xxh8j_d1f6824d-b614-4d99-a538-2917ad1c4bb3/pull/0.log" Dec 10 23:59:07 crc kubenswrapper[4791]: I1210 23:59:07.391117 4791 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_af69698b82fdf008f5ff9e195cf808a654240e16b13dcd924b74994f83xxh8j_d1f6824d-b614-4d99-a538-2917ad1c4bb3/util/0.log" Dec 10 23:59:07 crc kubenswrapper[4791]: I1210 23:59:07.396564 4791 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_af69698b82fdf008f5ff9e195cf808a654240e16b13dcd924b74994f83xxh8j_d1f6824d-b614-4d99-a538-2917ad1c4bb3/pull/0.log" Dec 10 23:59:07 crc kubenswrapper[4791]: I1210 23:59:07.619235 4791 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_af69698b82fdf008f5ff9e195cf808a654240e16b13dcd924b74994f83xxh8j_d1f6824d-b614-4d99-a538-2917ad1c4bb3/util/0.log" Dec 10 23:59:07 crc kubenswrapper[4791]: I1210 23:59:07.626615 4791 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_af69698b82fdf008f5ff9e195cf808a654240e16b13dcd924b74994f83xxh8j_d1f6824d-b614-4d99-a538-2917ad1c4bb3/extract/0.log" Dec 10 23:59:07 crc kubenswrapper[4791]: I1210 23:59:07.627145 4791 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_af69698b82fdf008f5ff9e195cf808a654240e16b13dcd924b74994f83xxh8j_d1f6824d-b614-4d99-a538-2917ad1c4bb3/pull/0.log" Dec 10 23:59:07 crc kubenswrapper[4791]: I1210 23:59:07.764148 4791 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_certified-operators-cmz9s_2b323bf3-c265-4df1-969d-eb4d52f13929/extract-utilities/0.log" Dec 10 23:59:07 crc kubenswrapper[4791]: I1210 23:59:07.931494 4791 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_certified-operators-cmz9s_2b323bf3-c265-4df1-969d-eb4d52f13929/extract-content/0.log" Dec 10 23:59:07 crc kubenswrapper[4791]: I1210 23:59:07.932538 4791 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_certified-operators-cmz9s_2b323bf3-c265-4df1-969d-eb4d52f13929/extract-utilities/0.log" Dec 10 23:59:07 crc kubenswrapper[4791]: I1210 23:59:07.972179 4791 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_certified-operators-cmz9s_2b323bf3-c265-4df1-969d-eb4d52f13929/extract-content/0.log" Dec 10 23:59:08 crc kubenswrapper[4791]: I1210 23:59:08.138242 4791 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_certified-operators-cmz9s_2b323bf3-c265-4df1-969d-eb4d52f13929/extract-utilities/0.log" Dec 10 23:59:08 crc kubenswrapper[4791]: I1210 23:59:08.151645 4791 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_certified-operators-cmz9s_2b323bf3-c265-4df1-969d-eb4d52f13929/extract-content/0.log" Dec 10 23:59:08 crc kubenswrapper[4791]: I1210 23:59:08.240951 4791 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_certified-operators-cmz9s_2b323bf3-c265-4df1-969d-eb4d52f13929/registry-server/0.log" Dec 10 23:59:08 crc kubenswrapper[4791]: I1210 23:59:08.325406 4791 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_community-operators-x582q_d02b36c2-1b51-4466-af31-2e69357acc9c/extract-utilities/0.log" Dec 10 23:59:08 crc kubenswrapper[4791]: I1210 23:59:08.602699 4791 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_community-operators-x582q_d02b36c2-1b51-4466-af31-2e69357acc9c/extract-utilities/0.log" Dec 10 23:59:08 crc kubenswrapper[4791]: I1210 23:59:08.635399 4791 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_community-operators-x582q_d02b36c2-1b51-4466-af31-2e69357acc9c/extract-content/0.log" Dec 10 23:59:08 crc kubenswrapper[4791]: I1210 23:59:08.659058 4791 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_community-operators-x582q_d02b36c2-1b51-4466-af31-2e69357acc9c/extract-content/0.log" Dec 10 23:59:08 crc kubenswrapper[4791]: I1210 23:59:08.800765 4791 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_community-operators-x582q_d02b36c2-1b51-4466-af31-2e69357acc9c/extract-utilities/0.log" Dec 10 23:59:08 crc kubenswrapper[4791]: I1210 23:59:08.808211 4791 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_community-operators-x582q_d02b36c2-1b51-4466-af31-2e69357acc9c/extract-content/0.log" Dec 10 23:59:08 crc kubenswrapper[4791]: I1210 23:59:08.822786 4791 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-marketplace-dlskm" Dec 10 23:59:08 crc kubenswrapper[4791]: I1210 23:59:08.822834 4791 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-marketplace-dlskm" Dec 10 23:59:08 crc kubenswrapper[4791]: I1210 23:59:08.950111 4791 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-marketplace-dlskm" Dec 10 23:59:09 crc kubenswrapper[4791]: I1210 23:59:09.057755 4791 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_marketplace-operator-79b997595-8b9dm_c2479901-785c-4a0a-b494-90675f64b71d/marketplace-operator/0.log" Dec 10 23:59:09 crc kubenswrapper[4791]: I1210 23:59:09.201222 4791 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-marketplace-dlskm_ceb2c559-f4b0-4176-bec1-8593df45d701/extract-utilities/0.log" Dec 10 23:59:09 crc kubenswrapper[4791]: I1210 23:59:09.332148 4791 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-marketplace-dlskm_ceb2c559-f4b0-4176-bec1-8593df45d701/extract-utilities/0.log" Dec 10 23:59:09 crc kubenswrapper[4791]: I1210 23:59:09.400722 4791 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-marketplace-dlskm_ceb2c559-f4b0-4176-bec1-8593df45d701/extract-content/0.log" Dec 10 23:59:09 crc kubenswrapper[4791]: I1210 23:59:09.441173 4791 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-marketplace-dlskm_ceb2c559-f4b0-4176-bec1-8593df45d701/extract-content/0.log" Dec 10 23:59:09 crc kubenswrapper[4791]: I1210 23:59:09.612032 4791 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-marketplace-dlskm" Dec 10 23:59:09 crc kubenswrapper[4791]: I1210 23:59:09.666563 4791 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-dlskm"] Dec 10 23:59:09 crc kubenswrapper[4791]: I1210 23:59:09.683298 4791 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_community-operators-x582q_d02b36c2-1b51-4466-af31-2e69357acc9c/registry-server/0.log" Dec 10 23:59:09 crc kubenswrapper[4791]: I1210 23:59:09.699750 4791 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-marketplace-dlskm_ceb2c559-f4b0-4176-bec1-8593df45d701/registry-server/0.log" Dec 10 23:59:09 crc kubenswrapper[4791]: I1210 23:59:09.727367 4791 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-marketplace-dlskm_ceb2c559-f4b0-4176-bec1-8593df45d701/extract-content/0.log" Dec 10 23:59:09 crc kubenswrapper[4791]: I1210 23:59:09.733075 4791 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-marketplace-dlskm_ceb2c559-f4b0-4176-bec1-8593df45d701/extract-utilities/0.log" Dec 10 23:59:09 crc kubenswrapper[4791]: I1210 23:59:09.865963 4791 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-marketplace-fcnzr_44085f7f-f0e9-4dec-a816-a687a0dd06ec/extract-utilities/0.log" Dec 10 23:59:10 crc kubenswrapper[4791]: I1210 23:59:10.020131 4791 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-marketplace-fcnzr_44085f7f-f0e9-4dec-a816-a687a0dd06ec/extract-content/0.log" Dec 10 23:59:10 crc kubenswrapper[4791]: I1210 23:59:10.020224 4791 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-marketplace-fcnzr_44085f7f-f0e9-4dec-a816-a687a0dd06ec/extract-content/0.log" Dec 10 23:59:10 crc kubenswrapper[4791]: I1210 23:59:10.035730 4791 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-marketplace-fcnzr_44085f7f-f0e9-4dec-a816-a687a0dd06ec/extract-utilities/0.log" Dec 10 23:59:10 crc kubenswrapper[4791]: I1210 23:59:10.248492 4791 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-marketplace-fcnzr_44085f7f-f0e9-4dec-a816-a687a0dd06ec/extract-utilities/0.log" Dec 10 23:59:10 crc kubenswrapper[4791]: I1210 23:59:10.265404 4791 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-marketplace-fcnzr_44085f7f-f0e9-4dec-a816-a687a0dd06ec/extract-content/0.log" Dec 10 23:59:10 crc kubenswrapper[4791]: I1210 23:59:10.307973 4791 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-operators-f8n6m_a68872ea-89d3-42a0-b2da-877e2d37218b/extract-utilities/0.log" Dec 10 23:59:10 crc kubenswrapper[4791]: I1210 23:59:10.432283 4791 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-marketplace-fcnzr_44085f7f-f0e9-4dec-a816-a687a0dd06ec/registry-server/0.log" Dec 10 23:59:10 crc kubenswrapper[4791]: I1210 23:59:10.459290 4791 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-operators-f8n6m_a68872ea-89d3-42a0-b2da-877e2d37218b/extract-content/0.log" Dec 10 23:59:10 crc kubenswrapper[4791]: I1210 23:59:10.498732 4791 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-operators-f8n6m_a68872ea-89d3-42a0-b2da-877e2d37218b/extract-content/0.log" Dec 10 23:59:10 crc kubenswrapper[4791]: I1210 23:59:10.528216 4791 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-operators-f8n6m_a68872ea-89d3-42a0-b2da-877e2d37218b/extract-utilities/0.log" Dec 10 23:59:10 crc kubenswrapper[4791]: I1210 23:59:10.667650 4791 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-operators-f8n6m_a68872ea-89d3-42a0-b2da-877e2d37218b/extract-utilities/0.log" Dec 10 23:59:10 crc kubenswrapper[4791]: I1210 23:59:10.704937 4791 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-operators-f8n6m_a68872ea-89d3-42a0-b2da-877e2d37218b/extract-content/0.log" Dec 10 23:59:11 crc kubenswrapper[4791]: I1210 23:59:11.214588 4791 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-operators-f8n6m_a68872ea-89d3-42a0-b2da-877e2d37218b/registry-server/0.log" Dec 10 23:59:11 crc kubenswrapper[4791]: I1210 23:59:11.571662 4791 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-marketplace-dlskm" podUID="ceb2c559-f4b0-4176-bec1-8593df45d701" containerName="registry-server" containerID="cri-o://57f27dc686b512de8ce34ec4890dc880c4c6725b15baee4c75d9547527e26dcf" gracePeriod=2 Dec 10 23:59:12 crc kubenswrapper[4791]: I1210 23:59:12.584061 4791 generic.go:334] "Generic (PLEG): container finished" podID="ceb2c559-f4b0-4176-bec1-8593df45d701" containerID="57f27dc686b512de8ce34ec4890dc880c4c6725b15baee4c75d9547527e26dcf" exitCode=0 Dec 10 23:59:12 crc kubenswrapper[4791]: I1210 23:59:12.584368 4791 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-dlskm" event={"ID":"ceb2c559-f4b0-4176-bec1-8593df45d701","Type":"ContainerDied","Data":"57f27dc686b512de8ce34ec4890dc880c4c6725b15baee4c75d9547527e26dcf"} Dec 10 23:59:12 crc kubenswrapper[4791]: I1210 23:59:12.694572 4791 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-dlskm" Dec 10 23:59:12 crc kubenswrapper[4791]: I1210 23:59:12.698756 4791 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/ceb2c559-f4b0-4176-bec1-8593df45d701-utilities\") pod \"ceb2c559-f4b0-4176-bec1-8593df45d701\" (UID: \"ceb2c559-f4b0-4176-bec1-8593df45d701\") " Dec 10 23:59:12 crc kubenswrapper[4791]: I1210 23:59:12.699084 4791 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/ceb2c559-f4b0-4176-bec1-8593df45d701-catalog-content\") pod \"ceb2c559-f4b0-4176-bec1-8593df45d701\" (UID: \"ceb2c559-f4b0-4176-bec1-8593df45d701\") " Dec 10 23:59:12 crc kubenswrapper[4791]: I1210 23:59:12.699134 4791 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-vlwn5\" (UniqueName: \"kubernetes.io/projected/ceb2c559-f4b0-4176-bec1-8593df45d701-kube-api-access-vlwn5\") pod \"ceb2c559-f4b0-4176-bec1-8593df45d701\" (UID: \"ceb2c559-f4b0-4176-bec1-8593df45d701\") " Dec 10 23:59:12 crc kubenswrapper[4791]: I1210 23:59:12.699822 4791 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/ceb2c559-f4b0-4176-bec1-8593df45d701-utilities" (OuterVolumeSpecName: "utilities") pod "ceb2c559-f4b0-4176-bec1-8593df45d701" (UID: "ceb2c559-f4b0-4176-bec1-8593df45d701"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 10 23:59:12 crc kubenswrapper[4791]: I1210 23:59:12.704301 4791 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/ceb2c559-f4b0-4176-bec1-8593df45d701-kube-api-access-vlwn5" (OuterVolumeSpecName: "kube-api-access-vlwn5") pod "ceb2c559-f4b0-4176-bec1-8593df45d701" (UID: "ceb2c559-f4b0-4176-bec1-8593df45d701"). InnerVolumeSpecName "kube-api-access-vlwn5". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 10 23:59:12 crc kubenswrapper[4791]: I1210 23:59:12.729798 4791 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/ceb2c559-f4b0-4176-bec1-8593df45d701-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "ceb2c559-f4b0-4176-bec1-8593df45d701" (UID: "ceb2c559-f4b0-4176-bec1-8593df45d701"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 10 23:59:12 crc kubenswrapper[4791]: I1210 23:59:12.801525 4791 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/ceb2c559-f4b0-4176-bec1-8593df45d701-catalog-content\") on node \"crc\" DevicePath \"\"" Dec 10 23:59:12 crc kubenswrapper[4791]: I1210 23:59:12.801600 4791 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-vlwn5\" (UniqueName: \"kubernetes.io/projected/ceb2c559-f4b0-4176-bec1-8593df45d701-kube-api-access-vlwn5\") on node \"crc\" DevicePath \"\"" Dec 10 23:59:12 crc kubenswrapper[4791]: I1210 23:59:12.801619 4791 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/ceb2c559-f4b0-4176-bec1-8593df45d701-utilities\") on node \"crc\" DevicePath \"\"" Dec 10 23:59:13 crc kubenswrapper[4791]: I1210 23:59:13.596327 4791 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-dlskm" event={"ID":"ceb2c559-f4b0-4176-bec1-8593df45d701","Type":"ContainerDied","Data":"67ebf486ffb65360086b5b3a466beb705d506c3bf2ab40b6ce171d091138fdd8"} Dec 10 23:59:13 crc kubenswrapper[4791]: I1210 23:59:13.596486 4791 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-dlskm" Dec 10 23:59:13 crc kubenswrapper[4791]: I1210 23:59:13.596646 4791 scope.go:117] "RemoveContainer" containerID="57f27dc686b512de8ce34ec4890dc880c4c6725b15baee4c75d9547527e26dcf" Dec 10 23:59:13 crc kubenswrapper[4791]: I1210 23:59:13.622955 4791 scope.go:117] "RemoveContainer" containerID="ab85b078fca3dc15e4d9db28fd30a8e088890ce8ca53e486265b3a185bf390aa" Dec 10 23:59:13 crc kubenswrapper[4791]: I1210 23:59:13.744191 4791 scope.go:117] "RemoveContainer" containerID="e552eb49d2997f23bf3a37553872ff0f4610d226bc5028c5dd98929046a25921" Dec 10 23:59:13 crc kubenswrapper[4791]: I1210 23:59:13.746625 4791 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-dlskm"] Dec 10 23:59:13 crc kubenswrapper[4791]: I1210 23:59:13.754166 4791 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-marketplace-dlskm"] Dec 10 23:59:13 crc kubenswrapper[4791]: I1210 23:59:13.896298 4791 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="ceb2c559-f4b0-4176-bec1-8593df45d701" path="/var/lib/kubelet/pods/ceb2c559-f4b0-4176-bec1-8593df45d701/volumes" Dec 10 23:59:27 crc kubenswrapper[4791]: I1210 23:59:27.781925 4791 prober.go:107] "Probe failed" probeType="Liveness" pod="openstack/openstack-cell1-galera-0" podUID="f04fa829-3ebe-445d-a4e5-f7592ac682f3" containerName="galera" probeResult="failure" output="command timed out" Dec 10 23:59:37 crc kubenswrapper[4791]: E1210 23:59:37.914589 4791 upgradeaware.go:427] Error proxying data from client to backend: readfrom tcp 38.102.83.30:55894->38.102.83.30:41701: write tcp 38.102.83.30:55894->38.102.83.30:41701: write: broken pipe Dec 11 00:00:00 crc kubenswrapper[4791]: I1211 00:00:00.194285 4791 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29423520-dkcj9"] Dec 11 00:00:00 crc kubenswrapper[4791]: E1211 00:00:00.198259 4791 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="ceb2c559-f4b0-4176-bec1-8593df45d701" containerName="registry-server" Dec 11 00:00:00 crc kubenswrapper[4791]: I1211 00:00:00.198293 4791 state_mem.go:107] "Deleted CPUSet assignment" podUID="ceb2c559-f4b0-4176-bec1-8593df45d701" containerName="registry-server" Dec 11 00:00:00 crc kubenswrapper[4791]: E1211 00:00:00.198373 4791 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="ceb2c559-f4b0-4176-bec1-8593df45d701" containerName="extract-utilities" Dec 11 00:00:00 crc kubenswrapper[4791]: I1211 00:00:00.198385 4791 state_mem.go:107] "Deleted CPUSet assignment" podUID="ceb2c559-f4b0-4176-bec1-8593df45d701" containerName="extract-utilities" Dec 11 00:00:00 crc kubenswrapper[4791]: E1211 00:00:00.198398 4791 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="ceb2c559-f4b0-4176-bec1-8593df45d701" containerName="extract-content" Dec 11 00:00:00 crc kubenswrapper[4791]: I1211 00:00:00.198406 4791 state_mem.go:107] "Deleted CPUSet assignment" podUID="ceb2c559-f4b0-4176-bec1-8593df45d701" containerName="extract-content" Dec 11 00:00:00 crc kubenswrapper[4791]: I1211 00:00:00.198922 4791 memory_manager.go:354] "RemoveStaleState removing state" podUID="ceb2c559-f4b0-4176-bec1-8593df45d701" containerName="registry-server" Dec 11 00:00:00 crc kubenswrapper[4791]: I1211 00:00:00.200086 4791 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29423520-dkcj9" Dec 11 00:00:00 crc kubenswrapper[4791]: I1211 00:00:00.204212 4791 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"collect-profiles-dockercfg-kzf4t" Dec 11 00:00:00 crc kubenswrapper[4791]: I1211 00:00:00.204406 4791 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-operator-lifecycle-manager"/"collect-profiles-config" Dec 11 00:00:00 crc kubenswrapper[4791]: I1211 00:00:00.212018 4791 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-cell1-db-purge-29423520-zzzkt"] Dec 11 00:00:00 crc kubenswrapper[4791]: I1211 00:00:00.213461 4791 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-db-purge-29423520-zzzkt" Dec 11 00:00:00 crc kubenswrapper[4791]: I1211 00:00:00.216668 4791 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell1-conductor-scripts" Dec 11 00:00:00 crc kubenswrapper[4791]: I1211 00:00:00.220618 4791 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-cell0-db-purge-29423520-vv2hp"] Dec 11 00:00:00 crc kubenswrapper[4791]: I1211 00:00:00.222327 4791 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-db-purge-29423520-vv2hp" Dec 11 00:00:00 crc kubenswrapper[4791]: I1211 00:00:00.225292 4791 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell0-conductor-scripts" Dec 11 00:00:00 crc kubenswrapper[4791]: I1211 00:00:00.255072 4791 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-image-registry/image-pruner-29423520-ghc92"] Dec 11 00:00:00 crc kubenswrapper[4791]: I1211 00:00:00.257214 4791 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-image-registry/image-pruner-29423520-ghc92" Dec 11 00:00:00 crc kubenswrapper[4791]: I1211 00:00:00.266908 4791 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-image-registry"/"pruner-dockercfg-p7bcw" Dec 11 00:00:00 crc kubenswrapper[4791]: I1211 00:00:00.270198 4791 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/d46e7c67-8829-47d6-909e-bcdbe292f092-config-volume\") pod \"collect-profiles-29423520-dkcj9\" (UID: \"d46e7c67-8829-47d6-909e-bcdbe292f092\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29423520-dkcj9" Dec 11 00:00:00 crc kubenswrapper[4791]: I1211 00:00:00.270235 4791 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-m4584\" (UniqueName: \"kubernetes.io/projected/9f973940-a85e-4ee3-9373-d1c7512d9f1a-kube-api-access-m4584\") pod \"nova-cell1-db-purge-29423520-zzzkt\" (UID: \"9f973940-a85e-4ee3-9373-d1c7512d9f1a\") " pod="openstack/nova-cell1-db-purge-29423520-zzzkt" Dec 11 00:00:00 crc kubenswrapper[4791]: I1211 00:00:00.270267 4791 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/9f973940-a85e-4ee3-9373-d1c7512d9f1a-combined-ca-bundle\") pod \"nova-cell1-db-purge-29423520-zzzkt\" (UID: \"9f973940-a85e-4ee3-9373-d1c7512d9f1a\") " pod="openstack/nova-cell1-db-purge-29423520-zzzkt" Dec 11 00:00:00 crc kubenswrapper[4791]: I1211 00:00:00.270286 4791 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-26p8v\" (UniqueName: \"kubernetes.io/projected/5309e85e-16c8-46fb-b694-1e16db7d7770-kube-api-access-26p8v\") pod \"nova-cell0-db-purge-29423520-vv2hp\" (UID: \"5309e85e-16c8-46fb-b694-1e16db7d7770\") " pod="openstack/nova-cell0-db-purge-29423520-vv2hp" Dec 11 00:00:00 crc kubenswrapper[4791]: I1211 00:00:00.270304 4791 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serviceca\" (UniqueName: \"kubernetes.io/configmap/11aa2b84-ba50-48f2-ab7a-0591bf6c3f44-serviceca\") pod \"image-pruner-29423520-ghc92\" (UID: \"11aa2b84-ba50-48f2-ab7a-0591bf6c3f44\") " pod="openshift-image-registry/image-pruner-29423520-ghc92" Dec 11 00:00:00 crc kubenswrapper[4791]: I1211 00:00:00.270322 4791 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-8pgk5\" (UniqueName: \"kubernetes.io/projected/11aa2b84-ba50-48f2-ab7a-0591bf6c3f44-kube-api-access-8pgk5\") pod \"image-pruner-29423520-ghc92\" (UID: \"11aa2b84-ba50-48f2-ab7a-0591bf6c3f44\") " pod="openshift-image-registry/image-pruner-29423520-ghc92" Dec 11 00:00:00 crc kubenswrapper[4791]: I1211 00:00:00.270360 4791 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/9f973940-a85e-4ee3-9373-d1c7512d9f1a-config-data\") pod \"nova-cell1-db-purge-29423520-zzzkt\" (UID: \"9f973940-a85e-4ee3-9373-d1c7512d9f1a\") " pod="openstack/nova-cell1-db-purge-29423520-zzzkt" Dec 11 00:00:00 crc kubenswrapper[4791]: I1211 00:00:00.270439 4791 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/d46e7c67-8829-47d6-909e-bcdbe292f092-secret-volume\") pod \"collect-profiles-29423520-dkcj9\" (UID: \"d46e7c67-8829-47d6-909e-bcdbe292f092\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29423520-dkcj9" Dec 11 00:00:00 crc kubenswrapper[4791]: I1211 00:00:00.270464 4791 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/5309e85e-16c8-46fb-b694-1e16db7d7770-scripts\") pod \"nova-cell0-db-purge-29423520-vv2hp\" (UID: \"5309e85e-16c8-46fb-b694-1e16db7d7770\") " pod="openstack/nova-cell0-db-purge-29423520-vv2hp" Dec 11 00:00:00 crc kubenswrapper[4791]: I1211 00:00:00.270517 4791 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/5309e85e-16c8-46fb-b694-1e16db7d7770-combined-ca-bundle\") pod \"nova-cell0-db-purge-29423520-vv2hp\" (UID: \"5309e85e-16c8-46fb-b694-1e16db7d7770\") " pod="openstack/nova-cell0-db-purge-29423520-vv2hp" Dec 11 00:00:00 crc kubenswrapper[4791]: I1211 00:00:00.270541 4791 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/9f973940-a85e-4ee3-9373-d1c7512d9f1a-scripts\") pod \"nova-cell1-db-purge-29423520-zzzkt\" (UID: \"9f973940-a85e-4ee3-9373-d1c7512d9f1a\") " pod="openstack/nova-cell1-db-purge-29423520-zzzkt" Dec 11 00:00:00 crc kubenswrapper[4791]: I1211 00:00:00.270577 4791 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/5309e85e-16c8-46fb-b694-1e16db7d7770-config-data\") pod \"nova-cell0-db-purge-29423520-vv2hp\" (UID: \"5309e85e-16c8-46fb-b694-1e16db7d7770\") " pod="openstack/nova-cell0-db-purge-29423520-vv2hp" Dec 11 00:00:00 crc kubenswrapper[4791]: I1211 00:00:00.270610 4791 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-f6dlk\" (UniqueName: \"kubernetes.io/projected/d46e7c67-8829-47d6-909e-bcdbe292f092-kube-api-access-f6dlk\") pod \"collect-profiles-29423520-dkcj9\" (UID: \"d46e7c67-8829-47d6-909e-bcdbe292f092\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29423520-dkcj9" Dec 11 00:00:00 crc kubenswrapper[4791]: I1211 00:00:00.271014 4791 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-image-registry"/"serviceca" Dec 11 00:00:00 crc kubenswrapper[4791]: I1211 00:00:00.284805 4791 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell0-db-purge-29423520-vv2hp"] Dec 11 00:00:00 crc kubenswrapper[4791]: I1211 00:00:00.295104 4791 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-image-registry/image-pruner-29423520-ghc92"] Dec 11 00:00:00 crc kubenswrapper[4791]: I1211 00:00:00.308540 4791 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29423520-dkcj9"] Dec 11 00:00:00 crc kubenswrapper[4791]: I1211 00:00:00.317870 4791 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-db-purge-29423520-zzzkt"] Dec 11 00:00:00 crc kubenswrapper[4791]: I1211 00:00:00.371912 4791 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/d46e7c67-8829-47d6-909e-bcdbe292f092-config-volume\") pod \"collect-profiles-29423520-dkcj9\" (UID: \"d46e7c67-8829-47d6-909e-bcdbe292f092\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29423520-dkcj9" Dec 11 00:00:00 crc kubenswrapper[4791]: I1211 00:00:00.372281 4791 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-m4584\" (UniqueName: \"kubernetes.io/projected/9f973940-a85e-4ee3-9373-d1c7512d9f1a-kube-api-access-m4584\") pod \"nova-cell1-db-purge-29423520-zzzkt\" (UID: \"9f973940-a85e-4ee3-9373-d1c7512d9f1a\") " pod="openstack/nova-cell1-db-purge-29423520-zzzkt" Dec 11 00:00:00 crc kubenswrapper[4791]: I1211 00:00:00.372489 4791 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/9f973940-a85e-4ee3-9373-d1c7512d9f1a-combined-ca-bundle\") pod \"nova-cell1-db-purge-29423520-zzzkt\" (UID: \"9f973940-a85e-4ee3-9373-d1c7512d9f1a\") " pod="openstack/nova-cell1-db-purge-29423520-zzzkt" Dec 11 00:00:00 crc kubenswrapper[4791]: I1211 00:00:00.372605 4791 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-26p8v\" (UniqueName: \"kubernetes.io/projected/5309e85e-16c8-46fb-b694-1e16db7d7770-kube-api-access-26p8v\") pod \"nova-cell0-db-purge-29423520-vv2hp\" (UID: \"5309e85e-16c8-46fb-b694-1e16db7d7770\") " pod="openstack/nova-cell0-db-purge-29423520-vv2hp" Dec 11 00:00:00 crc kubenswrapper[4791]: I1211 00:00:00.372750 4791 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serviceca\" (UniqueName: \"kubernetes.io/configmap/11aa2b84-ba50-48f2-ab7a-0591bf6c3f44-serviceca\") pod \"image-pruner-29423520-ghc92\" (UID: \"11aa2b84-ba50-48f2-ab7a-0591bf6c3f44\") " pod="openshift-image-registry/image-pruner-29423520-ghc92" Dec 11 00:00:00 crc kubenswrapper[4791]: I1211 00:00:00.372863 4791 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-8pgk5\" (UniqueName: \"kubernetes.io/projected/11aa2b84-ba50-48f2-ab7a-0591bf6c3f44-kube-api-access-8pgk5\") pod \"image-pruner-29423520-ghc92\" (UID: \"11aa2b84-ba50-48f2-ab7a-0591bf6c3f44\") " pod="openshift-image-registry/image-pruner-29423520-ghc92" Dec 11 00:00:00 crc kubenswrapper[4791]: I1211 00:00:00.372982 4791 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/9f973940-a85e-4ee3-9373-d1c7512d9f1a-config-data\") pod \"nova-cell1-db-purge-29423520-zzzkt\" (UID: \"9f973940-a85e-4ee3-9373-d1c7512d9f1a\") " pod="openstack/nova-cell1-db-purge-29423520-zzzkt" Dec 11 00:00:00 crc kubenswrapper[4791]: I1211 00:00:00.373212 4791 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/d46e7c67-8829-47d6-909e-bcdbe292f092-secret-volume\") pod \"collect-profiles-29423520-dkcj9\" (UID: \"d46e7c67-8829-47d6-909e-bcdbe292f092\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29423520-dkcj9" Dec 11 00:00:00 crc kubenswrapper[4791]: I1211 00:00:00.373355 4791 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/5309e85e-16c8-46fb-b694-1e16db7d7770-scripts\") pod \"nova-cell0-db-purge-29423520-vv2hp\" (UID: \"5309e85e-16c8-46fb-b694-1e16db7d7770\") " pod="openstack/nova-cell0-db-purge-29423520-vv2hp" Dec 11 00:00:00 crc kubenswrapper[4791]: I1211 00:00:00.373541 4791 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/5309e85e-16c8-46fb-b694-1e16db7d7770-combined-ca-bundle\") pod \"nova-cell0-db-purge-29423520-vv2hp\" (UID: \"5309e85e-16c8-46fb-b694-1e16db7d7770\") " pod="openstack/nova-cell0-db-purge-29423520-vv2hp" Dec 11 00:00:00 crc kubenswrapper[4791]: I1211 00:00:00.373659 4791 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/9f973940-a85e-4ee3-9373-d1c7512d9f1a-scripts\") pod \"nova-cell1-db-purge-29423520-zzzkt\" (UID: \"9f973940-a85e-4ee3-9373-d1c7512d9f1a\") " pod="openstack/nova-cell1-db-purge-29423520-zzzkt" Dec 11 00:00:00 crc kubenswrapper[4791]: I1211 00:00:00.373806 4791 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/5309e85e-16c8-46fb-b694-1e16db7d7770-config-data\") pod \"nova-cell0-db-purge-29423520-vv2hp\" (UID: \"5309e85e-16c8-46fb-b694-1e16db7d7770\") " pod="openstack/nova-cell0-db-purge-29423520-vv2hp" Dec 11 00:00:00 crc kubenswrapper[4791]: I1211 00:00:00.373926 4791 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-f6dlk\" (UniqueName: \"kubernetes.io/projected/d46e7c67-8829-47d6-909e-bcdbe292f092-kube-api-access-f6dlk\") pod \"collect-profiles-29423520-dkcj9\" (UID: \"d46e7c67-8829-47d6-909e-bcdbe292f092\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29423520-dkcj9" Dec 11 00:00:00 crc kubenswrapper[4791]: I1211 00:00:00.373973 4791 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serviceca\" (UniqueName: \"kubernetes.io/configmap/11aa2b84-ba50-48f2-ab7a-0591bf6c3f44-serviceca\") pod \"image-pruner-29423520-ghc92\" (UID: \"11aa2b84-ba50-48f2-ab7a-0591bf6c3f44\") " pod="openshift-image-registry/image-pruner-29423520-ghc92" Dec 11 00:00:00 crc kubenswrapper[4791]: I1211 00:00:00.372882 4791 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/d46e7c67-8829-47d6-909e-bcdbe292f092-config-volume\") pod \"collect-profiles-29423520-dkcj9\" (UID: \"d46e7c67-8829-47d6-909e-bcdbe292f092\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29423520-dkcj9" Dec 11 00:00:00 crc kubenswrapper[4791]: I1211 00:00:00.380888 4791 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/9f973940-a85e-4ee3-9373-d1c7512d9f1a-combined-ca-bundle\") pod \"nova-cell1-db-purge-29423520-zzzkt\" (UID: \"9f973940-a85e-4ee3-9373-d1c7512d9f1a\") " pod="openstack/nova-cell1-db-purge-29423520-zzzkt" Dec 11 00:00:00 crc kubenswrapper[4791]: I1211 00:00:00.383277 4791 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/d46e7c67-8829-47d6-909e-bcdbe292f092-secret-volume\") pod \"collect-profiles-29423520-dkcj9\" (UID: \"d46e7c67-8829-47d6-909e-bcdbe292f092\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29423520-dkcj9" Dec 11 00:00:00 crc kubenswrapper[4791]: I1211 00:00:00.383498 4791 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/5309e85e-16c8-46fb-b694-1e16db7d7770-combined-ca-bundle\") pod \"nova-cell0-db-purge-29423520-vv2hp\" (UID: \"5309e85e-16c8-46fb-b694-1e16db7d7770\") " pod="openstack/nova-cell0-db-purge-29423520-vv2hp" Dec 11 00:00:00 crc kubenswrapper[4791]: I1211 00:00:00.383978 4791 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/5309e85e-16c8-46fb-b694-1e16db7d7770-scripts\") pod \"nova-cell0-db-purge-29423520-vv2hp\" (UID: \"5309e85e-16c8-46fb-b694-1e16db7d7770\") " pod="openstack/nova-cell0-db-purge-29423520-vv2hp" Dec 11 00:00:00 crc kubenswrapper[4791]: I1211 00:00:00.384150 4791 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/5309e85e-16c8-46fb-b694-1e16db7d7770-config-data\") pod \"nova-cell0-db-purge-29423520-vv2hp\" (UID: \"5309e85e-16c8-46fb-b694-1e16db7d7770\") " pod="openstack/nova-cell0-db-purge-29423520-vv2hp" Dec 11 00:00:00 crc kubenswrapper[4791]: I1211 00:00:00.386953 4791 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/9f973940-a85e-4ee3-9373-d1c7512d9f1a-scripts\") pod \"nova-cell1-db-purge-29423520-zzzkt\" (UID: \"9f973940-a85e-4ee3-9373-d1c7512d9f1a\") " pod="openstack/nova-cell1-db-purge-29423520-zzzkt" Dec 11 00:00:00 crc kubenswrapper[4791]: I1211 00:00:00.389781 4791 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-m4584\" (UniqueName: \"kubernetes.io/projected/9f973940-a85e-4ee3-9373-d1c7512d9f1a-kube-api-access-m4584\") pod \"nova-cell1-db-purge-29423520-zzzkt\" (UID: \"9f973940-a85e-4ee3-9373-d1c7512d9f1a\") " pod="openstack/nova-cell1-db-purge-29423520-zzzkt" Dec 11 00:00:00 crc kubenswrapper[4791]: I1211 00:00:00.394189 4791 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-8pgk5\" (UniqueName: \"kubernetes.io/projected/11aa2b84-ba50-48f2-ab7a-0591bf6c3f44-kube-api-access-8pgk5\") pod \"image-pruner-29423520-ghc92\" (UID: \"11aa2b84-ba50-48f2-ab7a-0591bf6c3f44\") " pod="openshift-image-registry/image-pruner-29423520-ghc92" Dec 11 00:00:00 crc kubenswrapper[4791]: I1211 00:00:00.395136 4791 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/9f973940-a85e-4ee3-9373-d1c7512d9f1a-config-data\") pod \"nova-cell1-db-purge-29423520-zzzkt\" (UID: \"9f973940-a85e-4ee3-9373-d1c7512d9f1a\") " pod="openstack/nova-cell1-db-purge-29423520-zzzkt" Dec 11 00:00:00 crc kubenswrapper[4791]: I1211 00:00:00.395993 4791 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-26p8v\" (UniqueName: \"kubernetes.io/projected/5309e85e-16c8-46fb-b694-1e16db7d7770-kube-api-access-26p8v\") pod \"nova-cell0-db-purge-29423520-vv2hp\" (UID: \"5309e85e-16c8-46fb-b694-1e16db7d7770\") " pod="openstack/nova-cell0-db-purge-29423520-vv2hp" Dec 11 00:00:00 crc kubenswrapper[4791]: I1211 00:00:00.396617 4791 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-f6dlk\" (UniqueName: \"kubernetes.io/projected/d46e7c67-8829-47d6-909e-bcdbe292f092-kube-api-access-f6dlk\") pod \"collect-profiles-29423520-dkcj9\" (UID: \"d46e7c67-8829-47d6-909e-bcdbe292f092\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29423520-dkcj9" Dec 11 00:00:00 crc kubenswrapper[4791]: I1211 00:00:00.567207 4791 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29423520-dkcj9" Dec 11 00:00:00 crc kubenswrapper[4791]: I1211 00:00:00.576902 4791 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-db-purge-29423520-zzzkt" Dec 11 00:00:00 crc kubenswrapper[4791]: I1211 00:00:00.601565 4791 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-db-purge-29423520-vv2hp" Dec 11 00:00:00 crc kubenswrapper[4791]: I1211 00:00:00.602744 4791 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-image-registry/image-pruner-29423520-ghc92" Dec 11 00:00:01 crc kubenswrapper[4791]: I1211 00:00:01.163468 4791 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-db-purge-29423520-zzzkt"] Dec 11 00:00:01 crc kubenswrapper[4791]: I1211 00:00:01.171963 4791 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29423520-dkcj9"] Dec 11 00:00:01 crc kubenswrapper[4791]: W1211 00:00:01.176551 4791 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod9f973940_a85e_4ee3_9373_d1c7512d9f1a.slice/crio-675d786841c45328890a4cb8f18561cf49cdfbe477808c2c51bb5a28fad81f94 WatchSource:0}: Error finding container 675d786841c45328890a4cb8f18561cf49cdfbe477808c2c51bb5a28fad81f94: Status 404 returned error can't find the container with id 675d786841c45328890a4cb8f18561cf49cdfbe477808c2c51bb5a28fad81f94 Dec 11 00:00:01 crc kubenswrapper[4791]: I1211 00:00:01.235493 4791 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29423520-dkcj9" event={"ID":"d46e7c67-8829-47d6-909e-bcdbe292f092","Type":"ContainerStarted","Data":"4aca408e4b7b57897d0053d58f3ccda350a8e0f12f1bdfc8ee6563e94431d532"} Dec 11 00:00:01 crc kubenswrapper[4791]: I1211 00:00:01.238863 4791 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-db-purge-29423520-zzzkt" event={"ID":"9f973940-a85e-4ee3-9373-d1c7512d9f1a","Type":"ContainerStarted","Data":"675d786841c45328890a4cb8f18561cf49cdfbe477808c2c51bb5a28fad81f94"} Dec 11 00:00:01 crc kubenswrapper[4791]: I1211 00:00:01.332535 4791 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell0-db-purge-29423520-vv2hp"] Dec 11 00:00:01 crc kubenswrapper[4791]: W1211 00:00:01.345550 4791 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod5309e85e_16c8_46fb_b694_1e16db7d7770.slice/crio-c7af93b04d512a5d3be1d13cff4917554d358f6246276e6ab0a5e685b890961d WatchSource:0}: Error finding container c7af93b04d512a5d3be1d13cff4917554d358f6246276e6ab0a5e685b890961d: Status 404 returned error can't find the container with id c7af93b04d512a5d3be1d13cff4917554d358f6246276e6ab0a5e685b890961d Dec 11 00:00:01 crc kubenswrapper[4791]: I1211 00:00:01.346901 4791 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-image-registry/image-pruner-29423520-ghc92"] Dec 11 00:00:01 crc kubenswrapper[4791]: W1211 00:00:01.351090 4791 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod11aa2b84_ba50_48f2_ab7a_0591bf6c3f44.slice/crio-fc0c67d9de0e1e76df4d106a0c86049359f47c5d1690f6eb00e9ac6f9bf04aae WatchSource:0}: Error finding container fc0c67d9de0e1e76df4d106a0c86049359f47c5d1690f6eb00e9ac6f9bf04aae: Status 404 returned error can't find the container with id fc0c67d9de0e1e76df4d106a0c86049359f47c5d1690f6eb00e9ac6f9bf04aae Dec 11 00:00:02 crc kubenswrapper[4791]: I1211 00:00:02.251138 4791 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-db-purge-29423520-vv2hp" event={"ID":"5309e85e-16c8-46fb-b694-1e16db7d7770","Type":"ContainerStarted","Data":"55ecbc8ec3c4fa0d09f052f8360db6057b37a513669a12346f1885202c9bec3c"} Dec 11 00:00:02 crc kubenswrapper[4791]: I1211 00:00:02.251538 4791 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-db-purge-29423520-vv2hp" event={"ID":"5309e85e-16c8-46fb-b694-1e16db7d7770","Type":"ContainerStarted","Data":"c7af93b04d512a5d3be1d13cff4917554d358f6246276e6ab0a5e685b890961d"} Dec 11 00:00:02 crc kubenswrapper[4791]: I1211 00:00:02.255005 4791 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-image-registry/image-pruner-29423520-ghc92" event={"ID":"11aa2b84-ba50-48f2-ab7a-0591bf6c3f44","Type":"ContainerStarted","Data":"9eb72320ece0957ef08a89558ba259fb6398cab5d201d35bb6c592406ec8680a"} Dec 11 00:00:02 crc kubenswrapper[4791]: I1211 00:00:02.255060 4791 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-image-registry/image-pruner-29423520-ghc92" event={"ID":"11aa2b84-ba50-48f2-ab7a-0591bf6c3f44","Type":"ContainerStarted","Data":"fc0c67d9de0e1e76df4d106a0c86049359f47c5d1690f6eb00e9ac6f9bf04aae"} Dec 11 00:00:02 crc kubenswrapper[4791]: I1211 00:00:02.259608 4791 generic.go:334] "Generic (PLEG): container finished" podID="d46e7c67-8829-47d6-909e-bcdbe292f092" containerID="36b4f9c8acb2740df7cabe4f43d776a70e1a92bc5a0ffb868cf760f4b8d9ebab" exitCode=0 Dec 11 00:00:02 crc kubenswrapper[4791]: I1211 00:00:02.259670 4791 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29423520-dkcj9" event={"ID":"d46e7c67-8829-47d6-909e-bcdbe292f092","Type":"ContainerDied","Data":"36b4f9c8acb2740df7cabe4f43d776a70e1a92bc5a0ffb868cf760f4b8d9ebab"} Dec 11 00:00:02 crc kubenswrapper[4791]: I1211 00:00:02.261300 4791 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-db-purge-29423520-zzzkt" event={"ID":"9f973940-a85e-4ee3-9373-d1c7512d9f1a","Type":"ContainerStarted","Data":"ebc6f2b18ce08d6d74d55b8fae15182ebb5dfd60e4a65559e72547a7d3b1db9d"} Dec 11 00:00:02 crc kubenswrapper[4791]: I1211 00:00:02.290012 4791 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-cell0-db-purge-29423520-vv2hp" podStartSLOduration=2.289959166 podStartE2EDuration="2.289959166s" podCreationTimestamp="2025-12-11 00:00:00 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-11 00:00:02.274878551 +0000 UTC m=+4236.704496184" watchObservedRunningTime="2025-12-11 00:00:02.289959166 +0000 UTC m=+4236.719576779" Dec 11 00:00:02 crc kubenswrapper[4791]: I1211 00:00:02.303296 4791 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-cell1-db-purge-29423520-zzzkt" podStartSLOduration=2.3032764820000002 podStartE2EDuration="2.303276482s" podCreationTimestamp="2025-12-11 00:00:00 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-11 00:00:02.29080164 +0000 UTC m=+4236.720419253" watchObservedRunningTime="2025-12-11 00:00:02.303276482 +0000 UTC m=+4236.732894095" Dec 11 00:00:02 crc kubenswrapper[4791]: I1211 00:00:02.351246 4791 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-image-registry/image-pruner-29423520-ghc92" podStartSLOduration=2.351209405 podStartE2EDuration="2.351209405s" podCreationTimestamp="2025-12-11 00:00:00 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-11 00:00:02.329300177 +0000 UTC m=+4236.758917780" watchObservedRunningTime="2025-12-11 00:00:02.351209405 +0000 UTC m=+4236.780827018" Dec 11 00:00:03 crc kubenswrapper[4791]: I1211 00:00:03.279016 4791 generic.go:334] "Generic (PLEG): container finished" podID="11aa2b84-ba50-48f2-ab7a-0591bf6c3f44" containerID="9eb72320ece0957ef08a89558ba259fb6398cab5d201d35bb6c592406ec8680a" exitCode=0 Dec 11 00:00:03 crc kubenswrapper[4791]: I1211 00:00:03.280153 4791 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-image-registry/image-pruner-29423520-ghc92" event={"ID":"11aa2b84-ba50-48f2-ab7a-0591bf6c3f44","Type":"ContainerDied","Data":"9eb72320ece0957ef08a89558ba259fb6398cab5d201d35bb6c592406ec8680a"} Dec 11 00:00:03 crc kubenswrapper[4791]: I1211 00:00:03.606045 4791 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29423520-dkcj9" Dec 11 00:00:03 crc kubenswrapper[4791]: I1211 00:00:03.777652 4791 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/d46e7c67-8829-47d6-909e-bcdbe292f092-config-volume\") pod \"d46e7c67-8829-47d6-909e-bcdbe292f092\" (UID: \"d46e7c67-8829-47d6-909e-bcdbe292f092\") " Dec 11 00:00:03 crc kubenswrapper[4791]: I1211 00:00:03.778088 4791 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/d46e7c67-8829-47d6-909e-bcdbe292f092-secret-volume\") pod \"d46e7c67-8829-47d6-909e-bcdbe292f092\" (UID: \"d46e7c67-8829-47d6-909e-bcdbe292f092\") " Dec 11 00:00:03 crc kubenswrapper[4791]: I1211 00:00:03.778200 4791 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-f6dlk\" (UniqueName: \"kubernetes.io/projected/d46e7c67-8829-47d6-909e-bcdbe292f092-kube-api-access-f6dlk\") pod \"d46e7c67-8829-47d6-909e-bcdbe292f092\" (UID: \"d46e7c67-8829-47d6-909e-bcdbe292f092\") " Dec 11 00:00:03 crc kubenswrapper[4791]: I1211 00:00:03.778759 4791 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/d46e7c67-8829-47d6-909e-bcdbe292f092-config-volume" (OuterVolumeSpecName: "config-volume") pod "d46e7c67-8829-47d6-909e-bcdbe292f092" (UID: "d46e7c67-8829-47d6-909e-bcdbe292f092"). InnerVolumeSpecName "config-volume". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 11 00:00:03 crc kubenswrapper[4791]: I1211 00:00:03.799281 4791 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/d46e7c67-8829-47d6-909e-bcdbe292f092-kube-api-access-f6dlk" (OuterVolumeSpecName: "kube-api-access-f6dlk") pod "d46e7c67-8829-47d6-909e-bcdbe292f092" (UID: "d46e7c67-8829-47d6-909e-bcdbe292f092"). InnerVolumeSpecName "kube-api-access-f6dlk". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 11 00:00:03 crc kubenswrapper[4791]: I1211 00:00:03.799849 4791 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/d46e7c67-8829-47d6-909e-bcdbe292f092-secret-volume" (OuterVolumeSpecName: "secret-volume") pod "d46e7c67-8829-47d6-909e-bcdbe292f092" (UID: "d46e7c67-8829-47d6-909e-bcdbe292f092"). InnerVolumeSpecName "secret-volume". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 11 00:00:03 crc kubenswrapper[4791]: I1211 00:00:03.880176 4791 reconciler_common.go:293] "Volume detached for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/d46e7c67-8829-47d6-909e-bcdbe292f092-secret-volume\") on node \"crc\" DevicePath \"\"" Dec 11 00:00:03 crc kubenswrapper[4791]: I1211 00:00:03.880458 4791 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-f6dlk\" (UniqueName: \"kubernetes.io/projected/d46e7c67-8829-47d6-909e-bcdbe292f092-kube-api-access-f6dlk\") on node \"crc\" DevicePath \"\"" Dec 11 00:00:03 crc kubenswrapper[4791]: I1211 00:00:03.880533 4791 reconciler_common.go:293] "Volume detached for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/d46e7c67-8829-47d6-909e-bcdbe292f092-config-volume\") on node \"crc\" DevicePath \"\"" Dec 11 00:00:04 crc kubenswrapper[4791]: I1211 00:00:04.292276 4791 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29423520-dkcj9" Dec 11 00:00:04 crc kubenswrapper[4791]: I1211 00:00:04.292511 4791 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29423520-dkcj9" event={"ID":"d46e7c67-8829-47d6-909e-bcdbe292f092","Type":"ContainerDied","Data":"4aca408e4b7b57897d0053d58f3ccda350a8e0f12f1bdfc8ee6563e94431d532"} Dec 11 00:00:04 crc kubenswrapper[4791]: I1211 00:00:04.300564 4791 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="4aca408e4b7b57897d0053d58f3ccda350a8e0f12f1bdfc8ee6563e94431d532" Dec 11 00:00:04 crc kubenswrapper[4791]: I1211 00:00:04.688895 4791 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29423475-l2jwz"] Dec 11 00:00:04 crc kubenswrapper[4791]: I1211 00:00:04.699552 4791 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29423475-l2jwz"] Dec 11 00:00:04 crc kubenswrapper[4791]: I1211 00:00:04.726520 4791 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-image-registry/image-pruner-29423520-ghc92" Dec 11 00:00:04 crc kubenswrapper[4791]: I1211 00:00:04.803173 4791 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serviceca\" (UniqueName: \"kubernetes.io/configmap/11aa2b84-ba50-48f2-ab7a-0591bf6c3f44-serviceca\") pod \"11aa2b84-ba50-48f2-ab7a-0591bf6c3f44\" (UID: \"11aa2b84-ba50-48f2-ab7a-0591bf6c3f44\") " Dec 11 00:00:04 crc kubenswrapper[4791]: I1211 00:00:04.803380 4791 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-8pgk5\" (UniqueName: \"kubernetes.io/projected/11aa2b84-ba50-48f2-ab7a-0591bf6c3f44-kube-api-access-8pgk5\") pod \"11aa2b84-ba50-48f2-ab7a-0591bf6c3f44\" (UID: \"11aa2b84-ba50-48f2-ab7a-0591bf6c3f44\") " Dec 11 00:00:04 crc kubenswrapper[4791]: I1211 00:00:04.804165 4791 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/11aa2b84-ba50-48f2-ab7a-0591bf6c3f44-serviceca" (OuterVolumeSpecName: "serviceca") pod "11aa2b84-ba50-48f2-ab7a-0591bf6c3f44" (UID: "11aa2b84-ba50-48f2-ab7a-0591bf6c3f44"). InnerVolumeSpecName "serviceca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 11 00:00:04 crc kubenswrapper[4791]: I1211 00:00:04.818521 4791 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/11aa2b84-ba50-48f2-ab7a-0591bf6c3f44-kube-api-access-8pgk5" (OuterVolumeSpecName: "kube-api-access-8pgk5") pod "11aa2b84-ba50-48f2-ab7a-0591bf6c3f44" (UID: "11aa2b84-ba50-48f2-ab7a-0591bf6c3f44"). InnerVolumeSpecName "kube-api-access-8pgk5". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 11 00:00:04 crc kubenswrapper[4791]: I1211 00:00:04.905112 4791 reconciler_common.go:293] "Volume detached for volume \"serviceca\" (UniqueName: \"kubernetes.io/configmap/11aa2b84-ba50-48f2-ab7a-0591bf6c3f44-serviceca\") on node \"crc\" DevicePath \"\"" Dec 11 00:00:04 crc kubenswrapper[4791]: I1211 00:00:04.905141 4791 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-8pgk5\" (UniqueName: \"kubernetes.io/projected/11aa2b84-ba50-48f2-ab7a-0591bf6c3f44-kube-api-access-8pgk5\") on node \"crc\" DevicePath \"\"" Dec 11 00:00:05 crc kubenswrapper[4791]: I1211 00:00:05.314049 4791 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-image-registry/image-pruner-29423520-ghc92" event={"ID":"11aa2b84-ba50-48f2-ab7a-0591bf6c3f44","Type":"ContainerDied","Data":"fc0c67d9de0e1e76df4d106a0c86049359f47c5d1690f6eb00e9ac6f9bf04aae"} Dec 11 00:00:05 crc kubenswrapper[4791]: I1211 00:00:05.314090 4791 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="fc0c67d9de0e1e76df4d106a0c86049359f47c5d1690f6eb00e9ac6f9bf04aae" Dec 11 00:00:05 crc kubenswrapper[4791]: I1211 00:00:05.314108 4791 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-image-registry/image-pruner-29423520-ghc92" Dec 11 00:00:05 crc kubenswrapper[4791]: I1211 00:00:05.895538 4791 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="ce778cac-da93-4e3e-89ed-769568cb4b52" path="/var/lib/kubelet/pods/ce778cac-da93-4e3e-89ed-769568cb4b52/volumes" Dec 11 00:00:07 crc kubenswrapper[4791]: I1211 00:00:07.342914 4791 generic.go:334] "Generic (PLEG): container finished" podID="9f973940-a85e-4ee3-9373-d1c7512d9f1a" containerID="ebc6f2b18ce08d6d74d55b8fae15182ebb5dfd60e4a65559e72547a7d3b1db9d" exitCode=0 Dec 11 00:00:07 crc kubenswrapper[4791]: I1211 00:00:07.343171 4791 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-db-purge-29423520-zzzkt" event={"ID":"9f973940-a85e-4ee3-9373-d1c7512d9f1a","Type":"ContainerDied","Data":"ebc6f2b18ce08d6d74d55b8fae15182ebb5dfd60e4a65559e72547a7d3b1db9d"} Dec 11 00:00:07 crc kubenswrapper[4791]: I1211 00:00:07.348764 4791 generic.go:334] "Generic (PLEG): container finished" podID="5309e85e-16c8-46fb-b694-1e16db7d7770" containerID="55ecbc8ec3c4fa0d09f052f8360db6057b37a513669a12346f1885202c9bec3c" exitCode=0 Dec 11 00:00:07 crc kubenswrapper[4791]: I1211 00:00:07.348843 4791 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-db-purge-29423520-vv2hp" event={"ID":"5309e85e-16c8-46fb-b694-1e16db7d7770","Type":"ContainerDied","Data":"55ecbc8ec3c4fa0d09f052f8360db6057b37a513669a12346f1885202c9bec3c"} Dec 11 00:00:09 crc kubenswrapper[4791]: I1211 00:00:09.011615 4791 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-db-purge-29423520-vv2hp" Dec 11 00:00:09 crc kubenswrapper[4791]: I1211 00:00:09.020789 4791 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-db-purge-29423520-zzzkt" Dec 11 00:00:09 crc kubenswrapper[4791]: I1211 00:00:09.204144 4791 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/5309e85e-16c8-46fb-b694-1e16db7d7770-scripts\") pod \"5309e85e-16c8-46fb-b694-1e16db7d7770\" (UID: \"5309e85e-16c8-46fb-b694-1e16db7d7770\") " Dec 11 00:00:09 crc kubenswrapper[4791]: I1211 00:00:09.204225 4791 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/5309e85e-16c8-46fb-b694-1e16db7d7770-config-data\") pod \"5309e85e-16c8-46fb-b694-1e16db7d7770\" (UID: \"5309e85e-16c8-46fb-b694-1e16db7d7770\") " Dec 11 00:00:09 crc kubenswrapper[4791]: I1211 00:00:09.204287 4791 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/5309e85e-16c8-46fb-b694-1e16db7d7770-combined-ca-bundle\") pod \"5309e85e-16c8-46fb-b694-1e16db7d7770\" (UID: \"5309e85e-16c8-46fb-b694-1e16db7d7770\") " Dec 11 00:00:09 crc kubenswrapper[4791]: I1211 00:00:09.204427 4791 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/9f973940-a85e-4ee3-9373-d1c7512d9f1a-config-data\") pod \"9f973940-a85e-4ee3-9373-d1c7512d9f1a\" (UID: \"9f973940-a85e-4ee3-9373-d1c7512d9f1a\") " Dec 11 00:00:09 crc kubenswrapper[4791]: I1211 00:00:09.204532 4791 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/9f973940-a85e-4ee3-9373-d1c7512d9f1a-scripts\") pod \"9f973940-a85e-4ee3-9373-d1c7512d9f1a\" (UID: \"9f973940-a85e-4ee3-9373-d1c7512d9f1a\") " Dec 11 00:00:09 crc kubenswrapper[4791]: I1211 00:00:09.204560 4791 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/9f973940-a85e-4ee3-9373-d1c7512d9f1a-combined-ca-bundle\") pod \"9f973940-a85e-4ee3-9373-d1c7512d9f1a\" (UID: \"9f973940-a85e-4ee3-9373-d1c7512d9f1a\") " Dec 11 00:00:09 crc kubenswrapper[4791]: I1211 00:00:09.204588 4791 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-m4584\" (UniqueName: \"kubernetes.io/projected/9f973940-a85e-4ee3-9373-d1c7512d9f1a-kube-api-access-m4584\") pod \"9f973940-a85e-4ee3-9373-d1c7512d9f1a\" (UID: \"9f973940-a85e-4ee3-9373-d1c7512d9f1a\") " Dec 11 00:00:09 crc kubenswrapper[4791]: I1211 00:00:09.204618 4791 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-26p8v\" (UniqueName: \"kubernetes.io/projected/5309e85e-16c8-46fb-b694-1e16db7d7770-kube-api-access-26p8v\") pod \"5309e85e-16c8-46fb-b694-1e16db7d7770\" (UID: \"5309e85e-16c8-46fb-b694-1e16db7d7770\") " Dec 11 00:00:09 crc kubenswrapper[4791]: I1211 00:00:09.210983 4791 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/5309e85e-16c8-46fb-b694-1e16db7d7770-scripts" (OuterVolumeSpecName: "scripts") pod "5309e85e-16c8-46fb-b694-1e16db7d7770" (UID: "5309e85e-16c8-46fb-b694-1e16db7d7770"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 11 00:00:09 crc kubenswrapper[4791]: I1211 00:00:09.211263 4791 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/9f973940-a85e-4ee3-9373-d1c7512d9f1a-scripts" (OuterVolumeSpecName: "scripts") pod "9f973940-a85e-4ee3-9373-d1c7512d9f1a" (UID: "9f973940-a85e-4ee3-9373-d1c7512d9f1a"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 11 00:00:09 crc kubenswrapper[4791]: I1211 00:00:09.221151 4791 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/5309e85e-16c8-46fb-b694-1e16db7d7770-kube-api-access-26p8v" (OuterVolumeSpecName: "kube-api-access-26p8v") pod "5309e85e-16c8-46fb-b694-1e16db7d7770" (UID: "5309e85e-16c8-46fb-b694-1e16db7d7770"). InnerVolumeSpecName "kube-api-access-26p8v". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 11 00:00:09 crc kubenswrapper[4791]: I1211 00:00:09.221238 4791 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/9f973940-a85e-4ee3-9373-d1c7512d9f1a-kube-api-access-m4584" (OuterVolumeSpecName: "kube-api-access-m4584") pod "9f973940-a85e-4ee3-9373-d1c7512d9f1a" (UID: "9f973940-a85e-4ee3-9373-d1c7512d9f1a"). InnerVolumeSpecName "kube-api-access-m4584". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 11 00:00:09 crc kubenswrapper[4791]: I1211 00:00:09.235790 4791 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/9f973940-a85e-4ee3-9373-d1c7512d9f1a-config-data" (OuterVolumeSpecName: "config-data") pod "9f973940-a85e-4ee3-9373-d1c7512d9f1a" (UID: "9f973940-a85e-4ee3-9373-d1c7512d9f1a"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 11 00:00:09 crc kubenswrapper[4791]: I1211 00:00:09.239818 4791 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/5309e85e-16c8-46fb-b694-1e16db7d7770-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "5309e85e-16c8-46fb-b694-1e16db7d7770" (UID: "5309e85e-16c8-46fb-b694-1e16db7d7770"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 11 00:00:09 crc kubenswrapper[4791]: I1211 00:00:09.244976 4791 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/5309e85e-16c8-46fb-b694-1e16db7d7770-config-data" (OuterVolumeSpecName: "config-data") pod "5309e85e-16c8-46fb-b694-1e16db7d7770" (UID: "5309e85e-16c8-46fb-b694-1e16db7d7770"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 11 00:00:09 crc kubenswrapper[4791]: I1211 00:00:09.261975 4791 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/9f973940-a85e-4ee3-9373-d1c7512d9f1a-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "9f973940-a85e-4ee3-9373-d1c7512d9f1a" (UID: "9f973940-a85e-4ee3-9373-d1c7512d9f1a"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 11 00:00:09 crc kubenswrapper[4791]: I1211 00:00:09.307283 4791 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-26p8v\" (UniqueName: \"kubernetes.io/projected/5309e85e-16c8-46fb-b694-1e16db7d7770-kube-api-access-26p8v\") on node \"crc\" DevicePath \"\"" Dec 11 00:00:09 crc kubenswrapper[4791]: I1211 00:00:09.307328 4791 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/5309e85e-16c8-46fb-b694-1e16db7d7770-scripts\") on node \"crc\" DevicePath \"\"" Dec 11 00:00:09 crc kubenswrapper[4791]: I1211 00:00:09.307356 4791 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/5309e85e-16c8-46fb-b694-1e16db7d7770-config-data\") on node \"crc\" DevicePath \"\"" Dec 11 00:00:09 crc kubenswrapper[4791]: I1211 00:00:09.307372 4791 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/5309e85e-16c8-46fb-b694-1e16db7d7770-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 11 00:00:09 crc kubenswrapper[4791]: I1211 00:00:09.307385 4791 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/9f973940-a85e-4ee3-9373-d1c7512d9f1a-config-data\") on node \"crc\" DevicePath \"\"" Dec 11 00:00:09 crc kubenswrapper[4791]: I1211 00:00:09.307395 4791 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/9f973940-a85e-4ee3-9373-d1c7512d9f1a-scripts\") on node \"crc\" DevicePath \"\"" Dec 11 00:00:09 crc kubenswrapper[4791]: I1211 00:00:09.307406 4791 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/9f973940-a85e-4ee3-9373-d1c7512d9f1a-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 11 00:00:09 crc kubenswrapper[4791]: I1211 00:00:09.307417 4791 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-m4584\" (UniqueName: \"kubernetes.io/projected/9f973940-a85e-4ee3-9373-d1c7512d9f1a-kube-api-access-m4584\") on node \"crc\" DevicePath \"\"" Dec 11 00:00:09 crc kubenswrapper[4791]: I1211 00:00:09.368116 4791 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-db-purge-29423520-vv2hp" event={"ID":"5309e85e-16c8-46fb-b694-1e16db7d7770","Type":"ContainerDied","Data":"c7af93b04d512a5d3be1d13cff4917554d358f6246276e6ab0a5e685b890961d"} Dec 11 00:00:09 crc kubenswrapper[4791]: I1211 00:00:09.368151 4791 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-db-purge-29423520-vv2hp" Dec 11 00:00:09 crc kubenswrapper[4791]: I1211 00:00:09.368165 4791 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="c7af93b04d512a5d3be1d13cff4917554d358f6246276e6ab0a5e685b890961d" Dec 11 00:00:09 crc kubenswrapper[4791]: I1211 00:00:09.369969 4791 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-db-purge-29423520-zzzkt" event={"ID":"9f973940-a85e-4ee3-9373-d1c7512d9f1a","Type":"ContainerDied","Data":"675d786841c45328890a4cb8f18561cf49cdfbe477808c2c51bb5a28fad81f94"} Dec 11 00:00:09 crc kubenswrapper[4791]: I1211 00:00:09.370074 4791 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="675d786841c45328890a4cb8f18561cf49cdfbe477808c2c51bb5a28fad81f94" Dec 11 00:00:09 crc kubenswrapper[4791]: I1211 00:00:09.370008 4791 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-db-purge-29423520-zzzkt" Dec 11 00:00:54 crc kubenswrapper[4791]: I1211 00:00:54.871128 4791 generic.go:334] "Generic (PLEG): container finished" podID="6e53fefe-b61d-4ba3-a360-4ae9f1213d1c" containerID="fd094fcb02987fc8ae45b115d82c5662af63b3157267e481ca35913a767e24aa" exitCode=0 Dec 11 00:00:54 crc kubenswrapper[4791]: I1211 00:00:54.871538 4791 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-2dlqt/must-gather-rqrn8" event={"ID":"6e53fefe-b61d-4ba3-a360-4ae9f1213d1c","Type":"ContainerDied","Data":"fd094fcb02987fc8ae45b115d82c5662af63b3157267e481ca35913a767e24aa"} Dec 11 00:00:54 crc kubenswrapper[4791]: I1211 00:00:54.872497 4791 scope.go:117] "RemoveContainer" containerID="fd094fcb02987fc8ae45b115d82c5662af63b3157267e481ca35913a767e24aa" Dec 11 00:00:55 crc kubenswrapper[4791]: I1211 00:00:55.037747 4791 patch_prober.go:28] interesting pod/machine-config-daemon-5rb5l container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 11 00:00:55 crc kubenswrapper[4791]: I1211 00:00:55.037797 4791 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-5rb5l" podUID="ba35653c-6e06-4cee-a4d6-137764090d18" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 11 00:00:55 crc kubenswrapper[4791]: I1211 00:00:55.469531 4791 scope.go:117] "RemoveContainer" containerID="4f8e348e6239dcaf799551ffbdf8449ca88dd21d073c94cf6252f32b69d9ae1f" Dec 11 00:00:55 crc kubenswrapper[4791]: I1211 00:00:55.532310 4791 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-must-gather-2dlqt_must-gather-rqrn8_6e53fefe-b61d-4ba3-a360-4ae9f1213d1c/gather/0.log" Dec 11 00:01:00 crc kubenswrapper[4791]: I1211 00:01:00.180736 4791 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/glance-db-purge-29423521-ptqnd"] Dec 11 00:01:00 crc kubenswrapper[4791]: E1211 00:01:00.182028 4791 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="d46e7c67-8829-47d6-909e-bcdbe292f092" containerName="collect-profiles" Dec 11 00:01:00 crc kubenswrapper[4791]: I1211 00:01:00.182052 4791 state_mem.go:107] "Deleted CPUSet assignment" podUID="d46e7c67-8829-47d6-909e-bcdbe292f092" containerName="collect-profiles" Dec 11 00:01:00 crc kubenswrapper[4791]: E1211 00:01:00.182078 4791 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="9f973940-a85e-4ee3-9373-d1c7512d9f1a" containerName="nova-manage" Dec 11 00:01:00 crc kubenswrapper[4791]: I1211 00:01:00.182091 4791 state_mem.go:107] "Deleted CPUSet assignment" podUID="9f973940-a85e-4ee3-9373-d1c7512d9f1a" containerName="nova-manage" Dec 11 00:01:00 crc kubenswrapper[4791]: E1211 00:01:00.182119 4791 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="11aa2b84-ba50-48f2-ab7a-0591bf6c3f44" containerName="image-pruner" Dec 11 00:01:00 crc kubenswrapper[4791]: I1211 00:01:00.182133 4791 state_mem.go:107] "Deleted CPUSet assignment" podUID="11aa2b84-ba50-48f2-ab7a-0591bf6c3f44" containerName="image-pruner" Dec 11 00:01:00 crc kubenswrapper[4791]: E1211 00:01:00.182169 4791 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="5309e85e-16c8-46fb-b694-1e16db7d7770" containerName="nova-manage" Dec 11 00:01:00 crc kubenswrapper[4791]: I1211 00:01:00.182181 4791 state_mem.go:107] "Deleted CPUSet assignment" podUID="5309e85e-16c8-46fb-b694-1e16db7d7770" containerName="nova-manage" Dec 11 00:01:00 crc kubenswrapper[4791]: I1211 00:01:00.182534 4791 memory_manager.go:354] "RemoveStaleState removing state" podUID="9f973940-a85e-4ee3-9373-d1c7512d9f1a" containerName="nova-manage" Dec 11 00:01:00 crc kubenswrapper[4791]: I1211 00:01:00.182577 4791 memory_manager.go:354] "RemoveStaleState removing state" podUID="5309e85e-16c8-46fb-b694-1e16db7d7770" containerName="nova-manage" Dec 11 00:01:00 crc kubenswrapper[4791]: I1211 00:01:00.182598 4791 memory_manager.go:354] "RemoveStaleState removing state" podUID="11aa2b84-ba50-48f2-ab7a-0591bf6c3f44" containerName="image-pruner" Dec 11 00:01:00 crc kubenswrapper[4791]: I1211 00:01:00.182634 4791 memory_manager.go:354] "RemoveStaleState removing state" podUID="d46e7c67-8829-47d6-909e-bcdbe292f092" containerName="collect-profiles" Dec 11 00:01:00 crc kubenswrapper[4791]: I1211 00:01:00.183775 4791 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-db-purge-29423521-ptqnd" Dec 11 00:01:00 crc kubenswrapper[4791]: I1211 00:01:00.186313 4791 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"glance-config-data" Dec 11 00:01:00 crc kubenswrapper[4791]: I1211 00:01:00.189799 4791 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/keystone-cron-29423521-vlkhm"] Dec 11 00:01:00 crc kubenswrapper[4791]: I1211 00:01:00.191411 4791 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-cron-29423521-vlkhm" Dec 11 00:01:00 crc kubenswrapper[4791]: I1211 00:01:00.204768 4791 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/cinder-db-purge-29423521-h646b"] Dec 11 00:01:00 crc kubenswrapper[4791]: I1211 00:01:00.206299 4791 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-db-purge-29423521-h646b" Dec 11 00:01:00 crc kubenswrapper[4791]: I1211 00:01:00.216798 4791 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-db-purge-29423521-h646b"] Dec 11 00:01:00 crc kubenswrapper[4791]: I1211 00:01:00.242907 4791 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/keystone-cron-29423521-vlkhm"] Dec 11 00:01:00 crc kubenswrapper[4791]: I1211 00:01:00.277070 4791 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-db-purge-29423521-ptqnd"] Dec 11 00:01:00 crc kubenswrapper[4791]: I1211 00:01:00.307752 4791 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"db-purge-config-data\" (UniqueName: \"kubernetes.io/secret/d332f1db-3d1a-4e2d-b2d2-69498733ceaf-db-purge-config-data\") pod \"cinder-db-purge-29423521-h646b\" (UID: \"d332f1db-3d1a-4e2d-b2d2-69498733ceaf\") " pod="openstack/cinder-db-purge-29423521-h646b" Dec 11 00:01:00 crc kubenswrapper[4791]: I1211 00:01:00.308026 4791 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"db-purge-config-data\" (UniqueName: \"kubernetes.io/secret/ea21aaf5-23c0-47f7-86c0-1f92f0365633-db-purge-config-data\") pod \"glance-db-purge-29423521-ptqnd\" (UID: \"ea21aaf5-23c0-47f7-86c0-1f92f0365633\") " pod="openstack/glance-db-purge-29423521-ptqnd" Dec 11 00:01:00 crc kubenswrapper[4791]: I1211 00:01:00.308135 4791 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/d332f1db-3d1a-4e2d-b2d2-69498733ceaf-config-data\") pod \"cinder-db-purge-29423521-h646b\" (UID: \"d332f1db-3d1a-4e2d-b2d2-69498733ceaf\") " pod="openstack/cinder-db-purge-29423521-h646b" Dec 11 00:01:00 crc kubenswrapper[4791]: I1211 00:01:00.308228 4791 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-582zc\" (UniqueName: \"kubernetes.io/projected/ea21aaf5-23c0-47f7-86c0-1f92f0365633-kube-api-access-582zc\") pod \"glance-db-purge-29423521-ptqnd\" (UID: \"ea21aaf5-23c0-47f7-86c0-1f92f0365633\") " pod="openstack/glance-db-purge-29423521-ptqnd" Dec 11 00:01:00 crc kubenswrapper[4791]: I1211 00:01:00.308440 4791 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-p45qx\" (UniqueName: \"kubernetes.io/projected/d332f1db-3d1a-4e2d-b2d2-69498733ceaf-kube-api-access-p45qx\") pod \"cinder-db-purge-29423521-h646b\" (UID: \"d332f1db-3d1a-4e2d-b2d2-69498733ceaf\") " pod="openstack/cinder-db-purge-29423521-h646b" Dec 11 00:01:00 crc kubenswrapper[4791]: I1211 00:01:00.308549 4791 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/9349278c-a330-4119-9de6-ca3cb6f28ce7-fernet-keys\") pod \"keystone-cron-29423521-vlkhm\" (UID: \"9349278c-a330-4119-9de6-ca3cb6f28ce7\") " pod="openstack/keystone-cron-29423521-vlkhm" Dec 11 00:01:00 crc kubenswrapper[4791]: I1211 00:01:00.308635 4791 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d332f1db-3d1a-4e2d-b2d2-69498733ceaf-combined-ca-bundle\") pod \"cinder-db-purge-29423521-h646b\" (UID: \"d332f1db-3d1a-4e2d-b2d2-69498733ceaf\") " pod="openstack/cinder-db-purge-29423521-h646b" Dec 11 00:01:00 crc kubenswrapper[4791]: I1211 00:01:00.308751 4791 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-nrlcm\" (UniqueName: \"kubernetes.io/projected/9349278c-a330-4119-9de6-ca3cb6f28ce7-kube-api-access-nrlcm\") pod \"keystone-cron-29423521-vlkhm\" (UID: \"9349278c-a330-4119-9de6-ca3cb6f28ce7\") " pod="openstack/keystone-cron-29423521-vlkhm" Dec 11 00:01:00 crc kubenswrapper[4791]: I1211 00:01:00.308874 4791 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/9349278c-a330-4119-9de6-ca3cb6f28ce7-combined-ca-bundle\") pod \"keystone-cron-29423521-vlkhm\" (UID: \"9349278c-a330-4119-9de6-ca3cb6f28ce7\") " pod="openstack/keystone-cron-29423521-vlkhm" Dec 11 00:01:00 crc kubenswrapper[4791]: I1211 00:01:00.308976 4791 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/9349278c-a330-4119-9de6-ca3cb6f28ce7-config-data\") pod \"keystone-cron-29423521-vlkhm\" (UID: \"9349278c-a330-4119-9de6-ca3cb6f28ce7\") " pod="openstack/keystone-cron-29423521-vlkhm" Dec 11 00:01:00 crc kubenswrapper[4791]: I1211 00:01:00.309073 4791 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/ea21aaf5-23c0-47f7-86c0-1f92f0365633-config-data\") pod \"glance-db-purge-29423521-ptqnd\" (UID: \"ea21aaf5-23c0-47f7-86c0-1f92f0365633\") " pod="openstack/glance-db-purge-29423521-ptqnd" Dec 11 00:01:00 crc kubenswrapper[4791]: I1211 00:01:00.309161 4791 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ea21aaf5-23c0-47f7-86c0-1f92f0365633-combined-ca-bundle\") pod \"glance-db-purge-29423521-ptqnd\" (UID: \"ea21aaf5-23c0-47f7-86c0-1f92f0365633\") " pod="openstack/glance-db-purge-29423521-ptqnd" Dec 11 00:01:00 crc kubenswrapper[4791]: I1211 00:01:00.411266 4791 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"db-purge-config-data\" (UniqueName: \"kubernetes.io/secret/ea21aaf5-23c0-47f7-86c0-1f92f0365633-db-purge-config-data\") pod \"glance-db-purge-29423521-ptqnd\" (UID: \"ea21aaf5-23c0-47f7-86c0-1f92f0365633\") " pod="openstack/glance-db-purge-29423521-ptqnd" Dec 11 00:01:00 crc kubenswrapper[4791]: I1211 00:01:00.411589 4791 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/d332f1db-3d1a-4e2d-b2d2-69498733ceaf-config-data\") pod \"cinder-db-purge-29423521-h646b\" (UID: \"d332f1db-3d1a-4e2d-b2d2-69498733ceaf\") " pod="openstack/cinder-db-purge-29423521-h646b" Dec 11 00:01:00 crc kubenswrapper[4791]: I1211 00:01:00.411713 4791 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-582zc\" (UniqueName: \"kubernetes.io/projected/ea21aaf5-23c0-47f7-86c0-1f92f0365633-kube-api-access-582zc\") pod \"glance-db-purge-29423521-ptqnd\" (UID: \"ea21aaf5-23c0-47f7-86c0-1f92f0365633\") " pod="openstack/glance-db-purge-29423521-ptqnd" Dec 11 00:01:00 crc kubenswrapper[4791]: I1211 00:01:00.411828 4791 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-p45qx\" (UniqueName: \"kubernetes.io/projected/d332f1db-3d1a-4e2d-b2d2-69498733ceaf-kube-api-access-p45qx\") pod \"cinder-db-purge-29423521-h646b\" (UID: \"d332f1db-3d1a-4e2d-b2d2-69498733ceaf\") " pod="openstack/cinder-db-purge-29423521-h646b" Dec 11 00:01:00 crc kubenswrapper[4791]: I1211 00:01:00.411945 4791 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/9349278c-a330-4119-9de6-ca3cb6f28ce7-fernet-keys\") pod \"keystone-cron-29423521-vlkhm\" (UID: \"9349278c-a330-4119-9de6-ca3cb6f28ce7\") " pod="openstack/keystone-cron-29423521-vlkhm" Dec 11 00:01:00 crc kubenswrapper[4791]: I1211 00:01:00.412057 4791 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d332f1db-3d1a-4e2d-b2d2-69498733ceaf-combined-ca-bundle\") pod \"cinder-db-purge-29423521-h646b\" (UID: \"d332f1db-3d1a-4e2d-b2d2-69498733ceaf\") " pod="openstack/cinder-db-purge-29423521-h646b" Dec 11 00:01:00 crc kubenswrapper[4791]: I1211 00:01:00.412177 4791 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-nrlcm\" (UniqueName: \"kubernetes.io/projected/9349278c-a330-4119-9de6-ca3cb6f28ce7-kube-api-access-nrlcm\") pod \"keystone-cron-29423521-vlkhm\" (UID: \"9349278c-a330-4119-9de6-ca3cb6f28ce7\") " pod="openstack/keystone-cron-29423521-vlkhm" Dec 11 00:01:00 crc kubenswrapper[4791]: I1211 00:01:00.412301 4791 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/9349278c-a330-4119-9de6-ca3cb6f28ce7-combined-ca-bundle\") pod \"keystone-cron-29423521-vlkhm\" (UID: \"9349278c-a330-4119-9de6-ca3cb6f28ce7\") " pod="openstack/keystone-cron-29423521-vlkhm" Dec 11 00:01:00 crc kubenswrapper[4791]: I1211 00:01:00.412439 4791 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/9349278c-a330-4119-9de6-ca3cb6f28ce7-config-data\") pod \"keystone-cron-29423521-vlkhm\" (UID: \"9349278c-a330-4119-9de6-ca3cb6f28ce7\") " pod="openstack/keystone-cron-29423521-vlkhm" Dec 11 00:01:00 crc kubenswrapper[4791]: I1211 00:01:00.412556 4791 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/ea21aaf5-23c0-47f7-86c0-1f92f0365633-config-data\") pod \"glance-db-purge-29423521-ptqnd\" (UID: \"ea21aaf5-23c0-47f7-86c0-1f92f0365633\") " pod="openstack/glance-db-purge-29423521-ptqnd" Dec 11 00:01:00 crc kubenswrapper[4791]: I1211 00:01:00.412658 4791 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ea21aaf5-23c0-47f7-86c0-1f92f0365633-combined-ca-bundle\") pod \"glance-db-purge-29423521-ptqnd\" (UID: \"ea21aaf5-23c0-47f7-86c0-1f92f0365633\") " pod="openstack/glance-db-purge-29423521-ptqnd" Dec 11 00:01:00 crc kubenswrapper[4791]: I1211 00:01:00.412794 4791 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"db-purge-config-data\" (UniqueName: \"kubernetes.io/secret/d332f1db-3d1a-4e2d-b2d2-69498733ceaf-db-purge-config-data\") pod \"cinder-db-purge-29423521-h646b\" (UID: \"d332f1db-3d1a-4e2d-b2d2-69498733ceaf\") " pod="openstack/cinder-db-purge-29423521-h646b" Dec 11 00:01:00 crc kubenswrapper[4791]: I1211 00:01:00.418442 4791 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d332f1db-3d1a-4e2d-b2d2-69498733ceaf-combined-ca-bundle\") pod \"cinder-db-purge-29423521-h646b\" (UID: \"d332f1db-3d1a-4e2d-b2d2-69498733ceaf\") " pod="openstack/cinder-db-purge-29423521-h646b" Dec 11 00:01:00 crc kubenswrapper[4791]: I1211 00:01:00.418498 4791 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/9349278c-a330-4119-9de6-ca3cb6f28ce7-fernet-keys\") pod \"keystone-cron-29423521-vlkhm\" (UID: \"9349278c-a330-4119-9de6-ca3cb6f28ce7\") " pod="openstack/keystone-cron-29423521-vlkhm" Dec 11 00:01:00 crc kubenswrapper[4791]: I1211 00:01:00.423853 4791 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/d332f1db-3d1a-4e2d-b2d2-69498733ceaf-config-data\") pod \"cinder-db-purge-29423521-h646b\" (UID: \"d332f1db-3d1a-4e2d-b2d2-69498733ceaf\") " pod="openstack/cinder-db-purge-29423521-h646b" Dec 11 00:01:00 crc kubenswrapper[4791]: I1211 00:01:00.424097 4791 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/9349278c-a330-4119-9de6-ca3cb6f28ce7-combined-ca-bundle\") pod \"keystone-cron-29423521-vlkhm\" (UID: \"9349278c-a330-4119-9de6-ca3cb6f28ce7\") " pod="openstack/keystone-cron-29423521-vlkhm" Dec 11 00:01:00 crc kubenswrapper[4791]: I1211 00:01:00.425603 4791 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/9349278c-a330-4119-9de6-ca3cb6f28ce7-config-data\") pod \"keystone-cron-29423521-vlkhm\" (UID: \"9349278c-a330-4119-9de6-ca3cb6f28ce7\") " pod="openstack/keystone-cron-29423521-vlkhm" Dec 11 00:01:00 crc kubenswrapper[4791]: I1211 00:01:00.428576 4791 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ea21aaf5-23c0-47f7-86c0-1f92f0365633-combined-ca-bundle\") pod \"glance-db-purge-29423521-ptqnd\" (UID: \"ea21aaf5-23c0-47f7-86c0-1f92f0365633\") " pod="openstack/glance-db-purge-29423521-ptqnd" Dec 11 00:01:00 crc kubenswrapper[4791]: I1211 00:01:00.428804 4791 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"db-purge-config-data\" (UniqueName: \"kubernetes.io/secret/d332f1db-3d1a-4e2d-b2d2-69498733ceaf-db-purge-config-data\") pod \"cinder-db-purge-29423521-h646b\" (UID: \"d332f1db-3d1a-4e2d-b2d2-69498733ceaf\") " pod="openstack/cinder-db-purge-29423521-h646b" Dec 11 00:01:00 crc kubenswrapper[4791]: I1211 00:01:00.428764 4791 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"db-purge-config-data\" (UniqueName: \"kubernetes.io/secret/ea21aaf5-23c0-47f7-86c0-1f92f0365633-db-purge-config-data\") pod \"glance-db-purge-29423521-ptqnd\" (UID: \"ea21aaf5-23c0-47f7-86c0-1f92f0365633\") " pod="openstack/glance-db-purge-29423521-ptqnd" Dec 11 00:01:00 crc kubenswrapper[4791]: I1211 00:01:00.429542 4791 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/ea21aaf5-23c0-47f7-86c0-1f92f0365633-config-data\") pod \"glance-db-purge-29423521-ptqnd\" (UID: \"ea21aaf5-23c0-47f7-86c0-1f92f0365633\") " pod="openstack/glance-db-purge-29423521-ptqnd" Dec 11 00:01:00 crc kubenswrapper[4791]: I1211 00:01:00.431373 4791 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-582zc\" (UniqueName: \"kubernetes.io/projected/ea21aaf5-23c0-47f7-86c0-1f92f0365633-kube-api-access-582zc\") pod \"glance-db-purge-29423521-ptqnd\" (UID: \"ea21aaf5-23c0-47f7-86c0-1f92f0365633\") " pod="openstack/glance-db-purge-29423521-ptqnd" Dec 11 00:01:00 crc kubenswrapper[4791]: I1211 00:01:00.431579 4791 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-nrlcm\" (UniqueName: \"kubernetes.io/projected/9349278c-a330-4119-9de6-ca3cb6f28ce7-kube-api-access-nrlcm\") pod \"keystone-cron-29423521-vlkhm\" (UID: \"9349278c-a330-4119-9de6-ca3cb6f28ce7\") " pod="openstack/keystone-cron-29423521-vlkhm" Dec 11 00:01:00 crc kubenswrapper[4791]: I1211 00:01:00.434643 4791 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-p45qx\" (UniqueName: \"kubernetes.io/projected/d332f1db-3d1a-4e2d-b2d2-69498733ceaf-kube-api-access-p45qx\") pod \"cinder-db-purge-29423521-h646b\" (UID: \"d332f1db-3d1a-4e2d-b2d2-69498733ceaf\") " pod="openstack/cinder-db-purge-29423521-h646b" Dec 11 00:01:00 crc kubenswrapper[4791]: I1211 00:01:00.509741 4791 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-db-purge-29423521-ptqnd" Dec 11 00:01:00 crc kubenswrapper[4791]: I1211 00:01:00.522262 4791 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-cron-29423521-vlkhm" Dec 11 00:01:00 crc kubenswrapper[4791]: I1211 00:01:00.532645 4791 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-db-purge-29423521-h646b" Dec 11 00:01:01 crc kubenswrapper[4791]: I1211 00:01:01.011224 4791 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/keystone-cron-29423521-vlkhm"] Dec 11 00:01:01 crc kubenswrapper[4791]: I1211 00:01:01.587192 4791 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-db-purge-29423521-ptqnd"] Dec 11 00:01:01 crc kubenswrapper[4791]: W1211 00:01:01.589280 4791 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podea21aaf5_23c0_47f7_86c0_1f92f0365633.slice/crio-af15d0235a763b87568c830f3ec4e95ceec0d67f4a0cb72d27b9c65c54042546 WatchSource:0}: Error finding container af15d0235a763b87568c830f3ec4e95ceec0d67f4a0cb72d27b9c65c54042546: Status 404 returned error can't find the container with id af15d0235a763b87568c830f3ec4e95ceec0d67f4a0cb72d27b9c65c54042546 Dec 11 00:01:01 crc kubenswrapper[4791]: I1211 00:01:01.606187 4791 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-db-purge-29423521-h646b"] Dec 11 00:01:01 crc kubenswrapper[4791]: W1211 00:01:01.616166 4791 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podd332f1db_3d1a_4e2d_b2d2_69498733ceaf.slice/crio-1c45271ec45e4bb03950bb74f40d179fffca017c73806c2b42af752f0b668b2a WatchSource:0}: Error finding container 1c45271ec45e4bb03950bb74f40d179fffca017c73806c2b42af752f0b668b2a: Status 404 returned error can't find the container with id 1c45271ec45e4bb03950bb74f40d179fffca017c73806c2b42af752f0b668b2a Dec 11 00:01:02 crc kubenswrapper[4791]: I1211 00:01:02.049514 4791 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-db-purge-29423521-ptqnd" event={"ID":"ea21aaf5-23c0-47f7-86c0-1f92f0365633","Type":"ContainerStarted","Data":"af15d0235a763b87568c830f3ec4e95ceec0d67f4a0cb72d27b9c65c54042546"} Dec 11 00:01:02 crc kubenswrapper[4791]: I1211 00:01:02.054681 4791 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-cron-29423521-vlkhm" event={"ID":"9349278c-a330-4119-9de6-ca3cb6f28ce7","Type":"ContainerStarted","Data":"271ffb63e9f4ce72805a0ee33eb1e7b0f4d8b67af092d51e0098016937e2b5f5"} Dec 11 00:01:02 crc kubenswrapper[4791]: I1211 00:01:02.054728 4791 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-cron-29423521-vlkhm" event={"ID":"9349278c-a330-4119-9de6-ca3cb6f28ce7","Type":"ContainerStarted","Data":"dbfdbe3ccecf7f5629f52560483defd9b36567c5453c4513a876d63f967e562b"} Dec 11 00:01:02 crc kubenswrapper[4791]: I1211 00:01:02.057380 4791 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-db-purge-29423521-h646b" event={"ID":"d332f1db-3d1a-4e2d-b2d2-69498733ceaf","Type":"ContainerStarted","Data":"1c45271ec45e4bb03950bb74f40d179fffca017c73806c2b42af752f0b668b2a"} Dec 11 00:01:02 crc kubenswrapper[4791]: I1211 00:01:02.071521 4791 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/keystone-cron-29423521-vlkhm" podStartSLOduration=2.071498306 podStartE2EDuration="2.071498306s" podCreationTimestamp="2025-12-11 00:01:00 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-11 00:01:02.067525894 +0000 UTC m=+4296.497143507" watchObservedRunningTime="2025-12-11 00:01:02.071498306 +0000 UTC m=+4296.501115919" Dec 11 00:01:03 crc kubenswrapper[4791]: I1211 00:01:03.071328 4791 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-db-purge-29423521-h646b" event={"ID":"d332f1db-3d1a-4e2d-b2d2-69498733ceaf","Type":"ContainerStarted","Data":"81fc97267a261176f46584c284f15646ccc831fae9954f389f943186621c1c90"} Dec 11 00:01:03 crc kubenswrapper[4791]: I1211 00:01:03.075083 4791 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-db-purge-29423521-ptqnd" event={"ID":"ea21aaf5-23c0-47f7-86c0-1f92f0365633","Type":"ContainerStarted","Data":"d6e184fe9c12efaf5d63dc4ba4816602a0cac583e3191b65ad1c564600cc3bd9"} Dec 11 00:01:03 crc kubenswrapper[4791]: I1211 00:01:03.095383 4791 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/cinder-db-purge-29423521-h646b" podStartSLOduration=3.095368059 podStartE2EDuration="3.095368059s" podCreationTimestamp="2025-12-11 00:01:00 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-11 00:01:03.090219434 +0000 UTC m=+4297.519837077" watchObservedRunningTime="2025-12-11 00:01:03.095368059 +0000 UTC m=+4297.524985672" Dec 11 00:01:03 crc kubenswrapper[4791]: I1211 00:01:03.116972 4791 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/glance-db-purge-29423521-ptqnd" podStartSLOduration=3.116946238 podStartE2EDuration="3.116946238s" podCreationTimestamp="2025-12-11 00:01:00 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-11 00:01:03.107774069 +0000 UTC m=+4297.537391692" watchObservedRunningTime="2025-12-11 00:01:03.116946238 +0000 UTC m=+4297.546563871" Dec 11 00:01:04 crc kubenswrapper[4791]: I1211 00:01:04.084683 4791 generic.go:334] "Generic (PLEG): container finished" podID="9349278c-a330-4119-9de6-ca3cb6f28ce7" containerID="271ffb63e9f4ce72805a0ee33eb1e7b0f4d8b67af092d51e0098016937e2b5f5" exitCode=0 Dec 11 00:01:04 crc kubenswrapper[4791]: I1211 00:01:04.084800 4791 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-cron-29423521-vlkhm" event={"ID":"9349278c-a330-4119-9de6-ca3cb6f28ce7","Type":"ContainerDied","Data":"271ffb63e9f4ce72805a0ee33eb1e7b0f4d8b67af092d51e0098016937e2b5f5"} Dec 11 00:01:05 crc kubenswrapper[4791]: I1211 00:01:05.099692 4791 generic.go:334] "Generic (PLEG): container finished" podID="ea21aaf5-23c0-47f7-86c0-1f92f0365633" containerID="d6e184fe9c12efaf5d63dc4ba4816602a0cac583e3191b65ad1c564600cc3bd9" exitCode=0 Dec 11 00:01:05 crc kubenswrapper[4791]: I1211 00:01:05.099829 4791 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-db-purge-29423521-ptqnd" event={"ID":"ea21aaf5-23c0-47f7-86c0-1f92f0365633","Type":"ContainerDied","Data":"d6e184fe9c12efaf5d63dc4ba4816602a0cac583e3191b65ad1c564600cc3bd9"} Dec 11 00:01:05 crc kubenswrapper[4791]: I1211 00:01:05.101957 4791 generic.go:334] "Generic (PLEG): container finished" podID="d332f1db-3d1a-4e2d-b2d2-69498733ceaf" containerID="81fc97267a261176f46584c284f15646ccc831fae9954f389f943186621c1c90" exitCode=0 Dec 11 00:01:05 crc kubenswrapper[4791]: I1211 00:01:05.101995 4791 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-db-purge-29423521-h646b" event={"ID":"d332f1db-3d1a-4e2d-b2d2-69498733ceaf","Type":"ContainerDied","Data":"81fc97267a261176f46584c284f15646ccc831fae9954f389f943186621c1c90"} Dec 11 00:01:05 crc kubenswrapper[4791]: I1211 00:01:05.507233 4791 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-cron-29423521-vlkhm" Dec 11 00:01:05 crc kubenswrapper[4791]: I1211 00:01:05.570015 4791 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/9349278c-a330-4119-9de6-ca3cb6f28ce7-combined-ca-bundle\") pod \"9349278c-a330-4119-9de6-ca3cb6f28ce7\" (UID: \"9349278c-a330-4119-9de6-ca3cb6f28ce7\") " Dec 11 00:01:05 crc kubenswrapper[4791]: I1211 00:01:05.570094 4791 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/9349278c-a330-4119-9de6-ca3cb6f28ce7-config-data\") pod \"9349278c-a330-4119-9de6-ca3cb6f28ce7\" (UID: \"9349278c-a330-4119-9de6-ca3cb6f28ce7\") " Dec 11 00:01:05 crc kubenswrapper[4791]: I1211 00:01:05.570150 4791 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-nrlcm\" (UniqueName: \"kubernetes.io/projected/9349278c-a330-4119-9de6-ca3cb6f28ce7-kube-api-access-nrlcm\") pod \"9349278c-a330-4119-9de6-ca3cb6f28ce7\" (UID: \"9349278c-a330-4119-9de6-ca3cb6f28ce7\") " Dec 11 00:01:05 crc kubenswrapper[4791]: I1211 00:01:05.570206 4791 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/9349278c-a330-4119-9de6-ca3cb6f28ce7-fernet-keys\") pod \"9349278c-a330-4119-9de6-ca3cb6f28ce7\" (UID: \"9349278c-a330-4119-9de6-ca3cb6f28ce7\") " Dec 11 00:01:05 crc kubenswrapper[4791]: I1211 00:01:05.579322 4791 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/9349278c-a330-4119-9de6-ca3cb6f28ce7-kube-api-access-nrlcm" (OuterVolumeSpecName: "kube-api-access-nrlcm") pod "9349278c-a330-4119-9de6-ca3cb6f28ce7" (UID: "9349278c-a330-4119-9de6-ca3cb6f28ce7"). InnerVolumeSpecName "kube-api-access-nrlcm". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 11 00:01:05 crc kubenswrapper[4791]: I1211 00:01:05.582243 4791 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/9349278c-a330-4119-9de6-ca3cb6f28ce7-fernet-keys" (OuterVolumeSpecName: "fernet-keys") pod "9349278c-a330-4119-9de6-ca3cb6f28ce7" (UID: "9349278c-a330-4119-9de6-ca3cb6f28ce7"). InnerVolumeSpecName "fernet-keys". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 11 00:01:05 crc kubenswrapper[4791]: I1211 00:01:05.672395 4791 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-nrlcm\" (UniqueName: \"kubernetes.io/projected/9349278c-a330-4119-9de6-ca3cb6f28ce7-kube-api-access-nrlcm\") on node \"crc\" DevicePath \"\"" Dec 11 00:01:05 crc kubenswrapper[4791]: I1211 00:01:05.672438 4791 reconciler_common.go:293] "Volume detached for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/9349278c-a330-4119-9de6-ca3cb6f28ce7-fernet-keys\") on node \"crc\" DevicePath \"\"" Dec 11 00:01:05 crc kubenswrapper[4791]: I1211 00:01:05.791218 4791 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-must-gather-2dlqt/must-gather-rqrn8"] Dec 11 00:01:05 crc kubenswrapper[4791]: I1211 00:01:05.791521 4791 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-must-gather-2dlqt/must-gather-rqrn8" podUID="6e53fefe-b61d-4ba3-a360-4ae9f1213d1c" containerName="copy" containerID="cri-o://7fc47c42954af916ee253f740108cdb8db79306d0712fc787fe2aa1fe0310e3d" gracePeriod=2 Dec 11 00:01:05 crc kubenswrapper[4791]: I1211 00:01:05.800794 4791 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-must-gather-2dlqt/must-gather-rqrn8"] Dec 11 00:01:06 crc kubenswrapper[4791]: I1211 00:01:06.115686 4791 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-cron-29423521-vlkhm" Dec 11 00:01:06 crc kubenswrapper[4791]: I1211 00:01:06.277928 4791 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/9349278c-a330-4119-9de6-ca3cb6f28ce7-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "9349278c-a330-4119-9de6-ca3cb6f28ce7" (UID: "9349278c-a330-4119-9de6-ca3cb6f28ce7"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 11 00:01:06 crc kubenswrapper[4791]: I1211 00:01:06.279817 4791 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/9349278c-a330-4119-9de6-ca3cb6f28ce7-config-data" (OuterVolumeSpecName: "config-data") pod "9349278c-a330-4119-9de6-ca3cb6f28ce7" (UID: "9349278c-a330-4119-9de6-ca3cb6f28ce7"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 11 00:01:06 crc kubenswrapper[4791]: I1211 00:01:06.289661 4791 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/9349278c-a330-4119-9de6-ca3cb6f28ce7-config-data\") on node \"crc\" DevicePath \"\"" Dec 11 00:01:06 crc kubenswrapper[4791]: I1211 00:01:06.289691 4791 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/9349278c-a330-4119-9de6-ca3cb6f28ce7-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 11 00:01:06 crc kubenswrapper[4791]: I1211 00:01:06.416236 4791 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-cron-29423521-vlkhm" event={"ID":"9349278c-a330-4119-9de6-ca3cb6f28ce7","Type":"ContainerDied","Data":"dbfdbe3ccecf7f5629f52560483defd9b36567c5453c4513a876d63f967e562b"} Dec 11 00:01:06 crc kubenswrapper[4791]: I1211 00:01:06.416279 4791 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="dbfdbe3ccecf7f5629f52560483defd9b36567c5453c4513a876d63f967e562b" Dec 11 00:01:06 crc kubenswrapper[4791]: I1211 00:01:06.542592 4791 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-db-purge-29423521-h646b" Dec 11 00:01:06 crc kubenswrapper[4791]: I1211 00:01:06.557545 4791 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glance-db-purge-29423521-ptqnd" Dec 11 00:01:06 crc kubenswrapper[4791]: I1211 00:01:06.595675 4791 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"db-purge-config-data\" (UniqueName: \"kubernetes.io/secret/ea21aaf5-23c0-47f7-86c0-1f92f0365633-db-purge-config-data\") pod \"ea21aaf5-23c0-47f7-86c0-1f92f0365633\" (UID: \"ea21aaf5-23c0-47f7-86c0-1f92f0365633\") " Dec 11 00:01:06 crc kubenswrapper[4791]: I1211 00:01:06.595790 4791 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d332f1db-3d1a-4e2d-b2d2-69498733ceaf-combined-ca-bundle\") pod \"d332f1db-3d1a-4e2d-b2d2-69498733ceaf\" (UID: \"d332f1db-3d1a-4e2d-b2d2-69498733ceaf\") " Dec 11 00:01:06 crc kubenswrapper[4791]: I1211 00:01:06.595937 4791 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"db-purge-config-data\" (UniqueName: \"kubernetes.io/secret/d332f1db-3d1a-4e2d-b2d2-69498733ceaf-db-purge-config-data\") pod \"d332f1db-3d1a-4e2d-b2d2-69498733ceaf\" (UID: \"d332f1db-3d1a-4e2d-b2d2-69498733ceaf\") " Dec 11 00:01:06 crc kubenswrapper[4791]: I1211 00:01:06.596036 4791 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/ea21aaf5-23c0-47f7-86c0-1f92f0365633-config-data\") pod \"ea21aaf5-23c0-47f7-86c0-1f92f0365633\" (UID: \"ea21aaf5-23c0-47f7-86c0-1f92f0365633\") " Dec 11 00:01:06 crc kubenswrapper[4791]: I1211 00:01:06.596077 4791 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/d332f1db-3d1a-4e2d-b2d2-69498733ceaf-config-data\") pod \"d332f1db-3d1a-4e2d-b2d2-69498733ceaf\" (UID: \"d332f1db-3d1a-4e2d-b2d2-69498733ceaf\") " Dec 11 00:01:06 crc kubenswrapper[4791]: I1211 00:01:06.596130 4791 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ea21aaf5-23c0-47f7-86c0-1f92f0365633-combined-ca-bundle\") pod \"ea21aaf5-23c0-47f7-86c0-1f92f0365633\" (UID: \"ea21aaf5-23c0-47f7-86c0-1f92f0365633\") " Dec 11 00:01:06 crc kubenswrapper[4791]: I1211 00:01:06.596201 4791 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-582zc\" (UniqueName: \"kubernetes.io/projected/ea21aaf5-23c0-47f7-86c0-1f92f0365633-kube-api-access-582zc\") pod \"ea21aaf5-23c0-47f7-86c0-1f92f0365633\" (UID: \"ea21aaf5-23c0-47f7-86c0-1f92f0365633\") " Dec 11 00:01:06 crc kubenswrapper[4791]: I1211 00:01:06.596293 4791 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-p45qx\" (UniqueName: \"kubernetes.io/projected/d332f1db-3d1a-4e2d-b2d2-69498733ceaf-kube-api-access-p45qx\") pod \"d332f1db-3d1a-4e2d-b2d2-69498733ceaf\" (UID: \"d332f1db-3d1a-4e2d-b2d2-69498733ceaf\") " Dec 11 00:01:06 crc kubenswrapper[4791]: I1211 00:01:06.600471 4791 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/d332f1db-3d1a-4e2d-b2d2-69498733ceaf-db-purge-config-data" (OuterVolumeSpecName: "db-purge-config-data") pod "d332f1db-3d1a-4e2d-b2d2-69498733ceaf" (UID: "d332f1db-3d1a-4e2d-b2d2-69498733ceaf"). InnerVolumeSpecName "db-purge-config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 11 00:01:06 crc kubenswrapper[4791]: I1211 00:01:06.601070 4791 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/ea21aaf5-23c0-47f7-86c0-1f92f0365633-kube-api-access-582zc" (OuterVolumeSpecName: "kube-api-access-582zc") pod "ea21aaf5-23c0-47f7-86c0-1f92f0365633" (UID: "ea21aaf5-23c0-47f7-86c0-1f92f0365633"). InnerVolumeSpecName "kube-api-access-582zc". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 11 00:01:06 crc kubenswrapper[4791]: I1211 00:01:06.602984 4791 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/d332f1db-3d1a-4e2d-b2d2-69498733ceaf-kube-api-access-p45qx" (OuterVolumeSpecName: "kube-api-access-p45qx") pod "d332f1db-3d1a-4e2d-b2d2-69498733ceaf" (UID: "d332f1db-3d1a-4e2d-b2d2-69498733ceaf"). InnerVolumeSpecName "kube-api-access-p45qx". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 11 00:01:06 crc kubenswrapper[4791]: I1211 00:01:06.669742 4791 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-must-gather-2dlqt_must-gather-rqrn8_6e53fefe-b61d-4ba3-a360-4ae9f1213d1c/copy/0.log" Dec 11 00:01:06 crc kubenswrapper[4791]: I1211 00:01:06.670508 4791 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-2dlqt/must-gather-rqrn8" Dec 11 00:01:06 crc kubenswrapper[4791]: I1211 00:01:06.682636 4791 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/ea21aaf5-23c0-47f7-86c0-1f92f0365633-db-purge-config-data" (OuterVolumeSpecName: "db-purge-config-data") pod "ea21aaf5-23c0-47f7-86c0-1f92f0365633" (UID: "ea21aaf5-23c0-47f7-86c0-1f92f0365633"). InnerVolumeSpecName "db-purge-config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 11 00:01:06 crc kubenswrapper[4791]: I1211 00:01:06.698780 4791 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-ppqw4\" (UniqueName: \"kubernetes.io/projected/6e53fefe-b61d-4ba3-a360-4ae9f1213d1c-kube-api-access-ppqw4\") pod \"6e53fefe-b61d-4ba3-a360-4ae9f1213d1c\" (UID: \"6e53fefe-b61d-4ba3-a360-4ae9f1213d1c\") " Dec 11 00:01:06 crc kubenswrapper[4791]: I1211 00:01:06.698918 4791 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"must-gather-output\" (UniqueName: \"kubernetes.io/empty-dir/6e53fefe-b61d-4ba3-a360-4ae9f1213d1c-must-gather-output\") pod \"6e53fefe-b61d-4ba3-a360-4ae9f1213d1c\" (UID: \"6e53fefe-b61d-4ba3-a360-4ae9f1213d1c\") " Dec 11 00:01:06 crc kubenswrapper[4791]: I1211 00:01:06.700437 4791 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-582zc\" (UniqueName: \"kubernetes.io/projected/ea21aaf5-23c0-47f7-86c0-1f92f0365633-kube-api-access-582zc\") on node \"crc\" DevicePath \"\"" Dec 11 00:01:06 crc kubenswrapper[4791]: I1211 00:01:06.700459 4791 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-p45qx\" (UniqueName: \"kubernetes.io/projected/d332f1db-3d1a-4e2d-b2d2-69498733ceaf-kube-api-access-p45qx\") on node \"crc\" DevicePath \"\"" Dec 11 00:01:06 crc kubenswrapper[4791]: I1211 00:01:06.700468 4791 reconciler_common.go:293] "Volume detached for volume \"db-purge-config-data\" (UniqueName: \"kubernetes.io/secret/ea21aaf5-23c0-47f7-86c0-1f92f0365633-db-purge-config-data\") on node \"crc\" DevicePath \"\"" Dec 11 00:01:06 crc kubenswrapper[4791]: I1211 00:01:06.700476 4791 reconciler_common.go:293] "Volume detached for volume \"db-purge-config-data\" (UniqueName: \"kubernetes.io/secret/d332f1db-3d1a-4e2d-b2d2-69498733ceaf-db-purge-config-data\") on node \"crc\" DevicePath \"\"" Dec 11 00:01:06 crc kubenswrapper[4791]: I1211 00:01:06.703131 4791 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/6e53fefe-b61d-4ba3-a360-4ae9f1213d1c-kube-api-access-ppqw4" (OuterVolumeSpecName: "kube-api-access-ppqw4") pod "6e53fefe-b61d-4ba3-a360-4ae9f1213d1c" (UID: "6e53fefe-b61d-4ba3-a360-4ae9f1213d1c"). InnerVolumeSpecName "kube-api-access-ppqw4". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 11 00:01:06 crc kubenswrapper[4791]: I1211 00:01:06.710247 4791 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/d332f1db-3d1a-4e2d-b2d2-69498733ceaf-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "d332f1db-3d1a-4e2d-b2d2-69498733ceaf" (UID: "d332f1db-3d1a-4e2d-b2d2-69498733ceaf"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 11 00:01:06 crc kubenswrapper[4791]: I1211 00:01:06.715877 4791 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/ea21aaf5-23c0-47f7-86c0-1f92f0365633-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "ea21aaf5-23c0-47f7-86c0-1f92f0365633" (UID: "ea21aaf5-23c0-47f7-86c0-1f92f0365633"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 11 00:01:06 crc kubenswrapper[4791]: I1211 00:01:06.718060 4791 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/d332f1db-3d1a-4e2d-b2d2-69498733ceaf-config-data" (OuterVolumeSpecName: "config-data") pod "d332f1db-3d1a-4e2d-b2d2-69498733ceaf" (UID: "d332f1db-3d1a-4e2d-b2d2-69498733ceaf"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 11 00:01:06 crc kubenswrapper[4791]: I1211 00:01:06.721528 4791 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/ea21aaf5-23c0-47f7-86c0-1f92f0365633-config-data" (OuterVolumeSpecName: "config-data") pod "ea21aaf5-23c0-47f7-86c0-1f92f0365633" (UID: "ea21aaf5-23c0-47f7-86c0-1f92f0365633"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 11 00:01:06 crc kubenswrapper[4791]: I1211 00:01:06.801846 4791 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d332f1db-3d1a-4e2d-b2d2-69498733ceaf-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 11 00:01:06 crc kubenswrapper[4791]: I1211 00:01:06.801879 4791 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-ppqw4\" (UniqueName: \"kubernetes.io/projected/6e53fefe-b61d-4ba3-a360-4ae9f1213d1c-kube-api-access-ppqw4\") on node \"crc\" DevicePath \"\"" Dec 11 00:01:06 crc kubenswrapper[4791]: I1211 00:01:06.801890 4791 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/ea21aaf5-23c0-47f7-86c0-1f92f0365633-config-data\") on node \"crc\" DevicePath \"\"" Dec 11 00:01:06 crc kubenswrapper[4791]: I1211 00:01:06.801900 4791 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/d332f1db-3d1a-4e2d-b2d2-69498733ceaf-config-data\") on node \"crc\" DevicePath \"\"" Dec 11 00:01:06 crc kubenswrapper[4791]: I1211 00:01:06.801908 4791 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ea21aaf5-23c0-47f7-86c0-1f92f0365633-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 11 00:01:06 crc kubenswrapper[4791]: I1211 00:01:06.898607 4791 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/6e53fefe-b61d-4ba3-a360-4ae9f1213d1c-must-gather-output" (OuterVolumeSpecName: "must-gather-output") pod "6e53fefe-b61d-4ba3-a360-4ae9f1213d1c" (UID: "6e53fefe-b61d-4ba3-a360-4ae9f1213d1c"). InnerVolumeSpecName "must-gather-output". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 11 00:01:06 crc kubenswrapper[4791]: I1211 00:01:06.906484 4791 reconciler_common.go:293] "Volume detached for volume \"must-gather-output\" (UniqueName: \"kubernetes.io/empty-dir/6e53fefe-b61d-4ba3-a360-4ae9f1213d1c-must-gather-output\") on node \"crc\" DevicePath \"\"" Dec 11 00:01:07 crc kubenswrapper[4791]: I1211 00:01:07.124548 4791 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-must-gather-2dlqt_must-gather-rqrn8_6e53fefe-b61d-4ba3-a360-4ae9f1213d1c/copy/0.log" Dec 11 00:01:07 crc kubenswrapper[4791]: I1211 00:01:07.125064 4791 generic.go:334] "Generic (PLEG): container finished" podID="6e53fefe-b61d-4ba3-a360-4ae9f1213d1c" containerID="7fc47c42954af916ee253f740108cdb8db79306d0712fc787fe2aa1fe0310e3d" exitCode=143 Dec 11 00:01:07 crc kubenswrapper[4791]: I1211 00:01:07.125120 4791 scope.go:117] "RemoveContainer" containerID="7fc47c42954af916ee253f740108cdb8db79306d0712fc787fe2aa1fe0310e3d" Dec 11 00:01:07 crc kubenswrapper[4791]: I1211 00:01:07.125120 4791 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-2dlqt/must-gather-rqrn8" Dec 11 00:01:07 crc kubenswrapper[4791]: I1211 00:01:07.126774 4791 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-db-purge-29423521-h646b" event={"ID":"d332f1db-3d1a-4e2d-b2d2-69498733ceaf","Type":"ContainerDied","Data":"1c45271ec45e4bb03950bb74f40d179fffca017c73806c2b42af752f0b668b2a"} Dec 11 00:01:07 crc kubenswrapper[4791]: I1211 00:01:07.126805 4791 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="1c45271ec45e4bb03950bb74f40d179fffca017c73806c2b42af752f0b668b2a" Dec 11 00:01:07 crc kubenswrapper[4791]: I1211 00:01:07.126836 4791 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-db-purge-29423521-h646b" Dec 11 00:01:07 crc kubenswrapper[4791]: I1211 00:01:07.131609 4791 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-db-purge-29423521-ptqnd" event={"ID":"ea21aaf5-23c0-47f7-86c0-1f92f0365633","Type":"ContainerDied","Data":"af15d0235a763b87568c830f3ec4e95ceec0d67f4a0cb72d27b9c65c54042546"} Dec 11 00:01:07 crc kubenswrapper[4791]: I1211 00:01:07.131649 4791 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="af15d0235a763b87568c830f3ec4e95ceec0d67f4a0cb72d27b9c65c54042546" Dec 11 00:01:07 crc kubenswrapper[4791]: I1211 00:01:07.131705 4791 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glance-db-purge-29423521-ptqnd" Dec 11 00:01:07 crc kubenswrapper[4791]: I1211 00:01:07.147074 4791 scope.go:117] "RemoveContainer" containerID="fd094fcb02987fc8ae45b115d82c5662af63b3157267e481ca35913a767e24aa" Dec 11 00:01:07 crc kubenswrapper[4791]: I1211 00:01:07.223451 4791 scope.go:117] "RemoveContainer" containerID="7fc47c42954af916ee253f740108cdb8db79306d0712fc787fe2aa1fe0310e3d" Dec 11 00:01:07 crc kubenswrapper[4791]: E1211 00:01:07.223893 4791 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"7fc47c42954af916ee253f740108cdb8db79306d0712fc787fe2aa1fe0310e3d\": container with ID starting with 7fc47c42954af916ee253f740108cdb8db79306d0712fc787fe2aa1fe0310e3d not found: ID does not exist" containerID="7fc47c42954af916ee253f740108cdb8db79306d0712fc787fe2aa1fe0310e3d" Dec 11 00:01:07 crc kubenswrapper[4791]: I1211 00:01:07.223927 4791 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"7fc47c42954af916ee253f740108cdb8db79306d0712fc787fe2aa1fe0310e3d"} err="failed to get container status \"7fc47c42954af916ee253f740108cdb8db79306d0712fc787fe2aa1fe0310e3d\": rpc error: code = NotFound desc = could not find container \"7fc47c42954af916ee253f740108cdb8db79306d0712fc787fe2aa1fe0310e3d\": container with ID starting with 7fc47c42954af916ee253f740108cdb8db79306d0712fc787fe2aa1fe0310e3d not found: ID does not exist" Dec 11 00:01:07 crc kubenswrapper[4791]: I1211 00:01:07.223944 4791 scope.go:117] "RemoveContainer" containerID="fd094fcb02987fc8ae45b115d82c5662af63b3157267e481ca35913a767e24aa" Dec 11 00:01:07 crc kubenswrapper[4791]: E1211 00:01:07.224471 4791 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"fd094fcb02987fc8ae45b115d82c5662af63b3157267e481ca35913a767e24aa\": container with ID starting with fd094fcb02987fc8ae45b115d82c5662af63b3157267e481ca35913a767e24aa not found: ID does not exist" containerID="fd094fcb02987fc8ae45b115d82c5662af63b3157267e481ca35913a767e24aa" Dec 11 00:01:07 crc kubenswrapper[4791]: I1211 00:01:07.224531 4791 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"fd094fcb02987fc8ae45b115d82c5662af63b3157267e481ca35913a767e24aa"} err="failed to get container status \"fd094fcb02987fc8ae45b115d82c5662af63b3157267e481ca35913a767e24aa\": rpc error: code = NotFound desc = could not find container \"fd094fcb02987fc8ae45b115d82c5662af63b3157267e481ca35913a767e24aa\": container with ID starting with fd094fcb02987fc8ae45b115d82c5662af63b3157267e481ca35913a767e24aa not found: ID does not exist" Dec 11 00:01:07 crc kubenswrapper[4791]: I1211 00:01:07.899130 4791 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="6e53fefe-b61d-4ba3-a360-4ae9f1213d1c" path="/var/lib/kubelet/pods/6e53fefe-b61d-4ba3-a360-4ae9f1213d1c/volumes" Dec 11 00:01:25 crc kubenswrapper[4791]: I1211 00:01:25.038885 4791 patch_prober.go:28] interesting pod/machine-config-daemon-5rb5l container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 11 00:01:25 crc kubenswrapper[4791]: I1211 00:01:25.039571 4791 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-5rb5l" podUID="ba35653c-6e06-4cee-a4d6-137764090d18" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 11 00:01:25 crc kubenswrapper[4791]: I1211 00:01:25.761033 4791 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-operators-mqx65"] Dec 11 00:01:25 crc kubenswrapper[4791]: E1211 00:01:25.761788 4791 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="ea21aaf5-23c0-47f7-86c0-1f92f0365633" containerName="glance-dbpurge" Dec 11 00:01:25 crc kubenswrapper[4791]: I1211 00:01:25.761806 4791 state_mem.go:107] "Deleted CPUSet assignment" podUID="ea21aaf5-23c0-47f7-86c0-1f92f0365633" containerName="glance-dbpurge" Dec 11 00:01:25 crc kubenswrapper[4791]: E1211 00:01:25.761823 4791 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="d332f1db-3d1a-4e2d-b2d2-69498733ceaf" containerName="cinder-db-purge" Dec 11 00:01:25 crc kubenswrapper[4791]: I1211 00:01:25.761829 4791 state_mem.go:107] "Deleted CPUSet assignment" podUID="d332f1db-3d1a-4e2d-b2d2-69498733ceaf" containerName="cinder-db-purge" Dec 11 00:01:25 crc kubenswrapper[4791]: E1211 00:01:25.761843 4791 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="6e53fefe-b61d-4ba3-a360-4ae9f1213d1c" containerName="gather" Dec 11 00:01:25 crc kubenswrapper[4791]: I1211 00:01:25.761850 4791 state_mem.go:107] "Deleted CPUSet assignment" podUID="6e53fefe-b61d-4ba3-a360-4ae9f1213d1c" containerName="gather" Dec 11 00:01:25 crc kubenswrapper[4791]: E1211 00:01:25.761863 4791 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="9349278c-a330-4119-9de6-ca3cb6f28ce7" containerName="keystone-cron" Dec 11 00:01:25 crc kubenswrapper[4791]: I1211 00:01:25.761868 4791 state_mem.go:107] "Deleted CPUSet assignment" podUID="9349278c-a330-4119-9de6-ca3cb6f28ce7" containerName="keystone-cron" Dec 11 00:01:25 crc kubenswrapper[4791]: E1211 00:01:25.761889 4791 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="6e53fefe-b61d-4ba3-a360-4ae9f1213d1c" containerName="copy" Dec 11 00:01:25 crc kubenswrapper[4791]: I1211 00:01:25.761895 4791 state_mem.go:107] "Deleted CPUSet assignment" podUID="6e53fefe-b61d-4ba3-a360-4ae9f1213d1c" containerName="copy" Dec 11 00:01:25 crc kubenswrapper[4791]: I1211 00:01:25.762065 4791 memory_manager.go:354] "RemoveStaleState removing state" podUID="ea21aaf5-23c0-47f7-86c0-1f92f0365633" containerName="glance-dbpurge" Dec 11 00:01:25 crc kubenswrapper[4791]: I1211 00:01:25.762078 4791 memory_manager.go:354] "RemoveStaleState removing state" podUID="6e53fefe-b61d-4ba3-a360-4ae9f1213d1c" containerName="copy" Dec 11 00:01:25 crc kubenswrapper[4791]: I1211 00:01:25.762093 4791 memory_manager.go:354] "RemoveStaleState removing state" podUID="d332f1db-3d1a-4e2d-b2d2-69498733ceaf" containerName="cinder-db-purge" Dec 11 00:01:25 crc kubenswrapper[4791]: I1211 00:01:25.762107 4791 memory_manager.go:354] "RemoveStaleState removing state" podUID="6e53fefe-b61d-4ba3-a360-4ae9f1213d1c" containerName="gather" Dec 11 00:01:25 crc kubenswrapper[4791]: I1211 00:01:25.762117 4791 memory_manager.go:354] "RemoveStaleState removing state" podUID="9349278c-a330-4119-9de6-ca3cb6f28ce7" containerName="keystone-cron" Dec 11 00:01:25 crc kubenswrapper[4791]: I1211 00:01:25.763680 4791 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-mqx65" Dec 11 00:01:25 crc kubenswrapper[4791]: I1211 00:01:25.784533 4791 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-mqx65"] Dec 11 00:01:25 crc kubenswrapper[4791]: I1211 00:01:25.872247 4791 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/3605379e-37bc-4b93-9226-e031301f0b9c-utilities\") pod \"redhat-operators-mqx65\" (UID: \"3605379e-37bc-4b93-9226-e031301f0b9c\") " pod="openshift-marketplace/redhat-operators-mqx65" Dec 11 00:01:25 crc kubenswrapper[4791]: I1211 00:01:25.872543 4791 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-fspnb\" (UniqueName: \"kubernetes.io/projected/3605379e-37bc-4b93-9226-e031301f0b9c-kube-api-access-fspnb\") pod \"redhat-operators-mqx65\" (UID: \"3605379e-37bc-4b93-9226-e031301f0b9c\") " pod="openshift-marketplace/redhat-operators-mqx65" Dec 11 00:01:25 crc kubenswrapper[4791]: I1211 00:01:25.872758 4791 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/3605379e-37bc-4b93-9226-e031301f0b9c-catalog-content\") pod \"redhat-operators-mqx65\" (UID: \"3605379e-37bc-4b93-9226-e031301f0b9c\") " pod="openshift-marketplace/redhat-operators-mqx65" Dec 11 00:01:25 crc kubenswrapper[4791]: I1211 00:01:25.975032 4791 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-fspnb\" (UniqueName: \"kubernetes.io/projected/3605379e-37bc-4b93-9226-e031301f0b9c-kube-api-access-fspnb\") pod \"redhat-operators-mqx65\" (UID: \"3605379e-37bc-4b93-9226-e031301f0b9c\") " pod="openshift-marketplace/redhat-operators-mqx65" Dec 11 00:01:25 crc kubenswrapper[4791]: I1211 00:01:25.975149 4791 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/3605379e-37bc-4b93-9226-e031301f0b9c-catalog-content\") pod \"redhat-operators-mqx65\" (UID: \"3605379e-37bc-4b93-9226-e031301f0b9c\") " pod="openshift-marketplace/redhat-operators-mqx65" Dec 11 00:01:25 crc kubenswrapper[4791]: I1211 00:01:25.975284 4791 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/3605379e-37bc-4b93-9226-e031301f0b9c-utilities\") pod \"redhat-operators-mqx65\" (UID: \"3605379e-37bc-4b93-9226-e031301f0b9c\") " pod="openshift-marketplace/redhat-operators-mqx65" Dec 11 00:01:25 crc kubenswrapper[4791]: I1211 00:01:25.975892 4791 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/3605379e-37bc-4b93-9226-e031301f0b9c-catalog-content\") pod \"redhat-operators-mqx65\" (UID: \"3605379e-37bc-4b93-9226-e031301f0b9c\") " pod="openshift-marketplace/redhat-operators-mqx65" Dec 11 00:01:25 crc kubenswrapper[4791]: I1211 00:01:25.976095 4791 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/3605379e-37bc-4b93-9226-e031301f0b9c-utilities\") pod \"redhat-operators-mqx65\" (UID: \"3605379e-37bc-4b93-9226-e031301f0b9c\") " pod="openshift-marketplace/redhat-operators-mqx65" Dec 11 00:01:25 crc kubenswrapper[4791]: I1211 00:01:25.996647 4791 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-fspnb\" (UniqueName: \"kubernetes.io/projected/3605379e-37bc-4b93-9226-e031301f0b9c-kube-api-access-fspnb\") pod \"redhat-operators-mqx65\" (UID: \"3605379e-37bc-4b93-9226-e031301f0b9c\") " pod="openshift-marketplace/redhat-operators-mqx65" Dec 11 00:01:26 crc kubenswrapper[4791]: I1211 00:01:26.085812 4791 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-mqx65" Dec 11 00:01:26 crc kubenswrapper[4791]: I1211 00:01:26.574870 4791 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-mqx65"] Dec 11 00:01:27 crc kubenswrapper[4791]: I1211 00:01:27.352653 4791 generic.go:334] "Generic (PLEG): container finished" podID="3605379e-37bc-4b93-9226-e031301f0b9c" containerID="9c2703eb9c38c5b44fd9bef1329515bcb122fc7a9ade8048da650051cd5925be" exitCode=0 Dec 11 00:01:27 crc kubenswrapper[4791]: I1211 00:01:27.352773 4791 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-mqx65" event={"ID":"3605379e-37bc-4b93-9226-e031301f0b9c","Type":"ContainerDied","Data":"9c2703eb9c38c5b44fd9bef1329515bcb122fc7a9ade8048da650051cd5925be"} Dec 11 00:01:27 crc kubenswrapper[4791]: I1211 00:01:27.352993 4791 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-mqx65" event={"ID":"3605379e-37bc-4b93-9226-e031301f0b9c","Type":"ContainerStarted","Data":"da4836f362e6d923ffe0557ef502830f2320fe89e1973f3a2ab8b714589d257d"} Dec 11 00:01:29 crc kubenswrapper[4791]: I1211 00:01:29.384145 4791 generic.go:334] "Generic (PLEG): container finished" podID="3605379e-37bc-4b93-9226-e031301f0b9c" containerID="3c55cfa25df90a82b39de008ce4f443ca5e2f76a3e9964fb42c6c609789fe353" exitCode=0 Dec 11 00:01:29 crc kubenswrapper[4791]: I1211 00:01:29.384255 4791 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-mqx65" event={"ID":"3605379e-37bc-4b93-9226-e031301f0b9c","Type":"ContainerDied","Data":"3c55cfa25df90a82b39de008ce4f443ca5e2f76a3e9964fb42c6c609789fe353"} Dec 11 00:01:32 crc kubenswrapper[4791]: I1211 00:01:32.414953 4791 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-mqx65" event={"ID":"3605379e-37bc-4b93-9226-e031301f0b9c","Type":"ContainerStarted","Data":"a2d7df9cf387dce07d1b8c250c2dad57c696290c743b64666c35bbf4ef84cd9a"} Dec 11 00:01:32 crc kubenswrapper[4791]: I1211 00:01:32.429791 4791 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-operators-mqx65" podStartSLOduration=3.874355259 podStartE2EDuration="7.429772482s" podCreationTimestamp="2025-12-11 00:01:25 +0000 UTC" firstStartedPulling="2025-12-11 00:01:27.354257512 +0000 UTC m=+4321.783875125" lastFinishedPulling="2025-12-11 00:01:30.909674725 +0000 UTC m=+4325.339292348" observedRunningTime="2025-12-11 00:01:32.428387623 +0000 UTC m=+4326.858005246" watchObservedRunningTime="2025-12-11 00:01:32.429772482 +0000 UTC m=+4326.859390095" Dec 11 00:01:36 crc kubenswrapper[4791]: I1211 00:01:36.086551 4791 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-operators-mqx65" Dec 11 00:01:36 crc kubenswrapper[4791]: I1211 00:01:36.087562 4791 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-operators-mqx65" Dec 11 00:01:37 crc kubenswrapper[4791]: I1211 00:01:37.154309 4791 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-marketplace/redhat-operators-mqx65" podUID="3605379e-37bc-4b93-9226-e031301f0b9c" containerName="registry-server" probeResult="failure" output=< Dec 11 00:01:37 crc kubenswrapper[4791]: timeout: failed to connect service ":50051" within 1s Dec 11 00:01:37 crc kubenswrapper[4791]: > Dec 11 00:01:46 crc kubenswrapper[4791]: I1211 00:01:46.151423 4791 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-operators-mqx65" Dec 11 00:01:46 crc kubenswrapper[4791]: I1211 00:01:46.224733 4791 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-operators-mqx65" Dec 11 00:01:46 crc kubenswrapper[4791]: I1211 00:01:46.390532 4791 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-mqx65"] Dec 11 00:01:47 crc kubenswrapper[4791]: I1211 00:01:47.593408 4791 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-operators-mqx65" podUID="3605379e-37bc-4b93-9226-e031301f0b9c" containerName="registry-server" containerID="cri-o://a2d7df9cf387dce07d1b8c250c2dad57c696290c743b64666c35bbf4ef84cd9a" gracePeriod=2 Dec 11 00:01:48 crc kubenswrapper[4791]: I1211 00:01:48.048442 4791 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-mqx65" Dec 11 00:01:48 crc kubenswrapper[4791]: I1211 00:01:48.194990 4791 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-fspnb\" (UniqueName: \"kubernetes.io/projected/3605379e-37bc-4b93-9226-e031301f0b9c-kube-api-access-fspnb\") pod \"3605379e-37bc-4b93-9226-e031301f0b9c\" (UID: \"3605379e-37bc-4b93-9226-e031301f0b9c\") " Dec 11 00:01:48 crc kubenswrapper[4791]: I1211 00:01:48.195197 4791 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/3605379e-37bc-4b93-9226-e031301f0b9c-utilities\") pod \"3605379e-37bc-4b93-9226-e031301f0b9c\" (UID: \"3605379e-37bc-4b93-9226-e031301f0b9c\") " Dec 11 00:01:48 crc kubenswrapper[4791]: I1211 00:01:48.195298 4791 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/3605379e-37bc-4b93-9226-e031301f0b9c-catalog-content\") pod \"3605379e-37bc-4b93-9226-e031301f0b9c\" (UID: \"3605379e-37bc-4b93-9226-e031301f0b9c\") " Dec 11 00:01:48 crc kubenswrapper[4791]: I1211 00:01:48.196146 4791 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/3605379e-37bc-4b93-9226-e031301f0b9c-utilities" (OuterVolumeSpecName: "utilities") pod "3605379e-37bc-4b93-9226-e031301f0b9c" (UID: "3605379e-37bc-4b93-9226-e031301f0b9c"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 11 00:01:48 crc kubenswrapper[4791]: I1211 00:01:48.211332 4791 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/3605379e-37bc-4b93-9226-e031301f0b9c-kube-api-access-fspnb" (OuterVolumeSpecName: "kube-api-access-fspnb") pod "3605379e-37bc-4b93-9226-e031301f0b9c" (UID: "3605379e-37bc-4b93-9226-e031301f0b9c"). InnerVolumeSpecName "kube-api-access-fspnb". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 11 00:01:48 crc kubenswrapper[4791]: I1211 00:01:48.297808 4791 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-fspnb\" (UniqueName: \"kubernetes.io/projected/3605379e-37bc-4b93-9226-e031301f0b9c-kube-api-access-fspnb\") on node \"crc\" DevicePath \"\"" Dec 11 00:01:48 crc kubenswrapper[4791]: I1211 00:01:48.297845 4791 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/3605379e-37bc-4b93-9226-e031301f0b9c-utilities\") on node \"crc\" DevicePath \"\"" Dec 11 00:01:48 crc kubenswrapper[4791]: I1211 00:01:48.326472 4791 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/3605379e-37bc-4b93-9226-e031301f0b9c-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "3605379e-37bc-4b93-9226-e031301f0b9c" (UID: "3605379e-37bc-4b93-9226-e031301f0b9c"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 11 00:01:48 crc kubenswrapper[4791]: I1211 00:01:48.400834 4791 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/3605379e-37bc-4b93-9226-e031301f0b9c-catalog-content\") on node \"crc\" DevicePath \"\"" Dec 11 00:01:48 crc kubenswrapper[4791]: I1211 00:01:48.606506 4791 generic.go:334] "Generic (PLEG): container finished" podID="3605379e-37bc-4b93-9226-e031301f0b9c" containerID="a2d7df9cf387dce07d1b8c250c2dad57c696290c743b64666c35bbf4ef84cd9a" exitCode=0 Dec 11 00:01:48 crc kubenswrapper[4791]: I1211 00:01:48.606679 4791 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-mqx65" event={"ID":"3605379e-37bc-4b93-9226-e031301f0b9c","Type":"ContainerDied","Data":"a2d7df9cf387dce07d1b8c250c2dad57c696290c743b64666c35bbf4ef84cd9a"} Dec 11 00:01:48 crc kubenswrapper[4791]: I1211 00:01:48.606807 4791 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-mqx65" event={"ID":"3605379e-37bc-4b93-9226-e031301f0b9c","Type":"ContainerDied","Data":"da4836f362e6d923ffe0557ef502830f2320fe89e1973f3a2ab8b714589d257d"} Dec 11 00:01:48 crc kubenswrapper[4791]: I1211 00:01:48.606832 4791 scope.go:117] "RemoveContainer" containerID="a2d7df9cf387dce07d1b8c250c2dad57c696290c743b64666c35bbf4ef84cd9a" Dec 11 00:01:48 crc kubenswrapper[4791]: I1211 00:01:48.606750 4791 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-mqx65" Dec 11 00:01:48 crc kubenswrapper[4791]: I1211 00:01:48.648100 4791 scope.go:117] "RemoveContainer" containerID="3c55cfa25df90a82b39de008ce4f443ca5e2f76a3e9964fb42c6c609789fe353" Dec 11 00:01:48 crc kubenswrapper[4791]: I1211 00:01:48.650012 4791 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-mqx65"] Dec 11 00:01:48 crc kubenswrapper[4791]: I1211 00:01:48.690143 4791 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-operators-mqx65"] Dec 11 00:01:48 crc kubenswrapper[4791]: I1211 00:01:48.695429 4791 scope.go:117] "RemoveContainer" containerID="9c2703eb9c38c5b44fd9bef1329515bcb122fc7a9ade8048da650051cd5925be" Dec 11 00:01:48 crc kubenswrapper[4791]: I1211 00:01:48.718880 4791 scope.go:117] "RemoveContainer" containerID="a2d7df9cf387dce07d1b8c250c2dad57c696290c743b64666c35bbf4ef84cd9a" Dec 11 00:01:48 crc kubenswrapper[4791]: E1211 00:01:48.719224 4791 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"a2d7df9cf387dce07d1b8c250c2dad57c696290c743b64666c35bbf4ef84cd9a\": container with ID starting with a2d7df9cf387dce07d1b8c250c2dad57c696290c743b64666c35bbf4ef84cd9a not found: ID does not exist" containerID="a2d7df9cf387dce07d1b8c250c2dad57c696290c743b64666c35bbf4ef84cd9a" Dec 11 00:01:48 crc kubenswrapper[4791]: I1211 00:01:48.719264 4791 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"a2d7df9cf387dce07d1b8c250c2dad57c696290c743b64666c35bbf4ef84cd9a"} err="failed to get container status \"a2d7df9cf387dce07d1b8c250c2dad57c696290c743b64666c35bbf4ef84cd9a\": rpc error: code = NotFound desc = could not find container \"a2d7df9cf387dce07d1b8c250c2dad57c696290c743b64666c35bbf4ef84cd9a\": container with ID starting with a2d7df9cf387dce07d1b8c250c2dad57c696290c743b64666c35bbf4ef84cd9a not found: ID does not exist" Dec 11 00:01:48 crc kubenswrapper[4791]: I1211 00:01:48.719290 4791 scope.go:117] "RemoveContainer" containerID="3c55cfa25df90a82b39de008ce4f443ca5e2f76a3e9964fb42c6c609789fe353" Dec 11 00:01:48 crc kubenswrapper[4791]: E1211 00:01:48.720518 4791 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"3c55cfa25df90a82b39de008ce4f443ca5e2f76a3e9964fb42c6c609789fe353\": container with ID starting with 3c55cfa25df90a82b39de008ce4f443ca5e2f76a3e9964fb42c6c609789fe353 not found: ID does not exist" containerID="3c55cfa25df90a82b39de008ce4f443ca5e2f76a3e9964fb42c6c609789fe353" Dec 11 00:01:48 crc kubenswrapper[4791]: I1211 00:01:48.720554 4791 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"3c55cfa25df90a82b39de008ce4f443ca5e2f76a3e9964fb42c6c609789fe353"} err="failed to get container status \"3c55cfa25df90a82b39de008ce4f443ca5e2f76a3e9964fb42c6c609789fe353\": rpc error: code = NotFound desc = could not find container \"3c55cfa25df90a82b39de008ce4f443ca5e2f76a3e9964fb42c6c609789fe353\": container with ID starting with 3c55cfa25df90a82b39de008ce4f443ca5e2f76a3e9964fb42c6c609789fe353 not found: ID does not exist" Dec 11 00:01:48 crc kubenswrapper[4791]: I1211 00:01:48.720579 4791 scope.go:117] "RemoveContainer" containerID="9c2703eb9c38c5b44fd9bef1329515bcb122fc7a9ade8048da650051cd5925be" Dec 11 00:01:48 crc kubenswrapper[4791]: E1211 00:01:48.720884 4791 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"9c2703eb9c38c5b44fd9bef1329515bcb122fc7a9ade8048da650051cd5925be\": container with ID starting with 9c2703eb9c38c5b44fd9bef1329515bcb122fc7a9ade8048da650051cd5925be not found: ID does not exist" containerID="9c2703eb9c38c5b44fd9bef1329515bcb122fc7a9ade8048da650051cd5925be" Dec 11 00:01:48 crc kubenswrapper[4791]: I1211 00:01:48.720919 4791 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"9c2703eb9c38c5b44fd9bef1329515bcb122fc7a9ade8048da650051cd5925be"} err="failed to get container status \"9c2703eb9c38c5b44fd9bef1329515bcb122fc7a9ade8048da650051cd5925be\": rpc error: code = NotFound desc = could not find container \"9c2703eb9c38c5b44fd9bef1329515bcb122fc7a9ade8048da650051cd5925be\": container with ID starting with 9c2703eb9c38c5b44fd9bef1329515bcb122fc7a9ade8048da650051cd5925be not found: ID does not exist" Dec 11 00:01:49 crc kubenswrapper[4791]: I1211 00:01:49.899277 4791 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="3605379e-37bc-4b93-9226-e031301f0b9c" path="/var/lib/kubelet/pods/3605379e-37bc-4b93-9226-e031301f0b9c/volumes" Dec 11 00:01:55 crc kubenswrapper[4791]: I1211 00:01:55.040593 4791 patch_prober.go:28] interesting pod/machine-config-daemon-5rb5l container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 11 00:01:55 crc kubenswrapper[4791]: I1211 00:01:55.042595 4791 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-5rb5l" podUID="ba35653c-6e06-4cee-a4d6-137764090d18" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 11 00:01:55 crc kubenswrapper[4791]: I1211 00:01:55.042646 4791 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-5rb5l" Dec 11 00:01:55 crc kubenswrapper[4791]: I1211 00:01:55.043605 4791 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"174b0c37e74227b1037fd59658b8e726e56afcb2c113dde44e2136ad132c104b"} pod="openshift-machine-config-operator/machine-config-daemon-5rb5l" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Dec 11 00:01:55 crc kubenswrapper[4791]: I1211 00:01:55.043666 4791 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-5rb5l" podUID="ba35653c-6e06-4cee-a4d6-137764090d18" containerName="machine-config-daemon" containerID="cri-o://174b0c37e74227b1037fd59658b8e726e56afcb2c113dde44e2136ad132c104b" gracePeriod=600 Dec 11 00:01:55 crc kubenswrapper[4791]: E1211 00:01:55.172621 4791 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-5rb5l_openshift-machine-config-operator(ba35653c-6e06-4cee-a4d6-137764090d18)\"" pod="openshift-machine-config-operator/machine-config-daemon-5rb5l" podUID="ba35653c-6e06-4cee-a4d6-137764090d18" Dec 11 00:01:55 crc kubenswrapper[4791]: I1211 00:01:55.683609 4791 generic.go:334] "Generic (PLEG): container finished" podID="ba35653c-6e06-4cee-a4d6-137764090d18" containerID="174b0c37e74227b1037fd59658b8e726e56afcb2c113dde44e2136ad132c104b" exitCode=0 Dec 11 00:01:55 crc kubenswrapper[4791]: I1211 00:01:55.683687 4791 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-5rb5l" event={"ID":"ba35653c-6e06-4cee-a4d6-137764090d18","Type":"ContainerDied","Data":"174b0c37e74227b1037fd59658b8e726e56afcb2c113dde44e2136ad132c104b"} Dec 11 00:01:55 crc kubenswrapper[4791]: I1211 00:01:55.683905 4791 scope.go:117] "RemoveContainer" containerID="1a35e34c3589dfe3497551cd10aa8a1e1fa1f05668706b1469b66da53eca8dc3" Dec 11 00:01:55 crc kubenswrapper[4791]: I1211 00:01:55.684869 4791 scope.go:117] "RemoveContainer" containerID="174b0c37e74227b1037fd59658b8e726e56afcb2c113dde44e2136ad132c104b" Dec 11 00:01:55 crc kubenswrapper[4791]: E1211 00:01:55.685188 4791 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-5rb5l_openshift-machine-config-operator(ba35653c-6e06-4cee-a4d6-137764090d18)\"" pod="openshift-machine-config-operator/machine-config-daemon-5rb5l" podUID="ba35653c-6e06-4cee-a4d6-137764090d18" Dec 11 00:02:09 crc kubenswrapper[4791]: I1211 00:02:09.885075 4791 scope.go:117] "RemoveContainer" containerID="174b0c37e74227b1037fd59658b8e726e56afcb2c113dde44e2136ad132c104b" Dec 11 00:02:09 crc kubenswrapper[4791]: E1211 00:02:09.885733 4791 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-5rb5l_openshift-machine-config-operator(ba35653c-6e06-4cee-a4d6-137764090d18)\"" pod="openshift-machine-config-operator/machine-config-daemon-5rb5l" podUID="ba35653c-6e06-4cee-a4d6-137764090d18" Dec 11 00:02:22 crc kubenswrapper[4791]: I1211 00:02:22.885118 4791 scope.go:117] "RemoveContainer" containerID="174b0c37e74227b1037fd59658b8e726e56afcb2c113dde44e2136ad132c104b" Dec 11 00:02:22 crc kubenswrapper[4791]: E1211 00:02:22.887165 4791 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-5rb5l_openshift-machine-config-operator(ba35653c-6e06-4cee-a4d6-137764090d18)\"" pod="openshift-machine-config-operator/machine-config-daemon-5rb5l" podUID="ba35653c-6e06-4cee-a4d6-137764090d18" Dec 11 00:02:36 crc kubenswrapper[4791]: I1211 00:02:36.911069 4791 scope.go:117] "RemoveContainer" containerID="174b0c37e74227b1037fd59658b8e726e56afcb2c113dde44e2136ad132c104b" Dec 11 00:02:36 crc kubenswrapper[4791]: E1211 00:02:36.912501 4791 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-5rb5l_openshift-machine-config-operator(ba35653c-6e06-4cee-a4d6-137764090d18)\"" pod="openshift-machine-config-operator/machine-config-daemon-5rb5l" podUID="ba35653c-6e06-4cee-a4d6-137764090d18" Dec 11 00:02:49 crc kubenswrapper[4791]: I1211 00:02:49.885270 4791 scope.go:117] "RemoveContainer" containerID="174b0c37e74227b1037fd59658b8e726e56afcb2c113dde44e2136ad132c104b" Dec 11 00:02:49 crc kubenswrapper[4791]: E1211 00:02:49.886434 4791 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-5rb5l_openshift-machine-config-operator(ba35653c-6e06-4cee-a4d6-137764090d18)\"" pod="openshift-machine-config-operator/machine-config-daemon-5rb5l" podUID="ba35653c-6e06-4cee-a4d6-137764090d18" Dec 11 00:03:04 crc kubenswrapper[4791]: I1211 00:03:04.886573 4791 scope.go:117] "RemoveContainer" containerID="174b0c37e74227b1037fd59658b8e726e56afcb2c113dde44e2136ad132c104b" Dec 11 00:03:04 crc kubenswrapper[4791]: E1211 00:03:04.888079 4791 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-5rb5l_openshift-machine-config-operator(ba35653c-6e06-4cee-a4d6-137764090d18)\"" pod="openshift-machine-config-operator/machine-config-daemon-5rb5l" podUID="ba35653c-6e06-4cee-a4d6-137764090d18" Dec 11 00:03:19 crc kubenswrapper[4791]: I1211 00:03:19.885331 4791 scope.go:117] "RemoveContainer" containerID="174b0c37e74227b1037fd59658b8e726e56afcb2c113dde44e2136ad132c104b" Dec 11 00:03:19 crc kubenswrapper[4791]: E1211 00:03:19.886124 4791 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-5rb5l_openshift-machine-config-operator(ba35653c-6e06-4cee-a4d6-137764090d18)\"" pod="openshift-machine-config-operator/machine-config-daemon-5rb5l" podUID="ba35653c-6e06-4cee-a4d6-137764090d18" Dec 11 00:03:32 crc kubenswrapper[4791]: I1211 00:03:32.884958 4791 scope.go:117] "RemoveContainer" containerID="174b0c37e74227b1037fd59658b8e726e56afcb2c113dde44e2136ad132c104b" Dec 11 00:03:32 crc kubenswrapper[4791]: E1211 00:03:32.885543 4791 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-5rb5l_openshift-machine-config-operator(ba35653c-6e06-4cee-a4d6-137764090d18)\"" pod="openshift-machine-config-operator/machine-config-daemon-5rb5l" podUID="ba35653c-6e06-4cee-a4d6-137764090d18" Dec 11 00:03:43 crc kubenswrapper[4791]: I1211 00:03:43.884922 4791 scope.go:117] "RemoveContainer" containerID="174b0c37e74227b1037fd59658b8e726e56afcb2c113dde44e2136ad132c104b" Dec 11 00:03:43 crc kubenswrapper[4791]: E1211 00:03:43.885789 4791 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-5rb5l_openshift-machine-config-operator(ba35653c-6e06-4cee-a4d6-137764090d18)\"" pod="openshift-machine-config-operator/machine-config-daemon-5rb5l" podUID="ba35653c-6e06-4cee-a4d6-137764090d18" Dec 11 00:03:54 crc kubenswrapper[4791]: I1211 00:03:54.885327 4791 scope.go:117] "RemoveContainer" containerID="174b0c37e74227b1037fd59658b8e726e56afcb2c113dde44e2136ad132c104b" Dec 11 00:03:54 crc kubenswrapper[4791]: E1211 00:03:54.886209 4791 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-5rb5l_openshift-machine-config-operator(ba35653c-6e06-4cee-a4d6-137764090d18)\"" pod="openshift-machine-config-operator/machine-config-daemon-5rb5l" podUID="ba35653c-6e06-4cee-a4d6-137764090d18" Dec 11 00:04:08 crc kubenswrapper[4791]: I1211 00:04:08.885312 4791 scope.go:117] "RemoveContainer" containerID="174b0c37e74227b1037fd59658b8e726e56afcb2c113dde44e2136ad132c104b" Dec 11 00:04:08 crc kubenswrapper[4791]: E1211 00:04:08.886636 4791 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-5rb5l_openshift-machine-config-operator(ba35653c-6e06-4cee-a4d6-137764090d18)\"" pod="openshift-machine-config-operator/machine-config-daemon-5rb5l" podUID="ba35653c-6e06-4cee-a4d6-137764090d18" Dec 11 00:04:19 crc kubenswrapper[4791]: I1211 00:04:19.886120 4791 scope.go:117] "RemoveContainer" containerID="174b0c37e74227b1037fd59658b8e726e56afcb2c113dde44e2136ad132c104b" Dec 11 00:04:19 crc kubenswrapper[4791]: E1211 00:04:19.887276 4791 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-5rb5l_openshift-machine-config-operator(ba35653c-6e06-4cee-a4d6-137764090d18)\"" pod="openshift-machine-config-operator/machine-config-daemon-5rb5l" podUID="ba35653c-6e06-4cee-a4d6-137764090d18" Dec 11 00:04:32 crc kubenswrapper[4791]: I1211 00:04:32.886438 4791 scope.go:117] "RemoveContainer" containerID="174b0c37e74227b1037fd59658b8e726e56afcb2c113dde44e2136ad132c104b" Dec 11 00:04:32 crc kubenswrapper[4791]: E1211 00:04:32.887214 4791 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-5rb5l_openshift-machine-config-operator(ba35653c-6e06-4cee-a4d6-137764090d18)\"" pod="openshift-machine-config-operator/machine-config-daemon-5rb5l" podUID="ba35653c-6e06-4cee-a4d6-137764090d18" Dec 11 00:04:47 crc kubenswrapper[4791]: I1211 00:04:47.886114 4791 scope.go:117] "RemoveContainer" containerID="174b0c37e74227b1037fd59658b8e726e56afcb2c113dde44e2136ad132c104b" Dec 11 00:04:47 crc kubenswrapper[4791]: E1211 00:04:47.886903 4791 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-5rb5l_openshift-machine-config-operator(ba35653c-6e06-4cee-a4d6-137764090d18)\"" pod="openshift-machine-config-operator/machine-config-daemon-5rb5l" podUID="ba35653c-6e06-4cee-a4d6-137764090d18" Dec 11 00:04:58 crc kubenswrapper[4791]: I1211 00:04:58.885227 4791 scope.go:117] "RemoveContainer" containerID="174b0c37e74227b1037fd59658b8e726e56afcb2c113dde44e2136ad132c104b" Dec 11 00:04:58 crc kubenswrapper[4791]: E1211 00:04:58.888300 4791 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-5rb5l_openshift-machine-config-operator(ba35653c-6e06-4cee-a4d6-137764090d18)\"" pod="openshift-machine-config-operator/machine-config-daemon-5rb5l" podUID="ba35653c-6e06-4cee-a4d6-137764090d18" var/home/core/zuul-output/logs/crc-cloud-workdir-crc-all-logs.tar.gz0000644000175000000000000000005515116405274024452 0ustar coreroot‹íÁ  ÷Om7 €7šÞ'(var/home/core/zuul-output/logs/crc-cloud/0000755000175000000000000000000015116405275017370 5ustar corerootvar/home/core/zuul-output/artifacts/0000755000175000017500000000000015116374062016512 5ustar corecorevar/home/core/zuul-output/docs/0000755000175000017500000000000015116374063015463 5ustar corecore